[ 473.367827] env[63515]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'linux_bridge' {{(pid=63515) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 473.368681] env[63515]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'noop' {{(pid=63515) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 473.368681] env[63515]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'ovs' {{(pid=63515) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 473.368681] env[63515]: INFO os_vif [-] Loaded VIF plugins: linux_bridge, noop, ovs [ 473.482616] env[63515]: DEBUG oslo_concurrency.processutils [-] Running cmd (subprocess): grep -F node.session.scan /sbin/iscsiadm {{(pid=63515) execute /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/processutils.py:390}} [ 473.493182] env[63515]: DEBUG oslo_concurrency.processutils [-] CMD "grep -F node.session.scan /sbin/iscsiadm" returned: 0 in 0.011s {{(pid=63515) execute /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/processutils.py:428}} [ 474.093155] env[63515]: INFO nova.virt.driver [None req-ed3e6ee9-ccc8-41c3-a5c6-f068b99c8ed9 None None] Loading compute driver 'vmwareapi.VMwareVCDriver' [ 474.163964] env[63515]: DEBUG oslo_concurrency.lockutils [-] Acquiring lock "oslo_vmware_api_lock" by "oslo_vmware.api.VMwareAPISession._create_session" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 474.164153] env[63515]: DEBUG oslo_concurrency.lockutils [-] Lock "oslo_vmware_api_lock" acquired by "oslo_vmware.api.VMwareAPISession._create_session" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 474.164255] env[63515]: DEBUG oslo_vmware.service [-] Creating suds client with soap_url='https://vc1.osci.c.eu-de-1.cloud.sap:443/sdk' and wsdl_url='https://vc1.osci.c.eu-de-1.cloud.sap:443/sdk/vimService.wsdl' {{(pid=63515) __init__ /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:242}} [ 477.236089] env[63515]: DEBUG oslo_vmware.service [-] Invoking ServiceInstance.RetrieveServiceContent with opID=oslo.vmware-b2ea4cbb-c735-4d6c-9077-5fcf981c16d6 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 477.251963] env[63515]: DEBUG oslo_vmware.api [-] Logging into host: vc1.osci.c.eu-de-1.cloud.sap. {{(pid=63515) _create_session /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:242}} [ 477.252124] env[63515]: DEBUG oslo_vmware.service [-] Invoking SessionManager.Login with opID=oslo.vmware-6ed2b44d-d6ec-4696-a9f1-a80eecfde0aa {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 477.284200] env[63515]: INFO oslo_vmware.api [-] Successfully established new session; session ID is 93930. [ 477.284342] env[63515]: DEBUG oslo_concurrency.lockutils [-] Lock "oslo_vmware_api_lock" "released" by "oslo_vmware.api.VMwareAPISession._create_session" :: held 3.120s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 477.284869] env[63515]: INFO nova.virt.vmwareapi.driver [None req-ed3e6ee9-ccc8-41c3-a5c6-f068b99c8ed9 None None] VMware vCenter version: 7.0.3 [ 477.288190] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6858845b-b016-4593-8776-900e5729b633 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 477.305027] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d95720db-b076-4884-93ab-1b4066f048eb {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 477.310587] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d58cfe1-0704-4760-a0d0-1a0fcdcb5f4e {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 477.317217] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f85ffc7-f786-4be4-b632-6ff2d22462ee {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 477.330016] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d20ff37b-956f-4888-bdc6-f5971315457a {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 477.335649] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22579b45-e0ba-4560-8cff-52eb50b4f851 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 477.365401] env[63515]: DEBUG oslo_vmware.service [-] Invoking ExtensionManager.FindExtension with opID=oslo.vmware-a4e3f811-4542-4e25-ad13-b9425d5386df {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 477.370086] env[63515]: DEBUG nova.virt.vmwareapi.driver [None req-ed3e6ee9-ccc8-41c3-a5c6-f068b99c8ed9 None None] Extension org.openstack.compute already exists. {{(pid=63515) _register_openstack_extension /opt/stack/nova/nova/virt/vmwareapi/driver.py:225}} [ 477.372723] env[63515]: INFO nova.compute.provider_config [None req-ed3e6ee9-ccc8-41c3-a5c6-f068b99c8ed9 None None] No provider configs found in /etc/nova/provider_config/. If files are present, ensure the Nova process has access. [ 477.875651] env[63515]: DEBUG nova.context [None req-ed3e6ee9-ccc8-41c3-a5c6-f068b99c8ed9 None None] Found 2 cells: 00000000-0000-0000-0000-000000000000(cell0),ad8e06fe-7f8a-4f5f-91c3-86cd0617c15e(cell1) {{(pid=63515) load_cells /opt/stack/nova/nova/context.py:464}} [ 477.877775] env[63515]: DEBUG oslo_concurrency.lockutils [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 477.878010] env[63515]: DEBUG oslo_concurrency.lockutils [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 477.878676] env[63515]: DEBUG oslo_concurrency.lockutils [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.001s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 477.879117] env[63515]: DEBUG oslo_concurrency.lockutils [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] Acquiring lock "ad8e06fe-7f8a-4f5f-91c3-86cd0617c15e" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 477.879314] env[63515]: DEBUG oslo_concurrency.lockutils [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] Lock "ad8e06fe-7f8a-4f5f-91c3-86cd0617c15e" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 477.880360] env[63515]: DEBUG oslo_concurrency.lockutils [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] Lock "ad8e06fe-7f8a-4f5f-91c3-86cd0617c15e" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.001s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 477.900613] env[63515]: INFO dbcounter [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] Registered counter for database nova_cell0 [ 477.909079] env[63515]: INFO dbcounter [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] Registered counter for database nova_cell1 [ 477.912548] env[63515]: DEBUG oslo_db.sqlalchemy.engines [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] MySQL server mode set to STRICT_TRANS_TABLES,STRICT_ALL_TABLES,NO_ZERO_IN_DATE,NO_ZERO_DATE,ERROR_FOR_DIVISION_BY_ZERO,TRADITIONAL,NO_AUTO_CREATE_USER,NO_ENGINE_SUBSTITUTION {{(pid=63515) _check_effective_sql_mode /opt/stack/data/venv/lib/python3.10/site-packages/oslo_db/sqlalchemy/engines.py:342}} [ 477.913280] env[63515]: DEBUG oslo_db.sqlalchemy.engines [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] MySQL server mode set to STRICT_TRANS_TABLES,STRICT_ALL_TABLES,NO_ZERO_IN_DATE,NO_ZERO_DATE,ERROR_FOR_DIVISION_BY_ZERO,TRADITIONAL,NO_AUTO_CREATE_USER,NO_ENGINE_SUBSTITUTION {{(pid=63515) _check_effective_sql_mode /opt/stack/data/venv/lib/python3.10/site-packages/oslo_db/sqlalchemy/engines.py:342}} [ 477.917944] env[63515]: ERROR nova.db.main.api [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] No DB access allowed in nova-compute: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 477.917944] env[63515]: result = function(*args, **kwargs) [ 477.917944] env[63515]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 477.917944] env[63515]: return func(*args, **kwargs) [ 477.917944] env[63515]: File "/opt/stack/nova/nova/context.py", line 422, in gather_result [ 477.917944] env[63515]: result = fn(*args, **kwargs) [ 477.917944] env[63515]: File "/opt/stack/nova/nova/db/main/api.py", line 179, in wrapper [ 477.917944] env[63515]: return f(*args, **kwargs) [ 477.917944] env[63515]: File "/opt/stack/nova/nova/objects/service.py", line 553, in _db_service_get_minimum_version [ 477.917944] env[63515]: return db.service_get_minimum_version(context, binaries) [ 477.917944] env[63515]: File "/opt/stack/nova/nova/db/main/api.py", line 238, in wrapper [ 477.917944] env[63515]: _check_db_access() [ 477.917944] env[63515]: File "/opt/stack/nova/nova/db/main/api.py", line 188, in _check_db_access [ 477.917944] env[63515]: stacktrace = ''.join(traceback.format_stack()) [ 477.917944] env[63515]: [ 477.918734] env[63515]: ERROR nova.db.main.api [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] No DB access allowed in nova-compute: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 477.918734] env[63515]: result = function(*args, **kwargs) [ 477.918734] env[63515]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 477.918734] env[63515]: return func(*args, **kwargs) [ 477.918734] env[63515]: File "/opt/stack/nova/nova/context.py", line 422, in gather_result [ 477.918734] env[63515]: result = fn(*args, **kwargs) [ 477.918734] env[63515]: File "/opt/stack/nova/nova/db/main/api.py", line 179, in wrapper [ 477.918734] env[63515]: return f(*args, **kwargs) [ 477.918734] env[63515]: File "/opt/stack/nova/nova/objects/service.py", line 553, in _db_service_get_minimum_version [ 477.918734] env[63515]: return db.service_get_minimum_version(context, binaries) [ 477.918734] env[63515]: File "/opt/stack/nova/nova/db/main/api.py", line 238, in wrapper [ 477.918734] env[63515]: _check_db_access() [ 477.918734] env[63515]: File "/opt/stack/nova/nova/db/main/api.py", line 188, in _check_db_access [ 477.918734] env[63515]: stacktrace = ''.join(traceback.format_stack()) [ 477.918734] env[63515]: [ 477.919187] env[63515]: WARNING nova.objects.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] Failed to get minimum service version for cell ad8e06fe-7f8a-4f5f-91c3-86cd0617c15e [ 477.919269] env[63515]: WARNING nova.objects.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] Failed to get minimum service version for cell 00000000-0000-0000-0000-000000000000 [ 477.919698] env[63515]: DEBUG oslo_concurrency.lockutils [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] Acquiring lock "singleton_lock" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 477.919833] env[63515]: DEBUG oslo_concurrency.lockutils [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] Acquired lock "singleton_lock" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 477.920086] env[63515]: DEBUG oslo_concurrency.lockutils [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] Releasing lock "singleton_lock" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 477.920416] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] Full set of CONF: {{(pid=63515) _wait_for_exit_or_signal /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/service.py:363}} [ 477.920560] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] ******************************************************************************** {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2806}} [ 477.920687] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] Configuration options gathered from: {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2807}} [ 477.920820] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] command line args: ['--config-file', '/etc/nova/nova.conf', '--config-file', '/etc/nova/nova-cpu-common.conf', '--config-file', '/etc/nova/nova-cpu-1.conf'] {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2808}} [ 477.921015] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] config files: ['/etc/nova/nova.conf', '/etc/nova/nova-cpu-common.conf', '/etc/nova/nova-cpu-1.conf'] {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2809}} [ 477.921153] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] ================================================================================ {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2811}} [ 477.921367] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] allow_resize_to_same_host = True {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 477.921536] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] arq_binding_timeout = 300 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 477.921667] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] backdoor_port = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 477.921792] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] backdoor_socket = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 477.921985] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] block_device_allocate_retries = 60 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 477.922130] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] block_device_allocate_retries_interval = 3 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 477.922298] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] cert = self.pem {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 477.922462] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] compute_driver = vmwareapi.VMwareVCDriver {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 477.922629] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] compute_monitors = [] {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 477.922802] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] config_dir = [] {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 477.922987] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] config_drive_format = iso9660 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 477.923152] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] config_file = ['/etc/nova/nova.conf', '/etc/nova/nova-cpu-common.conf', '/etc/nova/nova-cpu-1.conf'] {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 477.923395] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] config_source = [] {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 477.923594] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] console_host = devstack {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 477.923764] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] control_exchange = nova {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 477.923927] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] cpu_allocation_ratio = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 477.924097] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] daemon = False {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 477.924270] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] debug = True {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 477.924429] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] default_access_ip_network_name = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 477.924594] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] default_availability_zone = nova {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 477.924749] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] default_ephemeral_format = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 477.924907] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] default_green_pool_size = 1000 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 477.925152] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] default_log_levels = ['amqp=WARN', 'amqplib=WARN', 'boto=WARN', 'qpid=WARN', 'sqlalchemy=WARN', 'suds=INFO', 'oslo.messaging=INFO', 'oslo_messaging=INFO', 'iso8601=WARN', 'requests.packages.urllib3.connectionpool=WARN', 'urllib3.connectionpool=WARN', 'websocket=WARN', 'requests.packages.urllib3.util.retry=WARN', 'urllib3.util.retry=WARN', 'keystonemiddleware=WARN', 'routes.middleware=WARN', 'stevedore=WARN', 'taskflow=WARN', 'keystoneauth=WARN', 'oslo.cache=INFO', 'oslo_policy=INFO', 'dogpile.core.dogpile=INFO', 'glanceclient=WARN', 'oslo.privsep.daemon=INFO'] {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 477.925321] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] default_schedule_zone = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 477.925604] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] disk_allocation_ratio = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 477.925808] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] enable_new_services = True {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 477.925996] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] enabled_apis = ['osapi_compute'] {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 477.926180] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] enabled_ssl_apis = [] {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 477.926342] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] flat_injected = False {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 477.926501] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] force_config_drive = False {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 477.926662] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] force_raw_images = True {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 477.926833] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] graceful_shutdown_timeout = 5 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 477.926995] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] heal_instance_info_cache_interval = 60 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 477.927226] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] host = cpu-1 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 477.927409] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] initial_cpu_allocation_ratio = 4.0 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 477.927575] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] initial_disk_allocation_ratio = 1.0 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 477.927741] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] initial_ram_allocation_ratio = 1.0 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 477.927953] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] injected_network_template = /opt/stack/nova/nova/virt/interfaces.template {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 477.928136] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] instance_build_timeout = 0 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 477.928300] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] instance_delete_interval = 300 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 477.928466] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] instance_format = [instance: %(uuid)s] {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 477.928629] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] instance_name_template = instance-%08x {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 477.928787] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] instance_usage_audit = False {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 477.928956] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] instance_usage_audit_period = month {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 477.929145] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] instance_uuid_format = [instance: %(uuid)s] {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 477.929315] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] instances_path = /opt/stack/data/nova/instances {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 477.929479] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] internal_service_availability_zone = internal {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 477.929635] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] key = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 477.929794] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] live_migration_retry_count = 30 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 477.929961] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] log_color = False {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 477.930140] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] log_config_append = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 477.930310] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] log_date_format = %Y-%m-%d %H:%M:%S {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 477.930470] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] log_dir = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 477.930626] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] log_file = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 477.930752] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] log_options = True {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 477.930915] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] log_rotate_interval = 1 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 477.931095] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] log_rotate_interval_type = days {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 477.931264] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] log_rotation_type = none {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 477.931432] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] logging_context_format_string = %(color)s%(levelname)s %(name)s [%(global_request_id)s %(request_id)s %(project_name)s %(user_name)s%(color)s] %(instance)s%(color)s%(message)s {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 477.931504] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] logging_debug_format_suffix = {{(pid=%(process)d) %(funcName)s %(pathname)s:%(lineno)d}} {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 477.931670] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] logging_default_format_string = %(color)s%(levelname)s %(name)s [-%(color)s] %(instance)s%(color)s%(message)s {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 477.931834] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] logging_exception_prefix = ERROR %(name)s %(instance)s {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 477.931972] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] logging_user_identity_format = %(user)s %(project)s %(domain)s %(system_scope)s %(user_domain)s %(project_domain)s {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 477.932182] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] long_rpc_timeout = 1800 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 477.932359] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] max_concurrent_builds = 10 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 477.932520] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] max_concurrent_live_migrations = 1 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 477.932682] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] max_concurrent_snapshots = 5 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 477.932841] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] max_local_block_devices = 3 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 477.933019] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] max_logfile_count = 30 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 477.933214] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] max_logfile_size_mb = 200 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 477.933393] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] maximum_instance_delete_attempts = 5 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 477.933648] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] metadata_listen = 0.0.0.0 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 477.933823] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] metadata_listen_port = 8775 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 477.933995] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] metadata_workers = 2 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 477.934174] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] migrate_max_retries = -1 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 477.934342] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] mkisofs_cmd = genisoimage {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 477.934550] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] my_block_storage_ip = 10.180.1.21 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 477.934685] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] my_ip = 10.180.1.21 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 477.934848] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] network_allocate_retries = 0 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 477.935034] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] non_inheritable_image_properties = ['cache_in_nova', 'bittorrent'] {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 477.935211] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] osapi_compute_listen = 0.0.0.0 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 477.935373] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] osapi_compute_listen_port = 8774 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 477.935538] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] osapi_compute_unique_server_name_scope = {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 477.935862] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] osapi_compute_workers = 2 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 477.936053] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] password_length = 12 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 477.936227] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] periodic_enable = True {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 477.936390] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] periodic_fuzzy_delay = 60 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 477.936561] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] pointer_model = usbtablet {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 477.936728] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] preallocate_images = none {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 477.936886] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] publish_errors = False {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 477.937021] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] pybasedir = /opt/stack/nova {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 477.937186] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] ram_allocation_ratio = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 477.937347] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] rate_limit_burst = 0 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 477.937514] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] rate_limit_except_level = CRITICAL {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 477.937672] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] rate_limit_interval = 0 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 477.937831] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] reboot_timeout = 0 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 477.937991] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] reclaim_instance_interval = 0 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 477.938163] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] record = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 477.938330] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] reimage_timeout_per_gb = 60 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 477.938495] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] report_interval = 120 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 477.938656] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] rescue_timeout = 0 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 477.938814] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] reserved_host_cpus = 0 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 477.938971] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] reserved_host_disk_mb = 0 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 477.939144] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] reserved_host_memory_mb = 512 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 477.939306] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] reserved_huge_pages = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 477.939467] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] resize_confirm_window = 0 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 477.939624] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] resize_fs_using_block_device = False {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 477.939779] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] resume_guests_state_on_host_boot = False {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 477.939943] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] rootwrap_config = /etc/nova/rootwrap.conf {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 477.940116] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] rpc_response_timeout = 60 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 477.940277] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] run_external_periodic_tasks = True {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 477.940443] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] running_deleted_instance_action = reap {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 477.940604] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] running_deleted_instance_poll_interval = 1800 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 477.940762] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] running_deleted_instance_timeout = 0 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 477.940917] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] scheduler_instance_sync_interval = 120 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 477.941098] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] service_down_time = 720 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 477.941272] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] servicegroup_driver = db {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 477.941429] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] shell_completion = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 477.941589] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] shelved_offload_time = 0 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 477.941747] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] shelved_poll_interval = 3600 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 477.941911] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] shutdown_timeout = 0 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 477.942107] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] source_is_ipv6 = False {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 477.942277] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] ssl_only = False {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 477.942520] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] state_path = /opt/stack/data/n-cpu-1 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 477.942689] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] sync_power_state_interval = 600 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 477.942850] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] sync_power_state_pool_size = 1000 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 477.943029] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] syslog_log_facility = LOG_USER {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 477.943218] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] tempdir = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 477.943387] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] timeout_nbd = 10 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 477.943630] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] transport_url = **** {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 477.943816] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] update_resources_interval = 0 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 477.943983] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] use_cow_images = True {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 477.944160] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] use_eventlog = False {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 477.944320] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] use_journal = False {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 477.944478] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] use_json = False {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 477.944635] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] use_rootwrap_daemon = False {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 477.944792] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] use_stderr = False {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 477.944947] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] use_syslog = False {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 477.945132] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] vcpu_pin_set = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 477.945321] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] vif_plugging_is_fatal = True {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 477.945493] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] vif_plugging_timeout = 300 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 477.945660] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] virt_mkfs = [] {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 477.945982] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] volume_usage_poll_interval = 0 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 477.946175] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] watch_log_file = False {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 477.946352] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] web = /usr/share/spice-html5 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 477.946535] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] oslo_concurrency.disable_process_locking = False {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.947100] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] oslo_concurrency.lock_path = /opt/stack/data/n-cpu-1 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.947303] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] oslo_messaging_metrics.metrics_buffer_size = 1000 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.947479] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] oslo_messaging_metrics.metrics_enabled = False {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.947657] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] oslo_messaging_metrics.metrics_process_name = {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.947833] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] oslo_messaging_metrics.metrics_socket_file = /var/tmp/metrics_collector.sock {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.948009] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] oslo_messaging_metrics.metrics_thread_stop_timeout = 10 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.948207] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] api.auth_strategy = keystone {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.948379] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] api.compute_link_prefix = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.948556] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] api.config_drive_skip_versions = 1.0 2007-01-19 2007-03-01 2007-08-29 2007-10-10 2007-12-15 2008-02-01 2008-09-01 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.948731] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] api.dhcp_domain = novalocal {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.948902] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] api.enable_instance_password = True {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.949089] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] api.glance_link_prefix = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.949269] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] api.instance_list_cells_batch_fixed_size = 100 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.949443] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] api.instance_list_cells_batch_strategy = distributed {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.949608] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] api.instance_list_per_project_cells = False {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.949772] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] api.list_records_by_skipping_down_cells = True {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.949936] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] api.local_metadata_per_cell = False {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.950120] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] api.max_limit = 1000 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.950296] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] api.metadata_cache_expiration = 15 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.950470] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] api.neutron_default_tenant_id = default {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.950669] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] api.response_validation = warn {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.950866] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] api.use_neutron_default_nets = False {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.951187] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] api.vendordata_dynamic_connect_timeout = 5 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.951402] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] api.vendordata_dynamic_failure_fatal = False {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.951632] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] api.vendordata_dynamic_read_timeout = 5 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.951795] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] api.vendordata_dynamic_ssl_certfile = {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.952114] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] api.vendordata_dynamic_targets = [] {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.952315] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] api.vendordata_jsonfile_path = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.952507] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] api.vendordata_providers = ['StaticJSON'] {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.952706] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] cache.backend = dogpile.cache.memcached {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.952881] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] cache.backend_argument = **** {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.953378] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] cache.config_prefix = cache.oslo {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.953653] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] cache.dead_timeout = 60.0 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.953919] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] cache.debug_cache_backend = False {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.954128] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] cache.enable_retry_client = False {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.954306] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] cache.enable_socket_keepalive = False {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.954483] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] cache.enabled = True {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.954650] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] cache.enforce_fips_mode = False {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.954821] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] cache.expiration_time = 600 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.954987] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] cache.hashclient_retry_attempts = 2 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.955180] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] cache.hashclient_retry_delay = 1.0 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.955348] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] cache.memcache_dead_retry = 300 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.955509] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] cache.memcache_password = **** {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.955675] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] cache.memcache_pool_connection_get_timeout = 10 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.956009] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] cache.memcache_pool_flush_on_reconnect = False {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.956222] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] cache.memcache_pool_maxsize = 10 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.956394] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] cache.memcache_pool_unused_timeout = 60 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.956561] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] cache.memcache_sasl_enabled = False {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.956744] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] cache.memcache_servers = ['localhost:11211'] {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.956915] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] cache.memcache_socket_timeout = 1.0 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.957090] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] cache.memcache_username = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.957287] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] cache.proxies = [] {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.957458] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] cache.redis_db = 0 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.957640] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] cache.redis_password = **** {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.957823] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] cache.redis_sentinel_service_name = mymaster {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.958015] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] cache.redis_sentinels = ['localhost:26379'] {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.958197] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] cache.redis_server = localhost:6379 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.958367] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] cache.redis_socket_timeout = 1.0 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.958528] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] cache.redis_username = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.958693] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] cache.retry_attempts = 2 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.958859] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] cache.retry_delay = 0.0 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.959034] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] cache.socket_keepalive_count = 1 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.959205] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] cache.socket_keepalive_idle = 1 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.959368] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] cache.socket_keepalive_interval = 1 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.959528] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] cache.tls_allowed_ciphers = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.959686] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] cache.tls_cafile = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.959841] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] cache.tls_certfile = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.960008] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] cache.tls_enabled = False {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.960176] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] cache.tls_keyfile = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.960347] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] cinder.auth_section = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.960522] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] cinder.auth_type = password {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.960683] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] cinder.cafile = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.960859] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] cinder.catalog_info = volumev3::publicURL {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.961029] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] cinder.certfile = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.961200] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] cinder.collect_timing = False {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.961362] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] cinder.cross_az_attach = True {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.961524] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] cinder.debug = False {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.961733] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] cinder.endpoint_template = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.961841] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] cinder.http_retries = 3 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.962033] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] cinder.insecure = False {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.962217] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] cinder.keyfile = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.962393] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] cinder.os_region_name = RegionOne {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.962557] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] cinder.split_loggers = False {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.962719] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] cinder.timeout = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.962907] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] compute.consecutive_build_service_disable_threshold = 10 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.963103] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] compute.cpu_dedicated_set = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.963289] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] compute.cpu_shared_set = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.963460] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] compute.image_type_exclude_list = [] {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.963627] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] compute.live_migration_wait_for_vif_plug = True {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.963882] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] compute.max_concurrent_disk_ops = 0 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.964080] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] compute.max_disk_devices_to_attach = -1 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.964256] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] compute.packing_host_numa_cells_allocation_strategy = False {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.964428] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] compute.provider_config_location = /etc/nova/provider_config/ {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.964594] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] compute.resource_provider_association_refresh = 300 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.964755] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] compute.sharing_providers_max_uuids_per_request = 200 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.964918] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] compute.shutdown_retry_interval = 10 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.965113] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] compute.vmdk_allowed_types = ['streamOptimized', 'monolithicSparse'] {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.965301] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] conductor.workers = 2 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.965483] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] console.allowed_origins = [] {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.965646] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] console.ssl_ciphers = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.965818] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] console.ssl_minimum_version = default {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.966144] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] consoleauth.enforce_session_timeout = False {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.966344] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] consoleauth.token_ttl = 600 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.966518] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] cyborg.cafile = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.966679] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] cyborg.certfile = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.966846] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] cyborg.collect_timing = False {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.967016] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] cyborg.connect_retries = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.967193] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] cyborg.connect_retry_delay = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.967357] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] cyborg.endpoint_override = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.967521] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] cyborg.insecure = False {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.967678] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] cyborg.keyfile = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.967837] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] cyborg.max_version = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.967998] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] cyborg.min_version = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.968173] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] cyborg.region_name = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.968332] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] cyborg.retriable_status_codes = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.968487] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] cyborg.service_name = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.968658] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] cyborg.service_type = accelerator {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.968817] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] cyborg.split_loggers = False {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.968975] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] cyborg.status_code_retries = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.969149] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] cyborg.status_code_retry_delay = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.969309] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] cyborg.timeout = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.969495] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] cyborg.valid_interfaces = ['internal', 'public'] {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.969760] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] cyborg.version = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.969978] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] database.backend = sqlalchemy {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.970175] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] database.connection = **** {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.970347] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] database.connection_debug = 0 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.970520] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] database.connection_parameters = {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.970687] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] database.connection_recycle_time = 3600 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.970851] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] database.connection_trace = False {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.971020] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] database.db_inc_retry_interval = True {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.971191] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] database.db_max_retries = 20 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.971360] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] database.db_max_retry_interval = 10 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.971519] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] database.db_retry_interval = 1 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.971682] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] database.max_overflow = 50 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.971844] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] database.max_pool_size = 5 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.972031] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] database.max_retries = 10 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.972220] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] database.mysql_sql_mode = TRADITIONAL {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.972386] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] database.mysql_wsrep_sync_wait = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.972548] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] database.pool_timeout = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.972716] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] database.retry_interval = 10 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.972876] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] database.slave_connection = **** {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.973070] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] database.sqlite_synchronous = True {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.973248] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] database.use_db_reconnect = False {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.973429] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] api_database.backend = sqlalchemy {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.973600] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] api_database.connection = **** {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.973764] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] api_database.connection_debug = 0 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.974036] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] api_database.connection_parameters = {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.974219] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] api_database.connection_recycle_time = 3600 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.974387] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] api_database.connection_trace = False {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.974550] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] api_database.db_inc_retry_interval = True {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.974716] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] api_database.db_max_retries = 20 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.974879] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] api_database.db_max_retry_interval = 10 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.975056] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] api_database.db_retry_interval = 1 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.975229] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] api_database.max_overflow = 50 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.975392] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] api_database.max_pool_size = 5 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.975558] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] api_database.max_retries = 10 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.975724] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] api_database.mysql_sql_mode = TRADITIONAL {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.975883] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] api_database.mysql_wsrep_sync_wait = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.976161] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] api_database.pool_timeout = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.976358] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] api_database.retry_interval = 10 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.976525] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] api_database.slave_connection = **** {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.976690] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] api_database.sqlite_synchronous = True {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.976870] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] devices.enabled_mdev_types = [] {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.977062] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] ephemeral_storage_encryption.cipher = aes-xts-plain64 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.977244] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] ephemeral_storage_encryption.default_format = luks {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.977411] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] ephemeral_storage_encryption.enabled = False {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.977597] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] ephemeral_storage_encryption.key_size = 512 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.977770] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] glance.api_servers = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.977935] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] glance.cafile = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.978112] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] glance.certfile = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.978287] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] glance.collect_timing = False {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.978458] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] glance.connect_retries = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.978621] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] glance.connect_retry_delay = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.978784] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] glance.debug = False {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.978953] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] glance.default_trusted_certificate_ids = [] {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.979135] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] glance.enable_certificate_validation = False {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.979300] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] glance.enable_rbd_download = False {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.979459] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] glance.endpoint_override = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.979625] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] glance.insecure = False {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.979788] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] glance.keyfile = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.979946] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] glance.max_version = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.980116] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] glance.min_version = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.980285] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] glance.num_retries = 3 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.980453] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] glance.rbd_ceph_conf = {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.980617] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] glance.rbd_connect_timeout = 5 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.980786] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] glance.rbd_pool = {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.980952] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] glance.rbd_user = {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.981128] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] glance.region_name = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.981289] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] glance.retriable_status_codes = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.981444] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] glance.service_name = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.981611] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] glance.service_type = image {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.981770] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] glance.split_loggers = False {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.981929] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] glance.status_code_retries = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.982122] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] glance.status_code_retry_delay = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.982287] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] glance.timeout = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.982474] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] glance.valid_interfaces = ['internal', 'public'] {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.982638] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] glance.verify_glance_signatures = False {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.982797] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] glance.version = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.982965] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] guestfs.debug = False {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.983159] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] mks.enabled = False {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.983518] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] mks.mksproxy_base_url = http://127.0.0.1:6090/ {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.983711] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] image_cache.manager_interval = 2400 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.983947] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] image_cache.precache_concurrency = 1 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.984163] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] image_cache.remove_unused_base_images = True {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.984340] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] image_cache.remove_unused_original_minimum_age_seconds = 86400 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.984513] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] image_cache.remove_unused_resized_minimum_age_seconds = 3600 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.984691] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] image_cache.subdirectory_name = _base {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.984870] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] ironic.api_max_retries = 60 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.985046] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] ironic.api_retry_interval = 2 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.985215] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] ironic.auth_section = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.985380] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] ironic.auth_type = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.985542] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] ironic.cafile = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.985705] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] ironic.certfile = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.985865] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] ironic.collect_timing = False {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.986037] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] ironic.conductor_group = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.986284] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] ironic.connect_retries = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.986469] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] ironic.connect_retry_delay = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.986630] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] ironic.endpoint_override = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.986792] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] ironic.insecure = False {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.986950] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] ironic.keyfile = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.987129] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] ironic.max_version = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.987287] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] ironic.min_version = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.987454] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] ironic.peer_list = [] {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.987611] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] ironic.region_name = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.987769] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] ironic.retriable_status_codes = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.987933] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] ironic.serial_console_state_timeout = 10 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.988103] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] ironic.service_name = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.988276] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] ironic.service_type = baremetal {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.988438] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] ironic.shard = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.988599] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] ironic.split_loggers = False {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.988756] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] ironic.status_code_retries = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.988913] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] ironic.status_code_retry_delay = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.989084] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] ironic.timeout = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.989267] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] ironic.valid_interfaces = ['internal', 'public'] {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.989430] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] ironic.version = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.989611] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] key_manager.backend = nova.keymgr.conf_key_mgr.ConfKeyManager {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.989784] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] key_manager.fixed_key = **** {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.989969] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] barbican.auth_endpoint = http://localhost/identity/v3 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.990145] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] barbican.barbican_api_version = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.990303] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] barbican.barbican_endpoint = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.990470] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] barbican.barbican_endpoint_type = public {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.990629] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] barbican.barbican_region_name = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.990786] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] barbican.cafile = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.990941] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] barbican.certfile = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.991114] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] barbican.collect_timing = False {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.991278] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] barbican.insecure = False {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.991433] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] barbican.keyfile = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.991592] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] barbican.number_of_retries = 60 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.991750] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] barbican.retry_delay = 1 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.991913] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] barbican.send_service_user_token = False {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.992107] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] barbican.split_loggers = False {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.992276] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] barbican.timeout = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.992437] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] barbican.verify_ssl = True {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.992594] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] barbican.verify_ssl_path = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.992762] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] barbican_service_user.auth_section = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.992924] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] barbican_service_user.auth_type = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.993112] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] barbican_service_user.cafile = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.993291] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] barbican_service_user.certfile = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.993456] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] barbican_service_user.collect_timing = False {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.993617] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] barbican_service_user.insecure = False {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.993775] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] barbican_service_user.keyfile = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.993935] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] barbican_service_user.split_loggers = False {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.994197] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] barbican_service_user.timeout = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.994382] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] vault.approle_role_id = **** {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.994542] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] vault.approle_secret_id = **** {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.994714] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] vault.kv_mountpoint = secret {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.994878] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] vault.kv_path = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.995058] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] vault.kv_version = 2 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.995226] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] vault.namespace = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.995385] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] vault.root_token_id = **** {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.995542] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] vault.ssl_ca_crt_file = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.995714] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] vault.timeout = 60.0 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.995875] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] vault.use_ssl = False {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.996051] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] vault.vault_url = http://127.0.0.1:8200 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.996235] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] keystone.auth_section = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.996497] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] keystone.auth_type = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.996667] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] keystone.cafile = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.996829] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] keystone.certfile = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.996995] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] keystone.collect_timing = False {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.997207] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] keystone.connect_retries = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.997427] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] keystone.connect_retry_delay = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.997602] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] keystone.endpoint_override = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.997770] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] keystone.insecure = False {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.997932] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] keystone.keyfile = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.998106] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] keystone.max_version = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.998275] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] keystone.min_version = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.998433] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] keystone.region_name = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.998594] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] keystone.retriable_status_codes = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.998749] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] keystone.service_name = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.998919] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] keystone.service_type = identity {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.999095] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] keystone.split_loggers = False {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.999262] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] keystone.status_code_retries = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.999424] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] keystone.status_code_retry_delay = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.999584] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] keystone.timeout = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 477.999798] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] keystone.valid_interfaces = ['internal', 'public'] {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.000023] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] keystone.version = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.000242] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] libvirt.connection_uri = {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.000409] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] libvirt.cpu_mode = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.000580] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] libvirt.cpu_model_extra_flags = [] {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.000749] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] libvirt.cpu_models = [] {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.000921] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] libvirt.cpu_power_governor_high = performance {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.001104] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] libvirt.cpu_power_governor_low = powersave {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.001277] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] libvirt.cpu_power_management = False {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.001449] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] libvirt.cpu_power_management_strategy = cpu_state {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.001616] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] libvirt.device_detach_attempts = 8 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.001779] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] libvirt.device_detach_timeout = 20 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.001944] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] libvirt.disk_cachemodes = [] {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.002144] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] libvirt.disk_prefix = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.002320] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] libvirt.enabled_perf_events = [] {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.002488] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] libvirt.file_backed_memory = 0 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.002656] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] libvirt.gid_maps = [] {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.002873] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] libvirt.hw_disk_discard = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.003060] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] libvirt.hw_machine_type = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.003259] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] libvirt.images_rbd_ceph_conf = {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.003431] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] libvirt.images_rbd_glance_copy_poll_interval = 15 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.003597] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] libvirt.images_rbd_glance_copy_timeout = 600 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.003774] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] libvirt.images_rbd_glance_store_name = {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.003948] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] libvirt.images_rbd_pool = rbd {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.004213] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] libvirt.images_type = default {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.004389] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] libvirt.images_volume_group = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.004552] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] libvirt.inject_key = False {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.004714] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] libvirt.inject_partition = -2 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.004877] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] libvirt.inject_password = False {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.005051] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] libvirt.iscsi_iface = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.005222] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] libvirt.iser_use_multipath = False {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.005387] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] libvirt.live_migration_bandwidth = 0 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.005553] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] libvirt.live_migration_completion_timeout = 800 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.005714] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] libvirt.live_migration_downtime = 500 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.005874] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] libvirt.live_migration_downtime_delay = 75 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.006045] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] libvirt.live_migration_downtime_steps = 10 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.006211] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] libvirt.live_migration_inbound_addr = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.006446] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] libvirt.live_migration_permit_auto_converge = False {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.006640] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] libvirt.live_migration_permit_post_copy = False {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.006804] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] libvirt.live_migration_scheme = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.006980] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] libvirt.live_migration_timeout_action = abort {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.007166] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] libvirt.live_migration_tunnelled = False {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.007331] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] libvirt.live_migration_uri = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.007500] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] libvirt.live_migration_with_native_tls = False {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.007664] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] libvirt.max_queues = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.007831] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] libvirt.mem_stats_period_seconds = 10 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.008075] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] libvirt.migration_inbound_addr = 10.180.1.21 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.008249] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] libvirt.nfs_mount_options = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.008545] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] libvirt.nfs_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.008721] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] libvirt.num_aoe_discover_tries = 3 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.008887] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] libvirt.num_iser_scan_tries = 5 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.009058] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] libvirt.num_memory_encrypted_guests = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.009255] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] libvirt.num_nvme_discover_tries = 5 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.009441] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] libvirt.num_pcie_ports = 0 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.009613] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] libvirt.num_volume_scan_tries = 5 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.009782] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] libvirt.pmem_namespaces = [] {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.009941] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] libvirt.quobyte_client_cfg = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.010241] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] libvirt.quobyte_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.010420] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] libvirt.rbd_connect_timeout = 5 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.010588] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] libvirt.rbd_destroy_volume_retries = 12 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.010753] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] libvirt.rbd_destroy_volume_retry_interval = 5 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.010916] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] libvirt.rbd_secret_uuid = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.011089] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] libvirt.rbd_user = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.011262] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] libvirt.realtime_scheduler_priority = 1 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.011435] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] libvirt.remote_filesystem_transport = ssh {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.011597] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] libvirt.rescue_image_id = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.011755] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] libvirt.rescue_kernel_id = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.011914] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] libvirt.rescue_ramdisk_id = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.012116] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] libvirt.rng_dev_path = /dev/urandom {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.012287] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] libvirt.rx_queue_size = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.012456] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] libvirt.smbfs_mount_options = {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.012734] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] libvirt.smbfs_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.012905] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] libvirt.snapshot_compression = False {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.013100] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] libvirt.snapshot_image_format = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.013344] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] libvirt.snapshots_directory = /opt/stack/data/nova/instances/snapshots {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.013514] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] libvirt.sparse_logical_volumes = False {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.013681] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] libvirt.swtpm_enabled = False {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.013848] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] libvirt.swtpm_group = tss {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.014030] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] libvirt.swtpm_user = tss {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.014284] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] libvirt.sysinfo_serial = unique {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.014460] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] libvirt.tb_cache_size = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.014622] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] libvirt.tx_queue_size = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.014792] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] libvirt.uid_maps = [] {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.014957] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] libvirt.use_virtio_for_bridges = True {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.015146] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] libvirt.virt_type = kvm {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.015321] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] libvirt.volume_clear = zero {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.015487] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] libvirt.volume_clear_size = 0 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.015654] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] libvirt.volume_use_multipath = False {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.015814] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] libvirt.vzstorage_cache_path = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.015982] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] libvirt.vzstorage_log_path = /var/log/vstorage/%(cluster_name)s/nova.log.gz {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.016166] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] libvirt.vzstorage_mount_group = qemu {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.016333] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] libvirt.vzstorage_mount_opts = [] {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.016582] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] libvirt.vzstorage_mount_perms = 0770 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.016880] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] libvirt.vzstorage_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.017073] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] libvirt.vzstorage_mount_user = stack {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.017315] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] libvirt.wait_soft_reboot_seconds = 120 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.017524] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] neutron.auth_section = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.017702] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] neutron.auth_type = password {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.017866] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] neutron.cafile = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.018038] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] neutron.certfile = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.018215] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] neutron.collect_timing = False {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.018375] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] neutron.connect_retries = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.018534] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] neutron.connect_retry_delay = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.018704] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] neutron.default_floating_pool = public {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.018861] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] neutron.endpoint_override = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.019035] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] neutron.extension_sync_interval = 600 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.019205] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] neutron.http_retries = 3 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.019366] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] neutron.insecure = False {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.019526] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] neutron.keyfile = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.019678] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] neutron.max_version = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.019845] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] neutron.metadata_proxy_shared_secret = **** {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.020010] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] neutron.min_version = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.020185] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] neutron.ovs_bridge = br-int {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.020347] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] neutron.physnets = [] {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.020515] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] neutron.region_name = RegionOne {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.020675] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] neutron.retriable_status_codes = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.020841] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] neutron.service_metadata_proxy = True {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.021008] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] neutron.service_name = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.021180] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] neutron.service_type = network {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.021377] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] neutron.split_loggers = False {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.021546] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] neutron.status_code_retries = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.021705] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] neutron.status_code_retry_delay = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.021861] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] neutron.timeout = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.022070] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] neutron.valid_interfaces = ['internal', 'public'] {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.022244] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] neutron.version = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.022416] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] notifications.bdms_in_notifications = False {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.022591] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] notifications.default_level = INFO {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.022764] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] notifications.notification_format = unversioned {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.022927] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] notifications.notify_on_state_change = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.023114] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] notifications.versioned_notifications_topics = ['versioned_notifications'] {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.023295] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] pci.alias = [] {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.023464] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] pci.device_spec = [] {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.023626] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] pci.report_in_placement = False {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.023801] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] placement.auth_section = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.023972] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] placement.auth_type = password {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.024155] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] placement.auth_url = http://10.180.1.21/identity {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.024401] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] placement.cafile = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.024569] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] placement.certfile = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.024733] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] placement.collect_timing = False {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.024891] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] placement.connect_retries = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.025062] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] placement.connect_retry_delay = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.025230] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] placement.default_domain_id = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.025386] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] placement.default_domain_name = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.025542] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] placement.domain_id = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.025698] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] placement.domain_name = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.025853] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] placement.endpoint_override = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.026021] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] placement.insecure = False {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.026183] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] placement.keyfile = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.026340] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] placement.max_version = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.026500] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] placement.min_version = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.026753] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] placement.password = **** {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.026929] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] placement.project_domain_id = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.027112] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] placement.project_domain_name = Default {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.027285] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] placement.project_id = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.027459] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] placement.project_name = service {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.027630] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] placement.region_name = RegionOne {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.027793] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] placement.retriable_status_codes = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.027957] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] placement.service_name = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.028140] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] placement.service_type = placement {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.028306] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] placement.split_loggers = False {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.028465] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] placement.status_code_retries = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.028630] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] placement.status_code_retry_delay = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.028791] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] placement.system_scope = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.028948] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] placement.timeout = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.029120] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] placement.trust_id = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.029286] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] placement.user_domain_id = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.029451] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] placement.user_domain_name = Default {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.029609] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] placement.user_id = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.029781] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] placement.username = nova {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.029961] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] placement.valid_interfaces = ['internal', 'public'] {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.030139] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] placement.version = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.030318] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] quota.cores = 20 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.030485] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] quota.count_usage_from_placement = False {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.030657] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] quota.driver = nova.quota.DbQuotaDriver {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.030829] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] quota.injected_file_content_bytes = 10240 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.030997] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] quota.injected_file_path_length = 255 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.031181] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] quota.injected_files = 5 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.031349] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] quota.instances = 10 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.031515] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] quota.key_pairs = 100 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.031681] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] quota.metadata_items = 128 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.031846] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] quota.ram = 51200 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.032031] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] quota.recheck_quota = True {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.032230] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] quota.server_group_members = 10 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.032418] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] quota.server_groups = 10 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.032594] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] scheduler.discover_hosts_in_cells_interval = -1 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.032758] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] scheduler.enable_isolated_aggregate_filtering = False {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.032919] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] scheduler.image_metadata_prefilter = False {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.033142] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] scheduler.limit_tenants_to_placement_aggregate = False {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.033336] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] scheduler.max_attempts = 3 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.033508] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] scheduler.max_placement_results = 1000 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.033670] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] scheduler.placement_aggregate_required_for_tenants = False {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.033832] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] scheduler.query_placement_for_image_type_support = False {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.033992] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] scheduler.query_placement_for_routed_network_aggregates = False {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.034185] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] scheduler.workers = 2 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.034427] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] filter_scheduler.aggregate_image_properties_isolation_namespace = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.034620] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] filter_scheduler.aggregate_image_properties_isolation_separator = . {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.034810] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] filter_scheduler.available_filters = ['nova.scheduler.filters.all_filters'] {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.034989] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] filter_scheduler.build_failure_weight_multiplier = 1000000.0 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.035176] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] filter_scheduler.cpu_weight_multiplier = 1.0 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.035345] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] filter_scheduler.cross_cell_move_weight_multiplier = 1000000.0 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.035512] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] filter_scheduler.disk_weight_multiplier = 1.0 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.035701] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] filter_scheduler.enabled_filters = ['ComputeFilter', 'ComputeCapabilitiesFilter', 'ImagePropertiesFilter', 'ServerGroupAntiAffinityFilter', 'ServerGroupAffinityFilter', 'SameHostFilter', 'DifferentHostFilter'] {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.035867] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] filter_scheduler.host_subset_size = 1 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.036044] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] filter_scheduler.hypervisor_version_weight_multiplier = 1.0 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.036211] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] filter_scheduler.image_properties_default_architecture = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.036377] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] filter_scheduler.io_ops_weight_multiplier = -1.0 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.036540] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] filter_scheduler.isolated_hosts = [] {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.036782] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] filter_scheduler.isolated_images = [] {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.036972] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] filter_scheduler.max_instances_per_host = 50 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.037155] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] filter_scheduler.max_io_ops_per_host = 8 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.037324] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] filter_scheduler.num_instances_weight_multiplier = 0.0 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.037490] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] filter_scheduler.pci_in_placement = False {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.037657] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] filter_scheduler.pci_weight_multiplier = 1.0 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.037825] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] filter_scheduler.ram_weight_multiplier = 1.0 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.037987] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] filter_scheduler.restrict_isolated_hosts_to_isolated_images = True {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.038165] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] filter_scheduler.shuffle_best_same_weighed_hosts = False {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.038330] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] filter_scheduler.soft_affinity_weight_multiplier = 1.0 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.038502] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] filter_scheduler.soft_anti_affinity_weight_multiplier = 1.0 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.038668] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] filter_scheduler.track_instance_changes = True {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.038845] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] filter_scheduler.weight_classes = ['nova.scheduler.weights.all_weighers'] {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.039025] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] metrics.required = True {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.039199] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] metrics.weight_multiplier = 1.0 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.039364] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] metrics.weight_of_unavailable = -10000.0 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.039532] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] metrics.weight_setting = [] {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.039848] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] serial_console.base_url = ws://127.0.0.1:6083/ {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.040035] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] serial_console.enabled = False {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.040220] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] serial_console.port_range = 10000:20000 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.040424] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] serial_console.proxyclient_address = 127.0.0.1 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.040601] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] serial_console.serialproxy_host = 0.0.0.0 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.040767] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] serial_console.serialproxy_port = 6083 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.040933] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] service_user.auth_section = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.041123] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] service_user.auth_type = password {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.041291] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] service_user.cafile = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.041450] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] service_user.certfile = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.041614] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] service_user.collect_timing = False {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.041773] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] service_user.insecure = False {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.041930] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] service_user.keyfile = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.042149] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] service_user.send_service_user_token = True {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.042321] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] service_user.split_loggers = False {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.042483] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] service_user.timeout = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.042653] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] spice.agent_enabled = True {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.042816] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] spice.enabled = False {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.043187] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] spice.html5proxy_base_url = http://127.0.0.1:6082/spice_auto.html {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.043396] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] spice.html5proxy_host = 0.0.0.0 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.043569] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] spice.html5proxy_port = 6082 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.043732] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] spice.image_compression = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.043892] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] spice.jpeg_compression = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.044063] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] spice.playback_compression = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.044231] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] spice.require_secure = False {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.044464] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] spice.server_listen = 127.0.0.1 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.044664] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] spice.server_proxyclient_address = 127.0.0.1 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.044830] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] spice.streaming_mode = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.044986] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] spice.zlib_compression = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.045193] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] upgrade_levels.baseapi = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.045383] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] upgrade_levels.compute = auto {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.045547] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] upgrade_levels.conductor = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.045710] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] upgrade_levels.scheduler = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.045876] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] vendordata_dynamic_auth.auth_section = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.046050] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] vendordata_dynamic_auth.auth_type = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.046219] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] vendordata_dynamic_auth.cafile = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.046374] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] vendordata_dynamic_auth.certfile = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.046533] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] vendordata_dynamic_auth.collect_timing = False {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.046689] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] vendordata_dynamic_auth.insecure = False {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.046928] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] vendordata_dynamic_auth.keyfile = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.047125] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] vendordata_dynamic_auth.split_loggers = False {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.047295] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] vendordata_dynamic_auth.timeout = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.047472] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] vmware.api_retry_count = 10 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.047634] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] vmware.ca_file = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.047804] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] vmware.cache_prefix = devstack-image-cache {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.047971] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] vmware.cluster_name = testcl1 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.048155] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] vmware.connection_pool_size = 10 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.048315] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] vmware.console_delay_seconds = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.048480] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] vmware.datastore_regex = ^datastore.* {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.048688] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] vmware.host_ip = vc1.osci.c.eu-de-1.cloud.sap {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.048864] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] vmware.host_password = **** {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.049042] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] vmware.host_port = 443 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.049218] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] vmware.host_username = administrator@vsphere.local {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.049387] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] vmware.insecure = True {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.049548] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] vmware.integration_bridge = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.049711] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] vmware.maximum_objects = 100 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.049868] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] vmware.pbm_default_policy = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.050040] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] vmware.pbm_enabled = False {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.050205] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] vmware.pbm_wsdl_location = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.050372] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] vmware.serial_log_dir = /opt/vmware/vspc {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.050530] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] vmware.serial_port_proxy_uri = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.050685] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] vmware.serial_port_service_uri = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.050851] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] vmware.task_poll_interval = 0.5 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.051033] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] vmware.use_linked_clone = False {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.051210] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] vmware.vnc_keymap = en-us {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.051376] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] vmware.vnc_port = 5900 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.051537] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] vmware.vnc_port_total = 10000 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.051722] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] vnc.auth_schemes = ['none'] {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.051900] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] vnc.enabled = False {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.052238] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] vnc.novncproxy_base_url = http://127.0.0.1:6080/vnc_auto.html {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.052489] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] vnc.novncproxy_host = 0.0.0.0 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.052708] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] vnc.novncproxy_port = 6080 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.052860] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] vnc.server_listen = 127.0.0.1 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.053051] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] vnc.server_proxyclient_address = 127.0.0.1 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.053257] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] vnc.vencrypt_ca_certs = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.053423] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] vnc.vencrypt_client_cert = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.053586] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] vnc.vencrypt_client_key = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.053768] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] workarounds.disable_compute_service_check_for_ffu = False {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.053933] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] workarounds.disable_deep_image_inspection = False {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.054110] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] workarounds.disable_fallback_pcpu_query = False {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.054279] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] workarounds.disable_group_policy_check_upcall = False {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.054439] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] workarounds.disable_libvirt_livesnapshot = False {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.054689] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] workarounds.disable_rootwrap = False {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.054861] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] workarounds.enable_numa_live_migration = False {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.055044] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] workarounds.enable_qemu_monitor_announce_self = False {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.055217] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] workarounds.ensure_libvirt_rbd_instance_dir_cleanup = False {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.055380] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] workarounds.handle_virt_lifecycle_events = True {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.055540] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] workarounds.libvirt_disable_apic = False {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.055699] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] workarounds.never_download_image_if_on_rbd = False {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.055861] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] workarounds.qemu_monitor_announce_self_count = 3 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.056034] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] workarounds.qemu_monitor_announce_self_interval = 1 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.056199] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] workarounds.reserve_disk_resource_for_image_cache = False {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.056360] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] workarounds.skip_cpu_compare_at_startup = False {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.056521] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] workarounds.skip_cpu_compare_on_dest = False {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.056682] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] workarounds.skip_hypervisor_version_check_on_lm = False {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.056839] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] workarounds.skip_reserve_in_use_ironic_nodes = False {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.057102] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] workarounds.unified_limits_count_pcpu_as_vcpu = False {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.057312] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] workarounds.wait_for_vif_plugged_event_during_hard_reboot = [] {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.057509] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] wsgi.api_paste_config = /etc/nova/api-paste.ini {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.057682] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] wsgi.client_socket_timeout = 900 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.057848] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] wsgi.default_pool_size = 1000 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.058027] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] wsgi.keep_alive = True {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.058196] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] wsgi.max_header_line = 16384 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.058364] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] wsgi.secure_proxy_ssl_header = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.058526] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] wsgi.ssl_ca_file = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.058688] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] wsgi.ssl_cert_file = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.058848] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] wsgi.ssl_key_file = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.059018] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] wsgi.tcp_keepidle = 600 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.059200] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] wsgi.wsgi_log_format = %(client_ip)s "%(request_line)s" status: %(status_code)s len: %(body_length)s time: %(wall_seconds).7f {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.059368] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] zvm.ca_file = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.059529] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] zvm.cloud_connector_url = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.059814] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] zvm.image_tmp_path = /opt/stack/data/n-cpu-1/images {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.059992] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] zvm.reachable_timeout = 300 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.060193] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] oslo_policy.enforce_new_defaults = True {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.060575] env[63515]: WARNING oslo_config.cfg [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] Deprecated: Option "enforce_scope" from group "oslo_policy" is deprecated for removal (This configuration was added temporarily to facilitate a smooth transition to the new RBAC. OpenStack will always enforce scope checks. This configuration option is deprecated and will be removed in the 2025.2 cycle.). Its value may be silently ignored in the future. [ 478.060759] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] oslo_policy.enforce_scope = True {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.060937] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] oslo_policy.policy_default_rule = default {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.061131] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] oslo_policy.policy_dirs = ['policy.d'] {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.061311] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] oslo_policy.policy_file = policy.yaml {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.061488] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] oslo_policy.remote_content_type = application/x-www-form-urlencoded {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.061653] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] oslo_policy.remote_ssl_ca_crt_file = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.061811] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] oslo_policy.remote_ssl_client_crt_file = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.062017] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] oslo_policy.remote_ssl_client_key_file = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.062180] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] oslo_policy.remote_ssl_verify_server_crt = False {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.062356] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] oslo_versionedobjects.fatal_exception_format_errors = False {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.062534] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] oslo_middleware.http_basic_auth_user_file = /etc/htpasswd {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.062711] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] profiler.connection_string = messaging:// {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.062879] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] profiler.enabled = False {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.063066] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] profiler.es_doc_type = notification {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.063252] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] profiler.es_scroll_size = 10000 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.063423] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] profiler.es_scroll_time = 2m {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.063586] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] profiler.filter_error_trace = False {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.063753] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] profiler.hmac_keys = **** {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.063918] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] profiler.sentinel_service_name = mymaster {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.064098] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] profiler.socket_timeout = 0.1 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.064266] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] profiler.trace_requests = False {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.064426] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] profiler.trace_sqlalchemy = False {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.064671] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] profiler_jaeger.process_tags = {} {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.064857] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] profiler_jaeger.service_name_prefix = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.065031] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] profiler_otlp.service_name_prefix = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.065206] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] remote_debug.host = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.065368] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] remote_debug.port = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.065548] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] oslo_messaging_rabbit.amqp_auto_delete = False {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.065710] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] oslo_messaging_rabbit.amqp_durable_queues = False {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.065874] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] oslo_messaging_rabbit.conn_pool_min_size = 2 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.066045] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] oslo_messaging_rabbit.conn_pool_ttl = 1200 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.066214] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] oslo_messaging_rabbit.direct_mandatory_flag = True {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.066373] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] oslo_messaging_rabbit.enable_cancel_on_failover = False {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.066533] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] oslo_messaging_rabbit.heartbeat_in_pthread = False {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.066694] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] oslo_messaging_rabbit.heartbeat_rate = 3 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.066854] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] oslo_messaging_rabbit.heartbeat_timeout_threshold = 60 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.067120] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] oslo_messaging_rabbit.hostname = devstack {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.067311] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] oslo_messaging_rabbit.kombu_compression = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.067485] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] oslo_messaging_rabbit.kombu_failover_strategy = round-robin {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.067654] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] oslo_messaging_rabbit.kombu_missing_consumer_retry_timeout = 60 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.067824] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] oslo_messaging_rabbit.kombu_reconnect_delay = 1.0 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.067994] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] oslo_messaging_rabbit.processname = nova-compute {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.068176] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] oslo_messaging_rabbit.rabbit_ha_queues = False {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.068341] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] oslo_messaging_rabbit.rabbit_interval_max = 30 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.068510] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] oslo_messaging_rabbit.rabbit_login_method = AMQPLAIN {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.068672] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] oslo_messaging_rabbit.rabbit_qos_prefetch_count = 0 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.068830] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] oslo_messaging_rabbit.rabbit_quorum_delivery_limit = 0 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.068993] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] oslo_messaging_rabbit.rabbit_quorum_max_memory_bytes = 0 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.069175] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] oslo_messaging_rabbit.rabbit_quorum_max_memory_length = 0 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.069366] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] oslo_messaging_rabbit.rabbit_quorum_queue = False {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.069537] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] oslo_messaging_rabbit.rabbit_retry_backoff = 2 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.069699] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] oslo_messaging_rabbit.rabbit_retry_interval = 1 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.069860] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] oslo_messaging_rabbit.rabbit_stream_fanout = False {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.070029] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] oslo_messaging_rabbit.rabbit_transient_queues_ttl = 1800 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.070193] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] oslo_messaging_rabbit.rabbit_transient_quorum_queue = False {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.070362] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] oslo_messaging_rabbit.rpc_conn_pool_size = 30 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.070527] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] oslo_messaging_rabbit.ssl = False {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.070699] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] oslo_messaging_rabbit.ssl_ca_file = {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.070869] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] oslo_messaging_rabbit.ssl_cert_file = {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.071040] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] oslo_messaging_rabbit.ssl_enforce_fips_mode = False {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.071217] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] oslo_messaging_rabbit.ssl_key_file = {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.071389] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] oslo_messaging_rabbit.ssl_version = {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.071547] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] oslo_messaging_rabbit.use_queue_manager = False {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.071736] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] oslo_messaging_notifications.driver = ['messagingv2'] {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.071904] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] oslo_messaging_notifications.retry = -1 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.072128] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] oslo_messaging_notifications.topics = ['notifications'] {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.072318] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] oslo_messaging_notifications.transport_url = **** {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.072529] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] oslo_limit.auth_section = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.072712] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] oslo_limit.auth_type = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.072897] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] oslo_limit.cafile = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.073045] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] oslo_limit.certfile = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.073245] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] oslo_limit.collect_timing = False {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.073411] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] oslo_limit.connect_retries = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.073573] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] oslo_limit.connect_retry_delay = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.073732] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] oslo_limit.endpoint_id = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.073889] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] oslo_limit.endpoint_override = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.074062] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] oslo_limit.insecure = False {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.074226] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] oslo_limit.keyfile = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.074382] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] oslo_limit.max_version = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.074536] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] oslo_limit.min_version = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.074757] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] oslo_limit.region_name = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.074945] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] oslo_limit.retriable_status_codes = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.075124] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] oslo_limit.service_name = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.075286] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] oslo_limit.service_type = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.075448] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] oslo_limit.split_loggers = False {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.075610] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] oslo_limit.status_code_retries = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.075767] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] oslo_limit.status_code_retry_delay = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.075923] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] oslo_limit.timeout = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.076091] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] oslo_limit.valid_interfaces = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.076256] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] oslo_limit.version = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.076419] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] oslo_reports.file_event_handler = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.076585] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] oslo_reports.file_event_handler_interval = 1 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.076744] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] oslo_reports.log_dir = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.076915] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] vif_plug_linux_bridge_privileged.capabilities = [12] {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.077084] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] vif_plug_linux_bridge_privileged.group = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.077342] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] vif_plug_linux_bridge_privileged.helper_command = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.077518] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] vif_plug_linux_bridge_privileged.logger_name = oslo_privsep.daemon {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.077686] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] vif_plug_linux_bridge_privileged.thread_pool_size = 8 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.077848] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] vif_plug_linux_bridge_privileged.user = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.078028] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] vif_plug_ovs_privileged.capabilities = [12, 1] {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.078197] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] vif_plug_ovs_privileged.group = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.078356] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] vif_plug_ovs_privileged.helper_command = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.078518] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] vif_plug_ovs_privileged.logger_name = oslo_privsep.daemon {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.078678] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] vif_plug_ovs_privileged.thread_pool_size = 8 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.078836] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] vif_plug_ovs_privileged.user = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.079011] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] os_vif_linux_bridge.flat_interface = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.079202] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] os_vif_linux_bridge.forward_bridge_interface = ['all'] {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.079374] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] os_vif_linux_bridge.iptables_bottom_regex = {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.079543] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] os_vif_linux_bridge.iptables_drop_action = DROP {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.079714] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] os_vif_linux_bridge.iptables_top_regex = {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.079879] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] os_vif_linux_bridge.network_device_mtu = 1500 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.080054] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] os_vif_linux_bridge.use_ipv6 = False {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.080223] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] os_vif_linux_bridge.vlan_interface = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.080413] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] os_vif_ovs.default_qos_type = linux-noop {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.080605] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] os_vif_ovs.isolate_vif = False {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.080831] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] os_vif_ovs.network_device_mtu = 1500 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.081016] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] os_vif_ovs.ovs_vsctl_timeout = 120 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.081196] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] os_vif_ovs.ovsdb_connection = tcp:127.0.0.1:6640 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.081403] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] os_vif_ovs.ovsdb_interface = native {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.081576] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] os_vif_ovs.per_port_bridge = False {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.081743] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] os_brick.lock_path = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.081910] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] os_brick.wait_mpath_device_attempts = 4 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.082114] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] os_brick.wait_mpath_device_interval = 1 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.082296] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] privsep_osbrick.capabilities = [21] {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.082461] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] privsep_osbrick.group = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.082617] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] privsep_osbrick.helper_command = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.082783] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] privsep_osbrick.logger_name = os_brick.privileged {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.082993] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] privsep_osbrick.thread_pool_size = 8 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.083115] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] privsep_osbrick.user = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.083291] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] nova_sys_admin.capabilities = [0, 1, 2, 3, 12, 21] {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.083451] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] nova_sys_admin.group = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.083607] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] nova_sys_admin.helper_command = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.083770] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] nova_sys_admin.logger_name = oslo_privsep.daemon {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.083934] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] nova_sys_admin.thread_pool_size = 8 {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.084099] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] nova_sys_admin.user = None {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 478.084234] env[63515]: DEBUG oslo_service.service [None req-3c4d1450-7990-4369-9037-1b1f3486b198 None None] ******************************************************************************** {{(pid=63515) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2830}} [ 478.084753] env[63515]: INFO nova.service [-] Starting compute node (version 0.0.1) [ 478.587701] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-f5b45b22-e349-44b3-b0d0-3cd006f760a1 None None] Getting list of instances from cluster (obj){ [ 478.587701] env[63515]: value = "domain-c8" [ 478.587701] env[63515]: _type = "ClusterComputeResource" [ 478.587701] env[63515]: } {{(pid=63515) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2116}} [ 478.588990] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-389aa093-11b6-4601-8005-4b33c6edea8e {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 478.598188] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-f5b45b22-e349-44b3-b0d0-3cd006f760a1 None None] Got total of 0 instances {{(pid=63515) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2125}} [ 478.598724] env[63515]: WARNING nova.virt.vmwareapi.driver [None req-f5b45b22-e349-44b3-b0d0-3cd006f760a1 None None] The vmwareapi driver is not tested by the OpenStack project nor does it have clear maintainer(s) and thus its quality can not be ensured. It should be considered experimental and may be removed in a future release. If you are using the driver in production please let us know via the openstack-discuss mailing list. [ 478.599206] env[63515]: INFO nova.virt.node [None req-f5b45b22-e349-44b3-b0d0-3cd006f760a1 None None] Generated node identity 2bd86232-2b6f-44d5-9057-1a3a6b27185a [ 478.599521] env[63515]: INFO nova.virt.node [None req-f5b45b22-e349-44b3-b0d0-3cd006f760a1 None None] Wrote node identity 2bd86232-2b6f-44d5-9057-1a3a6b27185a to /opt/stack/data/n-cpu-1/compute_id [ 479.102267] env[63515]: WARNING nova.compute.manager [None req-f5b45b22-e349-44b3-b0d0-3cd006f760a1 None None] Compute nodes ['2bd86232-2b6f-44d5-9057-1a3a6b27185a'] for host cpu-1 were not found in the database. If this is the first time this service is starting on this host, then you can ignore this warning. [ 480.107367] env[63515]: INFO nova.compute.manager [None req-f5b45b22-e349-44b3-b0d0-3cd006f760a1 None None] Looking for unclaimed instances stuck in BUILDING status for nodes managed by this host [ 481.112846] env[63515]: WARNING nova.compute.manager [None req-f5b45b22-e349-44b3-b0d0-3cd006f760a1 None None] No compute node record found for host cpu-1. If this is the first time this service is starting on this host, then you can ignore this warning.: nova.exception_Remote.ComputeHostNotFound_Remote: Compute host cpu-1 could not be found. [ 481.113222] env[63515]: DEBUG oslo_concurrency.lockutils [None req-f5b45b22-e349-44b3-b0d0-3cd006f760a1 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 481.113442] env[63515]: DEBUG oslo_concurrency.lockutils [None req-f5b45b22-e349-44b3-b0d0-3cd006f760a1 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 481.113603] env[63515]: DEBUG oslo_concurrency.lockutils [None req-f5b45b22-e349-44b3-b0d0-3cd006f760a1 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 481.113762] env[63515]: DEBUG nova.compute.resource_tracker [None req-f5b45b22-e349-44b3-b0d0-3cd006f760a1 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=63515) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 481.114704] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f536e0e0-bbd5-48b5-90a1-87d7146124fa {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 481.123093] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7241c329-4050-4e95-a4b1-52e4892c4fa9 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 481.136449] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba14faab-6668-44ed-8aeb-a111520fab9c {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 481.143447] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1860e91-5cdc-4c8b-97fd-78eac12dccc3 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 481.173190] env[63515]: DEBUG nova.compute.resource_tracker [None req-f5b45b22-e349-44b3-b0d0-3cd006f760a1 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181580MB free_disk=172GB free_vcpus=48 pci_devices=None {{(pid=63515) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 481.173352] env[63515]: DEBUG oslo_concurrency.lockutils [None req-f5b45b22-e349-44b3-b0d0-3cd006f760a1 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 481.173539] env[63515]: DEBUG oslo_concurrency.lockutils [None req-f5b45b22-e349-44b3-b0d0-3cd006f760a1 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 481.676609] env[63515]: WARNING nova.compute.resource_tracker [None req-f5b45b22-e349-44b3-b0d0-3cd006f760a1 None None] No compute node record for cpu-1:2bd86232-2b6f-44d5-9057-1a3a6b27185a: nova.exception_Remote.ComputeHostNotFound_Remote: Compute host 2bd86232-2b6f-44d5-9057-1a3a6b27185a could not be found. [ 482.681801] env[63515]: INFO nova.compute.resource_tracker [None req-f5b45b22-e349-44b3-b0d0-3cd006f760a1 None None] Compute node record created for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 with uuid: 2bd86232-2b6f-44d5-9057-1a3a6b27185a [ 484.190827] env[63515]: DEBUG nova.compute.resource_tracker [None req-f5b45b22-e349-44b3-b0d0-3cd006f760a1 None None] Total usable vcpus: 48, total allocated vcpus: 0 {{(pid=63515) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 484.191046] env[63515]: DEBUG nova.compute.resource_tracker [None req-f5b45b22-e349-44b3-b0d0-3cd006f760a1 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=512MB phys_disk=200GB used_disk=0GB total_vcpus=48 used_vcpus=0 pci_stats=[] {{(pid=63515) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 484.342477] env[63515]: INFO nova.scheduler.client.report [None req-f5b45b22-e349-44b3-b0d0-3cd006f760a1 None None] [req-9a26c7b9-5495-4cb2-9d95-75b72998edb0] Created resource provider record via placement API for resource provider with UUID 2bd86232-2b6f-44d5-9057-1a3a6b27185a and name domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28. [ 484.373910] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d37bf5f-e727-49a5-beca-cef4e8db6859 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 484.381751] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cfdb6b34-cfe8-4dfe-9dde-39021caef80c {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 484.411485] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-617ab69c-9e94-4584-a023-16d51416002f {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 484.418699] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3729bdfe-f6df-4395-9d09-6f41afdf545a {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 484.431229] env[63515]: DEBUG nova.compute.provider_tree [None req-f5b45b22-e349-44b3-b0d0-3cd006f760a1 None None] Updating inventory in ProviderTree for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 172, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 487.891418] env[63515]: DEBUG nova.scheduler.client.report [None req-f5b45b22-e349-44b3-b0d0-3cd006f760a1 None None] Updated inventory for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a with generation 0 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 172, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 487.891730] env[63515]: DEBUG nova.compute.provider_tree [None req-f5b45b22-e349-44b3-b0d0-3cd006f760a1 None None] Updating resource provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a generation from 0 to 1 during operation: update_inventory {{(pid=63515) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 487.891938] env[63515]: DEBUG nova.compute.provider_tree [None req-f5b45b22-e349-44b3-b0d0-3cd006f760a1 None None] Updating inventory in ProviderTree for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 172, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 487.986050] env[63515]: DEBUG nova.compute.provider_tree [None req-f5b45b22-e349-44b3-b0d0-3cd006f760a1 None None] Updating resource provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a generation from 1 to 2 during operation: update_traits {{(pid=63515) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 488.490979] env[63515]: DEBUG nova.compute.resource_tracker [None req-f5b45b22-e349-44b3-b0d0-3cd006f760a1 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=63515) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 488.491263] env[63515]: DEBUG oslo_concurrency.lockutils [None req-f5b45b22-e349-44b3-b0d0-3cd006f760a1 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 7.318s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 488.491379] env[63515]: DEBUG nova.service [None req-f5b45b22-e349-44b3-b0d0-3cd006f760a1 None None] Creating RPC server for service compute {{(pid=63515) start /opt/stack/nova/nova/service.py:186}} [ 488.504307] env[63515]: DEBUG nova.service [None req-f5b45b22-e349-44b3-b0d0-3cd006f760a1 None None] Join ServiceGroup membership for this service compute {{(pid=63515) start /opt/stack/nova/nova/service.py:203}} [ 488.504511] env[63515]: DEBUG nova.servicegroup.drivers.db [None req-f5b45b22-e349-44b3-b0d0-3cd006f760a1 None None] DB_Driver: join new ServiceGroup member cpu-1 to the compute group, service = {{(pid=63515) join /opt/stack/nova/nova/servicegroup/drivers/db.py:44}} [ 492.505999] env[63515]: DEBUG oslo_service.periodic_task [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Running periodic task ComputeManager._sync_power_states {{(pid=63515) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 493.009192] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Getting list of instances from cluster (obj){ [ 493.009192] env[63515]: value = "domain-c8" [ 493.009192] env[63515]: _type = "ClusterComputeResource" [ 493.009192] env[63515]: } {{(pid=63515) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2116}} [ 493.010406] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4021a1b5-e3df-4ede-8f93-6879f31ce18a {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 493.019131] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Got total of 0 instances {{(pid=63515) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2125}} [ 493.019356] env[63515]: DEBUG oslo_service.periodic_task [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Running periodic task ComputeManager._cleanup_running_deleted_instances {{(pid=63515) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 493.019647] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Getting list of instances from cluster (obj){ [ 493.019647] env[63515]: value = "domain-c8" [ 493.019647] env[63515]: _type = "ClusterComputeResource" [ 493.019647] env[63515]: } {{(pid=63515) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2116}} [ 493.020531] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90ba84e7-32f1-4456-9459-1ed0709c10fe {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 493.027530] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Got total of 0 instances {{(pid=63515) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2125}} [ 522.583704] env[63515]: DEBUG oslo_concurrency.lockutils [None req-9ee35a7a-94b5-4758-9c8f-a8f1e8f2f2cd tempest-DeleteServersAdminTestJSON-944542165 tempest-DeleteServersAdminTestJSON-944542165-project-member] Acquiring lock "d7fb514e-4ea9-4f32-9b00-06f728813aa3" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 522.584012] env[63515]: DEBUG oslo_concurrency.lockutils [None req-9ee35a7a-94b5-4758-9c8f-a8f1e8f2f2cd tempest-DeleteServersAdminTestJSON-944542165 tempest-DeleteServersAdminTestJSON-944542165-project-member] Lock "d7fb514e-4ea9-4f32-9b00-06f728813aa3" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 523.091377] env[63515]: DEBUG nova.compute.manager [None req-9ee35a7a-94b5-4758-9c8f-a8f1e8f2f2cd tempest-DeleteServersAdminTestJSON-944542165 tempest-DeleteServersAdminTestJSON-944542165-project-member] [instance: d7fb514e-4ea9-4f32-9b00-06f728813aa3] Starting instance... {{(pid=63515) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 523.631251] env[63515]: DEBUG oslo_concurrency.lockutils [None req-9ee35a7a-94b5-4758-9c8f-a8f1e8f2f2cd tempest-DeleteServersAdminTestJSON-944542165 tempest-DeleteServersAdminTestJSON-944542165-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 523.631567] env[63515]: DEBUG oslo_concurrency.lockutils [None req-9ee35a7a-94b5-4758-9c8f-a8f1e8f2f2cd tempest-DeleteServersAdminTestJSON-944542165 tempest-DeleteServersAdminTestJSON-944542165-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.001s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 523.633391] env[63515]: INFO nova.compute.claims [None req-9ee35a7a-94b5-4758-9c8f-a8f1e8f2f2cd tempest-DeleteServersAdminTestJSON-944542165 tempest-DeleteServersAdminTestJSON-944542165-project-member] [instance: d7fb514e-4ea9-4f32-9b00-06f728813aa3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 524.684589] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1cb49fef-5690-48e2-ad8a-91bb69e5f3b9 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 524.693384] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63eb6e38-6c87-4970-99bc-3fd7401bb3fa {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 524.728235] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf3f1819-e4ee-432c-99ee-09012906dc8f {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 524.736447] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1729530-14b6-40f2-93a3-73335fc3a455 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 524.750876] env[63515]: DEBUG nova.compute.provider_tree [None req-9ee35a7a-94b5-4758-9c8f-a8f1e8f2f2cd tempest-DeleteServersAdminTestJSON-944542165 tempest-DeleteServersAdminTestJSON-944542165-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 525.253878] env[63515]: DEBUG nova.scheduler.client.report [None req-9ee35a7a-94b5-4758-9c8f-a8f1e8f2f2cd tempest-DeleteServersAdminTestJSON-944542165 tempest-DeleteServersAdminTestJSON-944542165-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 172, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 525.764037] env[63515]: DEBUG oslo_concurrency.lockutils [None req-9ee35a7a-94b5-4758-9c8f-a8f1e8f2f2cd tempest-DeleteServersAdminTestJSON-944542165 tempest-DeleteServersAdminTestJSON-944542165-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.132s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 525.764671] env[63515]: DEBUG nova.compute.manager [None req-9ee35a7a-94b5-4758-9c8f-a8f1e8f2f2cd tempest-DeleteServersAdminTestJSON-944542165 tempest-DeleteServersAdminTestJSON-944542165-project-member] [instance: d7fb514e-4ea9-4f32-9b00-06f728813aa3] Start building networks asynchronously for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 526.141970] env[63515]: DEBUG oslo_concurrency.lockutils [None req-0d00ff28-5b1e-46b5-9fb9-df8b8bcd6a1a tempest-TenantUsagesTestJSON-427870877 tempest-TenantUsagesTestJSON-427870877-project-member] Acquiring lock "9f7ad2dd-b882-4d98-8576-708b3abfa505" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 526.142266] env[63515]: DEBUG oslo_concurrency.lockutils [None req-0d00ff28-5b1e-46b5-9fb9-df8b8bcd6a1a tempest-TenantUsagesTestJSON-427870877 tempest-TenantUsagesTestJSON-427870877-project-member] Lock "9f7ad2dd-b882-4d98-8576-708b3abfa505" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 526.272370] env[63515]: DEBUG nova.compute.utils [None req-9ee35a7a-94b5-4758-9c8f-a8f1e8f2f2cd tempest-DeleteServersAdminTestJSON-944542165 tempest-DeleteServersAdminTestJSON-944542165-project-member] Using /dev/sd instead of None {{(pid=63515) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 526.274654] env[63515]: DEBUG nova.compute.manager [None req-9ee35a7a-94b5-4758-9c8f-a8f1e8f2f2cd tempest-DeleteServersAdminTestJSON-944542165 tempest-DeleteServersAdminTestJSON-944542165-project-member] [instance: d7fb514e-4ea9-4f32-9b00-06f728813aa3] Allocating IP information in the background. {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 526.275277] env[63515]: DEBUG nova.network.neutron [None req-9ee35a7a-94b5-4758-9c8f-a8f1e8f2f2cd tempest-DeleteServersAdminTestJSON-944542165 tempest-DeleteServersAdminTestJSON-944542165-project-member] [instance: d7fb514e-4ea9-4f32-9b00-06f728813aa3] allocate_for_instance() {{(pid=63515) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 526.554238] env[63515]: DEBUG nova.policy [None req-9ee35a7a-94b5-4758-9c8f-a8f1e8f2f2cd tempest-DeleteServersAdminTestJSON-944542165 tempest-DeleteServersAdminTestJSON-944542165-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '664c0b774cd64a8d9d798e724c206fe4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '48d812be60fd41179b1d5d43c39c335b', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63515) authorize /opt/stack/nova/nova/policy.py:201}} [ 526.645273] env[63515]: DEBUG nova.compute.manager [None req-0d00ff28-5b1e-46b5-9fb9-df8b8bcd6a1a tempest-TenantUsagesTestJSON-427870877 tempest-TenantUsagesTestJSON-427870877-project-member] [instance: 9f7ad2dd-b882-4d98-8576-708b3abfa505] Starting instance... {{(pid=63515) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 526.783233] env[63515]: DEBUG nova.compute.manager [None req-9ee35a7a-94b5-4758-9c8f-a8f1e8f2f2cd tempest-DeleteServersAdminTestJSON-944542165 tempest-DeleteServersAdminTestJSON-944542165-project-member] [instance: d7fb514e-4ea9-4f32-9b00-06f728813aa3] Start building block device mappings for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 527.171947] env[63515]: DEBUG oslo_concurrency.lockutils [None req-0d00ff28-5b1e-46b5-9fb9-df8b8bcd6a1a tempest-TenantUsagesTestJSON-427870877 tempest-TenantUsagesTestJSON-427870877-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 527.172226] env[63515]: DEBUG oslo_concurrency.lockutils [None req-0d00ff28-5b1e-46b5-9fb9-df8b8bcd6a1a tempest-TenantUsagesTestJSON-427870877 tempest-TenantUsagesTestJSON-427870877-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 527.173888] env[63515]: INFO nova.compute.claims [None req-0d00ff28-5b1e-46b5-9fb9-df8b8bcd6a1a tempest-TenantUsagesTestJSON-427870877 tempest-TenantUsagesTestJSON-427870877-project-member] [instance: 9f7ad2dd-b882-4d98-8576-708b3abfa505] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 527.533755] env[63515]: DEBUG oslo_concurrency.lockutils [None req-919af50b-07d1-4794-942f-d99fa56afe84 tempest-ServerExternalEventsTest-1976219582 tempest-ServerExternalEventsTest-1976219582-project-member] Acquiring lock "6662da99-ba90-49ac-982e-d18f1c7834b6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 527.534553] env[63515]: DEBUG oslo_concurrency.lockutils [None req-919af50b-07d1-4794-942f-d99fa56afe84 tempest-ServerExternalEventsTest-1976219582 tempest-ServerExternalEventsTest-1976219582-project-member] Lock "6662da99-ba90-49ac-982e-d18f1c7834b6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 527.798824] env[63515]: DEBUG nova.compute.manager [None req-9ee35a7a-94b5-4758-9c8f-a8f1e8f2f2cd tempest-DeleteServersAdminTestJSON-944542165 tempest-DeleteServersAdminTestJSON-944542165-project-member] [instance: d7fb514e-4ea9-4f32-9b00-06f728813aa3] Start spawning the instance on the hypervisor. {{(pid=63515) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 527.842085] env[63515]: DEBUG nova.virt.hardware [None req-9ee35a7a-94b5-4758-9c8f-a8f1e8f2f2cd tempest-DeleteServersAdminTestJSON-944542165 tempest-DeleteServersAdminTestJSON-944542165-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T02:52:14Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T02:51:56Z,direct_url=,disk_format='vmdk',id=8a120570-cb06-4099-b262-554ca0ad15c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b569255cc43e42e7ae3f2b4ad37c6ef4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T02:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 527.842085] env[63515]: DEBUG nova.virt.hardware [None req-9ee35a7a-94b5-4758-9c8f-a8f1e8f2f2cd tempest-DeleteServersAdminTestJSON-944542165 tempest-DeleteServersAdminTestJSON-944542165-project-member] Flavor limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 527.842085] env[63515]: DEBUG nova.virt.hardware [None req-9ee35a7a-94b5-4758-9c8f-a8f1e8f2f2cd tempest-DeleteServersAdminTestJSON-944542165 tempest-DeleteServersAdminTestJSON-944542165-project-member] Image limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 527.843059] env[63515]: DEBUG nova.virt.hardware [None req-9ee35a7a-94b5-4758-9c8f-a8f1e8f2f2cd tempest-DeleteServersAdminTestJSON-944542165 tempest-DeleteServersAdminTestJSON-944542165-project-member] Flavor pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 527.843059] env[63515]: DEBUG nova.virt.hardware [None req-9ee35a7a-94b5-4758-9c8f-a8f1e8f2f2cd tempest-DeleteServersAdminTestJSON-944542165 tempest-DeleteServersAdminTestJSON-944542165-project-member] Image pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 527.843059] env[63515]: DEBUG nova.virt.hardware [None req-9ee35a7a-94b5-4758-9c8f-a8f1e8f2f2cd tempest-DeleteServersAdminTestJSON-944542165 tempest-DeleteServersAdminTestJSON-944542165-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 527.843308] env[63515]: DEBUG nova.virt.hardware [None req-9ee35a7a-94b5-4758-9c8f-a8f1e8f2f2cd tempest-DeleteServersAdminTestJSON-944542165 tempest-DeleteServersAdminTestJSON-944542165-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 527.844030] env[63515]: DEBUG nova.virt.hardware [None req-9ee35a7a-94b5-4758-9c8f-a8f1e8f2f2cd tempest-DeleteServersAdminTestJSON-944542165 tempest-DeleteServersAdminTestJSON-944542165-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 527.844152] env[63515]: DEBUG nova.virt.hardware [None req-9ee35a7a-94b5-4758-9c8f-a8f1e8f2f2cd tempest-DeleteServersAdminTestJSON-944542165 tempest-DeleteServersAdminTestJSON-944542165-project-member] Got 1 possible topologies {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 527.844405] env[63515]: DEBUG nova.virt.hardware [None req-9ee35a7a-94b5-4758-9c8f-a8f1e8f2f2cd tempest-DeleteServersAdminTestJSON-944542165 tempest-DeleteServersAdminTestJSON-944542165-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 527.844632] env[63515]: DEBUG nova.virt.hardware [None req-9ee35a7a-94b5-4758-9c8f-a8f1e8f2f2cd tempest-DeleteServersAdminTestJSON-944542165 tempest-DeleteServersAdminTestJSON-944542165-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 527.847711] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea6dcb5f-fcc7-492c-9cf4-ce1a2517dfc7 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 527.856602] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff5b3526-6c7f-4b4c-9976-58b3f8b04143 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 527.875973] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06a7a30e-834f-45d5-8be0-c802b90d363a {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 528.040908] env[63515]: DEBUG nova.compute.manager [None req-919af50b-07d1-4794-942f-d99fa56afe84 tempest-ServerExternalEventsTest-1976219582 tempest-ServerExternalEventsTest-1976219582-project-member] [instance: 6662da99-ba90-49ac-982e-d18f1c7834b6] Starting instance... {{(pid=63515) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 528.272240] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40e0e51b-c343-4d0a-b8bd-2acae5306fad {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 528.280247] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd5c9d02-1189-4f9e-8f4e-4828ee91ae49 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 528.318526] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10f730fd-5acc-46e6-b049-dc1ed81655bb {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 528.326944] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-826e96a6-dcc7-4c98-ac5f-ee1807b49506 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 528.343672] env[63515]: DEBUG nova.compute.provider_tree [None req-0d00ff28-5b1e-46b5-9fb9-df8b8bcd6a1a tempest-TenantUsagesTestJSON-427870877 tempest-TenantUsagesTestJSON-427870877-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 528.576595] env[63515]: DEBUG oslo_concurrency.lockutils [None req-919af50b-07d1-4794-942f-d99fa56afe84 tempest-ServerExternalEventsTest-1976219582 tempest-ServerExternalEventsTest-1976219582-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 528.631786] env[63515]: DEBUG nova.network.neutron [None req-9ee35a7a-94b5-4758-9c8f-a8f1e8f2f2cd tempest-DeleteServersAdminTestJSON-944542165 tempest-DeleteServersAdminTestJSON-944542165-project-member] [instance: d7fb514e-4ea9-4f32-9b00-06f728813aa3] Successfully created port: 71cd34ac-f70e-44ba-856e-b65afc2b509f {{(pid=63515) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 528.848139] env[63515]: DEBUG nova.scheduler.client.report [None req-0d00ff28-5b1e-46b5-9fb9-df8b8bcd6a1a tempest-TenantUsagesTestJSON-427870877 tempest-TenantUsagesTestJSON-427870877-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 172, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 529.353633] env[63515]: DEBUG oslo_concurrency.lockutils [None req-0d00ff28-5b1e-46b5-9fb9-df8b8bcd6a1a tempest-TenantUsagesTestJSON-427870877 tempest-TenantUsagesTestJSON-427870877-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.181s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 529.354251] env[63515]: DEBUG nova.compute.manager [None req-0d00ff28-5b1e-46b5-9fb9-df8b8bcd6a1a tempest-TenantUsagesTestJSON-427870877 tempest-TenantUsagesTestJSON-427870877-project-member] [instance: 9f7ad2dd-b882-4d98-8576-708b3abfa505] Start building networks asynchronously for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 529.356713] env[63515]: DEBUG oslo_concurrency.lockutils [None req-919af50b-07d1-4794-942f-d99fa56afe84 tempest-ServerExternalEventsTest-1976219582 tempest-ServerExternalEventsTest-1976219582-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.780s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 529.358070] env[63515]: INFO nova.compute.claims [None req-919af50b-07d1-4794-942f-d99fa56afe84 tempest-ServerExternalEventsTest-1976219582 tempest-ServerExternalEventsTest-1976219582-project-member] [instance: 6662da99-ba90-49ac-982e-d18f1c7834b6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 529.865969] env[63515]: DEBUG nova.compute.utils [None req-0d00ff28-5b1e-46b5-9fb9-df8b8bcd6a1a tempest-TenantUsagesTestJSON-427870877 tempest-TenantUsagesTestJSON-427870877-project-member] Using /dev/sd instead of None {{(pid=63515) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 529.867583] env[63515]: DEBUG nova.compute.manager [None req-0d00ff28-5b1e-46b5-9fb9-df8b8bcd6a1a tempest-TenantUsagesTestJSON-427870877 tempest-TenantUsagesTestJSON-427870877-project-member] [instance: 9f7ad2dd-b882-4d98-8576-708b3abfa505] Allocating IP information in the background. {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 529.867830] env[63515]: DEBUG nova.network.neutron [None req-0d00ff28-5b1e-46b5-9fb9-df8b8bcd6a1a tempest-TenantUsagesTestJSON-427870877 tempest-TenantUsagesTestJSON-427870877-project-member] [instance: 9f7ad2dd-b882-4d98-8576-708b3abfa505] allocate_for_instance() {{(pid=63515) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 530.168626] env[63515]: DEBUG nova.policy [None req-0d00ff28-5b1e-46b5-9fb9-df8b8bcd6a1a tempest-TenantUsagesTestJSON-427870877 tempest-TenantUsagesTestJSON-427870877-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '445aeb53a5544856bd2f11ea4e2c102c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e09cd8254ca3426e84ea20c8588d61bd', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63515) authorize /opt/stack/nova/nova/policy.py:201}} [ 530.373780] env[63515]: DEBUG nova.compute.manager [None req-0d00ff28-5b1e-46b5-9fb9-df8b8bcd6a1a tempest-TenantUsagesTestJSON-427870877 tempest-TenantUsagesTestJSON-427870877-project-member] [instance: 9f7ad2dd-b882-4d98-8576-708b3abfa505] Start building block device mappings for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 530.459392] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90b6de75-3ea2-4674-8998-bac711513059 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 530.469447] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68ea43c5-43a7-4db1-b136-c1296761a195 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 530.502350] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc46f359-95a6-4210-97ee-fc4842c1a8ad {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 530.510748] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-535fc37f-b1d6-4187-99ce-3126604ebb58 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 530.526331] env[63515]: DEBUG nova.compute.provider_tree [None req-919af50b-07d1-4794-942f-d99fa56afe84 tempest-ServerExternalEventsTest-1976219582 tempest-ServerExternalEventsTest-1976219582-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 531.030647] env[63515]: DEBUG nova.scheduler.client.report [None req-919af50b-07d1-4794-942f-d99fa56afe84 tempest-ServerExternalEventsTest-1976219582 tempest-ServerExternalEventsTest-1976219582-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 172, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 531.386983] env[63515]: DEBUG nova.compute.manager [None req-0d00ff28-5b1e-46b5-9fb9-df8b8bcd6a1a tempest-TenantUsagesTestJSON-427870877 tempest-TenantUsagesTestJSON-427870877-project-member] [instance: 9f7ad2dd-b882-4d98-8576-708b3abfa505] Start spawning the instance on the hypervisor. {{(pid=63515) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 531.411408] env[63515]: DEBUG nova.virt.hardware [None req-0d00ff28-5b1e-46b5-9fb9-df8b8bcd6a1a tempest-TenantUsagesTestJSON-427870877 tempest-TenantUsagesTestJSON-427870877-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T02:52:14Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T02:51:56Z,direct_url=,disk_format='vmdk',id=8a120570-cb06-4099-b262-554ca0ad15c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b569255cc43e42e7ae3f2b4ad37c6ef4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T02:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 531.411661] env[63515]: DEBUG nova.virt.hardware [None req-0d00ff28-5b1e-46b5-9fb9-df8b8bcd6a1a tempest-TenantUsagesTestJSON-427870877 tempest-TenantUsagesTestJSON-427870877-project-member] Flavor limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 531.411909] env[63515]: DEBUG nova.virt.hardware [None req-0d00ff28-5b1e-46b5-9fb9-df8b8bcd6a1a tempest-TenantUsagesTestJSON-427870877 tempest-TenantUsagesTestJSON-427870877-project-member] Image limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 531.412018] env[63515]: DEBUG nova.virt.hardware [None req-0d00ff28-5b1e-46b5-9fb9-df8b8bcd6a1a tempest-TenantUsagesTestJSON-427870877 tempest-TenantUsagesTestJSON-427870877-project-member] Flavor pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 531.413104] env[63515]: DEBUG nova.virt.hardware [None req-0d00ff28-5b1e-46b5-9fb9-df8b8bcd6a1a tempest-TenantUsagesTestJSON-427870877 tempest-TenantUsagesTestJSON-427870877-project-member] Image pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 531.413283] env[63515]: DEBUG nova.virt.hardware [None req-0d00ff28-5b1e-46b5-9fb9-df8b8bcd6a1a tempest-TenantUsagesTestJSON-427870877 tempest-TenantUsagesTestJSON-427870877-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 531.413510] env[63515]: DEBUG nova.virt.hardware [None req-0d00ff28-5b1e-46b5-9fb9-df8b8bcd6a1a tempest-TenantUsagesTestJSON-427870877 tempest-TenantUsagesTestJSON-427870877-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 531.413717] env[63515]: DEBUG nova.virt.hardware [None req-0d00ff28-5b1e-46b5-9fb9-df8b8bcd6a1a tempest-TenantUsagesTestJSON-427870877 tempest-TenantUsagesTestJSON-427870877-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 531.414069] env[63515]: DEBUG nova.virt.hardware [None req-0d00ff28-5b1e-46b5-9fb9-df8b8bcd6a1a tempest-TenantUsagesTestJSON-427870877 tempest-TenantUsagesTestJSON-427870877-project-member] Got 1 possible topologies {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 531.414069] env[63515]: DEBUG nova.virt.hardware [None req-0d00ff28-5b1e-46b5-9fb9-df8b8bcd6a1a tempest-TenantUsagesTestJSON-427870877 tempest-TenantUsagesTestJSON-427870877-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 531.414249] env[63515]: DEBUG nova.virt.hardware [None req-0d00ff28-5b1e-46b5-9fb9-df8b8bcd6a1a tempest-TenantUsagesTestJSON-427870877 tempest-TenantUsagesTestJSON-427870877-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 531.415256] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20beb337-af69-4c32-9ecb-4fb03e67eb85 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 531.426346] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e353551-2959-4a07-ba07-6652ba071fd9 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 531.536754] env[63515]: DEBUG oslo_concurrency.lockutils [None req-919af50b-07d1-4794-942f-d99fa56afe84 tempest-ServerExternalEventsTest-1976219582 tempest-ServerExternalEventsTest-1976219582-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.180s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 531.537288] env[63515]: DEBUG nova.compute.manager [None req-919af50b-07d1-4794-942f-d99fa56afe84 tempest-ServerExternalEventsTest-1976219582 tempest-ServerExternalEventsTest-1976219582-project-member] [instance: 6662da99-ba90-49ac-982e-d18f1c7834b6] Start building networks asynchronously for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 531.739816] env[63515]: DEBUG nova.network.neutron [None req-0d00ff28-5b1e-46b5-9fb9-df8b8bcd6a1a tempest-TenantUsagesTestJSON-427870877 tempest-TenantUsagesTestJSON-427870877-project-member] [instance: 9f7ad2dd-b882-4d98-8576-708b3abfa505] Successfully created port: 111c82a0-4d47-4505-8e27-6281465572af {{(pid=63515) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 532.047776] env[63515]: DEBUG nova.compute.utils [None req-919af50b-07d1-4794-942f-d99fa56afe84 tempest-ServerExternalEventsTest-1976219582 tempest-ServerExternalEventsTest-1976219582-project-member] Using /dev/sd instead of None {{(pid=63515) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 532.050017] env[63515]: DEBUG nova.compute.manager [None req-919af50b-07d1-4794-942f-d99fa56afe84 tempest-ServerExternalEventsTest-1976219582 tempest-ServerExternalEventsTest-1976219582-project-member] [instance: 6662da99-ba90-49ac-982e-d18f1c7834b6] Allocating IP information in the background. {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 532.050769] env[63515]: DEBUG nova.network.neutron [None req-919af50b-07d1-4794-942f-d99fa56afe84 tempest-ServerExternalEventsTest-1976219582 tempest-ServerExternalEventsTest-1976219582-project-member] [instance: 6662da99-ba90-49ac-982e-d18f1c7834b6] allocate_for_instance() {{(pid=63515) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 532.344057] env[63515]: DEBUG nova.policy [None req-919af50b-07d1-4794-942f-d99fa56afe84 tempest-ServerExternalEventsTest-1976219582 tempest-ServerExternalEventsTest-1976219582-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f3f0ed2463a648a0ae232a9a72c452b0', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd1104895be234ffb89174d1c2e56ca42', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63515) authorize /opt/stack/nova/nova/policy.py:201}} [ 532.555457] env[63515]: DEBUG nova.compute.manager [None req-919af50b-07d1-4794-942f-d99fa56afe84 tempest-ServerExternalEventsTest-1976219582 tempest-ServerExternalEventsTest-1976219582-project-member] [instance: 6662da99-ba90-49ac-982e-d18f1c7834b6] Start building block device mappings for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 532.622991] env[63515]: DEBUG oslo_concurrency.lockutils [None req-c96f19aa-9abc-4b12-99f8-980442cab640 tempest-ServersAdminTestJSON-710082096 tempest-ServersAdminTestJSON-710082096-project-member] Acquiring lock "29d199bc-aff3-4c62-af3a-1e6880b52446" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 532.622991] env[63515]: DEBUG oslo_concurrency.lockutils [None req-c96f19aa-9abc-4b12-99f8-980442cab640 tempest-ServersAdminTestJSON-710082096 tempest-ServersAdminTestJSON-710082096-project-member] Lock "29d199bc-aff3-4c62-af3a-1e6880b52446" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 533.125620] env[63515]: DEBUG nova.compute.manager [None req-c96f19aa-9abc-4b12-99f8-980442cab640 tempest-ServersAdminTestJSON-710082096 tempest-ServersAdminTestJSON-710082096-project-member] [instance: 29d199bc-aff3-4c62-af3a-1e6880b52446] Starting instance... {{(pid=63515) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 533.543397] env[63515]: DEBUG oslo_service.periodic_task [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=63515) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 533.543703] env[63515]: DEBUG oslo_service.periodic_task [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=63515) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 533.543979] env[63515]: DEBUG nova.compute.manager [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Starting heal instance info cache {{(pid=63515) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9926}} [ 533.544077] env[63515]: DEBUG nova.compute.manager [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Rebuilding the list of instances to heal {{(pid=63515) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9930}} [ 533.572266] env[63515]: DEBUG nova.compute.manager [None req-919af50b-07d1-4794-942f-d99fa56afe84 tempest-ServerExternalEventsTest-1976219582 tempest-ServerExternalEventsTest-1976219582-project-member] [instance: 6662da99-ba90-49ac-982e-d18f1c7834b6] Start spawning the instance on the hypervisor. {{(pid=63515) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 533.611016] env[63515]: DEBUG nova.virt.hardware [None req-919af50b-07d1-4794-942f-d99fa56afe84 tempest-ServerExternalEventsTest-1976219582 tempest-ServerExternalEventsTest-1976219582-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T02:52:14Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T02:51:56Z,direct_url=,disk_format='vmdk',id=8a120570-cb06-4099-b262-554ca0ad15c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b569255cc43e42e7ae3f2b4ad37c6ef4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T02:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 533.611016] env[63515]: DEBUG nova.virt.hardware [None req-919af50b-07d1-4794-942f-d99fa56afe84 tempest-ServerExternalEventsTest-1976219582 tempest-ServerExternalEventsTest-1976219582-project-member] Flavor limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 533.611016] env[63515]: DEBUG nova.virt.hardware [None req-919af50b-07d1-4794-942f-d99fa56afe84 tempest-ServerExternalEventsTest-1976219582 tempest-ServerExternalEventsTest-1976219582-project-member] Image limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 533.611375] env[63515]: DEBUG nova.virt.hardware [None req-919af50b-07d1-4794-942f-d99fa56afe84 tempest-ServerExternalEventsTest-1976219582 tempest-ServerExternalEventsTest-1976219582-project-member] Flavor pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 533.611375] env[63515]: DEBUG nova.virt.hardware [None req-919af50b-07d1-4794-942f-d99fa56afe84 tempest-ServerExternalEventsTest-1976219582 tempest-ServerExternalEventsTest-1976219582-project-member] Image pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 533.615022] env[63515]: DEBUG nova.virt.hardware [None req-919af50b-07d1-4794-942f-d99fa56afe84 tempest-ServerExternalEventsTest-1976219582 tempest-ServerExternalEventsTest-1976219582-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 533.615401] env[63515]: DEBUG nova.virt.hardware [None req-919af50b-07d1-4794-942f-d99fa56afe84 tempest-ServerExternalEventsTest-1976219582 tempest-ServerExternalEventsTest-1976219582-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 533.616237] env[63515]: DEBUG nova.virt.hardware [None req-919af50b-07d1-4794-942f-d99fa56afe84 tempest-ServerExternalEventsTest-1976219582 tempest-ServerExternalEventsTest-1976219582-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 533.616549] env[63515]: DEBUG nova.virt.hardware [None req-919af50b-07d1-4794-942f-d99fa56afe84 tempest-ServerExternalEventsTest-1976219582 tempest-ServerExternalEventsTest-1976219582-project-member] Got 1 possible topologies {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 533.616855] env[63515]: DEBUG nova.virt.hardware [None req-919af50b-07d1-4794-942f-d99fa56afe84 tempest-ServerExternalEventsTest-1976219582 tempest-ServerExternalEventsTest-1976219582-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 533.617154] env[63515]: DEBUG nova.virt.hardware [None req-919af50b-07d1-4794-942f-d99fa56afe84 tempest-ServerExternalEventsTest-1976219582 tempest-ServerExternalEventsTest-1976219582-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 533.618164] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6244cf45-cc18-4509-ae40-20bbcb401c41 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 533.627632] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6846e64a-039a-4e9c-934b-6ae16ec97c14 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 533.683795] env[63515]: DEBUG oslo_concurrency.lockutils [None req-c96f19aa-9abc-4b12-99f8-980442cab640 tempest-ServersAdminTestJSON-710082096 tempest-ServersAdminTestJSON-710082096-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 533.683795] env[63515]: DEBUG oslo_concurrency.lockutils [None req-c96f19aa-9abc-4b12-99f8-980442cab640 tempest-ServersAdminTestJSON-710082096 tempest-ServersAdminTestJSON-710082096-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 533.687024] env[63515]: INFO nova.compute.claims [None req-c96f19aa-9abc-4b12-99f8-980442cab640 tempest-ServersAdminTestJSON-710082096 tempest-ServersAdminTestJSON-710082096-project-member] [instance: 29d199bc-aff3-4c62-af3a-1e6880b52446] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 533.981883] env[63515]: DEBUG nova.network.neutron [None req-919af50b-07d1-4794-942f-d99fa56afe84 tempest-ServerExternalEventsTest-1976219582 tempest-ServerExternalEventsTest-1976219582-project-member] [instance: 6662da99-ba90-49ac-982e-d18f1c7834b6] Successfully created port: 39b0bc20-a4b4-4683-ac31-11d4a2fa8333 {{(pid=63515) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 534.048819] env[63515]: DEBUG nova.compute.manager [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] [instance: d7fb514e-4ea9-4f32-9b00-06f728813aa3] Skipping network cache update for instance because it is Building. {{(pid=63515) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 534.048992] env[63515]: DEBUG nova.compute.manager [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] [instance: 9f7ad2dd-b882-4d98-8576-708b3abfa505] Skipping network cache update for instance because it is Building. {{(pid=63515) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 534.049134] env[63515]: DEBUG nova.compute.manager [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] [instance: 6662da99-ba90-49ac-982e-d18f1c7834b6] Skipping network cache update for instance because it is Building. {{(pid=63515) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 534.049355] env[63515]: DEBUG nova.compute.manager [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Didn't find any instances for network info cache update. {{(pid=63515) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10012}} [ 534.049457] env[63515]: DEBUG oslo_service.periodic_task [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=63515) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 534.049693] env[63515]: DEBUG oslo_service.periodic_task [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=63515) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 534.049883] env[63515]: DEBUG oslo_service.periodic_task [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=63515) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 534.050079] env[63515]: DEBUG oslo_service.periodic_task [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=63515) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 534.050269] env[63515]: DEBUG oslo_service.periodic_task [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=63515) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 534.050450] env[63515]: DEBUG oslo_service.periodic_task [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=63515) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 534.050615] env[63515]: DEBUG nova.compute.manager [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=63515) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10545}} [ 534.050896] env[63515]: DEBUG oslo_service.periodic_task [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Running periodic task ComputeManager.update_available_resource {{(pid=63515) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 534.250028] env[63515]: ERROR nova.compute.manager [None req-9ee35a7a-94b5-4758-9c8f-a8f1e8f2f2cd tempest-DeleteServersAdminTestJSON-944542165 tempest-DeleteServersAdminTestJSON-944542165-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 71cd34ac-f70e-44ba-856e-b65afc2b509f, please check neutron logs for more information. [ 534.250028] env[63515]: ERROR nova.compute.manager Traceback (most recent call last): [ 534.250028] env[63515]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 534.250028] env[63515]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 534.250028] env[63515]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 534.250028] env[63515]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 534.250028] env[63515]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 534.250028] env[63515]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 534.250028] env[63515]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 534.250028] env[63515]: ERROR nova.compute.manager self.force_reraise() [ 534.250028] env[63515]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 534.250028] env[63515]: ERROR nova.compute.manager raise self.value [ 534.250028] env[63515]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 534.250028] env[63515]: ERROR nova.compute.manager updated_port = self._update_port( [ 534.250028] env[63515]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 534.250028] env[63515]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 534.250716] env[63515]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 534.250716] env[63515]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 534.250716] env[63515]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 71cd34ac-f70e-44ba-856e-b65afc2b509f, please check neutron logs for more information. [ 534.250716] env[63515]: ERROR nova.compute.manager [ 534.250716] env[63515]: Traceback (most recent call last): [ 534.250716] env[63515]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 534.250716] env[63515]: listener.cb(fileno) [ 534.250716] env[63515]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 534.250716] env[63515]: result = function(*args, **kwargs) [ 534.250716] env[63515]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 534.250716] env[63515]: return func(*args, **kwargs) [ 534.250716] env[63515]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 534.250716] env[63515]: raise e [ 534.250716] env[63515]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 534.250716] env[63515]: nwinfo = self.network_api.allocate_for_instance( [ 534.250716] env[63515]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 534.250716] env[63515]: created_port_ids = self._update_ports_for_instance( [ 534.250716] env[63515]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 534.250716] env[63515]: with excutils.save_and_reraise_exception(): [ 534.250716] env[63515]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 534.250716] env[63515]: self.force_reraise() [ 534.250716] env[63515]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 534.250716] env[63515]: raise self.value [ 534.250716] env[63515]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 534.250716] env[63515]: updated_port = self._update_port( [ 534.250716] env[63515]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 534.250716] env[63515]: _ensure_no_port_binding_failure(port) [ 534.250716] env[63515]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 534.250716] env[63515]: raise exception.PortBindingFailed(port_id=port['id']) [ 534.251820] env[63515]: nova.exception.PortBindingFailed: Binding failed for port 71cd34ac-f70e-44ba-856e-b65afc2b509f, please check neutron logs for more information. [ 534.251820] env[63515]: Removing descriptor: 14 [ 534.251820] env[63515]: ERROR nova.compute.manager [None req-9ee35a7a-94b5-4758-9c8f-a8f1e8f2f2cd tempest-DeleteServersAdminTestJSON-944542165 tempest-DeleteServersAdminTestJSON-944542165-project-member] [instance: d7fb514e-4ea9-4f32-9b00-06f728813aa3] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 71cd34ac-f70e-44ba-856e-b65afc2b509f, please check neutron logs for more information. [ 534.251820] env[63515]: ERROR nova.compute.manager [instance: d7fb514e-4ea9-4f32-9b00-06f728813aa3] Traceback (most recent call last): [ 534.251820] env[63515]: ERROR nova.compute.manager [instance: d7fb514e-4ea9-4f32-9b00-06f728813aa3] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 534.251820] env[63515]: ERROR nova.compute.manager [instance: d7fb514e-4ea9-4f32-9b00-06f728813aa3] yield resources [ 534.251820] env[63515]: ERROR nova.compute.manager [instance: d7fb514e-4ea9-4f32-9b00-06f728813aa3] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 534.251820] env[63515]: ERROR nova.compute.manager [instance: d7fb514e-4ea9-4f32-9b00-06f728813aa3] self.driver.spawn(context, instance, image_meta, [ 534.251820] env[63515]: ERROR nova.compute.manager [instance: d7fb514e-4ea9-4f32-9b00-06f728813aa3] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 534.251820] env[63515]: ERROR nova.compute.manager [instance: d7fb514e-4ea9-4f32-9b00-06f728813aa3] self._vmops.spawn(context, instance, image_meta, injected_files, [ 534.251820] env[63515]: ERROR nova.compute.manager [instance: d7fb514e-4ea9-4f32-9b00-06f728813aa3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 534.251820] env[63515]: ERROR nova.compute.manager [instance: d7fb514e-4ea9-4f32-9b00-06f728813aa3] vm_ref = self.build_virtual_machine(instance, [ 534.252188] env[63515]: ERROR nova.compute.manager [instance: d7fb514e-4ea9-4f32-9b00-06f728813aa3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 534.252188] env[63515]: ERROR nova.compute.manager [instance: d7fb514e-4ea9-4f32-9b00-06f728813aa3] vif_infos = vmwarevif.get_vif_info(self._session, [ 534.252188] env[63515]: ERROR nova.compute.manager [instance: d7fb514e-4ea9-4f32-9b00-06f728813aa3] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 534.252188] env[63515]: ERROR nova.compute.manager [instance: d7fb514e-4ea9-4f32-9b00-06f728813aa3] for vif in network_info: [ 534.252188] env[63515]: ERROR nova.compute.manager [instance: d7fb514e-4ea9-4f32-9b00-06f728813aa3] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 534.252188] env[63515]: ERROR nova.compute.manager [instance: d7fb514e-4ea9-4f32-9b00-06f728813aa3] return self._sync_wrapper(fn, *args, **kwargs) [ 534.252188] env[63515]: ERROR nova.compute.manager [instance: d7fb514e-4ea9-4f32-9b00-06f728813aa3] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 534.252188] env[63515]: ERROR nova.compute.manager [instance: d7fb514e-4ea9-4f32-9b00-06f728813aa3] self.wait() [ 534.252188] env[63515]: ERROR nova.compute.manager [instance: d7fb514e-4ea9-4f32-9b00-06f728813aa3] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 534.252188] env[63515]: ERROR nova.compute.manager [instance: d7fb514e-4ea9-4f32-9b00-06f728813aa3] self[:] = self._gt.wait() [ 534.252188] env[63515]: ERROR nova.compute.manager [instance: d7fb514e-4ea9-4f32-9b00-06f728813aa3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 534.252188] env[63515]: ERROR nova.compute.manager [instance: d7fb514e-4ea9-4f32-9b00-06f728813aa3] return self._exit_event.wait() [ 534.252188] env[63515]: ERROR nova.compute.manager [instance: d7fb514e-4ea9-4f32-9b00-06f728813aa3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 534.252591] env[63515]: ERROR nova.compute.manager [instance: d7fb514e-4ea9-4f32-9b00-06f728813aa3] result = hub.switch() [ 534.252591] env[63515]: ERROR nova.compute.manager [instance: d7fb514e-4ea9-4f32-9b00-06f728813aa3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 534.252591] env[63515]: ERROR nova.compute.manager [instance: d7fb514e-4ea9-4f32-9b00-06f728813aa3] return self.greenlet.switch() [ 534.252591] env[63515]: ERROR nova.compute.manager [instance: d7fb514e-4ea9-4f32-9b00-06f728813aa3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 534.252591] env[63515]: ERROR nova.compute.manager [instance: d7fb514e-4ea9-4f32-9b00-06f728813aa3] result = function(*args, **kwargs) [ 534.252591] env[63515]: ERROR nova.compute.manager [instance: d7fb514e-4ea9-4f32-9b00-06f728813aa3] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 534.252591] env[63515]: ERROR nova.compute.manager [instance: d7fb514e-4ea9-4f32-9b00-06f728813aa3] return func(*args, **kwargs) [ 534.252591] env[63515]: ERROR nova.compute.manager [instance: d7fb514e-4ea9-4f32-9b00-06f728813aa3] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 534.252591] env[63515]: ERROR nova.compute.manager [instance: d7fb514e-4ea9-4f32-9b00-06f728813aa3] raise e [ 534.252591] env[63515]: ERROR nova.compute.manager [instance: d7fb514e-4ea9-4f32-9b00-06f728813aa3] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 534.252591] env[63515]: ERROR nova.compute.manager [instance: d7fb514e-4ea9-4f32-9b00-06f728813aa3] nwinfo = self.network_api.allocate_for_instance( [ 534.252591] env[63515]: ERROR nova.compute.manager [instance: d7fb514e-4ea9-4f32-9b00-06f728813aa3] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 534.252591] env[63515]: ERROR nova.compute.manager [instance: d7fb514e-4ea9-4f32-9b00-06f728813aa3] created_port_ids = self._update_ports_for_instance( [ 534.252965] env[63515]: ERROR nova.compute.manager [instance: d7fb514e-4ea9-4f32-9b00-06f728813aa3] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 534.252965] env[63515]: ERROR nova.compute.manager [instance: d7fb514e-4ea9-4f32-9b00-06f728813aa3] with excutils.save_and_reraise_exception(): [ 534.252965] env[63515]: ERROR nova.compute.manager [instance: d7fb514e-4ea9-4f32-9b00-06f728813aa3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 534.252965] env[63515]: ERROR nova.compute.manager [instance: d7fb514e-4ea9-4f32-9b00-06f728813aa3] self.force_reraise() [ 534.252965] env[63515]: ERROR nova.compute.manager [instance: d7fb514e-4ea9-4f32-9b00-06f728813aa3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 534.252965] env[63515]: ERROR nova.compute.manager [instance: d7fb514e-4ea9-4f32-9b00-06f728813aa3] raise self.value [ 534.252965] env[63515]: ERROR nova.compute.manager [instance: d7fb514e-4ea9-4f32-9b00-06f728813aa3] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 534.252965] env[63515]: ERROR nova.compute.manager [instance: d7fb514e-4ea9-4f32-9b00-06f728813aa3] updated_port = self._update_port( [ 534.252965] env[63515]: ERROR nova.compute.manager [instance: d7fb514e-4ea9-4f32-9b00-06f728813aa3] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 534.252965] env[63515]: ERROR nova.compute.manager [instance: d7fb514e-4ea9-4f32-9b00-06f728813aa3] _ensure_no_port_binding_failure(port) [ 534.252965] env[63515]: ERROR nova.compute.manager [instance: d7fb514e-4ea9-4f32-9b00-06f728813aa3] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 534.252965] env[63515]: ERROR nova.compute.manager [instance: d7fb514e-4ea9-4f32-9b00-06f728813aa3] raise exception.PortBindingFailed(port_id=port['id']) [ 534.253317] env[63515]: ERROR nova.compute.manager [instance: d7fb514e-4ea9-4f32-9b00-06f728813aa3] nova.exception.PortBindingFailed: Binding failed for port 71cd34ac-f70e-44ba-856e-b65afc2b509f, please check neutron logs for more information. [ 534.253317] env[63515]: ERROR nova.compute.manager [instance: d7fb514e-4ea9-4f32-9b00-06f728813aa3] [ 534.253317] env[63515]: INFO nova.compute.manager [None req-9ee35a7a-94b5-4758-9c8f-a8f1e8f2f2cd tempest-DeleteServersAdminTestJSON-944542165 tempest-DeleteServersAdminTestJSON-944542165-project-member] [instance: d7fb514e-4ea9-4f32-9b00-06f728813aa3] Terminating instance [ 534.255752] env[63515]: DEBUG oslo_concurrency.lockutils [None req-9ee35a7a-94b5-4758-9c8f-a8f1e8f2f2cd tempest-DeleteServersAdminTestJSON-944542165 tempest-DeleteServersAdminTestJSON-944542165-project-member] Acquiring lock "refresh_cache-d7fb514e-4ea9-4f32-9b00-06f728813aa3" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 534.255752] env[63515]: DEBUG oslo_concurrency.lockutils [None req-9ee35a7a-94b5-4758-9c8f-a8f1e8f2f2cd tempest-DeleteServersAdminTestJSON-944542165 tempest-DeleteServersAdminTestJSON-944542165-project-member] Acquired lock "refresh_cache-d7fb514e-4ea9-4f32-9b00-06f728813aa3" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 534.255822] env[63515]: DEBUG nova.network.neutron [None req-9ee35a7a-94b5-4758-9c8f-a8f1e8f2f2cd tempest-DeleteServersAdminTestJSON-944542165 tempest-DeleteServersAdminTestJSON-944542165-project-member] [instance: d7fb514e-4ea9-4f32-9b00-06f728813aa3] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 534.556493] env[63515]: DEBUG oslo_concurrency.lockutils [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 534.803039] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a540c1a7-dfcb-49c6-b008-80fe85393680 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 534.810374] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3e0049b-b084-47a5-81bc-1e76fe710960 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 534.843238] env[63515]: DEBUG nova.network.neutron [None req-9ee35a7a-94b5-4758-9c8f-a8f1e8f2f2cd tempest-DeleteServersAdminTestJSON-944542165 tempest-DeleteServersAdminTestJSON-944542165-project-member] [instance: d7fb514e-4ea9-4f32-9b00-06f728813aa3] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 534.847048] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e54d5b49-7077-40dd-a604-f718dced3a9d {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 534.853732] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62fe31e6-b56b-4dc8-9e6f-8158c1c455d2 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 534.867148] env[63515]: DEBUG nova.compute.provider_tree [None req-c96f19aa-9abc-4b12-99f8-980442cab640 tempest-ServersAdminTestJSON-710082096 tempest-ServersAdminTestJSON-710082096-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 535.001226] env[63515]: DEBUG nova.network.neutron [None req-9ee35a7a-94b5-4758-9c8f-a8f1e8f2f2cd tempest-DeleteServersAdminTestJSON-944542165 tempest-DeleteServersAdminTestJSON-944542165-project-member] [instance: d7fb514e-4ea9-4f32-9b00-06f728813aa3] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 535.371659] env[63515]: DEBUG nova.scheduler.client.report [None req-c96f19aa-9abc-4b12-99f8-980442cab640 tempest-ServersAdminTestJSON-710082096 tempest-ServersAdminTestJSON-710082096-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 172, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 535.423066] env[63515]: DEBUG oslo_concurrency.lockutils [None req-ada3051f-c750-48c7-b591-0344c25b8fea tempest-ServersAdminTestJSON-710082096 tempest-ServersAdminTestJSON-710082096-project-member] Acquiring lock "d7fed938-f7cd-4013-b545-5d5350242084" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 535.423652] env[63515]: DEBUG oslo_concurrency.lockutils [None req-ada3051f-c750-48c7-b591-0344c25b8fea tempest-ServersAdminTestJSON-710082096 tempest-ServersAdminTestJSON-710082096-project-member] Lock "d7fed938-f7cd-4013-b545-5d5350242084" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 535.452879] env[63515]: ERROR nova.compute.manager [None req-0d00ff28-5b1e-46b5-9fb9-df8b8bcd6a1a tempest-TenantUsagesTestJSON-427870877 tempest-TenantUsagesTestJSON-427870877-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 111c82a0-4d47-4505-8e27-6281465572af, please check neutron logs for more information. [ 535.452879] env[63515]: ERROR nova.compute.manager Traceback (most recent call last): [ 535.452879] env[63515]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 535.452879] env[63515]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 535.452879] env[63515]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 535.452879] env[63515]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 535.452879] env[63515]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 535.452879] env[63515]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 535.452879] env[63515]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 535.452879] env[63515]: ERROR nova.compute.manager self.force_reraise() [ 535.452879] env[63515]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 535.452879] env[63515]: ERROR nova.compute.manager raise self.value [ 535.452879] env[63515]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 535.452879] env[63515]: ERROR nova.compute.manager updated_port = self._update_port( [ 535.452879] env[63515]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 535.452879] env[63515]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 535.453745] env[63515]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 535.453745] env[63515]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 535.453745] env[63515]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 111c82a0-4d47-4505-8e27-6281465572af, please check neutron logs for more information. [ 535.453745] env[63515]: ERROR nova.compute.manager [ 535.453745] env[63515]: Traceback (most recent call last): [ 535.453745] env[63515]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 535.453745] env[63515]: listener.cb(fileno) [ 535.453745] env[63515]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 535.453745] env[63515]: result = function(*args, **kwargs) [ 535.453745] env[63515]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 535.453745] env[63515]: return func(*args, **kwargs) [ 535.453745] env[63515]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 535.453745] env[63515]: raise e [ 535.453745] env[63515]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 535.453745] env[63515]: nwinfo = self.network_api.allocate_for_instance( [ 535.453745] env[63515]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 535.453745] env[63515]: created_port_ids = self._update_ports_for_instance( [ 535.453745] env[63515]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 535.453745] env[63515]: with excutils.save_and_reraise_exception(): [ 535.453745] env[63515]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 535.453745] env[63515]: self.force_reraise() [ 535.453745] env[63515]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 535.453745] env[63515]: raise self.value [ 535.453745] env[63515]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 535.453745] env[63515]: updated_port = self._update_port( [ 535.453745] env[63515]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 535.453745] env[63515]: _ensure_no_port_binding_failure(port) [ 535.453745] env[63515]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 535.453745] env[63515]: raise exception.PortBindingFailed(port_id=port['id']) [ 535.454673] env[63515]: nova.exception.PortBindingFailed: Binding failed for port 111c82a0-4d47-4505-8e27-6281465572af, please check neutron logs for more information. [ 535.454673] env[63515]: Removing descriptor: 16 [ 535.454673] env[63515]: ERROR nova.compute.manager [None req-0d00ff28-5b1e-46b5-9fb9-df8b8bcd6a1a tempest-TenantUsagesTestJSON-427870877 tempest-TenantUsagesTestJSON-427870877-project-member] [instance: 9f7ad2dd-b882-4d98-8576-708b3abfa505] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 111c82a0-4d47-4505-8e27-6281465572af, please check neutron logs for more information. [ 535.454673] env[63515]: ERROR nova.compute.manager [instance: 9f7ad2dd-b882-4d98-8576-708b3abfa505] Traceback (most recent call last): [ 535.454673] env[63515]: ERROR nova.compute.manager [instance: 9f7ad2dd-b882-4d98-8576-708b3abfa505] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 535.454673] env[63515]: ERROR nova.compute.manager [instance: 9f7ad2dd-b882-4d98-8576-708b3abfa505] yield resources [ 535.454673] env[63515]: ERROR nova.compute.manager [instance: 9f7ad2dd-b882-4d98-8576-708b3abfa505] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 535.454673] env[63515]: ERROR nova.compute.manager [instance: 9f7ad2dd-b882-4d98-8576-708b3abfa505] self.driver.spawn(context, instance, image_meta, [ 535.454673] env[63515]: ERROR nova.compute.manager [instance: 9f7ad2dd-b882-4d98-8576-708b3abfa505] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 535.454673] env[63515]: ERROR nova.compute.manager [instance: 9f7ad2dd-b882-4d98-8576-708b3abfa505] self._vmops.spawn(context, instance, image_meta, injected_files, [ 535.454673] env[63515]: ERROR nova.compute.manager [instance: 9f7ad2dd-b882-4d98-8576-708b3abfa505] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 535.454673] env[63515]: ERROR nova.compute.manager [instance: 9f7ad2dd-b882-4d98-8576-708b3abfa505] vm_ref = self.build_virtual_machine(instance, [ 535.455169] env[63515]: ERROR nova.compute.manager [instance: 9f7ad2dd-b882-4d98-8576-708b3abfa505] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 535.455169] env[63515]: ERROR nova.compute.manager [instance: 9f7ad2dd-b882-4d98-8576-708b3abfa505] vif_infos = vmwarevif.get_vif_info(self._session, [ 535.455169] env[63515]: ERROR nova.compute.manager [instance: 9f7ad2dd-b882-4d98-8576-708b3abfa505] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 535.455169] env[63515]: ERROR nova.compute.manager [instance: 9f7ad2dd-b882-4d98-8576-708b3abfa505] for vif in network_info: [ 535.455169] env[63515]: ERROR nova.compute.manager [instance: 9f7ad2dd-b882-4d98-8576-708b3abfa505] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 535.455169] env[63515]: ERROR nova.compute.manager [instance: 9f7ad2dd-b882-4d98-8576-708b3abfa505] return self._sync_wrapper(fn, *args, **kwargs) [ 535.455169] env[63515]: ERROR nova.compute.manager [instance: 9f7ad2dd-b882-4d98-8576-708b3abfa505] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 535.455169] env[63515]: ERROR nova.compute.manager [instance: 9f7ad2dd-b882-4d98-8576-708b3abfa505] self.wait() [ 535.455169] env[63515]: ERROR nova.compute.manager [instance: 9f7ad2dd-b882-4d98-8576-708b3abfa505] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 535.455169] env[63515]: ERROR nova.compute.manager [instance: 9f7ad2dd-b882-4d98-8576-708b3abfa505] self[:] = self._gt.wait() [ 535.455169] env[63515]: ERROR nova.compute.manager [instance: 9f7ad2dd-b882-4d98-8576-708b3abfa505] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 535.455169] env[63515]: ERROR nova.compute.manager [instance: 9f7ad2dd-b882-4d98-8576-708b3abfa505] return self._exit_event.wait() [ 535.455169] env[63515]: ERROR nova.compute.manager [instance: 9f7ad2dd-b882-4d98-8576-708b3abfa505] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 535.455616] env[63515]: ERROR nova.compute.manager [instance: 9f7ad2dd-b882-4d98-8576-708b3abfa505] result = hub.switch() [ 535.455616] env[63515]: ERROR nova.compute.manager [instance: 9f7ad2dd-b882-4d98-8576-708b3abfa505] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 535.455616] env[63515]: ERROR nova.compute.manager [instance: 9f7ad2dd-b882-4d98-8576-708b3abfa505] return self.greenlet.switch() [ 535.455616] env[63515]: ERROR nova.compute.manager [instance: 9f7ad2dd-b882-4d98-8576-708b3abfa505] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 535.455616] env[63515]: ERROR nova.compute.manager [instance: 9f7ad2dd-b882-4d98-8576-708b3abfa505] result = function(*args, **kwargs) [ 535.455616] env[63515]: ERROR nova.compute.manager [instance: 9f7ad2dd-b882-4d98-8576-708b3abfa505] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 535.455616] env[63515]: ERROR nova.compute.manager [instance: 9f7ad2dd-b882-4d98-8576-708b3abfa505] return func(*args, **kwargs) [ 535.455616] env[63515]: ERROR nova.compute.manager [instance: 9f7ad2dd-b882-4d98-8576-708b3abfa505] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 535.455616] env[63515]: ERROR nova.compute.manager [instance: 9f7ad2dd-b882-4d98-8576-708b3abfa505] raise e [ 535.455616] env[63515]: ERROR nova.compute.manager [instance: 9f7ad2dd-b882-4d98-8576-708b3abfa505] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 535.455616] env[63515]: ERROR nova.compute.manager [instance: 9f7ad2dd-b882-4d98-8576-708b3abfa505] nwinfo = self.network_api.allocate_for_instance( [ 535.455616] env[63515]: ERROR nova.compute.manager [instance: 9f7ad2dd-b882-4d98-8576-708b3abfa505] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 535.455616] env[63515]: ERROR nova.compute.manager [instance: 9f7ad2dd-b882-4d98-8576-708b3abfa505] created_port_ids = self._update_ports_for_instance( [ 535.455996] env[63515]: ERROR nova.compute.manager [instance: 9f7ad2dd-b882-4d98-8576-708b3abfa505] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 535.455996] env[63515]: ERROR nova.compute.manager [instance: 9f7ad2dd-b882-4d98-8576-708b3abfa505] with excutils.save_and_reraise_exception(): [ 535.455996] env[63515]: ERROR nova.compute.manager [instance: 9f7ad2dd-b882-4d98-8576-708b3abfa505] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 535.455996] env[63515]: ERROR nova.compute.manager [instance: 9f7ad2dd-b882-4d98-8576-708b3abfa505] self.force_reraise() [ 535.455996] env[63515]: ERROR nova.compute.manager [instance: 9f7ad2dd-b882-4d98-8576-708b3abfa505] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 535.455996] env[63515]: ERROR nova.compute.manager [instance: 9f7ad2dd-b882-4d98-8576-708b3abfa505] raise self.value [ 535.455996] env[63515]: ERROR nova.compute.manager [instance: 9f7ad2dd-b882-4d98-8576-708b3abfa505] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 535.455996] env[63515]: ERROR nova.compute.manager [instance: 9f7ad2dd-b882-4d98-8576-708b3abfa505] updated_port = self._update_port( [ 535.455996] env[63515]: ERROR nova.compute.manager [instance: 9f7ad2dd-b882-4d98-8576-708b3abfa505] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 535.455996] env[63515]: ERROR nova.compute.manager [instance: 9f7ad2dd-b882-4d98-8576-708b3abfa505] _ensure_no_port_binding_failure(port) [ 535.455996] env[63515]: ERROR nova.compute.manager [instance: 9f7ad2dd-b882-4d98-8576-708b3abfa505] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 535.455996] env[63515]: ERROR nova.compute.manager [instance: 9f7ad2dd-b882-4d98-8576-708b3abfa505] raise exception.PortBindingFailed(port_id=port['id']) [ 535.456495] env[63515]: ERROR nova.compute.manager [instance: 9f7ad2dd-b882-4d98-8576-708b3abfa505] nova.exception.PortBindingFailed: Binding failed for port 111c82a0-4d47-4505-8e27-6281465572af, please check neutron logs for more information. [ 535.456495] env[63515]: ERROR nova.compute.manager [instance: 9f7ad2dd-b882-4d98-8576-708b3abfa505] [ 535.456495] env[63515]: INFO nova.compute.manager [None req-0d00ff28-5b1e-46b5-9fb9-df8b8bcd6a1a tempest-TenantUsagesTestJSON-427870877 tempest-TenantUsagesTestJSON-427870877-project-member] [instance: 9f7ad2dd-b882-4d98-8576-708b3abfa505] Terminating instance [ 535.459998] env[63515]: DEBUG oslo_concurrency.lockutils [None req-0d00ff28-5b1e-46b5-9fb9-df8b8bcd6a1a tempest-TenantUsagesTestJSON-427870877 tempest-TenantUsagesTestJSON-427870877-project-member] Acquiring lock "refresh_cache-9f7ad2dd-b882-4d98-8576-708b3abfa505" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 535.459998] env[63515]: DEBUG oslo_concurrency.lockutils [None req-0d00ff28-5b1e-46b5-9fb9-df8b8bcd6a1a tempest-TenantUsagesTestJSON-427870877 tempest-TenantUsagesTestJSON-427870877-project-member] Acquired lock "refresh_cache-9f7ad2dd-b882-4d98-8576-708b3abfa505" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 535.460182] env[63515]: DEBUG nova.network.neutron [None req-0d00ff28-5b1e-46b5-9fb9-df8b8bcd6a1a tempest-TenantUsagesTestJSON-427870877 tempest-TenantUsagesTestJSON-427870877-project-member] [instance: 9f7ad2dd-b882-4d98-8576-708b3abfa505] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 535.510768] env[63515]: DEBUG oslo_concurrency.lockutils [None req-9ee35a7a-94b5-4758-9c8f-a8f1e8f2f2cd tempest-DeleteServersAdminTestJSON-944542165 tempest-DeleteServersAdminTestJSON-944542165-project-member] Releasing lock "refresh_cache-d7fb514e-4ea9-4f32-9b00-06f728813aa3" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 535.510768] env[63515]: DEBUG nova.compute.manager [None req-9ee35a7a-94b5-4758-9c8f-a8f1e8f2f2cd tempest-DeleteServersAdminTestJSON-944542165 tempest-DeleteServersAdminTestJSON-944542165-project-member] [instance: d7fb514e-4ea9-4f32-9b00-06f728813aa3] Start destroying the instance on the hypervisor. {{(pid=63515) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 535.510768] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-9ee35a7a-94b5-4758-9c8f-a8f1e8f2f2cd tempest-DeleteServersAdminTestJSON-944542165 tempest-DeleteServersAdminTestJSON-944542165-project-member] [instance: d7fb514e-4ea9-4f32-9b00-06f728813aa3] Destroying instance {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 535.510768] env[63515]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-eeed3388-c4fe-473e-9895-f2ce76487a98 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 535.520016] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c46fd58-b1db-4573-9e8a-4083396cb6e6 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 535.547561] env[63515]: WARNING nova.virt.vmwareapi.vmops [None req-9ee35a7a-94b5-4758-9c8f-a8f1e8f2f2cd tempest-DeleteServersAdminTestJSON-944542165 tempest-DeleteServersAdminTestJSON-944542165-project-member] [instance: d7fb514e-4ea9-4f32-9b00-06f728813aa3] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance d7fb514e-4ea9-4f32-9b00-06f728813aa3 could not be found. [ 535.548025] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-9ee35a7a-94b5-4758-9c8f-a8f1e8f2f2cd tempest-DeleteServersAdminTestJSON-944542165 tempest-DeleteServersAdminTestJSON-944542165-project-member] [instance: d7fb514e-4ea9-4f32-9b00-06f728813aa3] Instance destroyed {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 535.550030] env[63515]: INFO nova.compute.manager [None req-9ee35a7a-94b5-4758-9c8f-a8f1e8f2f2cd tempest-DeleteServersAdminTestJSON-944542165 tempest-DeleteServersAdminTestJSON-944542165-project-member] [instance: d7fb514e-4ea9-4f32-9b00-06f728813aa3] Took 0.04 seconds to destroy the instance on the hypervisor. [ 535.550360] env[63515]: DEBUG oslo.service.loopingcall [None req-9ee35a7a-94b5-4758-9c8f-a8f1e8f2f2cd tempest-DeleteServersAdminTestJSON-944542165 tempest-DeleteServersAdminTestJSON-944542165-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 535.550710] env[63515]: DEBUG nova.compute.manager [-] [instance: d7fb514e-4ea9-4f32-9b00-06f728813aa3] Deallocating network for instance {{(pid=63515) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 535.550781] env[63515]: DEBUG nova.network.neutron [-] [instance: d7fb514e-4ea9-4f32-9b00-06f728813aa3] deallocate_for_instance() {{(pid=63515) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 535.589458] env[63515]: DEBUG nova.network.neutron [-] [instance: d7fb514e-4ea9-4f32-9b00-06f728813aa3] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 535.667034] env[63515]: DEBUG oslo_concurrency.lockutils [None req-222754a5-7546-43c7-a99a-1c936b9e5035 tempest-ImagesOneServerTestJSON-1950372435 tempest-ImagesOneServerTestJSON-1950372435-project-member] Acquiring lock "aec66939-7ebc-4bfc-bd8b-bae22e011239" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 535.667034] env[63515]: DEBUG oslo_concurrency.lockutils [None req-222754a5-7546-43c7-a99a-1c936b9e5035 tempest-ImagesOneServerTestJSON-1950372435 tempest-ImagesOneServerTestJSON-1950372435-project-member] Lock "aec66939-7ebc-4bfc-bd8b-bae22e011239" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 535.878485] env[63515]: DEBUG oslo_concurrency.lockutils [None req-c96f19aa-9abc-4b12-99f8-980442cab640 tempest-ServersAdminTestJSON-710082096 tempest-ServersAdminTestJSON-710082096-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.194s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 535.878920] env[63515]: DEBUG nova.compute.manager [None req-c96f19aa-9abc-4b12-99f8-980442cab640 tempest-ServersAdminTestJSON-710082096 tempest-ServersAdminTestJSON-710082096-project-member] [instance: 29d199bc-aff3-4c62-af3a-1e6880b52446] Start building networks asynchronously for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 535.882666] env[63515]: DEBUG oslo_concurrency.lockutils [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 1.326s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 535.882822] env[63515]: DEBUG oslo_concurrency.lockutils [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 535.882977] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=63515) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 535.883854] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a99ed9c1-d61c-42c5-9d83-c045bc5025d6 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 535.892834] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-741031b5-c5a9-41c1-a0ae-9dacb0673014 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 535.916284] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb14c4b7-2fbd-426a-9026-8af4037db5f2 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 535.924695] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-406dd3f9-b916-4f22-9ef1-bf9ca39c8240 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 535.931390] env[63515]: DEBUG nova.compute.manager [None req-ada3051f-c750-48c7-b591-0344c25b8fea tempest-ServersAdminTestJSON-710082096 tempest-ServersAdminTestJSON-710082096-project-member] [instance: d7fed938-f7cd-4013-b545-5d5350242084] Starting instance... {{(pid=63515) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 535.968689] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181581MB free_disk=172GB free_vcpus=48 pci_devices=None {{(pid=63515) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 535.968689] env[63515]: DEBUG oslo_concurrency.lockutils [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 535.968689] env[63515]: DEBUG oslo_concurrency.lockutils [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 536.004029] env[63515]: DEBUG nova.network.neutron [None req-0d00ff28-5b1e-46b5-9fb9-df8b8bcd6a1a tempest-TenantUsagesTestJSON-427870877 tempest-TenantUsagesTestJSON-427870877-project-member] [instance: 9f7ad2dd-b882-4d98-8576-708b3abfa505] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 536.039136] env[63515]: ERROR nova.compute.manager [None req-919af50b-07d1-4794-942f-d99fa56afe84 tempest-ServerExternalEventsTest-1976219582 tempest-ServerExternalEventsTest-1976219582-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 39b0bc20-a4b4-4683-ac31-11d4a2fa8333, please check neutron logs for more information. [ 536.039136] env[63515]: ERROR nova.compute.manager Traceback (most recent call last): [ 536.039136] env[63515]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 536.039136] env[63515]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 536.039136] env[63515]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 536.039136] env[63515]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 536.039136] env[63515]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 536.039136] env[63515]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 536.039136] env[63515]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 536.039136] env[63515]: ERROR nova.compute.manager self.force_reraise() [ 536.039136] env[63515]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 536.039136] env[63515]: ERROR nova.compute.manager raise self.value [ 536.039136] env[63515]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 536.039136] env[63515]: ERROR nova.compute.manager updated_port = self._update_port( [ 536.039136] env[63515]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 536.039136] env[63515]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 536.040960] env[63515]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 536.040960] env[63515]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 536.040960] env[63515]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 39b0bc20-a4b4-4683-ac31-11d4a2fa8333, please check neutron logs for more information. [ 536.040960] env[63515]: ERROR nova.compute.manager [ 536.040960] env[63515]: Traceback (most recent call last): [ 536.040960] env[63515]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 536.040960] env[63515]: listener.cb(fileno) [ 536.040960] env[63515]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 536.040960] env[63515]: result = function(*args, **kwargs) [ 536.040960] env[63515]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 536.040960] env[63515]: return func(*args, **kwargs) [ 536.040960] env[63515]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 536.040960] env[63515]: raise e [ 536.040960] env[63515]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 536.040960] env[63515]: nwinfo = self.network_api.allocate_for_instance( [ 536.040960] env[63515]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 536.040960] env[63515]: created_port_ids = self._update_ports_for_instance( [ 536.040960] env[63515]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 536.040960] env[63515]: with excutils.save_and_reraise_exception(): [ 536.040960] env[63515]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 536.040960] env[63515]: self.force_reraise() [ 536.040960] env[63515]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 536.040960] env[63515]: raise self.value [ 536.040960] env[63515]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 536.040960] env[63515]: updated_port = self._update_port( [ 536.040960] env[63515]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 536.040960] env[63515]: _ensure_no_port_binding_failure(port) [ 536.040960] env[63515]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 536.040960] env[63515]: raise exception.PortBindingFailed(port_id=port['id']) [ 536.041872] env[63515]: nova.exception.PortBindingFailed: Binding failed for port 39b0bc20-a4b4-4683-ac31-11d4a2fa8333, please check neutron logs for more information. [ 536.041872] env[63515]: Removing descriptor: 17 [ 536.041872] env[63515]: ERROR nova.compute.manager [None req-919af50b-07d1-4794-942f-d99fa56afe84 tempest-ServerExternalEventsTest-1976219582 tempest-ServerExternalEventsTest-1976219582-project-member] [instance: 6662da99-ba90-49ac-982e-d18f1c7834b6] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 39b0bc20-a4b4-4683-ac31-11d4a2fa8333, please check neutron logs for more information. [ 536.041872] env[63515]: ERROR nova.compute.manager [instance: 6662da99-ba90-49ac-982e-d18f1c7834b6] Traceback (most recent call last): [ 536.041872] env[63515]: ERROR nova.compute.manager [instance: 6662da99-ba90-49ac-982e-d18f1c7834b6] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 536.041872] env[63515]: ERROR nova.compute.manager [instance: 6662da99-ba90-49ac-982e-d18f1c7834b6] yield resources [ 536.041872] env[63515]: ERROR nova.compute.manager [instance: 6662da99-ba90-49ac-982e-d18f1c7834b6] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 536.041872] env[63515]: ERROR nova.compute.manager [instance: 6662da99-ba90-49ac-982e-d18f1c7834b6] self.driver.spawn(context, instance, image_meta, [ 536.041872] env[63515]: ERROR nova.compute.manager [instance: 6662da99-ba90-49ac-982e-d18f1c7834b6] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 536.041872] env[63515]: ERROR nova.compute.manager [instance: 6662da99-ba90-49ac-982e-d18f1c7834b6] self._vmops.spawn(context, instance, image_meta, injected_files, [ 536.041872] env[63515]: ERROR nova.compute.manager [instance: 6662da99-ba90-49ac-982e-d18f1c7834b6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 536.041872] env[63515]: ERROR nova.compute.manager [instance: 6662da99-ba90-49ac-982e-d18f1c7834b6] vm_ref = self.build_virtual_machine(instance, [ 536.042302] env[63515]: ERROR nova.compute.manager [instance: 6662da99-ba90-49ac-982e-d18f1c7834b6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 536.042302] env[63515]: ERROR nova.compute.manager [instance: 6662da99-ba90-49ac-982e-d18f1c7834b6] vif_infos = vmwarevif.get_vif_info(self._session, [ 536.042302] env[63515]: ERROR nova.compute.manager [instance: 6662da99-ba90-49ac-982e-d18f1c7834b6] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 536.042302] env[63515]: ERROR nova.compute.manager [instance: 6662da99-ba90-49ac-982e-d18f1c7834b6] for vif in network_info: [ 536.042302] env[63515]: ERROR nova.compute.manager [instance: 6662da99-ba90-49ac-982e-d18f1c7834b6] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 536.042302] env[63515]: ERROR nova.compute.manager [instance: 6662da99-ba90-49ac-982e-d18f1c7834b6] return self._sync_wrapper(fn, *args, **kwargs) [ 536.042302] env[63515]: ERROR nova.compute.manager [instance: 6662da99-ba90-49ac-982e-d18f1c7834b6] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 536.042302] env[63515]: ERROR nova.compute.manager [instance: 6662da99-ba90-49ac-982e-d18f1c7834b6] self.wait() [ 536.042302] env[63515]: ERROR nova.compute.manager [instance: 6662da99-ba90-49ac-982e-d18f1c7834b6] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 536.042302] env[63515]: ERROR nova.compute.manager [instance: 6662da99-ba90-49ac-982e-d18f1c7834b6] self[:] = self._gt.wait() [ 536.042302] env[63515]: ERROR nova.compute.manager [instance: 6662da99-ba90-49ac-982e-d18f1c7834b6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 536.042302] env[63515]: ERROR nova.compute.manager [instance: 6662da99-ba90-49ac-982e-d18f1c7834b6] return self._exit_event.wait() [ 536.042302] env[63515]: ERROR nova.compute.manager [instance: 6662da99-ba90-49ac-982e-d18f1c7834b6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 536.042653] env[63515]: ERROR nova.compute.manager [instance: 6662da99-ba90-49ac-982e-d18f1c7834b6] result = hub.switch() [ 536.042653] env[63515]: ERROR nova.compute.manager [instance: 6662da99-ba90-49ac-982e-d18f1c7834b6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 536.042653] env[63515]: ERROR nova.compute.manager [instance: 6662da99-ba90-49ac-982e-d18f1c7834b6] return self.greenlet.switch() [ 536.042653] env[63515]: ERROR nova.compute.manager [instance: 6662da99-ba90-49ac-982e-d18f1c7834b6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 536.042653] env[63515]: ERROR nova.compute.manager [instance: 6662da99-ba90-49ac-982e-d18f1c7834b6] result = function(*args, **kwargs) [ 536.042653] env[63515]: ERROR nova.compute.manager [instance: 6662da99-ba90-49ac-982e-d18f1c7834b6] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 536.042653] env[63515]: ERROR nova.compute.manager [instance: 6662da99-ba90-49ac-982e-d18f1c7834b6] return func(*args, **kwargs) [ 536.042653] env[63515]: ERROR nova.compute.manager [instance: 6662da99-ba90-49ac-982e-d18f1c7834b6] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 536.042653] env[63515]: ERROR nova.compute.manager [instance: 6662da99-ba90-49ac-982e-d18f1c7834b6] raise e [ 536.042653] env[63515]: ERROR nova.compute.manager [instance: 6662da99-ba90-49ac-982e-d18f1c7834b6] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 536.042653] env[63515]: ERROR nova.compute.manager [instance: 6662da99-ba90-49ac-982e-d18f1c7834b6] nwinfo = self.network_api.allocate_for_instance( [ 536.042653] env[63515]: ERROR nova.compute.manager [instance: 6662da99-ba90-49ac-982e-d18f1c7834b6] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 536.042653] env[63515]: ERROR nova.compute.manager [instance: 6662da99-ba90-49ac-982e-d18f1c7834b6] created_port_ids = self._update_ports_for_instance( [ 536.043066] env[63515]: ERROR nova.compute.manager [instance: 6662da99-ba90-49ac-982e-d18f1c7834b6] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 536.043066] env[63515]: ERROR nova.compute.manager [instance: 6662da99-ba90-49ac-982e-d18f1c7834b6] with excutils.save_and_reraise_exception(): [ 536.043066] env[63515]: ERROR nova.compute.manager [instance: 6662da99-ba90-49ac-982e-d18f1c7834b6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 536.043066] env[63515]: ERROR nova.compute.manager [instance: 6662da99-ba90-49ac-982e-d18f1c7834b6] self.force_reraise() [ 536.043066] env[63515]: ERROR nova.compute.manager [instance: 6662da99-ba90-49ac-982e-d18f1c7834b6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 536.043066] env[63515]: ERROR nova.compute.manager [instance: 6662da99-ba90-49ac-982e-d18f1c7834b6] raise self.value [ 536.043066] env[63515]: ERROR nova.compute.manager [instance: 6662da99-ba90-49ac-982e-d18f1c7834b6] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 536.043066] env[63515]: ERROR nova.compute.manager [instance: 6662da99-ba90-49ac-982e-d18f1c7834b6] updated_port = self._update_port( [ 536.043066] env[63515]: ERROR nova.compute.manager [instance: 6662da99-ba90-49ac-982e-d18f1c7834b6] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 536.043066] env[63515]: ERROR nova.compute.manager [instance: 6662da99-ba90-49ac-982e-d18f1c7834b6] _ensure_no_port_binding_failure(port) [ 536.043066] env[63515]: ERROR nova.compute.manager [instance: 6662da99-ba90-49ac-982e-d18f1c7834b6] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 536.043066] env[63515]: ERROR nova.compute.manager [instance: 6662da99-ba90-49ac-982e-d18f1c7834b6] raise exception.PortBindingFailed(port_id=port['id']) [ 536.043359] env[63515]: ERROR nova.compute.manager [instance: 6662da99-ba90-49ac-982e-d18f1c7834b6] nova.exception.PortBindingFailed: Binding failed for port 39b0bc20-a4b4-4683-ac31-11d4a2fa8333, please check neutron logs for more information. [ 536.043359] env[63515]: ERROR nova.compute.manager [instance: 6662da99-ba90-49ac-982e-d18f1c7834b6] [ 536.043359] env[63515]: INFO nova.compute.manager [None req-919af50b-07d1-4794-942f-d99fa56afe84 tempest-ServerExternalEventsTest-1976219582 tempest-ServerExternalEventsTest-1976219582-project-member] [instance: 6662da99-ba90-49ac-982e-d18f1c7834b6] Terminating instance [ 536.048900] env[63515]: DEBUG oslo_concurrency.lockutils [None req-919af50b-07d1-4794-942f-d99fa56afe84 tempest-ServerExternalEventsTest-1976219582 tempest-ServerExternalEventsTest-1976219582-project-member] Acquiring lock "refresh_cache-6662da99-ba90-49ac-982e-d18f1c7834b6" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 536.048900] env[63515]: DEBUG oslo_concurrency.lockutils [None req-919af50b-07d1-4794-942f-d99fa56afe84 tempest-ServerExternalEventsTest-1976219582 tempest-ServerExternalEventsTest-1976219582-project-member] Acquired lock "refresh_cache-6662da99-ba90-49ac-982e-d18f1c7834b6" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 536.048900] env[63515]: DEBUG nova.network.neutron [None req-919af50b-07d1-4794-942f-d99fa56afe84 tempest-ServerExternalEventsTest-1976219582 tempest-ServerExternalEventsTest-1976219582-project-member] [instance: 6662da99-ba90-49ac-982e-d18f1c7834b6] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 536.097358] env[63515]: DEBUG nova.network.neutron [-] [instance: d7fb514e-4ea9-4f32-9b00-06f728813aa3] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 536.133340] env[63515]: DEBUG nova.network.neutron [None req-0d00ff28-5b1e-46b5-9fb9-df8b8bcd6a1a tempest-TenantUsagesTestJSON-427870877 tempest-TenantUsagesTestJSON-427870877-project-member] [instance: 9f7ad2dd-b882-4d98-8576-708b3abfa505] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 536.172020] env[63515]: DEBUG nova.compute.manager [None req-222754a5-7546-43c7-a99a-1c936b9e5035 tempest-ImagesOneServerTestJSON-1950372435 tempest-ImagesOneServerTestJSON-1950372435-project-member] [instance: aec66939-7ebc-4bfc-bd8b-bae22e011239] Starting instance... {{(pid=63515) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 536.384439] env[63515]: DEBUG nova.compute.utils [None req-c96f19aa-9abc-4b12-99f8-980442cab640 tempest-ServersAdminTestJSON-710082096 tempest-ServersAdminTestJSON-710082096-project-member] Using /dev/sd instead of None {{(pid=63515) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 536.385406] env[63515]: DEBUG nova.compute.manager [None req-c96f19aa-9abc-4b12-99f8-980442cab640 tempest-ServersAdminTestJSON-710082096 tempest-ServersAdminTestJSON-710082096-project-member] [instance: 29d199bc-aff3-4c62-af3a-1e6880b52446] Allocating IP information in the background. {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 536.385501] env[63515]: DEBUG nova.network.neutron [None req-c96f19aa-9abc-4b12-99f8-980442cab640 tempest-ServersAdminTestJSON-710082096 tempest-ServersAdminTestJSON-710082096-project-member] [instance: 29d199bc-aff3-4c62-af3a-1e6880b52446] allocate_for_instance() {{(pid=63515) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 536.465304] env[63515]: DEBUG oslo_concurrency.lockutils [None req-ada3051f-c750-48c7-b591-0344c25b8fea tempest-ServersAdminTestJSON-710082096 tempest-ServersAdminTestJSON-710082096-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 536.544223] env[63515]: DEBUG nova.policy [None req-c96f19aa-9abc-4b12-99f8-980442cab640 tempest-ServersAdminTestJSON-710082096 tempest-ServersAdminTestJSON-710082096-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'de4cb162ff284f25916fdc058f001516', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '25b7ee3325094edf93fa3aeec09e8ba8', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63515) authorize /opt/stack/nova/nova/policy.py:201}} [ 536.579173] env[63515]: DEBUG nova.network.neutron [None req-919af50b-07d1-4794-942f-d99fa56afe84 tempest-ServerExternalEventsTest-1976219582 tempest-ServerExternalEventsTest-1976219582-project-member] [instance: 6662da99-ba90-49ac-982e-d18f1c7834b6] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 536.599237] env[63515]: INFO nova.compute.manager [-] [instance: d7fb514e-4ea9-4f32-9b00-06f728813aa3] Took 1.05 seconds to deallocate network for instance. [ 536.602978] env[63515]: DEBUG nova.compute.claims [None req-9ee35a7a-94b5-4758-9c8f-a8f1e8f2f2cd tempest-DeleteServersAdminTestJSON-944542165 tempest-DeleteServersAdminTestJSON-944542165-project-member] [instance: d7fb514e-4ea9-4f32-9b00-06f728813aa3] Aborting claim: {{(pid=63515) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 536.603680] env[63515]: DEBUG oslo_concurrency.lockutils [None req-9ee35a7a-94b5-4758-9c8f-a8f1e8f2f2cd tempest-DeleteServersAdminTestJSON-944542165 tempest-DeleteServersAdminTestJSON-944542165-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 536.636263] env[63515]: DEBUG nova.compute.manager [req-1667dbcb-c43c-42b4-bdeb-30c7c19d52b5 req-7b997c38-2e81-42b3-85d1-6b913e93314e service nova] [instance: d7fb514e-4ea9-4f32-9b00-06f728813aa3] Received event network-changed-71cd34ac-f70e-44ba-856e-b65afc2b509f {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 536.636450] env[63515]: DEBUG nova.compute.manager [req-1667dbcb-c43c-42b4-bdeb-30c7c19d52b5 req-7b997c38-2e81-42b3-85d1-6b913e93314e service nova] [instance: d7fb514e-4ea9-4f32-9b00-06f728813aa3] Refreshing instance network info cache due to event network-changed-71cd34ac-f70e-44ba-856e-b65afc2b509f. {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 536.636715] env[63515]: DEBUG oslo_concurrency.lockutils [req-1667dbcb-c43c-42b4-bdeb-30c7c19d52b5 req-7b997c38-2e81-42b3-85d1-6b913e93314e service nova] Acquiring lock "refresh_cache-d7fb514e-4ea9-4f32-9b00-06f728813aa3" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 536.636857] env[63515]: DEBUG oslo_concurrency.lockutils [req-1667dbcb-c43c-42b4-bdeb-30c7c19d52b5 req-7b997c38-2e81-42b3-85d1-6b913e93314e service nova] Acquired lock "refresh_cache-d7fb514e-4ea9-4f32-9b00-06f728813aa3" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 536.639096] env[63515]: DEBUG nova.network.neutron [req-1667dbcb-c43c-42b4-bdeb-30c7c19d52b5 req-7b997c38-2e81-42b3-85d1-6b913e93314e service nova] [instance: d7fb514e-4ea9-4f32-9b00-06f728813aa3] Refreshing network info cache for port 71cd34ac-f70e-44ba-856e-b65afc2b509f {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 536.644432] env[63515]: DEBUG oslo_concurrency.lockutils [None req-0d00ff28-5b1e-46b5-9fb9-df8b8bcd6a1a tempest-TenantUsagesTestJSON-427870877 tempest-TenantUsagesTestJSON-427870877-project-member] Releasing lock "refresh_cache-9f7ad2dd-b882-4d98-8576-708b3abfa505" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 536.644854] env[63515]: DEBUG nova.compute.manager [None req-0d00ff28-5b1e-46b5-9fb9-df8b8bcd6a1a tempest-TenantUsagesTestJSON-427870877 tempest-TenantUsagesTestJSON-427870877-project-member] [instance: 9f7ad2dd-b882-4d98-8576-708b3abfa505] Start destroying the instance on the hypervisor. {{(pid=63515) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 536.645050] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-0d00ff28-5b1e-46b5-9fb9-df8b8bcd6a1a tempest-TenantUsagesTestJSON-427870877 tempest-TenantUsagesTestJSON-427870877-project-member] [instance: 9f7ad2dd-b882-4d98-8576-708b3abfa505] Destroying instance {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 536.646594] env[63515]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-7d39abb2-8d83-4628-8f74-8e1b3be41f8d {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 536.661467] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87f93da0-73b1-4287-869a-18cee4f435e4 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 536.690722] env[63515]: WARNING nova.virt.vmwareapi.vmops [None req-0d00ff28-5b1e-46b5-9fb9-df8b8bcd6a1a tempest-TenantUsagesTestJSON-427870877 tempest-TenantUsagesTestJSON-427870877-project-member] [instance: 9f7ad2dd-b882-4d98-8576-708b3abfa505] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 9f7ad2dd-b882-4d98-8576-708b3abfa505 could not be found. [ 536.692235] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-0d00ff28-5b1e-46b5-9fb9-df8b8bcd6a1a tempest-TenantUsagesTestJSON-427870877 tempest-TenantUsagesTestJSON-427870877-project-member] [instance: 9f7ad2dd-b882-4d98-8576-708b3abfa505] Instance destroyed {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 536.692734] env[63515]: INFO nova.compute.manager [None req-0d00ff28-5b1e-46b5-9fb9-df8b8bcd6a1a tempest-TenantUsagesTestJSON-427870877 tempest-TenantUsagesTestJSON-427870877-project-member] [instance: 9f7ad2dd-b882-4d98-8576-708b3abfa505] Took 0.05 seconds to destroy the instance on the hypervisor. [ 536.692825] env[63515]: DEBUG oslo.service.loopingcall [None req-0d00ff28-5b1e-46b5-9fb9-df8b8bcd6a1a tempest-TenantUsagesTestJSON-427870877 tempest-TenantUsagesTestJSON-427870877-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 536.693609] env[63515]: DEBUG nova.compute.manager [-] [instance: 9f7ad2dd-b882-4d98-8576-708b3abfa505] Deallocating network for instance {{(pid=63515) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 536.693715] env[63515]: DEBUG nova.network.neutron [-] [instance: 9f7ad2dd-b882-4d98-8576-708b3abfa505] deallocate_for_instance() {{(pid=63515) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 536.698298] env[63515]: DEBUG oslo_concurrency.lockutils [None req-222754a5-7546-43c7-a99a-1c936b9e5035 tempest-ImagesOneServerTestJSON-1950372435 tempest-ImagesOneServerTestJSON-1950372435-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 536.774130] env[63515]: DEBUG nova.network.neutron [-] [instance: 9f7ad2dd-b882-4d98-8576-708b3abfa505] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 536.806241] env[63515]: DEBUG nova.network.neutron [None req-919af50b-07d1-4794-942f-d99fa56afe84 tempest-ServerExternalEventsTest-1976219582 tempest-ServerExternalEventsTest-1976219582-project-member] [instance: 6662da99-ba90-49ac-982e-d18f1c7834b6] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 536.889239] env[63515]: DEBUG nova.compute.manager [None req-c96f19aa-9abc-4b12-99f8-980442cab640 tempest-ServersAdminTestJSON-710082096 tempest-ServersAdminTestJSON-710082096-project-member] [instance: 29d199bc-aff3-4c62-af3a-1e6880b52446] Start building block device mappings for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 537.008665] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Instance d7fb514e-4ea9-4f32-9b00-06f728813aa3 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63515) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 537.008817] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Instance 9f7ad2dd-b882-4d98-8576-708b3abfa505 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63515) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 537.008938] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Instance 6662da99-ba90-49ac-982e-d18f1c7834b6 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63515) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 537.009580] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Instance 29d199bc-aff3-4c62-af3a-1e6880b52446 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63515) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 537.177516] env[63515]: DEBUG nova.network.neutron [req-1667dbcb-c43c-42b4-bdeb-30c7c19d52b5 req-7b997c38-2e81-42b3-85d1-6b913e93314e service nova] [instance: d7fb514e-4ea9-4f32-9b00-06f728813aa3] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 537.246386] env[63515]: DEBUG oslo_concurrency.lockutils [None req-757a72d0-3dae-4296-830e-18ff5e948610 tempest-ServerDiagnosticsNegativeTest-1799249066 tempest-ServerDiagnosticsNegativeTest-1799249066-project-member] Acquiring lock "1c5d87c7-6747-45a7-8488-6f107eeaeb78" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 537.246598] env[63515]: DEBUG oslo_concurrency.lockutils [None req-757a72d0-3dae-4296-830e-18ff5e948610 tempest-ServerDiagnosticsNegativeTest-1799249066 tempest-ServerDiagnosticsNegativeTest-1799249066-project-member] Lock "1c5d87c7-6747-45a7-8488-6f107eeaeb78" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 537.280029] env[63515]: DEBUG nova.network.neutron [-] [instance: 9f7ad2dd-b882-4d98-8576-708b3abfa505] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 537.314809] env[63515]: DEBUG oslo_concurrency.lockutils [None req-919af50b-07d1-4794-942f-d99fa56afe84 tempest-ServerExternalEventsTest-1976219582 tempest-ServerExternalEventsTest-1976219582-project-member] Releasing lock "refresh_cache-6662da99-ba90-49ac-982e-d18f1c7834b6" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 537.315259] env[63515]: DEBUG nova.compute.manager [None req-919af50b-07d1-4794-942f-d99fa56afe84 tempest-ServerExternalEventsTest-1976219582 tempest-ServerExternalEventsTest-1976219582-project-member] [instance: 6662da99-ba90-49ac-982e-d18f1c7834b6] Start destroying the instance on the hypervisor. {{(pid=63515) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 537.315461] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-919af50b-07d1-4794-942f-d99fa56afe84 tempest-ServerExternalEventsTest-1976219582 tempest-ServerExternalEventsTest-1976219582-project-member] [instance: 6662da99-ba90-49ac-982e-d18f1c7834b6] Destroying instance {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 537.315765] env[63515]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-16ac0e87-40e2-4f79-9f9b-5c241593a037 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 537.327496] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34c3c6b6-54a9-49e3-920e-3f0ea37dbcf7 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 537.339399] env[63515]: DEBUG nova.network.neutron [req-1667dbcb-c43c-42b4-bdeb-30c7c19d52b5 req-7b997c38-2e81-42b3-85d1-6b913e93314e service nova] [instance: d7fb514e-4ea9-4f32-9b00-06f728813aa3] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 537.353865] env[63515]: WARNING nova.virt.vmwareapi.vmops [None req-919af50b-07d1-4794-942f-d99fa56afe84 tempest-ServerExternalEventsTest-1976219582 tempest-ServerExternalEventsTest-1976219582-project-member] [instance: 6662da99-ba90-49ac-982e-d18f1c7834b6] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 6662da99-ba90-49ac-982e-d18f1c7834b6 could not be found. [ 537.354127] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-919af50b-07d1-4794-942f-d99fa56afe84 tempest-ServerExternalEventsTest-1976219582 tempest-ServerExternalEventsTest-1976219582-project-member] [instance: 6662da99-ba90-49ac-982e-d18f1c7834b6] Instance destroyed {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 537.354311] env[63515]: INFO nova.compute.manager [None req-919af50b-07d1-4794-942f-d99fa56afe84 tempest-ServerExternalEventsTest-1976219582 tempest-ServerExternalEventsTest-1976219582-project-member] [instance: 6662da99-ba90-49ac-982e-d18f1c7834b6] Took 0.04 seconds to destroy the instance on the hypervisor. [ 537.354554] env[63515]: DEBUG oslo.service.loopingcall [None req-919af50b-07d1-4794-942f-d99fa56afe84 tempest-ServerExternalEventsTest-1976219582 tempest-ServerExternalEventsTest-1976219582-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 537.355594] env[63515]: DEBUG nova.compute.manager [-] [instance: 6662da99-ba90-49ac-982e-d18f1c7834b6] Deallocating network for instance {{(pid=63515) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 537.355647] env[63515]: DEBUG nova.network.neutron [-] [instance: 6662da99-ba90-49ac-982e-d18f1c7834b6] deallocate_for_instance() {{(pid=63515) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 537.383759] env[63515]: DEBUG nova.network.neutron [-] [instance: 6662da99-ba90-49ac-982e-d18f1c7834b6] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 537.442897] env[63515]: DEBUG nova.network.neutron [None req-c96f19aa-9abc-4b12-99f8-980442cab640 tempest-ServersAdminTestJSON-710082096 tempest-ServersAdminTestJSON-710082096-project-member] [instance: 29d199bc-aff3-4c62-af3a-1e6880b52446] Successfully created port: 75dd05f7-1fd2-44b3-858d-74b76d1a0780 {{(pid=63515) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 537.512145] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Instance d7fed938-f7cd-4013-b545-5d5350242084 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63515) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 537.749376] env[63515]: DEBUG nova.compute.manager [None req-757a72d0-3dae-4296-830e-18ff5e948610 tempest-ServerDiagnosticsNegativeTest-1799249066 tempest-ServerDiagnosticsNegativeTest-1799249066-project-member] [instance: 1c5d87c7-6747-45a7-8488-6f107eeaeb78] Starting instance... {{(pid=63515) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 537.783087] env[63515]: INFO nova.compute.manager [-] [instance: 9f7ad2dd-b882-4d98-8576-708b3abfa505] Took 1.09 seconds to deallocate network for instance. [ 537.788287] env[63515]: DEBUG nova.compute.claims [None req-0d00ff28-5b1e-46b5-9fb9-df8b8bcd6a1a tempest-TenantUsagesTestJSON-427870877 tempest-TenantUsagesTestJSON-427870877-project-member] [instance: 9f7ad2dd-b882-4d98-8576-708b3abfa505] Aborting claim: {{(pid=63515) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 537.788287] env[63515]: DEBUG oslo_concurrency.lockutils [None req-0d00ff28-5b1e-46b5-9fb9-df8b8bcd6a1a tempest-TenantUsagesTestJSON-427870877 tempest-TenantUsagesTestJSON-427870877-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 537.842945] env[63515]: DEBUG oslo_concurrency.lockutils [req-1667dbcb-c43c-42b4-bdeb-30c7c19d52b5 req-7b997c38-2e81-42b3-85d1-6b913e93314e service nova] Releasing lock "refresh_cache-d7fb514e-4ea9-4f32-9b00-06f728813aa3" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 537.889794] env[63515]: DEBUG nova.network.neutron [-] [instance: 6662da99-ba90-49ac-982e-d18f1c7834b6] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 537.902602] env[63515]: DEBUG nova.compute.manager [None req-c96f19aa-9abc-4b12-99f8-980442cab640 tempest-ServersAdminTestJSON-710082096 tempest-ServersAdminTestJSON-710082096-project-member] [instance: 29d199bc-aff3-4c62-af3a-1e6880b52446] Start spawning the instance on the hypervisor. {{(pid=63515) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 537.935481] env[63515]: DEBUG nova.virt.hardware [None req-c96f19aa-9abc-4b12-99f8-980442cab640 tempest-ServersAdminTestJSON-710082096 tempest-ServersAdminTestJSON-710082096-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T02:52:14Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T02:51:56Z,direct_url=,disk_format='vmdk',id=8a120570-cb06-4099-b262-554ca0ad15c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b569255cc43e42e7ae3f2b4ad37c6ef4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T02:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 537.935812] env[63515]: DEBUG nova.virt.hardware [None req-c96f19aa-9abc-4b12-99f8-980442cab640 tempest-ServersAdminTestJSON-710082096 tempest-ServersAdminTestJSON-710082096-project-member] Flavor limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 537.935937] env[63515]: DEBUG nova.virt.hardware [None req-c96f19aa-9abc-4b12-99f8-980442cab640 tempest-ServersAdminTestJSON-710082096 tempest-ServersAdminTestJSON-710082096-project-member] Image limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 537.936149] env[63515]: DEBUG nova.virt.hardware [None req-c96f19aa-9abc-4b12-99f8-980442cab640 tempest-ServersAdminTestJSON-710082096 tempest-ServersAdminTestJSON-710082096-project-member] Flavor pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 537.936277] env[63515]: DEBUG nova.virt.hardware [None req-c96f19aa-9abc-4b12-99f8-980442cab640 tempest-ServersAdminTestJSON-710082096 tempest-ServersAdminTestJSON-710082096-project-member] Image pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 537.936277] env[63515]: DEBUG nova.virt.hardware [None req-c96f19aa-9abc-4b12-99f8-980442cab640 tempest-ServersAdminTestJSON-710082096 tempest-ServersAdminTestJSON-710082096-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 537.936839] env[63515]: DEBUG nova.virt.hardware [None req-c96f19aa-9abc-4b12-99f8-980442cab640 tempest-ServersAdminTestJSON-710082096 tempest-ServersAdminTestJSON-710082096-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 537.937450] env[63515]: DEBUG nova.virt.hardware [None req-c96f19aa-9abc-4b12-99f8-980442cab640 tempest-ServersAdminTestJSON-710082096 tempest-ServersAdminTestJSON-710082096-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 537.937535] env[63515]: DEBUG nova.virt.hardware [None req-c96f19aa-9abc-4b12-99f8-980442cab640 tempest-ServersAdminTestJSON-710082096 tempest-ServersAdminTestJSON-710082096-project-member] Got 1 possible topologies {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 537.937710] env[63515]: DEBUG nova.virt.hardware [None req-c96f19aa-9abc-4b12-99f8-980442cab640 tempest-ServersAdminTestJSON-710082096 tempest-ServersAdminTestJSON-710082096-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 537.937921] env[63515]: DEBUG nova.virt.hardware [None req-c96f19aa-9abc-4b12-99f8-980442cab640 tempest-ServersAdminTestJSON-710082096 tempest-ServersAdminTestJSON-710082096-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 537.939380] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5060d9fa-94d9-4690-9d22-a5493b0df937 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 537.949085] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-261787f1-a2f4-4b4c-abeb-2594f6ae36b1 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 538.018153] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Instance aec66939-7ebc-4bfc-bd8b-bae22e011239 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63515) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 538.129951] env[63515]: DEBUG oslo_concurrency.lockutils [None req-a5c600bc-b1c1-4a6d-9386-2b0dbd0395e4 tempest-InstanceActionsNegativeTestJSON-1061405461 tempest-InstanceActionsNegativeTestJSON-1061405461-project-member] Acquiring lock "b641d2d1-de70-48b6-9137-fe18c3bc7511" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 538.130234] env[63515]: DEBUG oslo_concurrency.lockutils [None req-a5c600bc-b1c1-4a6d-9386-2b0dbd0395e4 tempest-InstanceActionsNegativeTestJSON-1061405461 tempest-InstanceActionsNegativeTestJSON-1061405461-project-member] Lock "b641d2d1-de70-48b6-9137-fe18c3bc7511" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 538.274178] env[63515]: DEBUG oslo_concurrency.lockutils [None req-757a72d0-3dae-4296-830e-18ff5e948610 tempest-ServerDiagnosticsNegativeTest-1799249066 tempest-ServerDiagnosticsNegativeTest-1799249066-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 538.392732] env[63515]: INFO nova.compute.manager [-] [instance: 6662da99-ba90-49ac-982e-d18f1c7834b6] Took 1.04 seconds to deallocate network for instance. [ 538.395381] env[63515]: DEBUG nova.compute.claims [None req-919af50b-07d1-4794-942f-d99fa56afe84 tempest-ServerExternalEventsTest-1976219582 tempest-ServerExternalEventsTest-1976219582-project-member] [instance: 6662da99-ba90-49ac-982e-d18f1c7834b6] Aborting claim: {{(pid=63515) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 538.396182] env[63515]: DEBUG oslo_concurrency.lockutils [None req-919af50b-07d1-4794-942f-d99fa56afe84 tempest-ServerExternalEventsTest-1976219582 tempest-ServerExternalEventsTest-1976219582-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 538.522421] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Instance 1c5d87c7-6747-45a7-8488-6f107eeaeb78 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63515) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 538.522726] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Total usable vcpus: 48, total allocated vcpus: 4 {{(pid=63515) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 538.522815] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1280MB phys_disk=200GB used_disk=4GB total_vcpus=48 used_vcpus=4 pci_stats=[] {{(pid=63515) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 538.634650] env[63515]: DEBUG nova.compute.manager [None req-a5c600bc-b1c1-4a6d-9386-2b0dbd0395e4 tempest-InstanceActionsNegativeTestJSON-1061405461 tempest-InstanceActionsNegativeTestJSON-1061405461-project-member] [instance: b641d2d1-de70-48b6-9137-fe18c3bc7511] Starting instance... {{(pid=63515) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 538.709428] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b886493e-1474-43be-8601-91318ad4d99e {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 538.724639] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-149ceeba-66ec-4f82-aeba-6ffc1d367da1 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 538.761836] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fcfbf658-960e-45a4-af0c-936ec646ec5d {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 538.770671] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c07415cf-136a-42bb-82ef-c2e4eb0ef892 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 538.787277] env[63515]: DEBUG nova.compute.provider_tree [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 539.164814] env[63515]: DEBUG oslo_concurrency.lockutils [None req-a5c600bc-b1c1-4a6d-9386-2b0dbd0395e4 tempest-InstanceActionsNegativeTestJSON-1061405461 tempest-InstanceActionsNegativeTestJSON-1061405461-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 539.293449] env[63515]: DEBUG nova.scheduler.client.report [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 172, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 539.488388] env[63515]: DEBUG oslo_concurrency.lockutils [None req-50ea3e37-671c-4a4b-9c1a-d782ff1b578c tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Acquiring lock "cfc6c8d5-c9e8-46fb-99e7-d245664f9652" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 539.488607] env[63515]: DEBUG oslo_concurrency.lockutils [None req-50ea3e37-671c-4a4b-9c1a-d782ff1b578c tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Lock "cfc6c8d5-c9e8-46fb-99e7-d245664f9652" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 539.791332] env[63515]: DEBUG nova.compute.manager [req-8ddcb176-6597-4dc2-90a9-1c825f36f084 req-9eb9f996-3f74-43c6-be18-d49b413bfb82 service nova] [instance: d7fb514e-4ea9-4f32-9b00-06f728813aa3] Received event network-vif-deleted-71cd34ac-f70e-44ba-856e-b65afc2b509f {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 539.791578] env[63515]: DEBUG nova.compute.manager [req-8ddcb176-6597-4dc2-90a9-1c825f36f084 req-9eb9f996-3f74-43c6-be18-d49b413bfb82 service nova] [instance: 9f7ad2dd-b882-4d98-8576-708b3abfa505] Received event network-changed-111c82a0-4d47-4505-8e27-6281465572af {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 539.791619] env[63515]: DEBUG nova.compute.manager [req-8ddcb176-6597-4dc2-90a9-1c825f36f084 req-9eb9f996-3f74-43c6-be18-d49b413bfb82 service nova] [instance: 9f7ad2dd-b882-4d98-8576-708b3abfa505] Refreshing instance network info cache due to event network-changed-111c82a0-4d47-4505-8e27-6281465572af. {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 539.791876] env[63515]: DEBUG oslo_concurrency.lockutils [req-8ddcb176-6597-4dc2-90a9-1c825f36f084 req-9eb9f996-3f74-43c6-be18-d49b413bfb82 service nova] Acquiring lock "refresh_cache-9f7ad2dd-b882-4d98-8576-708b3abfa505" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 539.792073] env[63515]: DEBUG oslo_concurrency.lockutils [req-8ddcb176-6597-4dc2-90a9-1c825f36f084 req-9eb9f996-3f74-43c6-be18-d49b413bfb82 service nova] Acquired lock "refresh_cache-9f7ad2dd-b882-4d98-8576-708b3abfa505" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 539.792264] env[63515]: DEBUG nova.network.neutron [req-8ddcb176-6597-4dc2-90a9-1c825f36f084 req-9eb9f996-3f74-43c6-be18-d49b413bfb82 service nova] [instance: 9f7ad2dd-b882-4d98-8576-708b3abfa505] Refreshing network info cache for port 111c82a0-4d47-4505-8e27-6281465572af {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 539.798196] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=63515) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 539.798196] env[63515]: DEBUG oslo_concurrency.lockutils [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 3.830s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 539.798502] env[63515]: DEBUG oslo_concurrency.lockutils [None req-ada3051f-c750-48c7-b591-0344c25b8fea tempest-ServersAdminTestJSON-710082096 tempest-ServersAdminTestJSON-710082096-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.333s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 539.800445] env[63515]: INFO nova.compute.claims [None req-ada3051f-c750-48c7-b591-0344c25b8fea tempest-ServersAdminTestJSON-710082096 tempest-ServersAdminTestJSON-710082096-project-member] [instance: d7fed938-f7cd-4013-b545-5d5350242084] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 539.834244] env[63515]: ERROR nova.compute.manager [None req-c96f19aa-9abc-4b12-99f8-980442cab640 tempest-ServersAdminTestJSON-710082096 tempest-ServersAdminTestJSON-710082096-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 75dd05f7-1fd2-44b3-858d-74b76d1a0780, please check neutron logs for more information. [ 539.834244] env[63515]: ERROR nova.compute.manager Traceback (most recent call last): [ 539.834244] env[63515]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 539.834244] env[63515]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 539.834244] env[63515]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 539.834244] env[63515]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 539.834244] env[63515]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 539.834244] env[63515]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 539.834244] env[63515]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 539.834244] env[63515]: ERROR nova.compute.manager self.force_reraise() [ 539.834244] env[63515]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 539.834244] env[63515]: ERROR nova.compute.manager raise self.value [ 539.834244] env[63515]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 539.834244] env[63515]: ERROR nova.compute.manager updated_port = self._update_port( [ 539.834244] env[63515]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 539.834244] env[63515]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 539.834715] env[63515]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 539.834715] env[63515]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 539.834715] env[63515]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 75dd05f7-1fd2-44b3-858d-74b76d1a0780, please check neutron logs for more information. [ 539.834715] env[63515]: ERROR nova.compute.manager [ 539.834715] env[63515]: Traceback (most recent call last): [ 539.834715] env[63515]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 539.834715] env[63515]: listener.cb(fileno) [ 539.834715] env[63515]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 539.834715] env[63515]: result = function(*args, **kwargs) [ 539.834715] env[63515]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 539.834715] env[63515]: return func(*args, **kwargs) [ 539.834715] env[63515]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 539.834715] env[63515]: raise e [ 539.834715] env[63515]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 539.834715] env[63515]: nwinfo = self.network_api.allocate_for_instance( [ 539.834715] env[63515]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 539.834715] env[63515]: created_port_ids = self._update_ports_for_instance( [ 539.834715] env[63515]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 539.834715] env[63515]: with excutils.save_and_reraise_exception(): [ 539.834715] env[63515]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 539.834715] env[63515]: self.force_reraise() [ 539.834715] env[63515]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 539.834715] env[63515]: raise self.value [ 539.834715] env[63515]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 539.834715] env[63515]: updated_port = self._update_port( [ 539.834715] env[63515]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 539.834715] env[63515]: _ensure_no_port_binding_failure(port) [ 539.834715] env[63515]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 539.834715] env[63515]: raise exception.PortBindingFailed(port_id=port['id']) [ 539.835458] env[63515]: nova.exception.PortBindingFailed: Binding failed for port 75dd05f7-1fd2-44b3-858d-74b76d1a0780, please check neutron logs for more information. [ 539.835458] env[63515]: Removing descriptor: 16 [ 539.835458] env[63515]: ERROR nova.compute.manager [None req-c96f19aa-9abc-4b12-99f8-980442cab640 tempest-ServersAdminTestJSON-710082096 tempest-ServersAdminTestJSON-710082096-project-member] [instance: 29d199bc-aff3-4c62-af3a-1e6880b52446] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 75dd05f7-1fd2-44b3-858d-74b76d1a0780, please check neutron logs for more information. [ 539.835458] env[63515]: ERROR nova.compute.manager [instance: 29d199bc-aff3-4c62-af3a-1e6880b52446] Traceback (most recent call last): [ 539.835458] env[63515]: ERROR nova.compute.manager [instance: 29d199bc-aff3-4c62-af3a-1e6880b52446] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 539.835458] env[63515]: ERROR nova.compute.manager [instance: 29d199bc-aff3-4c62-af3a-1e6880b52446] yield resources [ 539.835458] env[63515]: ERROR nova.compute.manager [instance: 29d199bc-aff3-4c62-af3a-1e6880b52446] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 539.835458] env[63515]: ERROR nova.compute.manager [instance: 29d199bc-aff3-4c62-af3a-1e6880b52446] self.driver.spawn(context, instance, image_meta, [ 539.835458] env[63515]: ERROR nova.compute.manager [instance: 29d199bc-aff3-4c62-af3a-1e6880b52446] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 539.835458] env[63515]: ERROR nova.compute.manager [instance: 29d199bc-aff3-4c62-af3a-1e6880b52446] self._vmops.spawn(context, instance, image_meta, injected_files, [ 539.835458] env[63515]: ERROR nova.compute.manager [instance: 29d199bc-aff3-4c62-af3a-1e6880b52446] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 539.835458] env[63515]: ERROR nova.compute.manager [instance: 29d199bc-aff3-4c62-af3a-1e6880b52446] vm_ref = self.build_virtual_machine(instance, [ 539.835788] env[63515]: ERROR nova.compute.manager [instance: 29d199bc-aff3-4c62-af3a-1e6880b52446] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 539.835788] env[63515]: ERROR nova.compute.manager [instance: 29d199bc-aff3-4c62-af3a-1e6880b52446] vif_infos = vmwarevif.get_vif_info(self._session, [ 539.835788] env[63515]: ERROR nova.compute.manager [instance: 29d199bc-aff3-4c62-af3a-1e6880b52446] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 539.835788] env[63515]: ERROR nova.compute.manager [instance: 29d199bc-aff3-4c62-af3a-1e6880b52446] for vif in network_info: [ 539.835788] env[63515]: ERROR nova.compute.manager [instance: 29d199bc-aff3-4c62-af3a-1e6880b52446] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 539.835788] env[63515]: ERROR nova.compute.manager [instance: 29d199bc-aff3-4c62-af3a-1e6880b52446] return self._sync_wrapper(fn, *args, **kwargs) [ 539.835788] env[63515]: ERROR nova.compute.manager [instance: 29d199bc-aff3-4c62-af3a-1e6880b52446] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 539.835788] env[63515]: ERROR nova.compute.manager [instance: 29d199bc-aff3-4c62-af3a-1e6880b52446] self.wait() [ 539.835788] env[63515]: ERROR nova.compute.manager [instance: 29d199bc-aff3-4c62-af3a-1e6880b52446] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 539.835788] env[63515]: ERROR nova.compute.manager [instance: 29d199bc-aff3-4c62-af3a-1e6880b52446] self[:] = self._gt.wait() [ 539.835788] env[63515]: ERROR nova.compute.manager [instance: 29d199bc-aff3-4c62-af3a-1e6880b52446] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 539.835788] env[63515]: ERROR nova.compute.manager [instance: 29d199bc-aff3-4c62-af3a-1e6880b52446] return self._exit_event.wait() [ 539.835788] env[63515]: ERROR nova.compute.manager [instance: 29d199bc-aff3-4c62-af3a-1e6880b52446] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 539.836210] env[63515]: ERROR nova.compute.manager [instance: 29d199bc-aff3-4c62-af3a-1e6880b52446] result = hub.switch() [ 539.836210] env[63515]: ERROR nova.compute.manager [instance: 29d199bc-aff3-4c62-af3a-1e6880b52446] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 539.836210] env[63515]: ERROR nova.compute.manager [instance: 29d199bc-aff3-4c62-af3a-1e6880b52446] return self.greenlet.switch() [ 539.836210] env[63515]: ERROR nova.compute.manager [instance: 29d199bc-aff3-4c62-af3a-1e6880b52446] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 539.836210] env[63515]: ERROR nova.compute.manager [instance: 29d199bc-aff3-4c62-af3a-1e6880b52446] result = function(*args, **kwargs) [ 539.836210] env[63515]: ERROR nova.compute.manager [instance: 29d199bc-aff3-4c62-af3a-1e6880b52446] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 539.836210] env[63515]: ERROR nova.compute.manager [instance: 29d199bc-aff3-4c62-af3a-1e6880b52446] return func(*args, **kwargs) [ 539.836210] env[63515]: ERROR nova.compute.manager [instance: 29d199bc-aff3-4c62-af3a-1e6880b52446] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 539.836210] env[63515]: ERROR nova.compute.manager [instance: 29d199bc-aff3-4c62-af3a-1e6880b52446] raise e [ 539.836210] env[63515]: ERROR nova.compute.manager [instance: 29d199bc-aff3-4c62-af3a-1e6880b52446] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 539.836210] env[63515]: ERROR nova.compute.manager [instance: 29d199bc-aff3-4c62-af3a-1e6880b52446] nwinfo = self.network_api.allocate_for_instance( [ 539.836210] env[63515]: ERROR nova.compute.manager [instance: 29d199bc-aff3-4c62-af3a-1e6880b52446] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 539.836210] env[63515]: ERROR nova.compute.manager [instance: 29d199bc-aff3-4c62-af3a-1e6880b52446] created_port_ids = self._update_ports_for_instance( [ 539.836593] env[63515]: ERROR nova.compute.manager [instance: 29d199bc-aff3-4c62-af3a-1e6880b52446] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 539.836593] env[63515]: ERROR nova.compute.manager [instance: 29d199bc-aff3-4c62-af3a-1e6880b52446] with excutils.save_and_reraise_exception(): [ 539.836593] env[63515]: ERROR nova.compute.manager [instance: 29d199bc-aff3-4c62-af3a-1e6880b52446] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 539.836593] env[63515]: ERROR nova.compute.manager [instance: 29d199bc-aff3-4c62-af3a-1e6880b52446] self.force_reraise() [ 539.836593] env[63515]: ERROR nova.compute.manager [instance: 29d199bc-aff3-4c62-af3a-1e6880b52446] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 539.836593] env[63515]: ERROR nova.compute.manager [instance: 29d199bc-aff3-4c62-af3a-1e6880b52446] raise self.value [ 539.836593] env[63515]: ERROR nova.compute.manager [instance: 29d199bc-aff3-4c62-af3a-1e6880b52446] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 539.836593] env[63515]: ERROR nova.compute.manager [instance: 29d199bc-aff3-4c62-af3a-1e6880b52446] updated_port = self._update_port( [ 539.836593] env[63515]: ERROR nova.compute.manager [instance: 29d199bc-aff3-4c62-af3a-1e6880b52446] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 539.836593] env[63515]: ERROR nova.compute.manager [instance: 29d199bc-aff3-4c62-af3a-1e6880b52446] _ensure_no_port_binding_failure(port) [ 539.836593] env[63515]: ERROR nova.compute.manager [instance: 29d199bc-aff3-4c62-af3a-1e6880b52446] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 539.836593] env[63515]: ERROR nova.compute.manager [instance: 29d199bc-aff3-4c62-af3a-1e6880b52446] raise exception.PortBindingFailed(port_id=port['id']) [ 539.836904] env[63515]: ERROR nova.compute.manager [instance: 29d199bc-aff3-4c62-af3a-1e6880b52446] nova.exception.PortBindingFailed: Binding failed for port 75dd05f7-1fd2-44b3-858d-74b76d1a0780, please check neutron logs for more information. [ 539.836904] env[63515]: ERROR nova.compute.manager [instance: 29d199bc-aff3-4c62-af3a-1e6880b52446] [ 539.836904] env[63515]: INFO nova.compute.manager [None req-c96f19aa-9abc-4b12-99f8-980442cab640 tempest-ServersAdminTestJSON-710082096 tempest-ServersAdminTestJSON-710082096-project-member] [instance: 29d199bc-aff3-4c62-af3a-1e6880b52446] Terminating instance [ 539.841847] env[63515]: DEBUG oslo_concurrency.lockutils [None req-c96f19aa-9abc-4b12-99f8-980442cab640 tempest-ServersAdminTestJSON-710082096 tempest-ServersAdminTestJSON-710082096-project-member] Acquiring lock "refresh_cache-29d199bc-aff3-4c62-af3a-1e6880b52446" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 539.841847] env[63515]: DEBUG oslo_concurrency.lockutils [None req-c96f19aa-9abc-4b12-99f8-980442cab640 tempest-ServersAdminTestJSON-710082096 tempest-ServersAdminTestJSON-710082096-project-member] Acquired lock "refresh_cache-29d199bc-aff3-4c62-af3a-1e6880b52446" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 539.841847] env[63515]: DEBUG nova.network.neutron [None req-c96f19aa-9abc-4b12-99f8-980442cab640 tempest-ServersAdminTestJSON-710082096 tempest-ServersAdminTestJSON-710082096-project-member] [instance: 29d199bc-aff3-4c62-af3a-1e6880b52446] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 539.991125] env[63515]: DEBUG nova.compute.manager [None req-50ea3e37-671c-4a4b-9c1a-d782ff1b578c tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] [instance: cfc6c8d5-c9e8-46fb-99e7-d245664f9652] Starting instance... {{(pid=63515) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 540.338141] env[63515]: DEBUG nova.network.neutron [req-8ddcb176-6597-4dc2-90a9-1c825f36f084 req-9eb9f996-3f74-43c6-be18-d49b413bfb82 service nova] [instance: 9f7ad2dd-b882-4d98-8576-708b3abfa505] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 540.410070] env[63515]: DEBUG nova.network.neutron [None req-c96f19aa-9abc-4b12-99f8-980442cab640 tempest-ServersAdminTestJSON-710082096 tempest-ServersAdminTestJSON-710082096-project-member] [instance: 29d199bc-aff3-4c62-af3a-1e6880b52446] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 540.523926] env[63515]: DEBUG oslo_concurrency.lockutils [None req-50ea3e37-671c-4a4b-9c1a-d782ff1b578c tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 540.572153] env[63515]: DEBUG nova.network.neutron [req-8ddcb176-6597-4dc2-90a9-1c825f36f084 req-9eb9f996-3f74-43c6-be18-d49b413bfb82 service nova] [instance: 9f7ad2dd-b882-4d98-8576-708b3abfa505] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 540.712091] env[63515]: DEBUG nova.network.neutron [None req-c96f19aa-9abc-4b12-99f8-980442cab640 tempest-ServersAdminTestJSON-710082096 tempest-ServersAdminTestJSON-710082096-project-member] [instance: 29d199bc-aff3-4c62-af3a-1e6880b52446] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 541.024886] env[63515]: DEBUG oslo_concurrency.lockutils [None req-44598345-8888-488a-9c24-527daf4d66df tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] Acquiring lock "93b28142-8454-43c0-b0a7-d61aa95c8fc8" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 541.025204] env[63515]: DEBUG oslo_concurrency.lockutils [None req-44598345-8888-488a-9c24-527daf4d66df tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] Lock "93b28142-8454-43c0-b0a7-d61aa95c8fc8" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 541.065140] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df5d3ddd-342e-4bd8-9c97-bb3bea3202e0 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.074115] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ced23880-1e3b-4004-9007-82d0885c9bfd {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.078239] env[63515]: DEBUG oslo_concurrency.lockutils [req-8ddcb176-6597-4dc2-90a9-1c825f36f084 req-9eb9f996-3f74-43c6-be18-d49b413bfb82 service nova] Releasing lock "refresh_cache-9f7ad2dd-b882-4d98-8576-708b3abfa505" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 541.078239] env[63515]: DEBUG nova.compute.manager [req-8ddcb176-6597-4dc2-90a9-1c825f36f084 req-9eb9f996-3f74-43c6-be18-d49b413bfb82 service nova] [instance: 9f7ad2dd-b882-4d98-8576-708b3abfa505] Received event network-vif-deleted-111c82a0-4d47-4505-8e27-6281465572af {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 541.078239] env[63515]: DEBUG nova.compute.manager [req-8ddcb176-6597-4dc2-90a9-1c825f36f084 req-9eb9f996-3f74-43c6-be18-d49b413bfb82 service nova] [instance: 6662da99-ba90-49ac-982e-d18f1c7834b6] Received event network-changed-39b0bc20-a4b4-4683-ac31-11d4a2fa8333 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 541.078239] env[63515]: DEBUG nova.compute.manager [req-8ddcb176-6597-4dc2-90a9-1c825f36f084 req-9eb9f996-3f74-43c6-be18-d49b413bfb82 service nova] [instance: 6662da99-ba90-49ac-982e-d18f1c7834b6] Refreshing instance network info cache due to event network-changed-39b0bc20-a4b4-4683-ac31-11d4a2fa8333. {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 541.078870] env[63515]: DEBUG oslo_concurrency.lockutils [req-8ddcb176-6597-4dc2-90a9-1c825f36f084 req-9eb9f996-3f74-43c6-be18-d49b413bfb82 service nova] Acquiring lock "refresh_cache-6662da99-ba90-49ac-982e-d18f1c7834b6" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 541.080172] env[63515]: DEBUG oslo_concurrency.lockutils [req-8ddcb176-6597-4dc2-90a9-1c825f36f084 req-9eb9f996-3f74-43c6-be18-d49b413bfb82 service nova] Acquired lock "refresh_cache-6662da99-ba90-49ac-982e-d18f1c7834b6" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 541.080172] env[63515]: DEBUG nova.network.neutron [req-8ddcb176-6597-4dc2-90a9-1c825f36f084 req-9eb9f996-3f74-43c6-be18-d49b413bfb82 service nova] [instance: 6662da99-ba90-49ac-982e-d18f1c7834b6] Refreshing network info cache for port 39b0bc20-a4b4-4683-ac31-11d4a2fa8333 {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 541.122582] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a3d2b02-6120-4490-b942-e8d4ac04d2e9 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.130683] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13f79ef6-dc75-4f6c-9c11-86e85878a4f7 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.146657] env[63515]: DEBUG nova.compute.provider_tree [None req-ada3051f-c750-48c7-b591-0344c25b8fea tempest-ServersAdminTestJSON-710082096 tempest-ServersAdminTestJSON-710082096-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 541.214470] env[63515]: DEBUG oslo_concurrency.lockutils [None req-c96f19aa-9abc-4b12-99f8-980442cab640 tempest-ServersAdminTestJSON-710082096 tempest-ServersAdminTestJSON-710082096-project-member] Releasing lock "refresh_cache-29d199bc-aff3-4c62-af3a-1e6880b52446" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 541.215249] env[63515]: DEBUG nova.compute.manager [None req-c96f19aa-9abc-4b12-99f8-980442cab640 tempest-ServersAdminTestJSON-710082096 tempest-ServersAdminTestJSON-710082096-project-member] [instance: 29d199bc-aff3-4c62-af3a-1e6880b52446] Start destroying the instance on the hypervisor. {{(pid=63515) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 541.215249] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-c96f19aa-9abc-4b12-99f8-980442cab640 tempest-ServersAdminTestJSON-710082096 tempest-ServersAdminTestJSON-710082096-project-member] [instance: 29d199bc-aff3-4c62-af3a-1e6880b52446] Destroying instance {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 541.215404] env[63515]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-3ef9bd9d-00a5-4611-9e72-a9a0c50a7582 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.227030] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-288382fc-a762-44ed-81b4-f9ed2b5e4d4f {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.253417] env[63515]: WARNING nova.virt.vmwareapi.vmops [None req-c96f19aa-9abc-4b12-99f8-980442cab640 tempest-ServersAdminTestJSON-710082096 tempest-ServersAdminTestJSON-710082096-project-member] [instance: 29d199bc-aff3-4c62-af3a-1e6880b52446] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 29d199bc-aff3-4c62-af3a-1e6880b52446 could not be found. [ 541.253506] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-c96f19aa-9abc-4b12-99f8-980442cab640 tempest-ServersAdminTestJSON-710082096 tempest-ServersAdminTestJSON-710082096-project-member] [instance: 29d199bc-aff3-4c62-af3a-1e6880b52446] Instance destroyed {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 541.253708] env[63515]: INFO nova.compute.manager [None req-c96f19aa-9abc-4b12-99f8-980442cab640 tempest-ServersAdminTestJSON-710082096 tempest-ServersAdminTestJSON-710082096-project-member] [instance: 29d199bc-aff3-4c62-af3a-1e6880b52446] Took 0.04 seconds to destroy the instance on the hypervisor. [ 541.254035] env[63515]: DEBUG oslo.service.loopingcall [None req-c96f19aa-9abc-4b12-99f8-980442cab640 tempest-ServersAdminTestJSON-710082096 tempest-ServersAdminTestJSON-710082096-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 541.254256] env[63515]: DEBUG nova.compute.manager [-] [instance: 29d199bc-aff3-4c62-af3a-1e6880b52446] Deallocating network for instance {{(pid=63515) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 541.254309] env[63515]: DEBUG nova.network.neutron [-] [instance: 29d199bc-aff3-4c62-af3a-1e6880b52446] deallocate_for_instance() {{(pid=63515) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 541.299352] env[63515]: DEBUG nova.network.neutron [-] [instance: 29d199bc-aff3-4c62-af3a-1e6880b52446] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 541.531204] env[63515]: DEBUG nova.compute.manager [None req-44598345-8888-488a-9c24-527daf4d66df tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] [instance: 93b28142-8454-43c0-b0a7-d61aa95c8fc8] Starting instance... {{(pid=63515) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 541.620327] env[63515]: DEBUG nova.network.neutron [req-8ddcb176-6597-4dc2-90a9-1c825f36f084 req-9eb9f996-3f74-43c6-be18-d49b413bfb82 service nova] [instance: 6662da99-ba90-49ac-982e-d18f1c7834b6] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 541.657068] env[63515]: DEBUG nova.scheduler.client.report [None req-ada3051f-c750-48c7-b591-0344c25b8fea tempest-ServersAdminTestJSON-710082096 tempest-ServersAdminTestJSON-710082096-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 172, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 541.716026] env[63515]: DEBUG nova.network.neutron [req-8ddcb176-6597-4dc2-90a9-1c825f36f084 req-9eb9f996-3f74-43c6-be18-d49b413bfb82 service nova] [instance: 6662da99-ba90-49ac-982e-d18f1c7834b6] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 541.801406] env[63515]: DEBUG nova.network.neutron [-] [instance: 29d199bc-aff3-4c62-af3a-1e6880b52446] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 542.059075] env[63515]: DEBUG oslo_concurrency.lockutils [None req-44598345-8888-488a-9c24-527daf4d66df tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 542.166142] env[63515]: DEBUG oslo_concurrency.lockutils [None req-ada3051f-c750-48c7-b591-0344c25b8fea tempest-ServersAdminTestJSON-710082096 tempest-ServersAdminTestJSON-710082096-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.367s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 542.166895] env[63515]: DEBUG nova.compute.manager [None req-ada3051f-c750-48c7-b591-0344c25b8fea tempest-ServersAdminTestJSON-710082096 tempest-ServersAdminTestJSON-710082096-project-member] [instance: d7fed938-f7cd-4013-b545-5d5350242084] Start building networks asynchronously for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 542.171017] env[63515]: DEBUG oslo_concurrency.lockutils [None req-9ee35a7a-94b5-4758-9c8f-a8f1e8f2f2cd tempest-DeleteServersAdminTestJSON-944542165 tempest-DeleteServersAdminTestJSON-944542165-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 5.566s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 542.222194] env[63515]: DEBUG oslo_concurrency.lockutils [req-8ddcb176-6597-4dc2-90a9-1c825f36f084 req-9eb9f996-3f74-43c6-be18-d49b413bfb82 service nova] Releasing lock "refresh_cache-6662da99-ba90-49ac-982e-d18f1c7834b6" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 542.222194] env[63515]: DEBUG nova.compute.manager [req-8ddcb176-6597-4dc2-90a9-1c825f36f084 req-9eb9f996-3f74-43c6-be18-d49b413bfb82 service nova] [instance: 6662da99-ba90-49ac-982e-d18f1c7834b6] Received event network-vif-deleted-39b0bc20-a4b4-4683-ac31-11d4a2fa8333 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 542.304757] env[63515]: INFO nova.compute.manager [-] [instance: 29d199bc-aff3-4c62-af3a-1e6880b52446] Took 1.05 seconds to deallocate network for instance. [ 542.310128] env[63515]: DEBUG nova.compute.claims [None req-c96f19aa-9abc-4b12-99f8-980442cab640 tempest-ServersAdminTestJSON-710082096 tempest-ServersAdminTestJSON-710082096-project-member] [instance: 29d199bc-aff3-4c62-af3a-1e6880b52446] Aborting claim: {{(pid=63515) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 542.311033] env[63515]: DEBUG oslo_concurrency.lockutils [None req-c96f19aa-9abc-4b12-99f8-980442cab640 tempest-ServersAdminTestJSON-710082096 tempest-ServersAdminTestJSON-710082096-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 542.311033] env[63515]: DEBUG oslo_concurrency.lockutils [None req-eaa7f189-a861-45f8-872f-29d870dc784b tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] Acquiring lock "e39bc55e-b677-4e2d-9aa4-7bc1d5ffadfa" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 542.311033] env[63515]: DEBUG oslo_concurrency.lockutils [None req-eaa7f189-a861-45f8-872f-29d870dc784b tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] Lock "e39bc55e-b677-4e2d-9aa4-7bc1d5ffadfa" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 542.674535] env[63515]: DEBUG nova.compute.utils [None req-ada3051f-c750-48c7-b591-0344c25b8fea tempest-ServersAdminTestJSON-710082096 tempest-ServersAdminTestJSON-710082096-project-member] Using /dev/sd instead of None {{(pid=63515) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 542.679356] env[63515]: DEBUG nova.compute.manager [None req-ada3051f-c750-48c7-b591-0344c25b8fea tempest-ServersAdminTestJSON-710082096 tempest-ServersAdminTestJSON-710082096-project-member] [instance: d7fed938-f7cd-4013-b545-5d5350242084] Allocating IP information in the background. {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 542.679986] env[63515]: DEBUG nova.network.neutron [None req-ada3051f-c750-48c7-b591-0344c25b8fea tempest-ServersAdminTestJSON-710082096 tempest-ServersAdminTestJSON-710082096-project-member] [instance: d7fed938-f7cd-4013-b545-5d5350242084] allocate_for_instance() {{(pid=63515) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 542.839049] env[63515]: DEBUG nova.policy [None req-ada3051f-c750-48c7-b591-0344c25b8fea tempest-ServersAdminTestJSON-710082096 tempest-ServersAdminTestJSON-710082096-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'de4cb162ff284f25916fdc058f001516', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '25b7ee3325094edf93fa3aeec09e8ba8', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63515) authorize /opt/stack/nova/nova/policy.py:201}} [ 543.028872] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a8bf1bb-32c8-4a29-8e98-a321a7eff0d9 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 543.036937] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-069565e7-b739-4d04-a38a-2b2a5b43f364 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 543.070980] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28bc6c7a-8bcd-4a9e-b0fb-2b9b73505248 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 543.078429] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-272dd75b-e7c7-4a98-ac98-6e762e9af0f9 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 543.092262] env[63515]: DEBUG nova.compute.provider_tree [None req-9ee35a7a-94b5-4758-9c8f-a8f1e8f2f2cd tempest-DeleteServersAdminTestJSON-944542165 tempest-DeleteServersAdminTestJSON-944542165-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 543.184023] env[63515]: DEBUG nova.compute.manager [None req-ada3051f-c750-48c7-b591-0344c25b8fea tempest-ServersAdminTestJSON-710082096 tempest-ServersAdminTestJSON-710082096-project-member] [instance: d7fed938-f7cd-4013-b545-5d5350242084] Start building block device mappings for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 543.416836] env[63515]: DEBUG oslo_concurrency.lockutils [None req-7f5043f7-e6d7-46dd-88fd-99ee049c2b4b tempest-VolumesAssistedSnapshotsTest-453349191 tempest-VolumesAssistedSnapshotsTest-453349191-project-member] Acquiring lock "db682708-36fa-4126-a848-bfb609df96d7" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 543.416996] env[63515]: DEBUG oslo_concurrency.lockutils [None req-7f5043f7-e6d7-46dd-88fd-99ee049c2b4b tempest-VolumesAssistedSnapshotsTest-453349191 tempest-VolumesAssistedSnapshotsTest-453349191-project-member] Lock "db682708-36fa-4126-a848-bfb609df96d7" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 543.477419] env[63515]: DEBUG nova.compute.manager [req-1d8f4c14-1483-4c94-884d-48d2a5bb57c2 req-528ffa5f-402f-436e-a11b-00b509fbbc96 service nova] [instance: 29d199bc-aff3-4c62-af3a-1e6880b52446] Received event network-changed-75dd05f7-1fd2-44b3-858d-74b76d1a0780 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 543.477598] env[63515]: DEBUG nova.compute.manager [req-1d8f4c14-1483-4c94-884d-48d2a5bb57c2 req-528ffa5f-402f-436e-a11b-00b509fbbc96 service nova] [instance: 29d199bc-aff3-4c62-af3a-1e6880b52446] Refreshing instance network info cache due to event network-changed-75dd05f7-1fd2-44b3-858d-74b76d1a0780. {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 543.477814] env[63515]: DEBUG oslo_concurrency.lockutils [req-1d8f4c14-1483-4c94-884d-48d2a5bb57c2 req-528ffa5f-402f-436e-a11b-00b509fbbc96 service nova] Acquiring lock "refresh_cache-29d199bc-aff3-4c62-af3a-1e6880b52446" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 543.477947] env[63515]: DEBUG oslo_concurrency.lockutils [req-1d8f4c14-1483-4c94-884d-48d2a5bb57c2 req-528ffa5f-402f-436e-a11b-00b509fbbc96 service nova] Acquired lock "refresh_cache-29d199bc-aff3-4c62-af3a-1e6880b52446" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 543.478124] env[63515]: DEBUG nova.network.neutron [req-1d8f4c14-1483-4c94-884d-48d2a5bb57c2 req-528ffa5f-402f-436e-a11b-00b509fbbc96 service nova] [instance: 29d199bc-aff3-4c62-af3a-1e6880b52446] Refreshing network info cache for port 75dd05f7-1fd2-44b3-858d-74b76d1a0780 {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 543.596059] env[63515]: DEBUG nova.scheduler.client.report [None req-9ee35a7a-94b5-4758-9c8f-a8f1e8f2f2cd tempest-DeleteServersAdminTestJSON-944542165 tempest-DeleteServersAdminTestJSON-944542165-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 172, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 543.844919] env[63515]: DEBUG oslo_concurrency.lockutils [None req-5e279136-7e72-43a2-85c4-9cd11158a72b tempest-ListImageFiltersTestJSON-1100101045 tempest-ListImageFiltersTestJSON-1100101045-project-member] Acquiring lock "48908cb0-a989-4bad-84a4-25a4bdd3baeb" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 543.845868] env[63515]: DEBUG oslo_concurrency.lockutils [None req-5e279136-7e72-43a2-85c4-9cd11158a72b tempest-ListImageFiltersTestJSON-1100101045 tempest-ListImageFiltersTestJSON-1100101045-project-member] Lock "48908cb0-a989-4bad-84a4-25a4bdd3baeb" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 544.012488] env[63515]: DEBUG nova.network.neutron [req-1d8f4c14-1483-4c94-884d-48d2a5bb57c2 req-528ffa5f-402f-436e-a11b-00b509fbbc96 service nova] [instance: 29d199bc-aff3-4c62-af3a-1e6880b52446] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 544.105489] env[63515]: DEBUG oslo_concurrency.lockutils [None req-9ee35a7a-94b5-4758-9c8f-a8f1e8f2f2cd tempest-DeleteServersAdminTestJSON-944542165 tempest-DeleteServersAdminTestJSON-944542165-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.936s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 544.106511] env[63515]: ERROR nova.compute.manager [None req-9ee35a7a-94b5-4758-9c8f-a8f1e8f2f2cd tempest-DeleteServersAdminTestJSON-944542165 tempest-DeleteServersAdminTestJSON-944542165-project-member] [instance: d7fb514e-4ea9-4f32-9b00-06f728813aa3] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 71cd34ac-f70e-44ba-856e-b65afc2b509f, please check neutron logs for more information. [ 544.106511] env[63515]: ERROR nova.compute.manager [instance: d7fb514e-4ea9-4f32-9b00-06f728813aa3] Traceback (most recent call last): [ 544.106511] env[63515]: ERROR nova.compute.manager [instance: d7fb514e-4ea9-4f32-9b00-06f728813aa3] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 544.106511] env[63515]: ERROR nova.compute.manager [instance: d7fb514e-4ea9-4f32-9b00-06f728813aa3] self.driver.spawn(context, instance, image_meta, [ 544.106511] env[63515]: ERROR nova.compute.manager [instance: d7fb514e-4ea9-4f32-9b00-06f728813aa3] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 544.106511] env[63515]: ERROR nova.compute.manager [instance: d7fb514e-4ea9-4f32-9b00-06f728813aa3] self._vmops.spawn(context, instance, image_meta, injected_files, [ 544.106511] env[63515]: ERROR nova.compute.manager [instance: d7fb514e-4ea9-4f32-9b00-06f728813aa3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 544.106511] env[63515]: ERROR nova.compute.manager [instance: d7fb514e-4ea9-4f32-9b00-06f728813aa3] vm_ref = self.build_virtual_machine(instance, [ 544.106511] env[63515]: ERROR nova.compute.manager [instance: d7fb514e-4ea9-4f32-9b00-06f728813aa3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 544.106511] env[63515]: ERROR nova.compute.manager [instance: d7fb514e-4ea9-4f32-9b00-06f728813aa3] vif_infos = vmwarevif.get_vif_info(self._session, [ 544.106511] env[63515]: ERROR nova.compute.manager [instance: d7fb514e-4ea9-4f32-9b00-06f728813aa3] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 544.106857] env[63515]: ERROR nova.compute.manager [instance: d7fb514e-4ea9-4f32-9b00-06f728813aa3] for vif in network_info: [ 544.106857] env[63515]: ERROR nova.compute.manager [instance: d7fb514e-4ea9-4f32-9b00-06f728813aa3] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 544.106857] env[63515]: ERROR nova.compute.manager [instance: d7fb514e-4ea9-4f32-9b00-06f728813aa3] return self._sync_wrapper(fn, *args, **kwargs) [ 544.106857] env[63515]: ERROR nova.compute.manager [instance: d7fb514e-4ea9-4f32-9b00-06f728813aa3] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 544.106857] env[63515]: ERROR nova.compute.manager [instance: d7fb514e-4ea9-4f32-9b00-06f728813aa3] self.wait() [ 544.106857] env[63515]: ERROR nova.compute.manager [instance: d7fb514e-4ea9-4f32-9b00-06f728813aa3] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 544.106857] env[63515]: ERROR nova.compute.manager [instance: d7fb514e-4ea9-4f32-9b00-06f728813aa3] self[:] = self._gt.wait() [ 544.106857] env[63515]: ERROR nova.compute.manager [instance: d7fb514e-4ea9-4f32-9b00-06f728813aa3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 544.106857] env[63515]: ERROR nova.compute.manager [instance: d7fb514e-4ea9-4f32-9b00-06f728813aa3] return self._exit_event.wait() [ 544.106857] env[63515]: ERROR nova.compute.manager [instance: d7fb514e-4ea9-4f32-9b00-06f728813aa3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 544.106857] env[63515]: ERROR nova.compute.manager [instance: d7fb514e-4ea9-4f32-9b00-06f728813aa3] result = hub.switch() [ 544.106857] env[63515]: ERROR nova.compute.manager [instance: d7fb514e-4ea9-4f32-9b00-06f728813aa3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 544.106857] env[63515]: ERROR nova.compute.manager [instance: d7fb514e-4ea9-4f32-9b00-06f728813aa3] return self.greenlet.switch() [ 544.107192] env[63515]: ERROR nova.compute.manager [instance: d7fb514e-4ea9-4f32-9b00-06f728813aa3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 544.107192] env[63515]: ERROR nova.compute.manager [instance: d7fb514e-4ea9-4f32-9b00-06f728813aa3] result = function(*args, **kwargs) [ 544.107192] env[63515]: ERROR nova.compute.manager [instance: d7fb514e-4ea9-4f32-9b00-06f728813aa3] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 544.107192] env[63515]: ERROR nova.compute.manager [instance: d7fb514e-4ea9-4f32-9b00-06f728813aa3] return func(*args, **kwargs) [ 544.107192] env[63515]: ERROR nova.compute.manager [instance: d7fb514e-4ea9-4f32-9b00-06f728813aa3] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 544.107192] env[63515]: ERROR nova.compute.manager [instance: d7fb514e-4ea9-4f32-9b00-06f728813aa3] raise e [ 544.107192] env[63515]: ERROR nova.compute.manager [instance: d7fb514e-4ea9-4f32-9b00-06f728813aa3] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 544.107192] env[63515]: ERROR nova.compute.manager [instance: d7fb514e-4ea9-4f32-9b00-06f728813aa3] nwinfo = self.network_api.allocate_for_instance( [ 544.107192] env[63515]: ERROR nova.compute.manager [instance: d7fb514e-4ea9-4f32-9b00-06f728813aa3] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 544.107192] env[63515]: ERROR nova.compute.manager [instance: d7fb514e-4ea9-4f32-9b00-06f728813aa3] created_port_ids = self._update_ports_for_instance( [ 544.107192] env[63515]: ERROR nova.compute.manager [instance: d7fb514e-4ea9-4f32-9b00-06f728813aa3] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 544.107192] env[63515]: ERROR nova.compute.manager [instance: d7fb514e-4ea9-4f32-9b00-06f728813aa3] with excutils.save_and_reraise_exception(): [ 544.107192] env[63515]: ERROR nova.compute.manager [instance: d7fb514e-4ea9-4f32-9b00-06f728813aa3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 544.107507] env[63515]: ERROR nova.compute.manager [instance: d7fb514e-4ea9-4f32-9b00-06f728813aa3] self.force_reraise() [ 544.107507] env[63515]: ERROR nova.compute.manager [instance: d7fb514e-4ea9-4f32-9b00-06f728813aa3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 544.107507] env[63515]: ERROR nova.compute.manager [instance: d7fb514e-4ea9-4f32-9b00-06f728813aa3] raise self.value [ 544.107507] env[63515]: ERROR nova.compute.manager [instance: d7fb514e-4ea9-4f32-9b00-06f728813aa3] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 544.107507] env[63515]: ERROR nova.compute.manager [instance: d7fb514e-4ea9-4f32-9b00-06f728813aa3] updated_port = self._update_port( [ 544.107507] env[63515]: ERROR nova.compute.manager [instance: d7fb514e-4ea9-4f32-9b00-06f728813aa3] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 544.107507] env[63515]: ERROR nova.compute.manager [instance: d7fb514e-4ea9-4f32-9b00-06f728813aa3] _ensure_no_port_binding_failure(port) [ 544.107507] env[63515]: ERROR nova.compute.manager [instance: d7fb514e-4ea9-4f32-9b00-06f728813aa3] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 544.107507] env[63515]: ERROR nova.compute.manager [instance: d7fb514e-4ea9-4f32-9b00-06f728813aa3] raise exception.PortBindingFailed(port_id=port['id']) [ 544.107507] env[63515]: ERROR nova.compute.manager [instance: d7fb514e-4ea9-4f32-9b00-06f728813aa3] nova.exception.PortBindingFailed: Binding failed for port 71cd34ac-f70e-44ba-856e-b65afc2b509f, please check neutron logs for more information. [ 544.107507] env[63515]: ERROR nova.compute.manager [instance: d7fb514e-4ea9-4f32-9b00-06f728813aa3] [ 544.107788] env[63515]: DEBUG nova.compute.utils [None req-9ee35a7a-94b5-4758-9c8f-a8f1e8f2f2cd tempest-DeleteServersAdminTestJSON-944542165 tempest-DeleteServersAdminTestJSON-944542165-project-member] [instance: d7fb514e-4ea9-4f32-9b00-06f728813aa3] Binding failed for port 71cd34ac-f70e-44ba-856e-b65afc2b509f, please check neutron logs for more information. {{(pid=63515) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 544.109458] env[63515]: DEBUG oslo_concurrency.lockutils [None req-222754a5-7546-43c7-a99a-1c936b9e5035 tempest-ImagesOneServerTestJSON-1950372435 tempest-ImagesOneServerTestJSON-1950372435-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 7.411s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 544.111142] env[63515]: INFO nova.compute.claims [None req-222754a5-7546-43c7-a99a-1c936b9e5035 tempest-ImagesOneServerTestJSON-1950372435 tempest-ImagesOneServerTestJSON-1950372435-project-member] [instance: aec66939-7ebc-4bfc-bd8b-bae22e011239] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 544.115632] env[63515]: DEBUG nova.compute.manager [None req-9ee35a7a-94b5-4758-9c8f-a8f1e8f2f2cd tempest-DeleteServersAdminTestJSON-944542165 tempest-DeleteServersAdminTestJSON-944542165-project-member] [instance: d7fb514e-4ea9-4f32-9b00-06f728813aa3] Build of instance d7fb514e-4ea9-4f32-9b00-06f728813aa3 was re-scheduled: Binding failed for port 71cd34ac-f70e-44ba-856e-b65afc2b509f, please check neutron logs for more information. {{(pid=63515) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 544.116147] env[63515]: DEBUG nova.compute.manager [None req-9ee35a7a-94b5-4758-9c8f-a8f1e8f2f2cd tempest-DeleteServersAdminTestJSON-944542165 tempest-DeleteServersAdminTestJSON-944542165-project-member] [instance: d7fb514e-4ea9-4f32-9b00-06f728813aa3] Unplugging VIFs for instance {{(pid=63515) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 544.116364] env[63515]: DEBUG oslo_concurrency.lockutils [None req-9ee35a7a-94b5-4758-9c8f-a8f1e8f2f2cd tempest-DeleteServersAdminTestJSON-944542165 tempest-DeleteServersAdminTestJSON-944542165-project-member] Acquiring lock "refresh_cache-d7fb514e-4ea9-4f32-9b00-06f728813aa3" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 544.116514] env[63515]: DEBUG oslo_concurrency.lockutils [None req-9ee35a7a-94b5-4758-9c8f-a8f1e8f2f2cd tempest-DeleteServersAdminTestJSON-944542165 tempest-DeleteServersAdminTestJSON-944542165-project-member] Acquired lock "refresh_cache-d7fb514e-4ea9-4f32-9b00-06f728813aa3" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 544.116609] env[63515]: DEBUG nova.network.neutron [None req-9ee35a7a-94b5-4758-9c8f-a8f1e8f2f2cd tempest-DeleteServersAdminTestJSON-944542165 tempest-DeleteServersAdminTestJSON-944542165-project-member] [instance: d7fb514e-4ea9-4f32-9b00-06f728813aa3] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 544.178596] env[63515]: DEBUG nova.network.neutron [None req-ada3051f-c750-48c7-b591-0344c25b8fea tempest-ServersAdminTestJSON-710082096 tempest-ServersAdminTestJSON-710082096-project-member] [instance: d7fed938-f7cd-4013-b545-5d5350242084] Successfully created port: abac7593-4174-4f4c-886d-f2653e2b4123 {{(pid=63515) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 544.191547] env[63515]: DEBUG nova.compute.manager [None req-ada3051f-c750-48c7-b591-0344c25b8fea tempest-ServersAdminTestJSON-710082096 tempest-ServersAdminTestJSON-710082096-project-member] [instance: d7fed938-f7cd-4013-b545-5d5350242084] Start spawning the instance on the hypervisor. {{(pid=63515) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 544.216272] env[63515]: DEBUG nova.virt.hardware [None req-ada3051f-c750-48c7-b591-0344c25b8fea tempest-ServersAdminTestJSON-710082096 tempest-ServersAdminTestJSON-710082096-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T02:52:14Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T02:51:56Z,direct_url=,disk_format='vmdk',id=8a120570-cb06-4099-b262-554ca0ad15c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b569255cc43e42e7ae3f2b4ad37c6ef4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T02:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 544.218131] env[63515]: DEBUG nova.virt.hardware [None req-ada3051f-c750-48c7-b591-0344c25b8fea tempest-ServersAdminTestJSON-710082096 tempest-ServersAdminTestJSON-710082096-project-member] Flavor limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 544.218131] env[63515]: DEBUG nova.virt.hardware [None req-ada3051f-c750-48c7-b591-0344c25b8fea tempest-ServersAdminTestJSON-710082096 tempest-ServersAdminTestJSON-710082096-project-member] Image limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 544.218131] env[63515]: DEBUG nova.virt.hardware [None req-ada3051f-c750-48c7-b591-0344c25b8fea tempest-ServersAdminTestJSON-710082096 tempest-ServersAdminTestJSON-710082096-project-member] Flavor pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 544.218131] env[63515]: DEBUG nova.virt.hardware [None req-ada3051f-c750-48c7-b591-0344c25b8fea tempest-ServersAdminTestJSON-710082096 tempest-ServersAdminTestJSON-710082096-project-member] Image pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 544.220144] env[63515]: DEBUG nova.virt.hardware [None req-ada3051f-c750-48c7-b591-0344c25b8fea tempest-ServersAdminTestJSON-710082096 tempest-ServersAdminTestJSON-710082096-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 544.220439] env[63515]: DEBUG nova.virt.hardware [None req-ada3051f-c750-48c7-b591-0344c25b8fea tempest-ServersAdminTestJSON-710082096 tempest-ServersAdminTestJSON-710082096-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 544.220634] env[63515]: DEBUG nova.virt.hardware [None req-ada3051f-c750-48c7-b591-0344c25b8fea tempest-ServersAdminTestJSON-710082096 tempest-ServersAdminTestJSON-710082096-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 544.220815] env[63515]: DEBUG nova.virt.hardware [None req-ada3051f-c750-48c7-b591-0344c25b8fea tempest-ServersAdminTestJSON-710082096 tempest-ServersAdminTestJSON-710082096-project-member] Got 1 possible topologies {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 544.220978] env[63515]: DEBUG nova.virt.hardware [None req-ada3051f-c750-48c7-b591-0344c25b8fea tempest-ServersAdminTestJSON-710082096 tempest-ServersAdminTestJSON-710082096-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 544.221171] env[63515]: DEBUG nova.virt.hardware [None req-ada3051f-c750-48c7-b591-0344c25b8fea tempest-ServersAdminTestJSON-710082096 tempest-ServersAdminTestJSON-710082096-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 544.222381] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd3171e8-a1b4-4c07-adaa-92f8c76bafe9 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 544.230660] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0aba8c19-f548-4ab5-8ba5-307285d66b4c {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 544.477758] env[63515]: DEBUG nova.network.neutron [req-1d8f4c14-1483-4c94-884d-48d2a5bb57c2 req-528ffa5f-402f-436e-a11b-00b509fbbc96 service nova] [instance: 29d199bc-aff3-4c62-af3a-1e6880b52446] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 544.683452] env[63515]: DEBUG nova.network.neutron [None req-9ee35a7a-94b5-4758-9c8f-a8f1e8f2f2cd tempest-DeleteServersAdminTestJSON-944542165 tempest-DeleteServersAdminTestJSON-944542165-project-member] [instance: d7fb514e-4ea9-4f32-9b00-06f728813aa3] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 544.981226] env[63515]: DEBUG oslo_concurrency.lockutils [req-1d8f4c14-1483-4c94-884d-48d2a5bb57c2 req-528ffa5f-402f-436e-a11b-00b509fbbc96 service nova] Releasing lock "refresh_cache-29d199bc-aff3-4c62-af3a-1e6880b52446" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 544.981226] env[63515]: DEBUG nova.compute.manager [req-1d8f4c14-1483-4c94-884d-48d2a5bb57c2 req-528ffa5f-402f-436e-a11b-00b509fbbc96 service nova] [instance: 29d199bc-aff3-4c62-af3a-1e6880b52446] Received event network-vif-deleted-75dd05f7-1fd2-44b3-858d-74b76d1a0780 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 545.143492] env[63515]: DEBUG nova.network.neutron [None req-9ee35a7a-94b5-4758-9c8f-a8f1e8f2f2cd tempest-DeleteServersAdminTestJSON-944542165 tempest-DeleteServersAdminTestJSON-944542165-project-member] [instance: d7fb514e-4ea9-4f32-9b00-06f728813aa3] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 545.385679] env[63515]: DEBUG oslo_concurrency.lockutils [None req-12d73cc4-118d-457e-b393-cc4362d15b0d tempest-ListImageFiltersTestJSON-1100101045 tempest-ListImageFiltersTestJSON-1100101045-project-member] Acquiring lock "ab1539ba-33a4-4139-a2f3-1f8c507beede" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 545.385979] env[63515]: DEBUG oslo_concurrency.lockutils [None req-12d73cc4-118d-457e-b393-cc4362d15b0d tempest-ListImageFiltersTestJSON-1100101045 tempest-ListImageFiltersTestJSON-1100101045-project-member] Lock "ab1539ba-33a4-4139-a2f3-1f8c507beede" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 545.436492] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11726f1e-08cf-42f1-bf5f-27d10fef9ccc {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 545.449750] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-283d27b8-002c-405f-83fb-56297de51ef0 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 545.491654] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be633e02-4a32-4f72-a7e3-b74dbea04a96 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 545.502989] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-198e478f-3300-44e6-aad2-7717b739e1bd {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 545.519937] env[63515]: DEBUG nova.compute.provider_tree [None req-222754a5-7546-43c7-a99a-1c936b9e5035 tempest-ImagesOneServerTestJSON-1950372435 tempest-ImagesOneServerTestJSON-1950372435-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 545.651308] env[63515]: DEBUG oslo_concurrency.lockutils [None req-9ee35a7a-94b5-4758-9c8f-a8f1e8f2f2cd tempest-DeleteServersAdminTestJSON-944542165 tempest-DeleteServersAdminTestJSON-944542165-project-member] Releasing lock "refresh_cache-d7fb514e-4ea9-4f32-9b00-06f728813aa3" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 545.651308] env[63515]: DEBUG nova.compute.manager [None req-9ee35a7a-94b5-4758-9c8f-a8f1e8f2f2cd tempest-DeleteServersAdminTestJSON-944542165 tempest-DeleteServersAdminTestJSON-944542165-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63515) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 545.651308] env[63515]: DEBUG nova.compute.manager [None req-9ee35a7a-94b5-4758-9c8f-a8f1e8f2f2cd tempest-DeleteServersAdminTestJSON-944542165 tempest-DeleteServersAdminTestJSON-944542165-project-member] [instance: d7fb514e-4ea9-4f32-9b00-06f728813aa3] Deallocating network for instance {{(pid=63515) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 545.651308] env[63515]: DEBUG nova.network.neutron [None req-9ee35a7a-94b5-4758-9c8f-a8f1e8f2f2cd tempest-DeleteServersAdminTestJSON-944542165 tempest-DeleteServersAdminTestJSON-944542165-project-member] [instance: d7fb514e-4ea9-4f32-9b00-06f728813aa3] deallocate_for_instance() {{(pid=63515) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 545.733103] env[63515]: DEBUG nova.network.neutron [None req-9ee35a7a-94b5-4758-9c8f-a8f1e8f2f2cd tempest-DeleteServersAdminTestJSON-944542165 tempest-DeleteServersAdminTestJSON-944542165-project-member] [instance: d7fb514e-4ea9-4f32-9b00-06f728813aa3] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 546.026878] env[63515]: DEBUG nova.scheduler.client.report [None req-222754a5-7546-43c7-a99a-1c936b9e5035 tempest-ImagesOneServerTestJSON-1950372435 tempest-ImagesOneServerTestJSON-1950372435-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 172, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 546.240953] env[63515]: DEBUG nova.network.neutron [None req-9ee35a7a-94b5-4758-9c8f-a8f1e8f2f2cd tempest-DeleteServersAdminTestJSON-944542165 tempest-DeleteServersAdminTestJSON-944542165-project-member] [instance: d7fb514e-4ea9-4f32-9b00-06f728813aa3] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 546.457402] env[63515]: DEBUG oslo_concurrency.lockutils [None req-c80cd50f-4127-4451-9298-83a67d8f0ab6 tempest-ServersTestFqdnHostnames-1452522528 tempest-ServersTestFqdnHostnames-1452522528-project-member] Acquiring lock "b84c6c8d-e358-4a14-8425-c4f1d8adf4ee" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 546.457510] env[63515]: DEBUG oslo_concurrency.lockutils [None req-c80cd50f-4127-4451-9298-83a67d8f0ab6 tempest-ServersTestFqdnHostnames-1452522528 tempest-ServersTestFqdnHostnames-1452522528-project-member] Lock "b84c6c8d-e358-4a14-8425-c4f1d8adf4ee" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 546.533958] env[63515]: DEBUG oslo_concurrency.lockutils [None req-222754a5-7546-43c7-a99a-1c936b9e5035 tempest-ImagesOneServerTestJSON-1950372435 tempest-ImagesOneServerTestJSON-1950372435-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.425s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 546.534600] env[63515]: DEBUG nova.compute.manager [None req-222754a5-7546-43c7-a99a-1c936b9e5035 tempest-ImagesOneServerTestJSON-1950372435 tempest-ImagesOneServerTestJSON-1950372435-project-member] [instance: aec66939-7ebc-4bfc-bd8b-bae22e011239] Start building networks asynchronously for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 546.537732] env[63515]: DEBUG oslo_concurrency.lockutils [None req-0d00ff28-5b1e-46b5-9fb9-df8b8bcd6a1a tempest-TenantUsagesTestJSON-427870877 tempest-TenantUsagesTestJSON-427870877-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 8.749s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 546.744372] env[63515]: INFO nova.compute.manager [None req-9ee35a7a-94b5-4758-9c8f-a8f1e8f2f2cd tempest-DeleteServersAdminTestJSON-944542165 tempest-DeleteServersAdminTestJSON-944542165-project-member] [instance: d7fb514e-4ea9-4f32-9b00-06f728813aa3] Took 1.09 seconds to deallocate network for instance. [ 547.043026] env[63515]: DEBUG nova.compute.utils [None req-222754a5-7546-43c7-a99a-1c936b9e5035 tempest-ImagesOneServerTestJSON-1950372435 tempest-ImagesOneServerTestJSON-1950372435-project-member] Using /dev/sd instead of None {{(pid=63515) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 547.051831] env[63515]: DEBUG nova.compute.manager [None req-222754a5-7546-43c7-a99a-1c936b9e5035 tempest-ImagesOneServerTestJSON-1950372435 tempest-ImagesOneServerTestJSON-1950372435-project-member] [instance: aec66939-7ebc-4bfc-bd8b-bae22e011239] Allocating IP information in the background. {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 547.051900] env[63515]: DEBUG nova.network.neutron [None req-222754a5-7546-43c7-a99a-1c936b9e5035 tempest-ImagesOneServerTestJSON-1950372435 tempest-ImagesOneServerTestJSON-1950372435-project-member] [instance: aec66939-7ebc-4bfc-bd8b-bae22e011239] allocate_for_instance() {{(pid=63515) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 547.239286] env[63515]: DEBUG nova.policy [None req-222754a5-7546-43c7-a99a-1c936b9e5035 tempest-ImagesOneServerTestJSON-1950372435 tempest-ImagesOneServerTestJSON-1950372435-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '16114e9f44d2421fa4ddae4ae303c1b2', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e9e7b788d2294f21a652a0c18286b690', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63515) authorize /opt/stack/nova/nova/policy.py:201}} [ 547.306069] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-395b60f4-6460-4571-8d66-15d6fcca87e4 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 547.317310] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-523a20fd-ea11-4262-97db-f61b6efbcd94 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 547.348168] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-996de885-2564-4ac8-b1df-cf0b0fc3b914 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 547.355827] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7dac1d31-444c-4d43-a784-cb7a0d20e84e {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 547.370185] env[63515]: DEBUG nova.compute.provider_tree [None req-0d00ff28-5b1e-46b5-9fb9-df8b8bcd6a1a tempest-TenantUsagesTestJSON-427870877 tempest-TenantUsagesTestJSON-427870877-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 547.551586] env[63515]: DEBUG nova.compute.manager [None req-222754a5-7546-43c7-a99a-1c936b9e5035 tempest-ImagesOneServerTestJSON-1950372435 tempest-ImagesOneServerTestJSON-1950372435-project-member] [instance: aec66939-7ebc-4bfc-bd8b-bae22e011239] Start building block device mappings for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 547.803786] env[63515]: INFO nova.scheduler.client.report [None req-9ee35a7a-94b5-4758-9c8f-a8f1e8f2f2cd tempest-DeleteServersAdminTestJSON-944542165 tempest-DeleteServersAdminTestJSON-944542165-project-member] Deleted allocations for instance d7fb514e-4ea9-4f32-9b00-06f728813aa3 [ 547.873597] env[63515]: DEBUG nova.scheduler.client.report [None req-0d00ff28-5b1e-46b5-9fb9-df8b8bcd6a1a tempest-TenantUsagesTestJSON-427870877 tempest-TenantUsagesTestJSON-427870877-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 172, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 548.325535] env[63515]: DEBUG oslo_concurrency.lockutils [None req-9ee35a7a-94b5-4758-9c8f-a8f1e8f2f2cd tempest-DeleteServersAdminTestJSON-944542165 tempest-DeleteServersAdminTestJSON-944542165-project-member] Lock "d7fb514e-4ea9-4f32-9b00-06f728813aa3" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 25.741s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 548.365221] env[63515]: DEBUG oslo_concurrency.lockutils [None req-22f02712-96d1-4031-8b8f-5bc880f34c33 tempest-ImagesNegativeTestJSON-2041328384 tempest-ImagesNegativeTestJSON-2041328384-project-member] Acquiring lock "d5f988f2-765b-4ecf-bad5-4a70fc95d74c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 548.365459] env[63515]: DEBUG oslo_concurrency.lockutils [None req-22f02712-96d1-4031-8b8f-5bc880f34c33 tempest-ImagesNegativeTestJSON-2041328384 tempest-ImagesNegativeTestJSON-2041328384-project-member] Lock "d5f988f2-765b-4ecf-bad5-4a70fc95d74c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 548.384371] env[63515]: DEBUG oslo_concurrency.lockutils [None req-0d00ff28-5b1e-46b5-9fb9-df8b8bcd6a1a tempest-TenantUsagesTestJSON-427870877 tempest-TenantUsagesTestJSON-427870877-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.847s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 548.384983] env[63515]: ERROR nova.compute.manager [None req-0d00ff28-5b1e-46b5-9fb9-df8b8bcd6a1a tempest-TenantUsagesTestJSON-427870877 tempest-TenantUsagesTestJSON-427870877-project-member] [instance: 9f7ad2dd-b882-4d98-8576-708b3abfa505] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 111c82a0-4d47-4505-8e27-6281465572af, please check neutron logs for more information. [ 548.384983] env[63515]: ERROR nova.compute.manager [instance: 9f7ad2dd-b882-4d98-8576-708b3abfa505] Traceback (most recent call last): [ 548.384983] env[63515]: ERROR nova.compute.manager [instance: 9f7ad2dd-b882-4d98-8576-708b3abfa505] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 548.384983] env[63515]: ERROR nova.compute.manager [instance: 9f7ad2dd-b882-4d98-8576-708b3abfa505] self.driver.spawn(context, instance, image_meta, [ 548.384983] env[63515]: ERROR nova.compute.manager [instance: 9f7ad2dd-b882-4d98-8576-708b3abfa505] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 548.384983] env[63515]: ERROR nova.compute.manager [instance: 9f7ad2dd-b882-4d98-8576-708b3abfa505] self._vmops.spawn(context, instance, image_meta, injected_files, [ 548.384983] env[63515]: ERROR nova.compute.manager [instance: 9f7ad2dd-b882-4d98-8576-708b3abfa505] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 548.384983] env[63515]: ERROR nova.compute.manager [instance: 9f7ad2dd-b882-4d98-8576-708b3abfa505] vm_ref = self.build_virtual_machine(instance, [ 548.384983] env[63515]: ERROR nova.compute.manager [instance: 9f7ad2dd-b882-4d98-8576-708b3abfa505] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 548.384983] env[63515]: ERROR nova.compute.manager [instance: 9f7ad2dd-b882-4d98-8576-708b3abfa505] vif_infos = vmwarevif.get_vif_info(self._session, [ 548.384983] env[63515]: ERROR nova.compute.manager [instance: 9f7ad2dd-b882-4d98-8576-708b3abfa505] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 548.385364] env[63515]: ERROR nova.compute.manager [instance: 9f7ad2dd-b882-4d98-8576-708b3abfa505] for vif in network_info: [ 548.385364] env[63515]: ERROR nova.compute.manager [instance: 9f7ad2dd-b882-4d98-8576-708b3abfa505] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 548.385364] env[63515]: ERROR nova.compute.manager [instance: 9f7ad2dd-b882-4d98-8576-708b3abfa505] return self._sync_wrapper(fn, *args, **kwargs) [ 548.385364] env[63515]: ERROR nova.compute.manager [instance: 9f7ad2dd-b882-4d98-8576-708b3abfa505] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 548.385364] env[63515]: ERROR nova.compute.manager [instance: 9f7ad2dd-b882-4d98-8576-708b3abfa505] self.wait() [ 548.385364] env[63515]: ERROR nova.compute.manager [instance: 9f7ad2dd-b882-4d98-8576-708b3abfa505] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 548.385364] env[63515]: ERROR nova.compute.manager [instance: 9f7ad2dd-b882-4d98-8576-708b3abfa505] self[:] = self._gt.wait() [ 548.385364] env[63515]: ERROR nova.compute.manager [instance: 9f7ad2dd-b882-4d98-8576-708b3abfa505] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 548.385364] env[63515]: ERROR nova.compute.manager [instance: 9f7ad2dd-b882-4d98-8576-708b3abfa505] return self._exit_event.wait() [ 548.385364] env[63515]: ERROR nova.compute.manager [instance: 9f7ad2dd-b882-4d98-8576-708b3abfa505] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 548.385364] env[63515]: ERROR nova.compute.manager [instance: 9f7ad2dd-b882-4d98-8576-708b3abfa505] result = hub.switch() [ 548.385364] env[63515]: ERROR nova.compute.manager [instance: 9f7ad2dd-b882-4d98-8576-708b3abfa505] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 548.385364] env[63515]: ERROR nova.compute.manager [instance: 9f7ad2dd-b882-4d98-8576-708b3abfa505] return self.greenlet.switch() [ 548.385709] env[63515]: ERROR nova.compute.manager [instance: 9f7ad2dd-b882-4d98-8576-708b3abfa505] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 548.385709] env[63515]: ERROR nova.compute.manager [instance: 9f7ad2dd-b882-4d98-8576-708b3abfa505] result = function(*args, **kwargs) [ 548.385709] env[63515]: ERROR nova.compute.manager [instance: 9f7ad2dd-b882-4d98-8576-708b3abfa505] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 548.385709] env[63515]: ERROR nova.compute.manager [instance: 9f7ad2dd-b882-4d98-8576-708b3abfa505] return func(*args, **kwargs) [ 548.385709] env[63515]: ERROR nova.compute.manager [instance: 9f7ad2dd-b882-4d98-8576-708b3abfa505] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 548.385709] env[63515]: ERROR nova.compute.manager [instance: 9f7ad2dd-b882-4d98-8576-708b3abfa505] raise e [ 548.385709] env[63515]: ERROR nova.compute.manager [instance: 9f7ad2dd-b882-4d98-8576-708b3abfa505] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 548.385709] env[63515]: ERROR nova.compute.manager [instance: 9f7ad2dd-b882-4d98-8576-708b3abfa505] nwinfo = self.network_api.allocate_for_instance( [ 548.385709] env[63515]: ERROR nova.compute.manager [instance: 9f7ad2dd-b882-4d98-8576-708b3abfa505] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 548.385709] env[63515]: ERROR nova.compute.manager [instance: 9f7ad2dd-b882-4d98-8576-708b3abfa505] created_port_ids = self._update_ports_for_instance( [ 548.385709] env[63515]: ERROR nova.compute.manager [instance: 9f7ad2dd-b882-4d98-8576-708b3abfa505] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 548.385709] env[63515]: ERROR nova.compute.manager [instance: 9f7ad2dd-b882-4d98-8576-708b3abfa505] with excutils.save_and_reraise_exception(): [ 548.385709] env[63515]: ERROR nova.compute.manager [instance: 9f7ad2dd-b882-4d98-8576-708b3abfa505] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 548.386082] env[63515]: ERROR nova.compute.manager [instance: 9f7ad2dd-b882-4d98-8576-708b3abfa505] self.force_reraise() [ 548.386082] env[63515]: ERROR nova.compute.manager [instance: 9f7ad2dd-b882-4d98-8576-708b3abfa505] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 548.386082] env[63515]: ERROR nova.compute.manager [instance: 9f7ad2dd-b882-4d98-8576-708b3abfa505] raise self.value [ 548.386082] env[63515]: ERROR nova.compute.manager [instance: 9f7ad2dd-b882-4d98-8576-708b3abfa505] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 548.386082] env[63515]: ERROR nova.compute.manager [instance: 9f7ad2dd-b882-4d98-8576-708b3abfa505] updated_port = self._update_port( [ 548.386082] env[63515]: ERROR nova.compute.manager [instance: 9f7ad2dd-b882-4d98-8576-708b3abfa505] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 548.386082] env[63515]: ERROR nova.compute.manager [instance: 9f7ad2dd-b882-4d98-8576-708b3abfa505] _ensure_no_port_binding_failure(port) [ 548.386082] env[63515]: ERROR nova.compute.manager [instance: 9f7ad2dd-b882-4d98-8576-708b3abfa505] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 548.386082] env[63515]: ERROR nova.compute.manager [instance: 9f7ad2dd-b882-4d98-8576-708b3abfa505] raise exception.PortBindingFailed(port_id=port['id']) [ 548.386082] env[63515]: ERROR nova.compute.manager [instance: 9f7ad2dd-b882-4d98-8576-708b3abfa505] nova.exception.PortBindingFailed: Binding failed for port 111c82a0-4d47-4505-8e27-6281465572af, please check neutron logs for more information. [ 548.386082] env[63515]: ERROR nova.compute.manager [instance: 9f7ad2dd-b882-4d98-8576-708b3abfa505] [ 548.386382] env[63515]: DEBUG nova.compute.utils [None req-0d00ff28-5b1e-46b5-9fb9-df8b8bcd6a1a tempest-TenantUsagesTestJSON-427870877 tempest-TenantUsagesTestJSON-427870877-project-member] [instance: 9f7ad2dd-b882-4d98-8576-708b3abfa505] Binding failed for port 111c82a0-4d47-4505-8e27-6281465572af, please check neutron logs for more information. {{(pid=63515) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 548.387719] env[63515]: DEBUG nova.compute.manager [None req-0d00ff28-5b1e-46b5-9fb9-df8b8bcd6a1a tempest-TenantUsagesTestJSON-427870877 tempest-TenantUsagesTestJSON-427870877-project-member] [instance: 9f7ad2dd-b882-4d98-8576-708b3abfa505] Build of instance 9f7ad2dd-b882-4d98-8576-708b3abfa505 was re-scheduled: Binding failed for port 111c82a0-4d47-4505-8e27-6281465572af, please check neutron logs for more information. {{(pid=63515) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 548.388499] env[63515]: DEBUG nova.compute.manager [None req-0d00ff28-5b1e-46b5-9fb9-df8b8bcd6a1a tempest-TenantUsagesTestJSON-427870877 tempest-TenantUsagesTestJSON-427870877-project-member] [instance: 9f7ad2dd-b882-4d98-8576-708b3abfa505] Unplugging VIFs for instance {{(pid=63515) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 548.388833] env[63515]: DEBUG oslo_concurrency.lockutils [None req-0d00ff28-5b1e-46b5-9fb9-df8b8bcd6a1a tempest-TenantUsagesTestJSON-427870877 tempest-TenantUsagesTestJSON-427870877-project-member] Acquiring lock "refresh_cache-9f7ad2dd-b882-4d98-8576-708b3abfa505" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 548.388996] env[63515]: DEBUG oslo_concurrency.lockutils [None req-0d00ff28-5b1e-46b5-9fb9-df8b8bcd6a1a tempest-TenantUsagesTestJSON-427870877 tempest-TenantUsagesTestJSON-427870877-project-member] Acquired lock "refresh_cache-9f7ad2dd-b882-4d98-8576-708b3abfa505" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 548.389183] env[63515]: DEBUG nova.network.neutron [None req-0d00ff28-5b1e-46b5-9fb9-df8b8bcd6a1a tempest-TenantUsagesTestJSON-427870877 tempest-TenantUsagesTestJSON-427870877-project-member] [instance: 9f7ad2dd-b882-4d98-8576-708b3abfa505] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 548.390627] env[63515]: DEBUG oslo_concurrency.lockutils [None req-757a72d0-3dae-4296-830e-18ff5e948610 tempest-ServerDiagnosticsNegativeTest-1799249066 tempest-ServerDiagnosticsNegativeTest-1799249066-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 10.116s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 548.392026] env[63515]: INFO nova.compute.claims [None req-757a72d0-3dae-4296-830e-18ff5e948610 tempest-ServerDiagnosticsNegativeTest-1799249066 tempest-ServerDiagnosticsNegativeTest-1799249066-project-member] [instance: 1c5d87c7-6747-45a7-8488-6f107eeaeb78] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 548.474221] env[63515]: DEBUG nova.network.neutron [None req-222754a5-7546-43c7-a99a-1c936b9e5035 tempest-ImagesOneServerTestJSON-1950372435 tempest-ImagesOneServerTestJSON-1950372435-project-member] [instance: aec66939-7ebc-4bfc-bd8b-bae22e011239] Successfully created port: b5d340e0-18c9-43c1-bcc4-7b682823c4c4 {{(pid=63515) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 548.475908] env[63515]: ERROR nova.compute.manager [None req-ada3051f-c750-48c7-b591-0344c25b8fea tempest-ServersAdminTestJSON-710082096 tempest-ServersAdminTestJSON-710082096-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port abac7593-4174-4f4c-886d-f2653e2b4123, please check neutron logs for more information. [ 548.475908] env[63515]: ERROR nova.compute.manager Traceback (most recent call last): [ 548.475908] env[63515]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 548.475908] env[63515]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 548.475908] env[63515]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 548.475908] env[63515]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 548.475908] env[63515]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 548.475908] env[63515]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 548.475908] env[63515]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 548.475908] env[63515]: ERROR nova.compute.manager self.force_reraise() [ 548.475908] env[63515]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 548.475908] env[63515]: ERROR nova.compute.manager raise self.value [ 548.475908] env[63515]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 548.475908] env[63515]: ERROR nova.compute.manager updated_port = self._update_port( [ 548.475908] env[63515]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 548.475908] env[63515]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 548.476462] env[63515]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 548.476462] env[63515]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 548.476462] env[63515]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port abac7593-4174-4f4c-886d-f2653e2b4123, please check neutron logs for more information. [ 548.476462] env[63515]: ERROR nova.compute.manager [ 548.478457] env[63515]: Traceback (most recent call last): [ 548.478457] env[63515]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 548.478457] env[63515]: listener.cb(fileno) [ 548.478457] env[63515]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 548.478457] env[63515]: result = function(*args, **kwargs) [ 548.478457] env[63515]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 548.478457] env[63515]: return func(*args, **kwargs) [ 548.478457] env[63515]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 548.478457] env[63515]: raise e [ 548.478457] env[63515]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 548.478457] env[63515]: nwinfo = self.network_api.allocate_for_instance( [ 548.478457] env[63515]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 548.478457] env[63515]: created_port_ids = self._update_ports_for_instance( [ 548.478457] env[63515]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 548.478457] env[63515]: with excutils.save_and_reraise_exception(): [ 548.478457] env[63515]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 548.478457] env[63515]: self.force_reraise() [ 548.478457] env[63515]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 548.478457] env[63515]: raise self.value [ 548.478457] env[63515]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 548.478457] env[63515]: updated_port = self._update_port( [ 548.478457] env[63515]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 548.478457] env[63515]: _ensure_no_port_binding_failure(port) [ 548.478457] env[63515]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 548.478457] env[63515]: raise exception.PortBindingFailed(port_id=port['id']) [ 548.478457] env[63515]: nova.exception.PortBindingFailed: Binding failed for port abac7593-4174-4f4c-886d-f2653e2b4123, please check neutron logs for more information. [ 548.478457] env[63515]: Removing descriptor: 17 [ 548.479458] env[63515]: ERROR nova.compute.manager [None req-ada3051f-c750-48c7-b591-0344c25b8fea tempest-ServersAdminTestJSON-710082096 tempest-ServersAdminTestJSON-710082096-project-member] [instance: d7fed938-f7cd-4013-b545-5d5350242084] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port abac7593-4174-4f4c-886d-f2653e2b4123, please check neutron logs for more information. [ 548.479458] env[63515]: ERROR nova.compute.manager [instance: d7fed938-f7cd-4013-b545-5d5350242084] Traceback (most recent call last): [ 548.479458] env[63515]: ERROR nova.compute.manager [instance: d7fed938-f7cd-4013-b545-5d5350242084] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 548.479458] env[63515]: ERROR nova.compute.manager [instance: d7fed938-f7cd-4013-b545-5d5350242084] yield resources [ 548.479458] env[63515]: ERROR nova.compute.manager [instance: d7fed938-f7cd-4013-b545-5d5350242084] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 548.479458] env[63515]: ERROR nova.compute.manager [instance: d7fed938-f7cd-4013-b545-5d5350242084] self.driver.spawn(context, instance, image_meta, [ 548.479458] env[63515]: ERROR nova.compute.manager [instance: d7fed938-f7cd-4013-b545-5d5350242084] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 548.479458] env[63515]: ERROR nova.compute.manager [instance: d7fed938-f7cd-4013-b545-5d5350242084] self._vmops.spawn(context, instance, image_meta, injected_files, [ 548.479458] env[63515]: ERROR nova.compute.manager [instance: d7fed938-f7cd-4013-b545-5d5350242084] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 548.479458] env[63515]: ERROR nova.compute.manager [instance: d7fed938-f7cd-4013-b545-5d5350242084] vm_ref = self.build_virtual_machine(instance, [ 548.479458] env[63515]: ERROR nova.compute.manager [instance: d7fed938-f7cd-4013-b545-5d5350242084] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 548.479805] env[63515]: ERROR nova.compute.manager [instance: d7fed938-f7cd-4013-b545-5d5350242084] vif_infos = vmwarevif.get_vif_info(self._session, [ 548.479805] env[63515]: ERROR nova.compute.manager [instance: d7fed938-f7cd-4013-b545-5d5350242084] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 548.479805] env[63515]: ERROR nova.compute.manager [instance: d7fed938-f7cd-4013-b545-5d5350242084] for vif in network_info: [ 548.479805] env[63515]: ERROR nova.compute.manager [instance: d7fed938-f7cd-4013-b545-5d5350242084] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 548.479805] env[63515]: ERROR nova.compute.manager [instance: d7fed938-f7cd-4013-b545-5d5350242084] return self._sync_wrapper(fn, *args, **kwargs) [ 548.479805] env[63515]: ERROR nova.compute.manager [instance: d7fed938-f7cd-4013-b545-5d5350242084] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 548.479805] env[63515]: ERROR nova.compute.manager [instance: d7fed938-f7cd-4013-b545-5d5350242084] self.wait() [ 548.479805] env[63515]: ERROR nova.compute.manager [instance: d7fed938-f7cd-4013-b545-5d5350242084] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 548.479805] env[63515]: ERROR nova.compute.manager [instance: d7fed938-f7cd-4013-b545-5d5350242084] self[:] = self._gt.wait() [ 548.479805] env[63515]: ERROR nova.compute.manager [instance: d7fed938-f7cd-4013-b545-5d5350242084] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 548.479805] env[63515]: ERROR nova.compute.manager [instance: d7fed938-f7cd-4013-b545-5d5350242084] return self._exit_event.wait() [ 548.479805] env[63515]: ERROR nova.compute.manager [instance: d7fed938-f7cd-4013-b545-5d5350242084] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 548.479805] env[63515]: ERROR nova.compute.manager [instance: d7fed938-f7cd-4013-b545-5d5350242084] result = hub.switch() [ 548.480211] env[63515]: ERROR nova.compute.manager [instance: d7fed938-f7cd-4013-b545-5d5350242084] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 548.480211] env[63515]: ERROR nova.compute.manager [instance: d7fed938-f7cd-4013-b545-5d5350242084] return self.greenlet.switch() [ 548.480211] env[63515]: ERROR nova.compute.manager [instance: d7fed938-f7cd-4013-b545-5d5350242084] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 548.480211] env[63515]: ERROR nova.compute.manager [instance: d7fed938-f7cd-4013-b545-5d5350242084] result = function(*args, **kwargs) [ 548.480211] env[63515]: ERROR nova.compute.manager [instance: d7fed938-f7cd-4013-b545-5d5350242084] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 548.480211] env[63515]: ERROR nova.compute.manager [instance: d7fed938-f7cd-4013-b545-5d5350242084] return func(*args, **kwargs) [ 548.480211] env[63515]: ERROR nova.compute.manager [instance: d7fed938-f7cd-4013-b545-5d5350242084] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 548.480211] env[63515]: ERROR nova.compute.manager [instance: d7fed938-f7cd-4013-b545-5d5350242084] raise e [ 548.480211] env[63515]: ERROR nova.compute.manager [instance: d7fed938-f7cd-4013-b545-5d5350242084] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 548.480211] env[63515]: ERROR nova.compute.manager [instance: d7fed938-f7cd-4013-b545-5d5350242084] nwinfo = self.network_api.allocate_for_instance( [ 548.480211] env[63515]: ERROR nova.compute.manager [instance: d7fed938-f7cd-4013-b545-5d5350242084] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 548.480211] env[63515]: ERROR nova.compute.manager [instance: d7fed938-f7cd-4013-b545-5d5350242084] created_port_ids = self._update_ports_for_instance( [ 548.480211] env[63515]: ERROR nova.compute.manager [instance: d7fed938-f7cd-4013-b545-5d5350242084] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 548.480605] env[63515]: ERROR nova.compute.manager [instance: d7fed938-f7cd-4013-b545-5d5350242084] with excutils.save_and_reraise_exception(): [ 548.480605] env[63515]: ERROR nova.compute.manager [instance: d7fed938-f7cd-4013-b545-5d5350242084] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 548.480605] env[63515]: ERROR nova.compute.manager [instance: d7fed938-f7cd-4013-b545-5d5350242084] self.force_reraise() [ 548.480605] env[63515]: ERROR nova.compute.manager [instance: d7fed938-f7cd-4013-b545-5d5350242084] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 548.480605] env[63515]: ERROR nova.compute.manager [instance: d7fed938-f7cd-4013-b545-5d5350242084] raise self.value [ 548.480605] env[63515]: ERROR nova.compute.manager [instance: d7fed938-f7cd-4013-b545-5d5350242084] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 548.480605] env[63515]: ERROR nova.compute.manager [instance: d7fed938-f7cd-4013-b545-5d5350242084] updated_port = self._update_port( [ 548.480605] env[63515]: ERROR nova.compute.manager [instance: d7fed938-f7cd-4013-b545-5d5350242084] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 548.480605] env[63515]: ERROR nova.compute.manager [instance: d7fed938-f7cd-4013-b545-5d5350242084] _ensure_no_port_binding_failure(port) [ 548.480605] env[63515]: ERROR nova.compute.manager [instance: d7fed938-f7cd-4013-b545-5d5350242084] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 548.480605] env[63515]: ERROR nova.compute.manager [instance: d7fed938-f7cd-4013-b545-5d5350242084] raise exception.PortBindingFailed(port_id=port['id']) [ 548.480605] env[63515]: ERROR nova.compute.manager [instance: d7fed938-f7cd-4013-b545-5d5350242084] nova.exception.PortBindingFailed: Binding failed for port abac7593-4174-4f4c-886d-f2653e2b4123, please check neutron logs for more information. [ 548.480605] env[63515]: ERROR nova.compute.manager [instance: d7fed938-f7cd-4013-b545-5d5350242084] [ 548.481031] env[63515]: INFO nova.compute.manager [None req-ada3051f-c750-48c7-b591-0344c25b8fea tempest-ServersAdminTestJSON-710082096 tempest-ServersAdminTestJSON-710082096-project-member] [instance: d7fed938-f7cd-4013-b545-5d5350242084] Terminating instance [ 548.482328] env[63515]: DEBUG oslo_concurrency.lockutils [None req-ada3051f-c750-48c7-b591-0344c25b8fea tempest-ServersAdminTestJSON-710082096 tempest-ServersAdminTestJSON-710082096-project-member] Acquiring lock "refresh_cache-d7fed938-f7cd-4013-b545-5d5350242084" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 548.482485] env[63515]: DEBUG oslo_concurrency.lockutils [None req-ada3051f-c750-48c7-b591-0344c25b8fea tempest-ServersAdminTestJSON-710082096 tempest-ServersAdminTestJSON-710082096-project-member] Acquired lock "refresh_cache-d7fed938-f7cd-4013-b545-5d5350242084" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 548.482640] env[63515]: DEBUG nova.network.neutron [None req-ada3051f-c750-48c7-b591-0344c25b8fea tempest-ServersAdminTestJSON-710082096 tempest-ServersAdminTestJSON-710082096-project-member] [instance: d7fed938-f7cd-4013-b545-5d5350242084] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 548.571323] env[63515]: DEBUG nova.compute.manager [None req-222754a5-7546-43c7-a99a-1c936b9e5035 tempest-ImagesOneServerTestJSON-1950372435 tempest-ImagesOneServerTestJSON-1950372435-project-member] [instance: aec66939-7ebc-4bfc-bd8b-bae22e011239] Start spawning the instance on the hypervisor. {{(pid=63515) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 548.608955] env[63515]: DEBUG nova.virt.hardware [None req-222754a5-7546-43c7-a99a-1c936b9e5035 tempest-ImagesOneServerTestJSON-1950372435 tempest-ImagesOneServerTestJSON-1950372435-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T02:52:14Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T02:51:56Z,direct_url=,disk_format='vmdk',id=8a120570-cb06-4099-b262-554ca0ad15c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b569255cc43e42e7ae3f2b4ad37c6ef4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T02:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 548.610324] env[63515]: DEBUG nova.virt.hardware [None req-222754a5-7546-43c7-a99a-1c936b9e5035 tempest-ImagesOneServerTestJSON-1950372435 tempest-ImagesOneServerTestJSON-1950372435-project-member] Flavor limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 548.610409] env[63515]: DEBUG nova.virt.hardware [None req-222754a5-7546-43c7-a99a-1c936b9e5035 tempest-ImagesOneServerTestJSON-1950372435 tempest-ImagesOneServerTestJSON-1950372435-project-member] Image limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 548.610578] env[63515]: DEBUG nova.virt.hardware [None req-222754a5-7546-43c7-a99a-1c936b9e5035 tempest-ImagesOneServerTestJSON-1950372435 tempest-ImagesOneServerTestJSON-1950372435-project-member] Flavor pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 548.610724] env[63515]: DEBUG nova.virt.hardware [None req-222754a5-7546-43c7-a99a-1c936b9e5035 tempest-ImagesOneServerTestJSON-1950372435 tempest-ImagesOneServerTestJSON-1950372435-project-member] Image pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 548.610870] env[63515]: DEBUG nova.virt.hardware [None req-222754a5-7546-43c7-a99a-1c936b9e5035 tempest-ImagesOneServerTestJSON-1950372435 tempest-ImagesOneServerTestJSON-1950372435-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 548.611128] env[63515]: DEBUG nova.virt.hardware [None req-222754a5-7546-43c7-a99a-1c936b9e5035 tempest-ImagesOneServerTestJSON-1950372435 tempest-ImagesOneServerTestJSON-1950372435-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 548.611316] env[63515]: DEBUG nova.virt.hardware [None req-222754a5-7546-43c7-a99a-1c936b9e5035 tempest-ImagesOneServerTestJSON-1950372435 tempest-ImagesOneServerTestJSON-1950372435-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 548.611496] env[63515]: DEBUG nova.virt.hardware [None req-222754a5-7546-43c7-a99a-1c936b9e5035 tempest-ImagesOneServerTestJSON-1950372435 tempest-ImagesOneServerTestJSON-1950372435-project-member] Got 1 possible topologies {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 548.611659] env[63515]: DEBUG nova.virt.hardware [None req-222754a5-7546-43c7-a99a-1c936b9e5035 tempest-ImagesOneServerTestJSON-1950372435 tempest-ImagesOneServerTestJSON-1950372435-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 548.611857] env[63515]: DEBUG nova.virt.hardware [None req-222754a5-7546-43c7-a99a-1c936b9e5035 tempest-ImagesOneServerTestJSON-1950372435 tempest-ImagesOneServerTestJSON-1950372435-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 548.612729] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8872abe0-2b9a-42fb-90c1-6bf668a024d2 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.632228] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1245461e-bf60-4a24-a035-94fcb22fad4c {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.833112] env[63515]: DEBUG nova.compute.manager [None req-eaa7f189-a861-45f8-872f-29d870dc784b tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] [instance: e39bc55e-b677-4e2d-9aa4-7bc1d5ffadfa] Starting instance... {{(pid=63515) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 548.977569] env[63515]: DEBUG nova.network.neutron [None req-0d00ff28-5b1e-46b5-9fb9-df8b8bcd6a1a tempest-TenantUsagesTestJSON-427870877 tempest-TenantUsagesTestJSON-427870877-project-member] [instance: 9f7ad2dd-b882-4d98-8576-708b3abfa505] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 549.079840] env[63515]: DEBUG nova.network.neutron [None req-ada3051f-c750-48c7-b591-0344c25b8fea tempest-ServersAdminTestJSON-710082096 tempest-ServersAdminTestJSON-710082096-project-member] [instance: d7fed938-f7cd-4013-b545-5d5350242084] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 549.275929] env[63515]: DEBUG nova.compute.manager [req-444ceb66-ff6d-4f78-8399-ed8bd93940e8 req-a5b631bb-fce1-4e1c-aaf7-115b98f1a188 service nova] [instance: d7fed938-f7cd-4013-b545-5d5350242084] Received event network-changed-abac7593-4174-4f4c-886d-f2653e2b4123 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 549.276260] env[63515]: DEBUG nova.compute.manager [req-444ceb66-ff6d-4f78-8399-ed8bd93940e8 req-a5b631bb-fce1-4e1c-aaf7-115b98f1a188 service nova] [instance: d7fed938-f7cd-4013-b545-5d5350242084] Refreshing instance network info cache due to event network-changed-abac7593-4174-4f4c-886d-f2653e2b4123. {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 549.276548] env[63515]: DEBUG oslo_concurrency.lockutils [req-444ceb66-ff6d-4f78-8399-ed8bd93940e8 req-a5b631bb-fce1-4e1c-aaf7-115b98f1a188 service nova] Acquiring lock "refresh_cache-d7fed938-f7cd-4013-b545-5d5350242084" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 549.287386] env[63515]: DEBUG oslo_concurrency.lockutils [None req-9e973647-d524-4c71-b05a-f1a253758bab tempest-ServerAddressesTestJSON-9863250 tempest-ServerAddressesTestJSON-9863250-project-member] Acquiring lock "7fee0afb-88c9-4e93-a4d9-94312394f927" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 549.287688] env[63515]: DEBUG oslo_concurrency.lockutils [None req-9e973647-d524-4c71-b05a-f1a253758bab tempest-ServerAddressesTestJSON-9863250 tempest-ServerAddressesTestJSON-9863250-project-member] Lock "7fee0afb-88c9-4e93-a4d9-94312394f927" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 549.301829] env[63515]: DEBUG nova.network.neutron [None req-0d00ff28-5b1e-46b5-9fb9-df8b8bcd6a1a tempest-TenantUsagesTestJSON-427870877 tempest-TenantUsagesTestJSON-427870877-project-member] [instance: 9f7ad2dd-b882-4d98-8576-708b3abfa505] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 549.356125] env[63515]: DEBUG oslo_concurrency.lockutils [None req-eaa7f189-a861-45f8-872f-29d870dc784b tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 549.445073] env[63515]: DEBUG nova.network.neutron [None req-ada3051f-c750-48c7-b591-0344c25b8fea tempest-ServersAdminTestJSON-710082096 tempest-ServersAdminTestJSON-710082096-project-member] [instance: d7fed938-f7cd-4013-b545-5d5350242084] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 549.733099] env[63515]: DEBUG oslo_concurrency.lockutils [None req-24636c84-3869-4a1b-acfb-9c10d33a9826 tempest-ServersWithSpecificFlavorTestJSON-691988195 tempest-ServersWithSpecificFlavorTestJSON-691988195-project-member] Acquiring lock "ebc737a3-e6cd-4f4c-a4b6-fce5924de4c6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 549.733099] env[63515]: DEBUG oslo_concurrency.lockutils [None req-24636c84-3869-4a1b-acfb-9c10d33a9826 tempest-ServersWithSpecificFlavorTestJSON-691988195 tempest-ServersWithSpecificFlavorTestJSON-691988195-project-member] Lock "ebc737a3-e6cd-4f4c-a4b6-fce5924de4c6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 549.757589] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a36c5fca-7679-4cd6-8db0-dc590278c60a {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 549.766483] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d5d4cfe-99f7-4b91-af50-eb09b06c030e {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 549.805912] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44bf72ba-acc4-467f-a29f-f4519f46205c {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 549.809060] env[63515]: DEBUG oslo_concurrency.lockutils [None req-0d00ff28-5b1e-46b5-9fb9-df8b8bcd6a1a tempest-TenantUsagesTestJSON-427870877 tempest-TenantUsagesTestJSON-427870877-project-member] Releasing lock "refresh_cache-9f7ad2dd-b882-4d98-8576-708b3abfa505" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 549.809273] env[63515]: DEBUG nova.compute.manager [None req-0d00ff28-5b1e-46b5-9fb9-df8b8bcd6a1a tempest-TenantUsagesTestJSON-427870877 tempest-TenantUsagesTestJSON-427870877-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63515) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 549.809444] env[63515]: DEBUG nova.compute.manager [None req-0d00ff28-5b1e-46b5-9fb9-df8b8bcd6a1a tempest-TenantUsagesTestJSON-427870877 tempest-TenantUsagesTestJSON-427870877-project-member] [instance: 9f7ad2dd-b882-4d98-8576-708b3abfa505] Deallocating network for instance {{(pid=63515) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 549.809587] env[63515]: DEBUG nova.network.neutron [None req-0d00ff28-5b1e-46b5-9fb9-df8b8bcd6a1a tempest-TenantUsagesTestJSON-427870877 tempest-TenantUsagesTestJSON-427870877-project-member] [instance: 9f7ad2dd-b882-4d98-8576-708b3abfa505] deallocate_for_instance() {{(pid=63515) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 549.818407] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aab733c3-d490-4c2e-a7ae-ab0a8e0d5a3e {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 549.833611] env[63515]: DEBUG nova.compute.provider_tree [None req-757a72d0-3dae-4296-830e-18ff5e948610 tempest-ServerDiagnosticsNegativeTest-1799249066 tempest-ServerDiagnosticsNegativeTest-1799249066-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 549.850362] env[63515]: DEBUG nova.network.neutron [None req-0d00ff28-5b1e-46b5-9fb9-df8b8bcd6a1a tempest-TenantUsagesTestJSON-427870877 tempest-TenantUsagesTestJSON-427870877-project-member] [instance: 9f7ad2dd-b882-4d98-8576-708b3abfa505] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 549.950440] env[63515]: DEBUG oslo_concurrency.lockutils [None req-ada3051f-c750-48c7-b591-0344c25b8fea tempest-ServersAdminTestJSON-710082096 tempest-ServersAdminTestJSON-710082096-project-member] Releasing lock "refresh_cache-d7fed938-f7cd-4013-b545-5d5350242084" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 549.953298] env[63515]: DEBUG nova.compute.manager [None req-ada3051f-c750-48c7-b591-0344c25b8fea tempest-ServersAdminTestJSON-710082096 tempest-ServersAdminTestJSON-710082096-project-member] [instance: d7fed938-f7cd-4013-b545-5d5350242084] Start destroying the instance on the hypervisor. {{(pid=63515) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 549.953298] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-ada3051f-c750-48c7-b591-0344c25b8fea tempest-ServersAdminTestJSON-710082096 tempest-ServersAdminTestJSON-710082096-project-member] [instance: d7fed938-f7cd-4013-b545-5d5350242084] Destroying instance {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 549.953298] env[63515]: DEBUG oslo_concurrency.lockutils [req-444ceb66-ff6d-4f78-8399-ed8bd93940e8 req-a5b631bb-fce1-4e1c-aaf7-115b98f1a188 service nova] Acquired lock "refresh_cache-d7fed938-f7cd-4013-b545-5d5350242084" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 549.953298] env[63515]: DEBUG nova.network.neutron [req-444ceb66-ff6d-4f78-8399-ed8bd93940e8 req-a5b631bb-fce1-4e1c-aaf7-115b98f1a188 service nova] [instance: d7fed938-f7cd-4013-b545-5d5350242084] Refreshing network info cache for port abac7593-4174-4f4c-886d-f2653e2b4123 {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 549.953298] env[63515]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-01b0e5d9-a257-4a93-8206-6f29f33f791e {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 549.963352] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7c42aba-946e-4dd2-82ae-018facccb89c {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 549.991785] env[63515]: WARNING nova.virt.vmwareapi.vmops [None req-ada3051f-c750-48c7-b591-0344c25b8fea tempest-ServersAdminTestJSON-710082096 tempest-ServersAdminTestJSON-710082096-project-member] [instance: d7fed938-f7cd-4013-b545-5d5350242084] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance d7fed938-f7cd-4013-b545-5d5350242084 could not be found. [ 549.992067] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-ada3051f-c750-48c7-b591-0344c25b8fea tempest-ServersAdminTestJSON-710082096 tempest-ServersAdminTestJSON-710082096-project-member] [instance: d7fed938-f7cd-4013-b545-5d5350242084] Instance destroyed {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 549.992589] env[63515]: INFO nova.compute.manager [None req-ada3051f-c750-48c7-b591-0344c25b8fea tempest-ServersAdminTestJSON-710082096 tempest-ServersAdminTestJSON-710082096-project-member] [instance: d7fed938-f7cd-4013-b545-5d5350242084] Took 0.04 seconds to destroy the instance on the hypervisor. [ 549.992877] env[63515]: DEBUG oslo.service.loopingcall [None req-ada3051f-c750-48c7-b591-0344c25b8fea tempest-ServersAdminTestJSON-710082096 tempest-ServersAdminTestJSON-710082096-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 549.993125] env[63515]: DEBUG nova.compute.manager [-] [instance: d7fed938-f7cd-4013-b545-5d5350242084] Deallocating network for instance {{(pid=63515) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 549.993275] env[63515]: DEBUG nova.network.neutron [-] [instance: d7fed938-f7cd-4013-b545-5d5350242084] deallocate_for_instance() {{(pid=63515) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 550.028952] env[63515]: DEBUG nova.network.neutron [-] [instance: d7fed938-f7cd-4013-b545-5d5350242084] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 550.337437] env[63515]: DEBUG nova.scheduler.client.report [None req-757a72d0-3dae-4296-830e-18ff5e948610 tempest-ServerDiagnosticsNegativeTest-1799249066 tempest-ServerDiagnosticsNegativeTest-1799249066-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 172, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 550.352408] env[63515]: DEBUG nova.network.neutron [None req-0d00ff28-5b1e-46b5-9fb9-df8b8bcd6a1a tempest-TenantUsagesTestJSON-427870877 tempest-TenantUsagesTestJSON-427870877-project-member] [instance: 9f7ad2dd-b882-4d98-8576-708b3abfa505] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 550.492495] env[63515]: DEBUG nova.network.neutron [req-444ceb66-ff6d-4f78-8399-ed8bd93940e8 req-a5b631bb-fce1-4e1c-aaf7-115b98f1a188 service nova] [instance: d7fed938-f7cd-4013-b545-5d5350242084] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 550.536156] env[63515]: DEBUG nova.network.neutron [-] [instance: d7fed938-f7cd-4013-b545-5d5350242084] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 550.704261] env[63515]: DEBUG nova.network.neutron [req-444ceb66-ff6d-4f78-8399-ed8bd93940e8 req-a5b631bb-fce1-4e1c-aaf7-115b98f1a188 service nova] [instance: d7fed938-f7cd-4013-b545-5d5350242084] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 550.845125] env[63515]: DEBUG oslo_concurrency.lockutils [None req-757a72d0-3dae-4296-830e-18ff5e948610 tempest-ServerDiagnosticsNegativeTest-1799249066 tempest-ServerDiagnosticsNegativeTest-1799249066-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.454s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 550.845664] env[63515]: DEBUG nova.compute.manager [None req-757a72d0-3dae-4296-830e-18ff5e948610 tempest-ServerDiagnosticsNegativeTest-1799249066 tempest-ServerDiagnosticsNegativeTest-1799249066-project-member] [instance: 1c5d87c7-6747-45a7-8488-6f107eeaeb78] Start building networks asynchronously for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 550.849187] env[63515]: DEBUG oslo_concurrency.lockutils [None req-919af50b-07d1-4794-942f-d99fa56afe84 tempest-ServerExternalEventsTest-1976219582 tempest-ServerExternalEventsTest-1976219582-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 12.453s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 550.858663] env[63515]: INFO nova.compute.manager [None req-0d00ff28-5b1e-46b5-9fb9-df8b8bcd6a1a tempest-TenantUsagesTestJSON-427870877 tempest-TenantUsagesTestJSON-427870877-project-member] [instance: 9f7ad2dd-b882-4d98-8576-708b3abfa505] Took 1.05 seconds to deallocate network for instance. [ 551.036271] env[63515]: INFO nova.compute.manager [-] [instance: d7fed938-f7cd-4013-b545-5d5350242084] Took 1.04 seconds to deallocate network for instance. [ 551.042466] env[63515]: DEBUG nova.compute.claims [None req-ada3051f-c750-48c7-b591-0344c25b8fea tempest-ServersAdminTestJSON-710082096 tempest-ServersAdminTestJSON-710082096-project-member] [instance: d7fed938-f7cd-4013-b545-5d5350242084] Aborting claim: {{(pid=63515) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 551.042746] env[63515]: DEBUG oslo_concurrency.lockutils [None req-ada3051f-c750-48c7-b591-0344c25b8fea tempest-ServersAdminTestJSON-710082096 tempest-ServersAdminTestJSON-710082096-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 551.208049] env[63515]: DEBUG oslo_concurrency.lockutils [req-444ceb66-ff6d-4f78-8399-ed8bd93940e8 req-a5b631bb-fce1-4e1c-aaf7-115b98f1a188 service nova] Releasing lock "refresh_cache-d7fed938-f7cd-4013-b545-5d5350242084" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 551.294696] env[63515]: ERROR nova.compute.manager [None req-222754a5-7546-43c7-a99a-1c936b9e5035 tempest-ImagesOneServerTestJSON-1950372435 tempest-ImagesOneServerTestJSON-1950372435-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port b5d340e0-18c9-43c1-bcc4-7b682823c4c4, please check neutron logs for more information. [ 551.294696] env[63515]: ERROR nova.compute.manager Traceback (most recent call last): [ 551.294696] env[63515]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 551.294696] env[63515]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 551.294696] env[63515]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 551.294696] env[63515]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 551.294696] env[63515]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 551.294696] env[63515]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 551.294696] env[63515]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 551.294696] env[63515]: ERROR nova.compute.manager self.force_reraise() [ 551.294696] env[63515]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 551.294696] env[63515]: ERROR nova.compute.manager raise self.value [ 551.294696] env[63515]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 551.294696] env[63515]: ERROR nova.compute.manager updated_port = self._update_port( [ 551.294696] env[63515]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 551.294696] env[63515]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 551.295602] env[63515]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 551.295602] env[63515]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 551.295602] env[63515]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port b5d340e0-18c9-43c1-bcc4-7b682823c4c4, please check neutron logs for more information. [ 551.295602] env[63515]: ERROR nova.compute.manager [ 551.295602] env[63515]: Traceback (most recent call last): [ 551.295602] env[63515]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 551.295602] env[63515]: listener.cb(fileno) [ 551.295602] env[63515]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 551.295602] env[63515]: result = function(*args, **kwargs) [ 551.295602] env[63515]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 551.295602] env[63515]: return func(*args, **kwargs) [ 551.295602] env[63515]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 551.295602] env[63515]: raise e [ 551.295602] env[63515]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 551.295602] env[63515]: nwinfo = self.network_api.allocate_for_instance( [ 551.295602] env[63515]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 551.295602] env[63515]: created_port_ids = self._update_ports_for_instance( [ 551.295602] env[63515]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 551.295602] env[63515]: with excutils.save_and_reraise_exception(): [ 551.295602] env[63515]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 551.295602] env[63515]: self.force_reraise() [ 551.295602] env[63515]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 551.295602] env[63515]: raise self.value [ 551.295602] env[63515]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 551.295602] env[63515]: updated_port = self._update_port( [ 551.295602] env[63515]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 551.295602] env[63515]: _ensure_no_port_binding_failure(port) [ 551.295602] env[63515]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 551.295602] env[63515]: raise exception.PortBindingFailed(port_id=port['id']) [ 551.296336] env[63515]: nova.exception.PortBindingFailed: Binding failed for port b5d340e0-18c9-43c1-bcc4-7b682823c4c4, please check neutron logs for more information. [ 551.296336] env[63515]: Removing descriptor: 16 [ 551.296336] env[63515]: ERROR nova.compute.manager [None req-222754a5-7546-43c7-a99a-1c936b9e5035 tempest-ImagesOneServerTestJSON-1950372435 tempest-ImagesOneServerTestJSON-1950372435-project-member] [instance: aec66939-7ebc-4bfc-bd8b-bae22e011239] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port b5d340e0-18c9-43c1-bcc4-7b682823c4c4, please check neutron logs for more information. [ 551.296336] env[63515]: ERROR nova.compute.manager [instance: aec66939-7ebc-4bfc-bd8b-bae22e011239] Traceback (most recent call last): [ 551.296336] env[63515]: ERROR nova.compute.manager [instance: aec66939-7ebc-4bfc-bd8b-bae22e011239] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 551.296336] env[63515]: ERROR nova.compute.manager [instance: aec66939-7ebc-4bfc-bd8b-bae22e011239] yield resources [ 551.296336] env[63515]: ERROR nova.compute.manager [instance: aec66939-7ebc-4bfc-bd8b-bae22e011239] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 551.296336] env[63515]: ERROR nova.compute.manager [instance: aec66939-7ebc-4bfc-bd8b-bae22e011239] self.driver.spawn(context, instance, image_meta, [ 551.296336] env[63515]: ERROR nova.compute.manager [instance: aec66939-7ebc-4bfc-bd8b-bae22e011239] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 551.296336] env[63515]: ERROR nova.compute.manager [instance: aec66939-7ebc-4bfc-bd8b-bae22e011239] self._vmops.spawn(context, instance, image_meta, injected_files, [ 551.296336] env[63515]: ERROR nova.compute.manager [instance: aec66939-7ebc-4bfc-bd8b-bae22e011239] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 551.296336] env[63515]: ERROR nova.compute.manager [instance: aec66939-7ebc-4bfc-bd8b-bae22e011239] vm_ref = self.build_virtual_machine(instance, [ 551.296643] env[63515]: ERROR nova.compute.manager [instance: aec66939-7ebc-4bfc-bd8b-bae22e011239] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 551.296643] env[63515]: ERROR nova.compute.manager [instance: aec66939-7ebc-4bfc-bd8b-bae22e011239] vif_infos = vmwarevif.get_vif_info(self._session, [ 551.296643] env[63515]: ERROR nova.compute.manager [instance: aec66939-7ebc-4bfc-bd8b-bae22e011239] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 551.296643] env[63515]: ERROR nova.compute.manager [instance: aec66939-7ebc-4bfc-bd8b-bae22e011239] for vif in network_info: [ 551.296643] env[63515]: ERROR nova.compute.manager [instance: aec66939-7ebc-4bfc-bd8b-bae22e011239] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 551.296643] env[63515]: ERROR nova.compute.manager [instance: aec66939-7ebc-4bfc-bd8b-bae22e011239] return self._sync_wrapper(fn, *args, **kwargs) [ 551.296643] env[63515]: ERROR nova.compute.manager [instance: aec66939-7ebc-4bfc-bd8b-bae22e011239] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 551.296643] env[63515]: ERROR nova.compute.manager [instance: aec66939-7ebc-4bfc-bd8b-bae22e011239] self.wait() [ 551.296643] env[63515]: ERROR nova.compute.manager [instance: aec66939-7ebc-4bfc-bd8b-bae22e011239] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 551.296643] env[63515]: ERROR nova.compute.manager [instance: aec66939-7ebc-4bfc-bd8b-bae22e011239] self[:] = self._gt.wait() [ 551.296643] env[63515]: ERROR nova.compute.manager [instance: aec66939-7ebc-4bfc-bd8b-bae22e011239] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 551.296643] env[63515]: ERROR nova.compute.manager [instance: aec66939-7ebc-4bfc-bd8b-bae22e011239] return self._exit_event.wait() [ 551.296643] env[63515]: ERROR nova.compute.manager [instance: aec66939-7ebc-4bfc-bd8b-bae22e011239] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 551.296961] env[63515]: ERROR nova.compute.manager [instance: aec66939-7ebc-4bfc-bd8b-bae22e011239] result = hub.switch() [ 551.296961] env[63515]: ERROR nova.compute.manager [instance: aec66939-7ebc-4bfc-bd8b-bae22e011239] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 551.296961] env[63515]: ERROR nova.compute.manager [instance: aec66939-7ebc-4bfc-bd8b-bae22e011239] return self.greenlet.switch() [ 551.296961] env[63515]: ERROR nova.compute.manager [instance: aec66939-7ebc-4bfc-bd8b-bae22e011239] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 551.296961] env[63515]: ERROR nova.compute.manager [instance: aec66939-7ebc-4bfc-bd8b-bae22e011239] result = function(*args, **kwargs) [ 551.296961] env[63515]: ERROR nova.compute.manager [instance: aec66939-7ebc-4bfc-bd8b-bae22e011239] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 551.296961] env[63515]: ERROR nova.compute.manager [instance: aec66939-7ebc-4bfc-bd8b-bae22e011239] return func(*args, **kwargs) [ 551.296961] env[63515]: ERROR nova.compute.manager [instance: aec66939-7ebc-4bfc-bd8b-bae22e011239] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 551.296961] env[63515]: ERROR nova.compute.manager [instance: aec66939-7ebc-4bfc-bd8b-bae22e011239] raise e [ 551.296961] env[63515]: ERROR nova.compute.manager [instance: aec66939-7ebc-4bfc-bd8b-bae22e011239] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 551.296961] env[63515]: ERROR nova.compute.manager [instance: aec66939-7ebc-4bfc-bd8b-bae22e011239] nwinfo = self.network_api.allocate_for_instance( [ 551.296961] env[63515]: ERROR nova.compute.manager [instance: aec66939-7ebc-4bfc-bd8b-bae22e011239] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 551.296961] env[63515]: ERROR nova.compute.manager [instance: aec66939-7ebc-4bfc-bd8b-bae22e011239] created_port_ids = self._update_ports_for_instance( [ 551.297319] env[63515]: ERROR nova.compute.manager [instance: aec66939-7ebc-4bfc-bd8b-bae22e011239] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 551.297319] env[63515]: ERROR nova.compute.manager [instance: aec66939-7ebc-4bfc-bd8b-bae22e011239] with excutils.save_and_reraise_exception(): [ 551.297319] env[63515]: ERROR nova.compute.manager [instance: aec66939-7ebc-4bfc-bd8b-bae22e011239] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 551.297319] env[63515]: ERROR nova.compute.manager [instance: aec66939-7ebc-4bfc-bd8b-bae22e011239] self.force_reraise() [ 551.297319] env[63515]: ERROR nova.compute.manager [instance: aec66939-7ebc-4bfc-bd8b-bae22e011239] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 551.297319] env[63515]: ERROR nova.compute.manager [instance: aec66939-7ebc-4bfc-bd8b-bae22e011239] raise self.value [ 551.297319] env[63515]: ERROR nova.compute.manager [instance: aec66939-7ebc-4bfc-bd8b-bae22e011239] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 551.297319] env[63515]: ERROR nova.compute.manager [instance: aec66939-7ebc-4bfc-bd8b-bae22e011239] updated_port = self._update_port( [ 551.297319] env[63515]: ERROR nova.compute.manager [instance: aec66939-7ebc-4bfc-bd8b-bae22e011239] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 551.297319] env[63515]: ERROR nova.compute.manager [instance: aec66939-7ebc-4bfc-bd8b-bae22e011239] _ensure_no_port_binding_failure(port) [ 551.297319] env[63515]: ERROR nova.compute.manager [instance: aec66939-7ebc-4bfc-bd8b-bae22e011239] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 551.297319] env[63515]: ERROR nova.compute.manager [instance: aec66939-7ebc-4bfc-bd8b-bae22e011239] raise exception.PortBindingFailed(port_id=port['id']) [ 551.297610] env[63515]: ERROR nova.compute.manager [instance: aec66939-7ebc-4bfc-bd8b-bae22e011239] nova.exception.PortBindingFailed: Binding failed for port b5d340e0-18c9-43c1-bcc4-7b682823c4c4, please check neutron logs for more information. [ 551.297610] env[63515]: ERROR nova.compute.manager [instance: aec66939-7ebc-4bfc-bd8b-bae22e011239] [ 551.297610] env[63515]: INFO nova.compute.manager [None req-222754a5-7546-43c7-a99a-1c936b9e5035 tempest-ImagesOneServerTestJSON-1950372435 tempest-ImagesOneServerTestJSON-1950372435-project-member] [instance: aec66939-7ebc-4bfc-bd8b-bae22e011239] Terminating instance [ 551.299889] env[63515]: DEBUG oslo_concurrency.lockutils [None req-222754a5-7546-43c7-a99a-1c936b9e5035 tempest-ImagesOneServerTestJSON-1950372435 tempest-ImagesOneServerTestJSON-1950372435-project-member] Acquiring lock "refresh_cache-aec66939-7ebc-4bfc-bd8b-bae22e011239" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 551.299889] env[63515]: DEBUG oslo_concurrency.lockutils [None req-222754a5-7546-43c7-a99a-1c936b9e5035 tempest-ImagesOneServerTestJSON-1950372435 tempest-ImagesOneServerTestJSON-1950372435-project-member] Acquired lock "refresh_cache-aec66939-7ebc-4bfc-bd8b-bae22e011239" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 551.299889] env[63515]: DEBUG nova.network.neutron [None req-222754a5-7546-43c7-a99a-1c936b9e5035 tempest-ImagesOneServerTestJSON-1950372435 tempest-ImagesOneServerTestJSON-1950372435-project-member] [instance: aec66939-7ebc-4bfc-bd8b-bae22e011239] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 551.366539] env[63515]: DEBUG nova.compute.utils [None req-757a72d0-3dae-4296-830e-18ff5e948610 tempest-ServerDiagnosticsNegativeTest-1799249066 tempest-ServerDiagnosticsNegativeTest-1799249066-project-member] Using /dev/sd instead of None {{(pid=63515) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 551.372064] env[63515]: DEBUG nova.compute.manager [None req-757a72d0-3dae-4296-830e-18ff5e948610 tempest-ServerDiagnosticsNegativeTest-1799249066 tempest-ServerDiagnosticsNegativeTest-1799249066-project-member] [instance: 1c5d87c7-6747-45a7-8488-6f107eeaeb78] Allocating IP information in the background. {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 551.372169] env[63515]: DEBUG nova.network.neutron [None req-757a72d0-3dae-4296-830e-18ff5e948610 tempest-ServerDiagnosticsNegativeTest-1799249066 tempest-ServerDiagnosticsNegativeTest-1799249066-project-member] [instance: 1c5d87c7-6747-45a7-8488-6f107eeaeb78] allocate_for_instance() {{(pid=63515) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 551.635603] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f405c974-cf87-4dc0-b7e8-81e8fd2fff5c {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 551.646339] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f55ca563-5ef7-4b5c-8fbb-5c725f29fbac {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 551.679692] env[63515]: DEBUG nova.policy [None req-757a72d0-3dae-4296-830e-18ff5e948610 tempest-ServerDiagnosticsNegativeTest-1799249066 tempest-ServerDiagnosticsNegativeTest-1799249066-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c5640b609694471ca393845b6ab68f58', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5266018d331a44fa860f77499f3d82c1', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63515) authorize /opt/stack/nova/nova/policy.py:201}} [ 551.681732] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a53fc35-1747-4c3e-8d82-0d5d5138897f {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 551.692760] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93d5fa75-aaff-45cb-a5ce-e5e3fee285b9 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 551.705309] env[63515]: DEBUG nova.compute.provider_tree [None req-919af50b-07d1-4794-942f-d99fa56afe84 tempest-ServerExternalEventsTest-1976219582 tempest-ServerExternalEventsTest-1976219582-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 551.834620] env[63515]: DEBUG nova.network.neutron [None req-222754a5-7546-43c7-a99a-1c936b9e5035 tempest-ImagesOneServerTestJSON-1950372435 tempest-ImagesOneServerTestJSON-1950372435-project-member] [instance: aec66939-7ebc-4bfc-bd8b-bae22e011239] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 551.882453] env[63515]: DEBUG nova.compute.manager [None req-757a72d0-3dae-4296-830e-18ff5e948610 tempest-ServerDiagnosticsNegativeTest-1799249066 tempest-ServerDiagnosticsNegativeTest-1799249066-project-member] [instance: 1c5d87c7-6747-45a7-8488-6f107eeaeb78] Start building block device mappings for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 551.917675] env[63515]: INFO nova.scheduler.client.report [None req-0d00ff28-5b1e-46b5-9fb9-df8b8bcd6a1a tempest-TenantUsagesTestJSON-427870877 tempest-TenantUsagesTestJSON-427870877-project-member] Deleted allocations for instance 9f7ad2dd-b882-4d98-8576-708b3abfa505 [ 552.209733] env[63515]: DEBUG nova.scheduler.client.report [None req-919af50b-07d1-4794-942f-d99fa56afe84 tempest-ServerExternalEventsTest-1976219582 tempest-ServerExternalEventsTest-1976219582-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 172, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 552.278118] env[63515]: DEBUG nova.network.neutron [None req-222754a5-7546-43c7-a99a-1c936b9e5035 tempest-ImagesOneServerTestJSON-1950372435 tempest-ImagesOneServerTestJSON-1950372435-project-member] [instance: aec66939-7ebc-4bfc-bd8b-bae22e011239] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 552.428502] env[63515]: DEBUG oslo_concurrency.lockutils [None req-0d00ff28-5b1e-46b5-9fb9-df8b8bcd6a1a tempest-TenantUsagesTestJSON-427870877 tempest-TenantUsagesTestJSON-427870877-project-member] Lock "9f7ad2dd-b882-4d98-8576-708b3abfa505" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 26.286s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 552.628214] env[63515]: DEBUG nova.compute.manager [req-92a44237-b15a-4b1b-a150-2c436746549b req-6c72c87a-4089-4b7b-9532-084434eb1fdd service nova] [instance: d7fed938-f7cd-4013-b545-5d5350242084] Received event network-vif-deleted-abac7593-4174-4f4c-886d-f2653e2b4123 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 552.628412] env[63515]: DEBUG nova.compute.manager [req-92a44237-b15a-4b1b-a150-2c436746549b req-6c72c87a-4089-4b7b-9532-084434eb1fdd service nova] [instance: aec66939-7ebc-4bfc-bd8b-bae22e011239] Received event network-changed-b5d340e0-18c9-43c1-bcc4-7b682823c4c4 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 552.628569] env[63515]: DEBUG nova.compute.manager [req-92a44237-b15a-4b1b-a150-2c436746549b req-6c72c87a-4089-4b7b-9532-084434eb1fdd service nova] [instance: aec66939-7ebc-4bfc-bd8b-bae22e011239] Refreshing instance network info cache due to event network-changed-b5d340e0-18c9-43c1-bcc4-7b682823c4c4. {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 552.628752] env[63515]: DEBUG oslo_concurrency.lockutils [req-92a44237-b15a-4b1b-a150-2c436746549b req-6c72c87a-4089-4b7b-9532-084434eb1fdd service nova] Acquiring lock "refresh_cache-aec66939-7ebc-4bfc-bd8b-bae22e011239" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 552.716625] env[63515]: DEBUG oslo_concurrency.lockutils [None req-919af50b-07d1-4794-942f-d99fa56afe84 tempest-ServerExternalEventsTest-1976219582 tempest-ServerExternalEventsTest-1976219582-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.867s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 552.717328] env[63515]: ERROR nova.compute.manager [None req-919af50b-07d1-4794-942f-d99fa56afe84 tempest-ServerExternalEventsTest-1976219582 tempest-ServerExternalEventsTest-1976219582-project-member] [instance: 6662da99-ba90-49ac-982e-d18f1c7834b6] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 39b0bc20-a4b4-4683-ac31-11d4a2fa8333, please check neutron logs for more information. [ 552.717328] env[63515]: ERROR nova.compute.manager [instance: 6662da99-ba90-49ac-982e-d18f1c7834b6] Traceback (most recent call last): [ 552.717328] env[63515]: ERROR nova.compute.manager [instance: 6662da99-ba90-49ac-982e-d18f1c7834b6] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 552.717328] env[63515]: ERROR nova.compute.manager [instance: 6662da99-ba90-49ac-982e-d18f1c7834b6] self.driver.spawn(context, instance, image_meta, [ 552.717328] env[63515]: ERROR nova.compute.manager [instance: 6662da99-ba90-49ac-982e-d18f1c7834b6] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 552.717328] env[63515]: ERROR nova.compute.manager [instance: 6662da99-ba90-49ac-982e-d18f1c7834b6] self._vmops.spawn(context, instance, image_meta, injected_files, [ 552.717328] env[63515]: ERROR nova.compute.manager [instance: 6662da99-ba90-49ac-982e-d18f1c7834b6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 552.717328] env[63515]: ERROR nova.compute.manager [instance: 6662da99-ba90-49ac-982e-d18f1c7834b6] vm_ref = self.build_virtual_machine(instance, [ 552.717328] env[63515]: ERROR nova.compute.manager [instance: 6662da99-ba90-49ac-982e-d18f1c7834b6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 552.717328] env[63515]: ERROR nova.compute.manager [instance: 6662da99-ba90-49ac-982e-d18f1c7834b6] vif_infos = vmwarevif.get_vif_info(self._session, [ 552.717328] env[63515]: ERROR nova.compute.manager [instance: 6662da99-ba90-49ac-982e-d18f1c7834b6] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 552.718030] env[63515]: ERROR nova.compute.manager [instance: 6662da99-ba90-49ac-982e-d18f1c7834b6] for vif in network_info: [ 552.718030] env[63515]: ERROR nova.compute.manager [instance: 6662da99-ba90-49ac-982e-d18f1c7834b6] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 552.718030] env[63515]: ERROR nova.compute.manager [instance: 6662da99-ba90-49ac-982e-d18f1c7834b6] return self._sync_wrapper(fn, *args, **kwargs) [ 552.718030] env[63515]: ERROR nova.compute.manager [instance: 6662da99-ba90-49ac-982e-d18f1c7834b6] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 552.718030] env[63515]: ERROR nova.compute.manager [instance: 6662da99-ba90-49ac-982e-d18f1c7834b6] self.wait() [ 552.718030] env[63515]: ERROR nova.compute.manager [instance: 6662da99-ba90-49ac-982e-d18f1c7834b6] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 552.718030] env[63515]: ERROR nova.compute.manager [instance: 6662da99-ba90-49ac-982e-d18f1c7834b6] self[:] = self._gt.wait() [ 552.718030] env[63515]: ERROR nova.compute.manager [instance: 6662da99-ba90-49ac-982e-d18f1c7834b6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 552.718030] env[63515]: ERROR nova.compute.manager [instance: 6662da99-ba90-49ac-982e-d18f1c7834b6] return self._exit_event.wait() [ 552.718030] env[63515]: ERROR nova.compute.manager [instance: 6662da99-ba90-49ac-982e-d18f1c7834b6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 552.718030] env[63515]: ERROR nova.compute.manager [instance: 6662da99-ba90-49ac-982e-d18f1c7834b6] result = hub.switch() [ 552.718030] env[63515]: ERROR nova.compute.manager [instance: 6662da99-ba90-49ac-982e-d18f1c7834b6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 552.718030] env[63515]: ERROR nova.compute.manager [instance: 6662da99-ba90-49ac-982e-d18f1c7834b6] return self.greenlet.switch() [ 552.718565] env[63515]: ERROR nova.compute.manager [instance: 6662da99-ba90-49ac-982e-d18f1c7834b6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 552.718565] env[63515]: ERROR nova.compute.manager [instance: 6662da99-ba90-49ac-982e-d18f1c7834b6] result = function(*args, **kwargs) [ 552.718565] env[63515]: ERROR nova.compute.manager [instance: 6662da99-ba90-49ac-982e-d18f1c7834b6] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 552.718565] env[63515]: ERROR nova.compute.manager [instance: 6662da99-ba90-49ac-982e-d18f1c7834b6] return func(*args, **kwargs) [ 552.718565] env[63515]: ERROR nova.compute.manager [instance: 6662da99-ba90-49ac-982e-d18f1c7834b6] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 552.718565] env[63515]: ERROR nova.compute.manager [instance: 6662da99-ba90-49ac-982e-d18f1c7834b6] raise e [ 552.718565] env[63515]: ERROR nova.compute.manager [instance: 6662da99-ba90-49ac-982e-d18f1c7834b6] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 552.718565] env[63515]: ERROR nova.compute.manager [instance: 6662da99-ba90-49ac-982e-d18f1c7834b6] nwinfo = self.network_api.allocate_for_instance( [ 552.718565] env[63515]: ERROR nova.compute.manager [instance: 6662da99-ba90-49ac-982e-d18f1c7834b6] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 552.718565] env[63515]: ERROR nova.compute.manager [instance: 6662da99-ba90-49ac-982e-d18f1c7834b6] created_port_ids = self._update_ports_for_instance( [ 552.718565] env[63515]: ERROR nova.compute.manager [instance: 6662da99-ba90-49ac-982e-d18f1c7834b6] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 552.718565] env[63515]: ERROR nova.compute.manager [instance: 6662da99-ba90-49ac-982e-d18f1c7834b6] with excutils.save_and_reraise_exception(): [ 552.718565] env[63515]: ERROR nova.compute.manager [instance: 6662da99-ba90-49ac-982e-d18f1c7834b6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 552.719119] env[63515]: ERROR nova.compute.manager [instance: 6662da99-ba90-49ac-982e-d18f1c7834b6] self.force_reraise() [ 552.719119] env[63515]: ERROR nova.compute.manager [instance: 6662da99-ba90-49ac-982e-d18f1c7834b6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 552.719119] env[63515]: ERROR nova.compute.manager [instance: 6662da99-ba90-49ac-982e-d18f1c7834b6] raise self.value [ 552.719119] env[63515]: ERROR nova.compute.manager [instance: 6662da99-ba90-49ac-982e-d18f1c7834b6] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 552.719119] env[63515]: ERROR nova.compute.manager [instance: 6662da99-ba90-49ac-982e-d18f1c7834b6] updated_port = self._update_port( [ 552.719119] env[63515]: ERROR nova.compute.manager [instance: 6662da99-ba90-49ac-982e-d18f1c7834b6] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 552.719119] env[63515]: ERROR nova.compute.manager [instance: 6662da99-ba90-49ac-982e-d18f1c7834b6] _ensure_no_port_binding_failure(port) [ 552.719119] env[63515]: ERROR nova.compute.manager [instance: 6662da99-ba90-49ac-982e-d18f1c7834b6] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 552.719119] env[63515]: ERROR nova.compute.manager [instance: 6662da99-ba90-49ac-982e-d18f1c7834b6] raise exception.PortBindingFailed(port_id=port['id']) [ 552.719119] env[63515]: ERROR nova.compute.manager [instance: 6662da99-ba90-49ac-982e-d18f1c7834b6] nova.exception.PortBindingFailed: Binding failed for port 39b0bc20-a4b4-4683-ac31-11d4a2fa8333, please check neutron logs for more information. [ 552.719119] env[63515]: ERROR nova.compute.manager [instance: 6662da99-ba90-49ac-982e-d18f1c7834b6] [ 552.719568] env[63515]: DEBUG nova.compute.utils [None req-919af50b-07d1-4794-942f-d99fa56afe84 tempest-ServerExternalEventsTest-1976219582 tempest-ServerExternalEventsTest-1976219582-project-member] [instance: 6662da99-ba90-49ac-982e-d18f1c7834b6] Binding failed for port 39b0bc20-a4b4-4683-ac31-11d4a2fa8333, please check neutron logs for more information. {{(pid=63515) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 552.720479] env[63515]: DEBUG oslo_concurrency.lockutils [None req-a5c600bc-b1c1-4a6d-9386-2b0dbd0395e4 tempest-InstanceActionsNegativeTestJSON-1061405461 tempest-InstanceActionsNegativeTestJSON-1061405461-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.556s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 552.723680] env[63515]: INFO nova.compute.claims [None req-a5c600bc-b1c1-4a6d-9386-2b0dbd0395e4 tempest-InstanceActionsNegativeTestJSON-1061405461 tempest-InstanceActionsNegativeTestJSON-1061405461-project-member] [instance: b641d2d1-de70-48b6-9137-fe18c3bc7511] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 552.727901] env[63515]: DEBUG nova.compute.manager [None req-919af50b-07d1-4794-942f-d99fa56afe84 tempest-ServerExternalEventsTest-1976219582 tempest-ServerExternalEventsTest-1976219582-project-member] [instance: 6662da99-ba90-49ac-982e-d18f1c7834b6] Build of instance 6662da99-ba90-49ac-982e-d18f1c7834b6 was re-scheduled: Binding failed for port 39b0bc20-a4b4-4683-ac31-11d4a2fa8333, please check neutron logs for more information. {{(pid=63515) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 552.728635] env[63515]: DEBUG nova.compute.manager [None req-919af50b-07d1-4794-942f-d99fa56afe84 tempest-ServerExternalEventsTest-1976219582 tempest-ServerExternalEventsTest-1976219582-project-member] [instance: 6662da99-ba90-49ac-982e-d18f1c7834b6] Unplugging VIFs for instance {{(pid=63515) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 552.729262] env[63515]: DEBUG oslo_concurrency.lockutils [None req-919af50b-07d1-4794-942f-d99fa56afe84 tempest-ServerExternalEventsTest-1976219582 tempest-ServerExternalEventsTest-1976219582-project-member] Acquiring lock "refresh_cache-6662da99-ba90-49ac-982e-d18f1c7834b6" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 552.729441] env[63515]: DEBUG oslo_concurrency.lockutils [None req-919af50b-07d1-4794-942f-d99fa56afe84 tempest-ServerExternalEventsTest-1976219582 tempest-ServerExternalEventsTest-1976219582-project-member] Acquired lock "refresh_cache-6662da99-ba90-49ac-982e-d18f1c7834b6" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 552.729606] env[63515]: DEBUG nova.network.neutron [None req-919af50b-07d1-4794-942f-d99fa56afe84 tempest-ServerExternalEventsTest-1976219582 tempest-ServerExternalEventsTest-1976219582-project-member] [instance: 6662da99-ba90-49ac-982e-d18f1c7834b6] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 552.782287] env[63515]: DEBUG oslo_concurrency.lockutils [None req-222754a5-7546-43c7-a99a-1c936b9e5035 tempest-ImagesOneServerTestJSON-1950372435 tempest-ImagesOneServerTestJSON-1950372435-project-member] Releasing lock "refresh_cache-aec66939-7ebc-4bfc-bd8b-bae22e011239" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 552.782687] env[63515]: DEBUG nova.compute.manager [None req-222754a5-7546-43c7-a99a-1c936b9e5035 tempest-ImagesOneServerTestJSON-1950372435 tempest-ImagesOneServerTestJSON-1950372435-project-member] [instance: aec66939-7ebc-4bfc-bd8b-bae22e011239] Start destroying the instance on the hypervisor. {{(pid=63515) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 552.782911] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-222754a5-7546-43c7-a99a-1c936b9e5035 tempest-ImagesOneServerTestJSON-1950372435 tempest-ImagesOneServerTestJSON-1950372435-project-member] [instance: aec66939-7ebc-4bfc-bd8b-bae22e011239] Destroying instance {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 552.783435] env[63515]: DEBUG oslo_concurrency.lockutils [req-92a44237-b15a-4b1b-a150-2c436746549b req-6c72c87a-4089-4b7b-9532-084434eb1fdd service nova] Acquired lock "refresh_cache-aec66939-7ebc-4bfc-bd8b-bae22e011239" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 552.783435] env[63515]: DEBUG nova.network.neutron [req-92a44237-b15a-4b1b-a150-2c436746549b req-6c72c87a-4089-4b7b-9532-084434eb1fdd service nova] [instance: aec66939-7ebc-4bfc-bd8b-bae22e011239] Refreshing network info cache for port b5d340e0-18c9-43c1-bcc4-7b682823c4c4 {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 552.785155] env[63515]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a500e45a-79b9-4812-b298-c7eec7559d7b {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 552.795097] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-abce5f3b-0c8f-4359-a69b-606275823feb {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 552.819336] env[63515]: WARNING nova.virt.vmwareapi.vmops [None req-222754a5-7546-43c7-a99a-1c936b9e5035 tempest-ImagesOneServerTestJSON-1950372435 tempest-ImagesOneServerTestJSON-1950372435-project-member] [instance: aec66939-7ebc-4bfc-bd8b-bae22e011239] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance aec66939-7ebc-4bfc-bd8b-bae22e011239 could not be found. [ 552.819605] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-222754a5-7546-43c7-a99a-1c936b9e5035 tempest-ImagesOneServerTestJSON-1950372435 tempest-ImagesOneServerTestJSON-1950372435-project-member] [instance: aec66939-7ebc-4bfc-bd8b-bae22e011239] Instance destroyed {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 552.819787] env[63515]: INFO nova.compute.manager [None req-222754a5-7546-43c7-a99a-1c936b9e5035 tempest-ImagesOneServerTestJSON-1950372435 tempest-ImagesOneServerTestJSON-1950372435-project-member] [instance: aec66939-7ebc-4bfc-bd8b-bae22e011239] Took 0.04 seconds to destroy the instance on the hypervisor. [ 552.820645] env[63515]: DEBUG oslo.service.loopingcall [None req-222754a5-7546-43c7-a99a-1c936b9e5035 tempest-ImagesOneServerTestJSON-1950372435 tempest-ImagesOneServerTestJSON-1950372435-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 552.820645] env[63515]: DEBUG nova.compute.manager [-] [instance: aec66939-7ebc-4bfc-bd8b-bae22e011239] Deallocating network for instance {{(pid=63515) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 552.820770] env[63515]: DEBUG nova.network.neutron [-] [instance: aec66939-7ebc-4bfc-bd8b-bae22e011239] deallocate_for_instance() {{(pid=63515) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 552.893820] env[63515]: DEBUG nova.compute.manager [None req-757a72d0-3dae-4296-830e-18ff5e948610 tempest-ServerDiagnosticsNegativeTest-1799249066 tempest-ServerDiagnosticsNegativeTest-1799249066-project-member] [instance: 1c5d87c7-6747-45a7-8488-6f107eeaeb78] Start spawning the instance on the hypervisor. {{(pid=63515) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 552.902843] env[63515]: DEBUG nova.network.neutron [-] [instance: aec66939-7ebc-4bfc-bd8b-bae22e011239] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 552.924059] env[63515]: DEBUG nova.virt.hardware [None req-757a72d0-3dae-4296-830e-18ff5e948610 tempest-ServerDiagnosticsNegativeTest-1799249066 tempest-ServerDiagnosticsNegativeTest-1799249066-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T02:52:14Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T02:51:56Z,direct_url=,disk_format='vmdk',id=8a120570-cb06-4099-b262-554ca0ad15c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b569255cc43e42e7ae3f2b4ad37c6ef4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T02:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 552.924838] env[63515]: DEBUG nova.virt.hardware [None req-757a72d0-3dae-4296-830e-18ff5e948610 tempest-ServerDiagnosticsNegativeTest-1799249066 tempest-ServerDiagnosticsNegativeTest-1799249066-project-member] Flavor limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 552.924838] env[63515]: DEBUG nova.virt.hardware [None req-757a72d0-3dae-4296-830e-18ff5e948610 tempest-ServerDiagnosticsNegativeTest-1799249066 tempest-ServerDiagnosticsNegativeTest-1799249066-project-member] Image limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 552.924838] env[63515]: DEBUG nova.virt.hardware [None req-757a72d0-3dae-4296-830e-18ff5e948610 tempest-ServerDiagnosticsNegativeTest-1799249066 tempest-ServerDiagnosticsNegativeTest-1799249066-project-member] Flavor pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 552.924838] env[63515]: DEBUG nova.virt.hardware [None req-757a72d0-3dae-4296-830e-18ff5e948610 tempest-ServerDiagnosticsNegativeTest-1799249066 tempest-ServerDiagnosticsNegativeTest-1799249066-project-member] Image pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 552.925170] env[63515]: DEBUG nova.virt.hardware [None req-757a72d0-3dae-4296-830e-18ff5e948610 tempest-ServerDiagnosticsNegativeTest-1799249066 tempest-ServerDiagnosticsNegativeTest-1799249066-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 552.926165] env[63515]: DEBUG nova.virt.hardware [None req-757a72d0-3dae-4296-830e-18ff5e948610 tempest-ServerDiagnosticsNegativeTest-1799249066 tempest-ServerDiagnosticsNegativeTest-1799249066-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 552.926338] env[63515]: DEBUG nova.virt.hardware [None req-757a72d0-3dae-4296-830e-18ff5e948610 tempest-ServerDiagnosticsNegativeTest-1799249066 tempest-ServerDiagnosticsNegativeTest-1799249066-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 552.926610] env[63515]: DEBUG nova.virt.hardware [None req-757a72d0-3dae-4296-830e-18ff5e948610 tempest-ServerDiagnosticsNegativeTest-1799249066 tempest-ServerDiagnosticsNegativeTest-1799249066-project-member] Got 1 possible topologies {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 552.926726] env[63515]: DEBUG nova.virt.hardware [None req-757a72d0-3dae-4296-830e-18ff5e948610 tempest-ServerDiagnosticsNegativeTest-1799249066 tempest-ServerDiagnosticsNegativeTest-1799249066-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 552.926833] env[63515]: DEBUG nova.virt.hardware [None req-757a72d0-3dae-4296-830e-18ff5e948610 tempest-ServerDiagnosticsNegativeTest-1799249066 tempest-ServerDiagnosticsNegativeTest-1799249066-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 552.928857] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e0b97a6-7f77-4849-997a-60cc34842c90 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 552.934035] env[63515]: DEBUG nova.compute.manager [None req-7f5043f7-e6d7-46dd-88fd-99ee049c2b4b tempest-VolumesAssistedSnapshotsTest-453349191 tempest-VolumesAssistedSnapshotsTest-453349191-project-member] [instance: db682708-36fa-4126-a848-bfb609df96d7] Starting instance... {{(pid=63515) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 552.946187] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-210a99cb-f84f-4a81-a937-695326e72c3d {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 553.155932] env[63515]: DEBUG nova.network.neutron [None req-757a72d0-3dae-4296-830e-18ff5e948610 tempest-ServerDiagnosticsNegativeTest-1799249066 tempest-ServerDiagnosticsNegativeTest-1799249066-project-member] [instance: 1c5d87c7-6747-45a7-8488-6f107eeaeb78] Successfully created port: bea3bc2e-ac36-4c4f-8398-abe757f5b66f {{(pid=63515) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 553.290563] env[63515]: DEBUG nova.network.neutron [None req-919af50b-07d1-4794-942f-d99fa56afe84 tempest-ServerExternalEventsTest-1976219582 tempest-ServerExternalEventsTest-1976219582-project-member] [instance: 6662da99-ba90-49ac-982e-d18f1c7834b6] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 553.350059] env[63515]: DEBUG nova.network.neutron [req-92a44237-b15a-4b1b-a150-2c436746549b req-6c72c87a-4089-4b7b-9532-084434eb1fdd service nova] [instance: aec66939-7ebc-4bfc-bd8b-bae22e011239] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 553.407994] env[63515]: DEBUG nova.network.neutron [-] [instance: aec66939-7ebc-4bfc-bd8b-bae22e011239] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 553.463411] env[63515]: DEBUG oslo_concurrency.lockutils [None req-7f5043f7-e6d7-46dd-88fd-99ee049c2b4b tempest-VolumesAssistedSnapshotsTest-453349191 tempest-VolumesAssistedSnapshotsTest-453349191-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 553.693483] env[63515]: DEBUG oslo_concurrency.lockutils [None req-d9ffa9d5-6856-47e5-8926-d090bbea2c09 tempest-DeleteServersAdminTestJSON-944542165 tempest-DeleteServersAdminTestJSON-944542165-project-member] Acquiring lock "ae5c6e69-a885-48e3-9c9e-017f33430edd" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 553.693767] env[63515]: DEBUG oslo_concurrency.lockutils [None req-d9ffa9d5-6856-47e5-8926-d090bbea2c09 tempest-DeleteServersAdminTestJSON-944542165 tempest-DeleteServersAdminTestJSON-944542165-project-member] Lock "ae5c6e69-a885-48e3-9c9e-017f33430edd" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 553.718902] env[63515]: DEBUG nova.network.neutron [None req-919af50b-07d1-4794-942f-d99fa56afe84 tempest-ServerExternalEventsTest-1976219582 tempest-ServerExternalEventsTest-1976219582-project-member] [instance: 6662da99-ba90-49ac-982e-d18f1c7834b6] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 553.873125] env[63515]: DEBUG nova.network.neutron [req-92a44237-b15a-4b1b-a150-2c436746549b req-6c72c87a-4089-4b7b-9532-084434eb1fdd service nova] [instance: aec66939-7ebc-4bfc-bd8b-bae22e011239] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 553.914624] env[63515]: INFO nova.compute.manager [-] [instance: aec66939-7ebc-4bfc-bd8b-bae22e011239] Took 1.09 seconds to deallocate network for instance. [ 553.918378] env[63515]: DEBUG nova.compute.claims [None req-222754a5-7546-43c7-a99a-1c936b9e5035 tempest-ImagesOneServerTestJSON-1950372435 tempest-ImagesOneServerTestJSON-1950372435-project-member] [instance: aec66939-7ebc-4bfc-bd8b-bae22e011239] Aborting claim: {{(pid=63515) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 553.918486] env[63515]: DEBUG oslo_concurrency.lockutils [None req-222754a5-7546-43c7-a99a-1c936b9e5035 tempest-ImagesOneServerTestJSON-1950372435 tempest-ImagesOneServerTestJSON-1950372435-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 554.031147] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b822771-2e23-4854-9b23-576e1b157016 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 554.039601] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d614b9b2-c133-4893-900d-99be85f024c2 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 554.076495] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7f961d5-c1ec-463e-b1de-65d35edf0ca4 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 554.085544] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55efcc2e-c7b6-4311-a0b5-6b5db88487e4 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 554.100461] env[63515]: DEBUG nova.compute.provider_tree [None req-a5c600bc-b1c1-4a6d-9386-2b0dbd0395e4 tempest-InstanceActionsNegativeTestJSON-1061405461 tempest-InstanceActionsNegativeTestJSON-1061405461-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 554.221656] env[63515]: DEBUG oslo_concurrency.lockutils [None req-919af50b-07d1-4794-942f-d99fa56afe84 tempest-ServerExternalEventsTest-1976219582 tempest-ServerExternalEventsTest-1976219582-project-member] Releasing lock "refresh_cache-6662da99-ba90-49ac-982e-d18f1c7834b6" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 554.221979] env[63515]: DEBUG nova.compute.manager [None req-919af50b-07d1-4794-942f-d99fa56afe84 tempest-ServerExternalEventsTest-1976219582 tempest-ServerExternalEventsTest-1976219582-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63515) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 554.222077] env[63515]: DEBUG nova.compute.manager [None req-919af50b-07d1-4794-942f-d99fa56afe84 tempest-ServerExternalEventsTest-1976219582 tempest-ServerExternalEventsTest-1976219582-project-member] [instance: 6662da99-ba90-49ac-982e-d18f1c7834b6] Deallocating network for instance {{(pid=63515) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 554.222251] env[63515]: DEBUG nova.network.neutron [None req-919af50b-07d1-4794-942f-d99fa56afe84 tempest-ServerExternalEventsTest-1976219582 tempest-ServerExternalEventsTest-1976219582-project-member] [instance: 6662da99-ba90-49ac-982e-d18f1c7834b6] deallocate_for_instance() {{(pid=63515) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 554.247393] env[63515]: DEBUG nova.network.neutron [None req-919af50b-07d1-4794-942f-d99fa56afe84 tempest-ServerExternalEventsTest-1976219582 tempest-ServerExternalEventsTest-1976219582-project-member] [instance: 6662da99-ba90-49ac-982e-d18f1c7834b6] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 554.378736] env[63515]: DEBUG oslo_concurrency.lockutils [req-92a44237-b15a-4b1b-a150-2c436746549b req-6c72c87a-4089-4b7b-9532-084434eb1fdd service nova] Releasing lock "refresh_cache-aec66939-7ebc-4bfc-bd8b-bae22e011239" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 554.604844] env[63515]: DEBUG nova.scheduler.client.report [None req-a5c600bc-b1c1-4a6d-9386-2b0dbd0395e4 tempest-InstanceActionsNegativeTestJSON-1061405461 tempest-InstanceActionsNegativeTestJSON-1061405461-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 172, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 554.751678] env[63515]: DEBUG nova.network.neutron [None req-919af50b-07d1-4794-942f-d99fa56afe84 tempest-ServerExternalEventsTest-1976219582 tempest-ServerExternalEventsTest-1976219582-project-member] [instance: 6662da99-ba90-49ac-982e-d18f1c7834b6] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 555.110148] env[63515]: DEBUG oslo_concurrency.lockutils [None req-a5c600bc-b1c1-4a6d-9386-2b0dbd0395e4 tempest-InstanceActionsNegativeTestJSON-1061405461 tempest-InstanceActionsNegativeTestJSON-1061405461-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.390s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 555.111189] env[63515]: DEBUG nova.compute.manager [None req-a5c600bc-b1c1-4a6d-9386-2b0dbd0395e4 tempest-InstanceActionsNegativeTestJSON-1061405461 tempest-InstanceActionsNegativeTestJSON-1061405461-project-member] [instance: b641d2d1-de70-48b6-9137-fe18c3bc7511] Start building networks asynchronously for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 555.114036] env[63515]: DEBUG oslo_concurrency.lockutils [None req-50ea3e37-671c-4a4b-9c1a-d782ff1b578c tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.591s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 555.115517] env[63515]: INFO nova.compute.claims [None req-50ea3e37-671c-4a4b-9c1a-d782ff1b578c tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] [instance: cfc6c8d5-c9e8-46fb-99e7-d245664f9652] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 555.255141] env[63515]: INFO nova.compute.manager [None req-919af50b-07d1-4794-942f-d99fa56afe84 tempest-ServerExternalEventsTest-1976219582 tempest-ServerExternalEventsTest-1976219582-project-member] [instance: 6662da99-ba90-49ac-982e-d18f1c7834b6] Took 1.03 seconds to deallocate network for instance. [ 555.423580] env[63515]: DEBUG oslo_concurrency.lockutils [None req-9f5c50b0-6bbe-4982-8818-c7255a8c23ab tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Acquiring lock "b53754b1-ec31-4dc3-a839-771f71affe1e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 555.423894] env[63515]: DEBUG oslo_concurrency.lockutils [None req-9f5c50b0-6bbe-4982-8818-c7255a8c23ab tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Lock "b53754b1-ec31-4dc3-a839-771f71affe1e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 555.624226] env[63515]: DEBUG nova.compute.utils [None req-a5c600bc-b1c1-4a6d-9386-2b0dbd0395e4 tempest-InstanceActionsNegativeTestJSON-1061405461 tempest-InstanceActionsNegativeTestJSON-1061405461-project-member] Using /dev/sd instead of None {{(pid=63515) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 555.626104] env[63515]: DEBUG nova.compute.manager [None req-a5c600bc-b1c1-4a6d-9386-2b0dbd0395e4 tempest-InstanceActionsNegativeTestJSON-1061405461 tempest-InstanceActionsNegativeTestJSON-1061405461-project-member] [instance: b641d2d1-de70-48b6-9137-fe18c3bc7511] Allocating IP information in the background. {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 555.626359] env[63515]: DEBUG nova.network.neutron [None req-a5c600bc-b1c1-4a6d-9386-2b0dbd0395e4 tempest-InstanceActionsNegativeTestJSON-1061405461 tempest-InstanceActionsNegativeTestJSON-1061405461-project-member] [instance: b641d2d1-de70-48b6-9137-fe18c3bc7511] allocate_for_instance() {{(pid=63515) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 555.721994] env[63515]: DEBUG nova.policy [None req-a5c600bc-b1c1-4a6d-9386-2b0dbd0395e4 tempest-InstanceActionsNegativeTestJSON-1061405461 tempest-InstanceActionsNegativeTestJSON-1061405461-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '07455cabdfb34353b7d67ee15e9f9684', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b4cc505efc394b568c804d92778978af', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63515) authorize /opt/stack/nova/nova/policy.py:201}} [ 555.879625] env[63515]: DEBUG oslo_concurrency.lockutils [None req-886adfc0-007d-4f83-bc30-b755d00d04c7 tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] Acquiring lock "8b65c223-03a8-41e0-917e-45dd76b6e57f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 555.880165] env[63515]: DEBUG oslo_concurrency.lockutils [None req-886adfc0-007d-4f83-bc30-b755d00d04c7 tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] Lock "8b65c223-03a8-41e0-917e-45dd76b6e57f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 556.131588] env[63515]: DEBUG nova.compute.manager [None req-a5c600bc-b1c1-4a6d-9386-2b0dbd0395e4 tempest-InstanceActionsNegativeTestJSON-1061405461 tempest-InstanceActionsNegativeTestJSON-1061405461-project-member] [instance: b641d2d1-de70-48b6-9137-fe18c3bc7511] Start building block device mappings for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 556.305847] env[63515]: INFO nova.scheduler.client.report [None req-919af50b-07d1-4794-942f-d99fa56afe84 tempest-ServerExternalEventsTest-1976219582 tempest-ServerExternalEventsTest-1976219582-project-member] Deleted allocations for instance 6662da99-ba90-49ac-982e-d18f1c7834b6 [ 556.528938] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2662f4bb-2860-46c4-8f22-54a4bccbdcf5 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 556.537712] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9c4e666-1a54-4bcd-9c82-137f1a84b316 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 556.582370] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b144261-ca77-49c4-9764-9c0b9b51dbb4 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 556.587309] env[63515]: ERROR nova.compute.manager [None req-757a72d0-3dae-4296-830e-18ff5e948610 tempest-ServerDiagnosticsNegativeTest-1799249066 tempest-ServerDiagnosticsNegativeTest-1799249066-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port bea3bc2e-ac36-4c4f-8398-abe757f5b66f, please check neutron logs for more information. [ 556.587309] env[63515]: ERROR nova.compute.manager Traceback (most recent call last): [ 556.587309] env[63515]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 556.587309] env[63515]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 556.587309] env[63515]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 556.587309] env[63515]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 556.587309] env[63515]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 556.587309] env[63515]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 556.587309] env[63515]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 556.587309] env[63515]: ERROR nova.compute.manager self.force_reraise() [ 556.587309] env[63515]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 556.587309] env[63515]: ERROR nova.compute.manager raise self.value [ 556.587309] env[63515]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 556.587309] env[63515]: ERROR nova.compute.manager updated_port = self._update_port( [ 556.587309] env[63515]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 556.587309] env[63515]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 556.587966] env[63515]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 556.587966] env[63515]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 556.587966] env[63515]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port bea3bc2e-ac36-4c4f-8398-abe757f5b66f, please check neutron logs for more information. [ 556.587966] env[63515]: ERROR nova.compute.manager [ 556.587966] env[63515]: Traceback (most recent call last): [ 556.587966] env[63515]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 556.587966] env[63515]: listener.cb(fileno) [ 556.587966] env[63515]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 556.587966] env[63515]: result = function(*args, **kwargs) [ 556.587966] env[63515]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 556.587966] env[63515]: return func(*args, **kwargs) [ 556.587966] env[63515]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 556.587966] env[63515]: raise e [ 556.587966] env[63515]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 556.587966] env[63515]: nwinfo = self.network_api.allocate_for_instance( [ 556.587966] env[63515]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 556.587966] env[63515]: created_port_ids = self._update_ports_for_instance( [ 556.587966] env[63515]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 556.587966] env[63515]: with excutils.save_and_reraise_exception(): [ 556.587966] env[63515]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 556.587966] env[63515]: self.force_reraise() [ 556.587966] env[63515]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 556.587966] env[63515]: raise self.value [ 556.587966] env[63515]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 556.587966] env[63515]: updated_port = self._update_port( [ 556.587966] env[63515]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 556.587966] env[63515]: _ensure_no_port_binding_failure(port) [ 556.587966] env[63515]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 556.587966] env[63515]: raise exception.PortBindingFailed(port_id=port['id']) [ 556.589378] env[63515]: nova.exception.PortBindingFailed: Binding failed for port bea3bc2e-ac36-4c4f-8398-abe757f5b66f, please check neutron logs for more information. [ 556.589378] env[63515]: Removing descriptor: 16 [ 556.590896] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7bfb8a44-87c4-43b0-a55a-21c3a3f7acd5 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 556.595621] env[63515]: ERROR nova.compute.manager [None req-757a72d0-3dae-4296-830e-18ff5e948610 tempest-ServerDiagnosticsNegativeTest-1799249066 tempest-ServerDiagnosticsNegativeTest-1799249066-project-member] [instance: 1c5d87c7-6747-45a7-8488-6f107eeaeb78] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port bea3bc2e-ac36-4c4f-8398-abe757f5b66f, please check neutron logs for more information. [ 556.595621] env[63515]: ERROR nova.compute.manager [instance: 1c5d87c7-6747-45a7-8488-6f107eeaeb78] Traceback (most recent call last): [ 556.595621] env[63515]: ERROR nova.compute.manager [instance: 1c5d87c7-6747-45a7-8488-6f107eeaeb78] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 556.595621] env[63515]: ERROR nova.compute.manager [instance: 1c5d87c7-6747-45a7-8488-6f107eeaeb78] yield resources [ 556.595621] env[63515]: ERROR nova.compute.manager [instance: 1c5d87c7-6747-45a7-8488-6f107eeaeb78] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 556.595621] env[63515]: ERROR nova.compute.manager [instance: 1c5d87c7-6747-45a7-8488-6f107eeaeb78] self.driver.spawn(context, instance, image_meta, [ 556.595621] env[63515]: ERROR nova.compute.manager [instance: 1c5d87c7-6747-45a7-8488-6f107eeaeb78] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 556.595621] env[63515]: ERROR nova.compute.manager [instance: 1c5d87c7-6747-45a7-8488-6f107eeaeb78] self._vmops.spawn(context, instance, image_meta, injected_files, [ 556.595621] env[63515]: ERROR nova.compute.manager [instance: 1c5d87c7-6747-45a7-8488-6f107eeaeb78] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 556.595621] env[63515]: ERROR nova.compute.manager [instance: 1c5d87c7-6747-45a7-8488-6f107eeaeb78] vm_ref = self.build_virtual_machine(instance, [ 556.595621] env[63515]: ERROR nova.compute.manager [instance: 1c5d87c7-6747-45a7-8488-6f107eeaeb78] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 556.596214] env[63515]: ERROR nova.compute.manager [instance: 1c5d87c7-6747-45a7-8488-6f107eeaeb78] vif_infos = vmwarevif.get_vif_info(self._session, [ 556.596214] env[63515]: ERROR nova.compute.manager [instance: 1c5d87c7-6747-45a7-8488-6f107eeaeb78] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 556.596214] env[63515]: ERROR nova.compute.manager [instance: 1c5d87c7-6747-45a7-8488-6f107eeaeb78] for vif in network_info: [ 556.596214] env[63515]: ERROR nova.compute.manager [instance: 1c5d87c7-6747-45a7-8488-6f107eeaeb78] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 556.596214] env[63515]: ERROR nova.compute.manager [instance: 1c5d87c7-6747-45a7-8488-6f107eeaeb78] return self._sync_wrapper(fn, *args, **kwargs) [ 556.596214] env[63515]: ERROR nova.compute.manager [instance: 1c5d87c7-6747-45a7-8488-6f107eeaeb78] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 556.596214] env[63515]: ERROR nova.compute.manager [instance: 1c5d87c7-6747-45a7-8488-6f107eeaeb78] self.wait() [ 556.596214] env[63515]: ERROR nova.compute.manager [instance: 1c5d87c7-6747-45a7-8488-6f107eeaeb78] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 556.596214] env[63515]: ERROR nova.compute.manager [instance: 1c5d87c7-6747-45a7-8488-6f107eeaeb78] self[:] = self._gt.wait() [ 556.596214] env[63515]: ERROR nova.compute.manager [instance: 1c5d87c7-6747-45a7-8488-6f107eeaeb78] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 556.596214] env[63515]: ERROR nova.compute.manager [instance: 1c5d87c7-6747-45a7-8488-6f107eeaeb78] return self._exit_event.wait() [ 556.596214] env[63515]: ERROR nova.compute.manager [instance: 1c5d87c7-6747-45a7-8488-6f107eeaeb78] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 556.596214] env[63515]: ERROR nova.compute.manager [instance: 1c5d87c7-6747-45a7-8488-6f107eeaeb78] result = hub.switch() [ 556.596851] env[63515]: ERROR nova.compute.manager [instance: 1c5d87c7-6747-45a7-8488-6f107eeaeb78] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 556.596851] env[63515]: ERROR nova.compute.manager [instance: 1c5d87c7-6747-45a7-8488-6f107eeaeb78] return self.greenlet.switch() [ 556.596851] env[63515]: ERROR nova.compute.manager [instance: 1c5d87c7-6747-45a7-8488-6f107eeaeb78] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 556.596851] env[63515]: ERROR nova.compute.manager [instance: 1c5d87c7-6747-45a7-8488-6f107eeaeb78] result = function(*args, **kwargs) [ 556.596851] env[63515]: ERROR nova.compute.manager [instance: 1c5d87c7-6747-45a7-8488-6f107eeaeb78] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 556.596851] env[63515]: ERROR nova.compute.manager [instance: 1c5d87c7-6747-45a7-8488-6f107eeaeb78] return func(*args, **kwargs) [ 556.596851] env[63515]: ERROR nova.compute.manager [instance: 1c5d87c7-6747-45a7-8488-6f107eeaeb78] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 556.596851] env[63515]: ERROR nova.compute.manager [instance: 1c5d87c7-6747-45a7-8488-6f107eeaeb78] raise e [ 556.596851] env[63515]: ERROR nova.compute.manager [instance: 1c5d87c7-6747-45a7-8488-6f107eeaeb78] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 556.596851] env[63515]: ERROR nova.compute.manager [instance: 1c5d87c7-6747-45a7-8488-6f107eeaeb78] nwinfo = self.network_api.allocate_for_instance( [ 556.596851] env[63515]: ERROR nova.compute.manager [instance: 1c5d87c7-6747-45a7-8488-6f107eeaeb78] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 556.596851] env[63515]: ERROR nova.compute.manager [instance: 1c5d87c7-6747-45a7-8488-6f107eeaeb78] created_port_ids = self._update_ports_for_instance( [ 556.596851] env[63515]: ERROR nova.compute.manager [instance: 1c5d87c7-6747-45a7-8488-6f107eeaeb78] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 556.597466] env[63515]: ERROR nova.compute.manager [instance: 1c5d87c7-6747-45a7-8488-6f107eeaeb78] with excutils.save_and_reraise_exception(): [ 556.597466] env[63515]: ERROR nova.compute.manager [instance: 1c5d87c7-6747-45a7-8488-6f107eeaeb78] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 556.597466] env[63515]: ERROR nova.compute.manager [instance: 1c5d87c7-6747-45a7-8488-6f107eeaeb78] self.force_reraise() [ 556.597466] env[63515]: ERROR nova.compute.manager [instance: 1c5d87c7-6747-45a7-8488-6f107eeaeb78] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 556.597466] env[63515]: ERROR nova.compute.manager [instance: 1c5d87c7-6747-45a7-8488-6f107eeaeb78] raise self.value [ 556.597466] env[63515]: ERROR nova.compute.manager [instance: 1c5d87c7-6747-45a7-8488-6f107eeaeb78] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 556.597466] env[63515]: ERROR nova.compute.manager [instance: 1c5d87c7-6747-45a7-8488-6f107eeaeb78] updated_port = self._update_port( [ 556.597466] env[63515]: ERROR nova.compute.manager [instance: 1c5d87c7-6747-45a7-8488-6f107eeaeb78] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 556.597466] env[63515]: ERROR nova.compute.manager [instance: 1c5d87c7-6747-45a7-8488-6f107eeaeb78] _ensure_no_port_binding_failure(port) [ 556.597466] env[63515]: ERROR nova.compute.manager [instance: 1c5d87c7-6747-45a7-8488-6f107eeaeb78] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 556.597466] env[63515]: ERROR nova.compute.manager [instance: 1c5d87c7-6747-45a7-8488-6f107eeaeb78] raise exception.PortBindingFailed(port_id=port['id']) [ 556.597466] env[63515]: ERROR nova.compute.manager [instance: 1c5d87c7-6747-45a7-8488-6f107eeaeb78] nova.exception.PortBindingFailed: Binding failed for port bea3bc2e-ac36-4c4f-8398-abe757f5b66f, please check neutron logs for more information. [ 556.597466] env[63515]: ERROR nova.compute.manager [instance: 1c5d87c7-6747-45a7-8488-6f107eeaeb78] [ 556.599212] env[63515]: INFO nova.compute.manager [None req-757a72d0-3dae-4296-830e-18ff5e948610 tempest-ServerDiagnosticsNegativeTest-1799249066 tempest-ServerDiagnosticsNegativeTest-1799249066-project-member] [instance: 1c5d87c7-6747-45a7-8488-6f107eeaeb78] Terminating instance [ 556.599212] env[63515]: DEBUG oslo_concurrency.lockutils [None req-757a72d0-3dae-4296-830e-18ff5e948610 tempest-ServerDiagnosticsNegativeTest-1799249066 tempest-ServerDiagnosticsNegativeTest-1799249066-project-member] Acquiring lock "refresh_cache-1c5d87c7-6747-45a7-8488-6f107eeaeb78" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 556.599329] env[63515]: DEBUG oslo_concurrency.lockutils [None req-757a72d0-3dae-4296-830e-18ff5e948610 tempest-ServerDiagnosticsNegativeTest-1799249066 tempest-ServerDiagnosticsNegativeTest-1799249066-project-member] Acquired lock "refresh_cache-1c5d87c7-6747-45a7-8488-6f107eeaeb78" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 556.599545] env[63515]: DEBUG nova.network.neutron [None req-757a72d0-3dae-4296-830e-18ff5e948610 tempest-ServerDiagnosticsNegativeTest-1799249066 tempest-ServerDiagnosticsNegativeTest-1799249066-project-member] [instance: 1c5d87c7-6747-45a7-8488-6f107eeaeb78] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 556.609659] env[63515]: DEBUG nova.compute.provider_tree [None req-50ea3e37-671c-4a4b-9c1a-d782ff1b578c tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 556.794787] env[63515]: DEBUG nova.network.neutron [None req-a5c600bc-b1c1-4a6d-9386-2b0dbd0395e4 tempest-InstanceActionsNegativeTestJSON-1061405461 tempest-InstanceActionsNegativeTestJSON-1061405461-project-member] [instance: b641d2d1-de70-48b6-9137-fe18c3bc7511] Successfully created port: cab05064-e039-47ee-9e97-1d390704eedd {{(pid=63515) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 556.819495] env[63515]: DEBUG oslo_concurrency.lockutils [None req-919af50b-07d1-4794-942f-d99fa56afe84 tempest-ServerExternalEventsTest-1976219582 tempest-ServerExternalEventsTest-1976219582-project-member] Lock "6662da99-ba90-49ac-982e-d18f1c7834b6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 29.285s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 556.824468] env[63515]: DEBUG nova.compute.manager [req-7533914b-d68e-4edb-9f9a-2a11bcae7103 req-e8ea5893-7c8b-46c8-bd8b-84054ea74e23 service nova] [instance: aec66939-7ebc-4bfc-bd8b-bae22e011239] Received event network-vif-deleted-b5d340e0-18c9-43c1-bcc4-7b682823c4c4 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 557.093505] env[63515]: DEBUG oslo_concurrency.lockutils [None req-9131a685-703c-44a1-81f8-7c157d95e037 tempest-ServerAddressesNegativeTestJSON-1262066083 tempest-ServerAddressesNegativeTestJSON-1262066083-project-member] Acquiring lock "5b201ab6-c7ec-4a5c-a310-982649f34be4" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 557.093770] env[63515]: DEBUG oslo_concurrency.lockutils [None req-9131a685-703c-44a1-81f8-7c157d95e037 tempest-ServerAddressesNegativeTestJSON-1262066083 tempest-ServerAddressesNegativeTestJSON-1262066083-project-member] Lock "5b201ab6-c7ec-4a5c-a310-982649f34be4" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 557.114731] env[63515]: DEBUG nova.scheduler.client.report [None req-50ea3e37-671c-4a4b-9c1a-d782ff1b578c tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 172, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 557.145693] env[63515]: DEBUG nova.compute.manager [None req-a5c600bc-b1c1-4a6d-9386-2b0dbd0395e4 tempest-InstanceActionsNegativeTestJSON-1061405461 tempest-InstanceActionsNegativeTestJSON-1061405461-project-member] [instance: b641d2d1-de70-48b6-9137-fe18c3bc7511] Start spawning the instance on the hypervisor. {{(pid=63515) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 557.150520] env[63515]: DEBUG nova.network.neutron [None req-757a72d0-3dae-4296-830e-18ff5e948610 tempest-ServerDiagnosticsNegativeTest-1799249066 tempest-ServerDiagnosticsNegativeTest-1799249066-project-member] [instance: 1c5d87c7-6747-45a7-8488-6f107eeaeb78] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 557.191636] env[63515]: DEBUG nova.virt.hardware [None req-a5c600bc-b1c1-4a6d-9386-2b0dbd0395e4 tempest-InstanceActionsNegativeTestJSON-1061405461 tempest-InstanceActionsNegativeTestJSON-1061405461-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T02:52:14Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T02:51:56Z,direct_url=,disk_format='vmdk',id=8a120570-cb06-4099-b262-554ca0ad15c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b569255cc43e42e7ae3f2b4ad37c6ef4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T02:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 557.192561] env[63515]: DEBUG nova.virt.hardware [None req-a5c600bc-b1c1-4a6d-9386-2b0dbd0395e4 tempest-InstanceActionsNegativeTestJSON-1061405461 tempest-InstanceActionsNegativeTestJSON-1061405461-project-member] Flavor limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 557.192561] env[63515]: DEBUG nova.virt.hardware [None req-a5c600bc-b1c1-4a6d-9386-2b0dbd0395e4 tempest-InstanceActionsNegativeTestJSON-1061405461 tempest-InstanceActionsNegativeTestJSON-1061405461-project-member] Image limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 557.192561] env[63515]: DEBUG nova.virt.hardware [None req-a5c600bc-b1c1-4a6d-9386-2b0dbd0395e4 tempest-InstanceActionsNegativeTestJSON-1061405461 tempest-InstanceActionsNegativeTestJSON-1061405461-project-member] Flavor pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 557.192561] env[63515]: DEBUG nova.virt.hardware [None req-a5c600bc-b1c1-4a6d-9386-2b0dbd0395e4 tempest-InstanceActionsNegativeTestJSON-1061405461 tempest-InstanceActionsNegativeTestJSON-1061405461-project-member] Image pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 557.192561] env[63515]: DEBUG nova.virt.hardware [None req-a5c600bc-b1c1-4a6d-9386-2b0dbd0395e4 tempest-InstanceActionsNegativeTestJSON-1061405461 tempest-InstanceActionsNegativeTestJSON-1061405461-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 557.192786] env[63515]: DEBUG nova.virt.hardware [None req-a5c600bc-b1c1-4a6d-9386-2b0dbd0395e4 tempest-InstanceActionsNegativeTestJSON-1061405461 tempest-InstanceActionsNegativeTestJSON-1061405461-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 557.192871] env[63515]: DEBUG nova.virt.hardware [None req-a5c600bc-b1c1-4a6d-9386-2b0dbd0395e4 tempest-InstanceActionsNegativeTestJSON-1061405461 tempest-InstanceActionsNegativeTestJSON-1061405461-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 557.194402] env[63515]: DEBUG nova.virt.hardware [None req-a5c600bc-b1c1-4a6d-9386-2b0dbd0395e4 tempest-InstanceActionsNegativeTestJSON-1061405461 tempest-InstanceActionsNegativeTestJSON-1061405461-project-member] Got 1 possible topologies {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 557.195527] env[63515]: DEBUG nova.virt.hardware [None req-a5c600bc-b1c1-4a6d-9386-2b0dbd0395e4 tempest-InstanceActionsNegativeTestJSON-1061405461 tempest-InstanceActionsNegativeTestJSON-1061405461-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 557.197027] env[63515]: DEBUG nova.virt.hardware [None req-a5c600bc-b1c1-4a6d-9386-2b0dbd0395e4 tempest-InstanceActionsNegativeTestJSON-1061405461 tempest-InstanceActionsNegativeTestJSON-1061405461-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 557.197938] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01c6499a-3499-4f8a-b3f2-ab95f5c1960f {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 557.207289] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-159c3e03-7c59-428f-a953-7394c584fcbb {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 557.234467] env[63515]: DEBUG nova.network.neutron [None req-757a72d0-3dae-4296-830e-18ff5e948610 tempest-ServerDiagnosticsNegativeTest-1799249066 tempest-ServerDiagnosticsNegativeTest-1799249066-project-member] [instance: 1c5d87c7-6747-45a7-8488-6f107eeaeb78] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 557.323713] env[63515]: DEBUG nova.compute.manager [None req-5e279136-7e72-43a2-85c4-9cd11158a72b tempest-ListImageFiltersTestJSON-1100101045 tempest-ListImageFiltersTestJSON-1100101045-project-member] [instance: 48908cb0-a989-4bad-84a4-25a4bdd3baeb] Starting instance... {{(pid=63515) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 557.329199] env[63515]: DEBUG oslo_concurrency.lockutils [None req-292f2bc3-2911-4498-95fb-4e4a03681601 tempest-FloatingIPsAssociationTestJSON-629096942 tempest-FloatingIPsAssociationTestJSON-629096942-project-member] Acquiring lock "2728c9f9-0d03-4313-ba2a-3dc22aff8538" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 557.329433] env[63515]: DEBUG oslo_concurrency.lockutils [None req-292f2bc3-2911-4498-95fb-4e4a03681601 tempest-FloatingIPsAssociationTestJSON-629096942 tempest-FloatingIPsAssociationTestJSON-629096942-project-member] Lock "2728c9f9-0d03-4313-ba2a-3dc22aff8538" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 557.622665] env[63515]: DEBUG oslo_concurrency.lockutils [None req-50ea3e37-671c-4a4b-9c1a-d782ff1b578c tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.508s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 557.623346] env[63515]: DEBUG nova.compute.manager [None req-50ea3e37-671c-4a4b-9c1a-d782ff1b578c tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] [instance: cfc6c8d5-c9e8-46fb-99e7-d245664f9652] Start building networks asynchronously for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 557.627026] env[63515]: DEBUG oslo_concurrency.lockutils [None req-44598345-8888-488a-9c24-527daf4d66df tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.568s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 557.629150] env[63515]: INFO nova.compute.claims [None req-44598345-8888-488a-9c24-527daf4d66df tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] [instance: 93b28142-8454-43c0-b0a7-d61aa95c8fc8] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 557.739470] env[63515]: DEBUG oslo_concurrency.lockutils [None req-757a72d0-3dae-4296-830e-18ff5e948610 tempest-ServerDiagnosticsNegativeTest-1799249066 tempest-ServerDiagnosticsNegativeTest-1799249066-project-member] Releasing lock "refresh_cache-1c5d87c7-6747-45a7-8488-6f107eeaeb78" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 557.739981] env[63515]: DEBUG nova.compute.manager [None req-757a72d0-3dae-4296-830e-18ff5e948610 tempest-ServerDiagnosticsNegativeTest-1799249066 tempest-ServerDiagnosticsNegativeTest-1799249066-project-member] [instance: 1c5d87c7-6747-45a7-8488-6f107eeaeb78] Start destroying the instance on the hypervisor. {{(pid=63515) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 557.740197] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-757a72d0-3dae-4296-830e-18ff5e948610 tempest-ServerDiagnosticsNegativeTest-1799249066 tempest-ServerDiagnosticsNegativeTest-1799249066-project-member] [instance: 1c5d87c7-6747-45a7-8488-6f107eeaeb78] Destroying instance {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 557.740411] env[63515]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-7d48502c-8d86-4c1d-96fa-36c4a3cc9405 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 557.752671] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3d2633b-f3b5-46e8-b9b6-2e400e363842 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 557.775669] env[63515]: WARNING nova.virt.vmwareapi.vmops [None req-757a72d0-3dae-4296-830e-18ff5e948610 tempest-ServerDiagnosticsNegativeTest-1799249066 tempest-ServerDiagnosticsNegativeTest-1799249066-project-member] [instance: 1c5d87c7-6747-45a7-8488-6f107eeaeb78] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 1c5d87c7-6747-45a7-8488-6f107eeaeb78 could not be found. [ 557.775906] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-757a72d0-3dae-4296-830e-18ff5e948610 tempest-ServerDiagnosticsNegativeTest-1799249066 tempest-ServerDiagnosticsNegativeTest-1799249066-project-member] [instance: 1c5d87c7-6747-45a7-8488-6f107eeaeb78] Instance destroyed {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 557.776101] env[63515]: INFO nova.compute.manager [None req-757a72d0-3dae-4296-830e-18ff5e948610 tempest-ServerDiagnosticsNegativeTest-1799249066 tempest-ServerDiagnosticsNegativeTest-1799249066-project-member] [instance: 1c5d87c7-6747-45a7-8488-6f107eeaeb78] Took 0.04 seconds to destroy the instance on the hypervisor. [ 557.776363] env[63515]: DEBUG oslo.service.loopingcall [None req-757a72d0-3dae-4296-830e-18ff5e948610 tempest-ServerDiagnosticsNegativeTest-1799249066 tempest-ServerDiagnosticsNegativeTest-1799249066-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 557.776564] env[63515]: DEBUG nova.compute.manager [-] [instance: 1c5d87c7-6747-45a7-8488-6f107eeaeb78] Deallocating network for instance {{(pid=63515) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 557.776655] env[63515]: DEBUG nova.network.neutron [-] [instance: 1c5d87c7-6747-45a7-8488-6f107eeaeb78] deallocate_for_instance() {{(pid=63515) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 557.802088] env[63515]: DEBUG nova.network.neutron [-] [instance: 1c5d87c7-6747-45a7-8488-6f107eeaeb78] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 557.852686] env[63515]: DEBUG oslo_concurrency.lockutils [None req-5e279136-7e72-43a2-85c4-9cd11158a72b tempest-ListImageFiltersTestJSON-1100101045 tempest-ListImageFiltersTestJSON-1100101045-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 558.135143] env[63515]: DEBUG nova.compute.utils [None req-50ea3e37-671c-4a4b-9c1a-d782ff1b578c tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Using /dev/sd instead of None {{(pid=63515) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 558.141355] env[63515]: DEBUG nova.compute.manager [None req-50ea3e37-671c-4a4b-9c1a-d782ff1b578c tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] [instance: cfc6c8d5-c9e8-46fb-99e7-d245664f9652] Allocating IP information in the background. {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 558.141355] env[63515]: DEBUG nova.network.neutron [None req-50ea3e37-671c-4a4b-9c1a-d782ff1b578c tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] [instance: cfc6c8d5-c9e8-46fb-99e7-d245664f9652] allocate_for_instance() {{(pid=63515) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 558.255794] env[63515]: DEBUG nova.policy [None req-50ea3e37-671c-4a4b-9c1a-d782ff1b578c tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '59a67d99419c47e9b4ff92c6d647656f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd3d0957a488b47559794495a451f5437', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63515) authorize /opt/stack/nova/nova/policy.py:201}} [ 558.305450] env[63515]: DEBUG nova.network.neutron [-] [instance: 1c5d87c7-6747-45a7-8488-6f107eeaeb78] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 558.341404] env[63515]: ERROR nova.compute.manager [None req-a5c600bc-b1c1-4a6d-9386-2b0dbd0395e4 tempest-InstanceActionsNegativeTestJSON-1061405461 tempest-InstanceActionsNegativeTestJSON-1061405461-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port cab05064-e039-47ee-9e97-1d390704eedd, please check neutron logs for more information. [ 558.341404] env[63515]: ERROR nova.compute.manager Traceback (most recent call last): [ 558.341404] env[63515]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 558.341404] env[63515]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 558.341404] env[63515]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 558.341404] env[63515]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 558.341404] env[63515]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 558.341404] env[63515]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 558.341404] env[63515]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 558.341404] env[63515]: ERROR nova.compute.manager self.force_reraise() [ 558.341404] env[63515]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 558.341404] env[63515]: ERROR nova.compute.manager raise self.value [ 558.341404] env[63515]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 558.341404] env[63515]: ERROR nova.compute.manager updated_port = self._update_port( [ 558.341404] env[63515]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 558.341404] env[63515]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 558.345568] env[63515]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 558.345568] env[63515]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 558.345568] env[63515]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port cab05064-e039-47ee-9e97-1d390704eedd, please check neutron logs for more information. [ 558.345568] env[63515]: ERROR nova.compute.manager [ 558.345568] env[63515]: Traceback (most recent call last): [ 558.345568] env[63515]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 558.345568] env[63515]: listener.cb(fileno) [ 558.345568] env[63515]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 558.345568] env[63515]: result = function(*args, **kwargs) [ 558.345568] env[63515]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 558.345568] env[63515]: return func(*args, **kwargs) [ 558.345568] env[63515]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 558.345568] env[63515]: raise e [ 558.345568] env[63515]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 558.345568] env[63515]: nwinfo = self.network_api.allocate_for_instance( [ 558.345568] env[63515]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 558.345568] env[63515]: created_port_ids = self._update_ports_for_instance( [ 558.345568] env[63515]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 558.345568] env[63515]: with excutils.save_and_reraise_exception(): [ 558.345568] env[63515]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 558.345568] env[63515]: self.force_reraise() [ 558.345568] env[63515]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 558.345568] env[63515]: raise self.value [ 558.345568] env[63515]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 558.345568] env[63515]: updated_port = self._update_port( [ 558.345568] env[63515]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 558.345568] env[63515]: _ensure_no_port_binding_failure(port) [ 558.345568] env[63515]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 558.345568] env[63515]: raise exception.PortBindingFailed(port_id=port['id']) [ 558.346498] env[63515]: nova.exception.PortBindingFailed: Binding failed for port cab05064-e039-47ee-9e97-1d390704eedd, please check neutron logs for more information. [ 558.346498] env[63515]: Removing descriptor: 17 [ 558.346498] env[63515]: ERROR nova.compute.manager [None req-a5c600bc-b1c1-4a6d-9386-2b0dbd0395e4 tempest-InstanceActionsNegativeTestJSON-1061405461 tempest-InstanceActionsNegativeTestJSON-1061405461-project-member] [instance: b641d2d1-de70-48b6-9137-fe18c3bc7511] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port cab05064-e039-47ee-9e97-1d390704eedd, please check neutron logs for more information. [ 558.346498] env[63515]: ERROR nova.compute.manager [instance: b641d2d1-de70-48b6-9137-fe18c3bc7511] Traceback (most recent call last): [ 558.346498] env[63515]: ERROR nova.compute.manager [instance: b641d2d1-de70-48b6-9137-fe18c3bc7511] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 558.346498] env[63515]: ERROR nova.compute.manager [instance: b641d2d1-de70-48b6-9137-fe18c3bc7511] yield resources [ 558.346498] env[63515]: ERROR nova.compute.manager [instance: b641d2d1-de70-48b6-9137-fe18c3bc7511] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 558.346498] env[63515]: ERROR nova.compute.manager [instance: b641d2d1-de70-48b6-9137-fe18c3bc7511] self.driver.spawn(context, instance, image_meta, [ 558.346498] env[63515]: ERROR nova.compute.manager [instance: b641d2d1-de70-48b6-9137-fe18c3bc7511] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 558.346498] env[63515]: ERROR nova.compute.manager [instance: b641d2d1-de70-48b6-9137-fe18c3bc7511] self._vmops.spawn(context, instance, image_meta, injected_files, [ 558.346498] env[63515]: ERROR nova.compute.manager [instance: b641d2d1-de70-48b6-9137-fe18c3bc7511] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 558.346498] env[63515]: ERROR nova.compute.manager [instance: b641d2d1-de70-48b6-9137-fe18c3bc7511] vm_ref = self.build_virtual_machine(instance, [ 558.346963] env[63515]: ERROR nova.compute.manager [instance: b641d2d1-de70-48b6-9137-fe18c3bc7511] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 558.346963] env[63515]: ERROR nova.compute.manager [instance: b641d2d1-de70-48b6-9137-fe18c3bc7511] vif_infos = vmwarevif.get_vif_info(self._session, [ 558.346963] env[63515]: ERROR nova.compute.manager [instance: b641d2d1-de70-48b6-9137-fe18c3bc7511] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 558.346963] env[63515]: ERROR nova.compute.manager [instance: b641d2d1-de70-48b6-9137-fe18c3bc7511] for vif in network_info: [ 558.346963] env[63515]: ERROR nova.compute.manager [instance: b641d2d1-de70-48b6-9137-fe18c3bc7511] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 558.346963] env[63515]: ERROR nova.compute.manager [instance: b641d2d1-de70-48b6-9137-fe18c3bc7511] return self._sync_wrapper(fn, *args, **kwargs) [ 558.346963] env[63515]: ERROR nova.compute.manager [instance: b641d2d1-de70-48b6-9137-fe18c3bc7511] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 558.346963] env[63515]: ERROR nova.compute.manager [instance: b641d2d1-de70-48b6-9137-fe18c3bc7511] self.wait() [ 558.346963] env[63515]: ERROR nova.compute.manager [instance: b641d2d1-de70-48b6-9137-fe18c3bc7511] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 558.346963] env[63515]: ERROR nova.compute.manager [instance: b641d2d1-de70-48b6-9137-fe18c3bc7511] self[:] = self._gt.wait() [ 558.346963] env[63515]: ERROR nova.compute.manager [instance: b641d2d1-de70-48b6-9137-fe18c3bc7511] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 558.346963] env[63515]: ERROR nova.compute.manager [instance: b641d2d1-de70-48b6-9137-fe18c3bc7511] return self._exit_event.wait() [ 558.346963] env[63515]: ERROR nova.compute.manager [instance: b641d2d1-de70-48b6-9137-fe18c3bc7511] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 558.347359] env[63515]: ERROR nova.compute.manager [instance: b641d2d1-de70-48b6-9137-fe18c3bc7511] result = hub.switch() [ 558.347359] env[63515]: ERROR nova.compute.manager [instance: b641d2d1-de70-48b6-9137-fe18c3bc7511] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 558.347359] env[63515]: ERROR nova.compute.manager [instance: b641d2d1-de70-48b6-9137-fe18c3bc7511] return self.greenlet.switch() [ 558.347359] env[63515]: ERROR nova.compute.manager [instance: b641d2d1-de70-48b6-9137-fe18c3bc7511] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 558.347359] env[63515]: ERROR nova.compute.manager [instance: b641d2d1-de70-48b6-9137-fe18c3bc7511] result = function(*args, **kwargs) [ 558.347359] env[63515]: ERROR nova.compute.manager [instance: b641d2d1-de70-48b6-9137-fe18c3bc7511] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 558.347359] env[63515]: ERROR nova.compute.manager [instance: b641d2d1-de70-48b6-9137-fe18c3bc7511] return func(*args, **kwargs) [ 558.347359] env[63515]: ERROR nova.compute.manager [instance: b641d2d1-de70-48b6-9137-fe18c3bc7511] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 558.347359] env[63515]: ERROR nova.compute.manager [instance: b641d2d1-de70-48b6-9137-fe18c3bc7511] raise e [ 558.347359] env[63515]: ERROR nova.compute.manager [instance: b641d2d1-de70-48b6-9137-fe18c3bc7511] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 558.347359] env[63515]: ERROR nova.compute.manager [instance: b641d2d1-de70-48b6-9137-fe18c3bc7511] nwinfo = self.network_api.allocate_for_instance( [ 558.347359] env[63515]: ERROR nova.compute.manager [instance: b641d2d1-de70-48b6-9137-fe18c3bc7511] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 558.347359] env[63515]: ERROR nova.compute.manager [instance: b641d2d1-de70-48b6-9137-fe18c3bc7511] created_port_ids = self._update_ports_for_instance( [ 558.347713] env[63515]: ERROR nova.compute.manager [instance: b641d2d1-de70-48b6-9137-fe18c3bc7511] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 558.347713] env[63515]: ERROR nova.compute.manager [instance: b641d2d1-de70-48b6-9137-fe18c3bc7511] with excutils.save_and_reraise_exception(): [ 558.347713] env[63515]: ERROR nova.compute.manager [instance: b641d2d1-de70-48b6-9137-fe18c3bc7511] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 558.347713] env[63515]: ERROR nova.compute.manager [instance: b641d2d1-de70-48b6-9137-fe18c3bc7511] self.force_reraise() [ 558.347713] env[63515]: ERROR nova.compute.manager [instance: b641d2d1-de70-48b6-9137-fe18c3bc7511] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 558.347713] env[63515]: ERROR nova.compute.manager [instance: b641d2d1-de70-48b6-9137-fe18c3bc7511] raise self.value [ 558.347713] env[63515]: ERROR nova.compute.manager [instance: b641d2d1-de70-48b6-9137-fe18c3bc7511] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 558.347713] env[63515]: ERROR nova.compute.manager [instance: b641d2d1-de70-48b6-9137-fe18c3bc7511] updated_port = self._update_port( [ 558.347713] env[63515]: ERROR nova.compute.manager [instance: b641d2d1-de70-48b6-9137-fe18c3bc7511] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 558.347713] env[63515]: ERROR nova.compute.manager [instance: b641d2d1-de70-48b6-9137-fe18c3bc7511] _ensure_no_port_binding_failure(port) [ 558.347713] env[63515]: ERROR nova.compute.manager [instance: b641d2d1-de70-48b6-9137-fe18c3bc7511] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 558.347713] env[63515]: ERROR nova.compute.manager [instance: b641d2d1-de70-48b6-9137-fe18c3bc7511] raise exception.PortBindingFailed(port_id=port['id']) [ 558.348038] env[63515]: ERROR nova.compute.manager [instance: b641d2d1-de70-48b6-9137-fe18c3bc7511] nova.exception.PortBindingFailed: Binding failed for port cab05064-e039-47ee-9e97-1d390704eedd, please check neutron logs for more information. [ 558.348038] env[63515]: ERROR nova.compute.manager [instance: b641d2d1-de70-48b6-9137-fe18c3bc7511] [ 558.348038] env[63515]: INFO nova.compute.manager [None req-a5c600bc-b1c1-4a6d-9386-2b0dbd0395e4 tempest-InstanceActionsNegativeTestJSON-1061405461 tempest-InstanceActionsNegativeTestJSON-1061405461-project-member] [instance: b641d2d1-de70-48b6-9137-fe18c3bc7511] Terminating instance [ 558.348038] env[63515]: DEBUG oslo_concurrency.lockutils [None req-a5c600bc-b1c1-4a6d-9386-2b0dbd0395e4 tempest-InstanceActionsNegativeTestJSON-1061405461 tempest-InstanceActionsNegativeTestJSON-1061405461-project-member] Acquiring lock "refresh_cache-b641d2d1-de70-48b6-9137-fe18c3bc7511" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 558.348038] env[63515]: DEBUG oslo_concurrency.lockutils [None req-a5c600bc-b1c1-4a6d-9386-2b0dbd0395e4 tempest-InstanceActionsNegativeTestJSON-1061405461 tempest-InstanceActionsNegativeTestJSON-1061405461-project-member] Acquired lock "refresh_cache-b641d2d1-de70-48b6-9137-fe18c3bc7511" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 558.348038] env[63515]: DEBUG nova.network.neutron [None req-a5c600bc-b1c1-4a6d-9386-2b0dbd0395e4 tempest-InstanceActionsNegativeTestJSON-1061405461 tempest-InstanceActionsNegativeTestJSON-1061405461-project-member] [instance: b641d2d1-de70-48b6-9137-fe18c3bc7511] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 558.642521] env[63515]: DEBUG nova.compute.manager [None req-50ea3e37-671c-4a4b-9c1a-d782ff1b578c tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] [instance: cfc6c8d5-c9e8-46fb-99e7-d245664f9652] Start building block device mappings for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 558.807940] env[63515]: INFO nova.compute.manager [-] [instance: 1c5d87c7-6747-45a7-8488-6f107eeaeb78] Took 1.03 seconds to deallocate network for instance. [ 558.812579] env[63515]: DEBUG nova.compute.claims [None req-757a72d0-3dae-4296-830e-18ff5e948610 tempest-ServerDiagnosticsNegativeTest-1799249066 tempest-ServerDiagnosticsNegativeTest-1799249066-project-member] [instance: 1c5d87c7-6747-45a7-8488-6f107eeaeb78] Aborting claim: {{(pid=63515) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 558.812758] env[63515]: DEBUG oslo_concurrency.lockutils [None req-757a72d0-3dae-4296-830e-18ff5e948610 tempest-ServerDiagnosticsNegativeTest-1799249066 tempest-ServerDiagnosticsNegativeTest-1799249066-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 558.878097] env[63515]: DEBUG nova.network.neutron [None req-a5c600bc-b1c1-4a6d-9386-2b0dbd0395e4 tempest-InstanceActionsNegativeTestJSON-1061405461 tempest-InstanceActionsNegativeTestJSON-1061405461-project-member] [instance: b641d2d1-de70-48b6-9137-fe18c3bc7511] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 558.881798] env[63515]: DEBUG nova.network.neutron [None req-50ea3e37-671c-4a4b-9c1a-d782ff1b578c tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] [instance: cfc6c8d5-c9e8-46fb-99e7-d245664f9652] Successfully created port: 589525f2-8185-44ca-9cdc-7c1b03dd8ee3 {{(pid=63515) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 559.004306] env[63515]: DEBUG oslo_concurrency.lockutils [None req-4334f93e-6967-4c78-b6ec-145141dc41da tempest-AttachInterfacesV270Test-352099301 tempest-AttachInterfacesV270Test-352099301-project-member] Acquiring lock "6876f61c-2315-4e49-9944-4a5caddede5b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 559.004602] env[63515]: DEBUG oslo_concurrency.lockutils [None req-4334f93e-6967-4c78-b6ec-145141dc41da tempest-AttachInterfacesV270Test-352099301 tempest-AttachInterfacesV270Test-352099301-project-member] Lock "6876f61c-2315-4e49-9944-4a5caddede5b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 559.020657] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e451f3b7-a64e-41b2-a458-908697f7dff4 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 559.033562] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ee91fc0-5540-400e-9e11-ca67b83ee658 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 559.070026] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b62016ee-a22f-4d97-9596-67edeb03b90b {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 559.077984] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3364bbd-bbd4-4bd9-8595-47f3037bf0cd {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 559.094060] env[63515]: DEBUG nova.compute.provider_tree [None req-44598345-8888-488a-9c24-527daf4d66df tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 559.142524] env[63515]: DEBUG nova.network.neutron [None req-a5c600bc-b1c1-4a6d-9386-2b0dbd0395e4 tempest-InstanceActionsNegativeTestJSON-1061405461 tempest-InstanceActionsNegativeTestJSON-1061405461-project-member] [instance: b641d2d1-de70-48b6-9137-fe18c3bc7511] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 559.597094] env[63515]: DEBUG nova.scheduler.client.report [None req-44598345-8888-488a-9c24-527daf4d66df tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 172, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 559.645755] env[63515]: DEBUG oslo_concurrency.lockutils [None req-a5c600bc-b1c1-4a6d-9386-2b0dbd0395e4 tempest-InstanceActionsNegativeTestJSON-1061405461 tempest-InstanceActionsNegativeTestJSON-1061405461-project-member] Releasing lock "refresh_cache-b641d2d1-de70-48b6-9137-fe18c3bc7511" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 559.646323] env[63515]: DEBUG nova.compute.manager [None req-a5c600bc-b1c1-4a6d-9386-2b0dbd0395e4 tempest-InstanceActionsNegativeTestJSON-1061405461 tempest-InstanceActionsNegativeTestJSON-1061405461-project-member] [instance: b641d2d1-de70-48b6-9137-fe18c3bc7511] Start destroying the instance on the hypervisor. {{(pid=63515) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 559.646578] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-a5c600bc-b1c1-4a6d-9386-2b0dbd0395e4 tempest-InstanceActionsNegativeTestJSON-1061405461 tempest-InstanceActionsNegativeTestJSON-1061405461-project-member] [instance: b641d2d1-de70-48b6-9137-fe18c3bc7511] Destroying instance {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 559.646968] env[63515]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-8b0784e1-9263-45ac-a658-9f1c01010497 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 559.658375] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0243d02f-e922-4126-808b-a3f4c7feff56 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 559.674290] env[63515]: DEBUG nova.compute.manager [None req-50ea3e37-671c-4a4b-9c1a-d782ff1b578c tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] [instance: cfc6c8d5-c9e8-46fb-99e7-d245664f9652] Start spawning the instance on the hypervisor. {{(pid=63515) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 559.690399] env[63515]: WARNING nova.virt.vmwareapi.vmops [None req-a5c600bc-b1c1-4a6d-9386-2b0dbd0395e4 tempest-InstanceActionsNegativeTestJSON-1061405461 tempest-InstanceActionsNegativeTestJSON-1061405461-project-member] [instance: b641d2d1-de70-48b6-9137-fe18c3bc7511] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance b641d2d1-de70-48b6-9137-fe18c3bc7511 could not be found. [ 559.690942] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-a5c600bc-b1c1-4a6d-9386-2b0dbd0395e4 tempest-InstanceActionsNegativeTestJSON-1061405461 tempest-InstanceActionsNegativeTestJSON-1061405461-project-member] [instance: b641d2d1-de70-48b6-9137-fe18c3bc7511] Instance destroyed {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 559.691217] env[63515]: INFO nova.compute.manager [None req-a5c600bc-b1c1-4a6d-9386-2b0dbd0395e4 tempest-InstanceActionsNegativeTestJSON-1061405461 tempest-InstanceActionsNegativeTestJSON-1061405461-project-member] [instance: b641d2d1-de70-48b6-9137-fe18c3bc7511] Took 0.04 seconds to destroy the instance on the hypervisor. [ 559.691553] env[63515]: DEBUG oslo.service.loopingcall [None req-a5c600bc-b1c1-4a6d-9386-2b0dbd0395e4 tempest-InstanceActionsNegativeTestJSON-1061405461 tempest-InstanceActionsNegativeTestJSON-1061405461-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 559.691833] env[63515]: DEBUG nova.compute.manager [-] [instance: b641d2d1-de70-48b6-9137-fe18c3bc7511] Deallocating network for instance {{(pid=63515) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 559.692135] env[63515]: DEBUG nova.network.neutron [-] [instance: b641d2d1-de70-48b6-9137-fe18c3bc7511] deallocate_for_instance() {{(pid=63515) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 559.705652] env[63515]: DEBUG nova.virt.hardware [None req-50ea3e37-671c-4a4b-9c1a-d782ff1b578c tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T02:52:14Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T02:51:56Z,direct_url=,disk_format='vmdk',id=8a120570-cb06-4099-b262-554ca0ad15c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b569255cc43e42e7ae3f2b4ad37c6ef4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T02:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 559.705880] env[63515]: DEBUG nova.virt.hardware [None req-50ea3e37-671c-4a4b-9c1a-d782ff1b578c tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Flavor limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 559.706116] env[63515]: DEBUG nova.virt.hardware [None req-50ea3e37-671c-4a4b-9c1a-d782ff1b578c tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Image limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 559.706228] env[63515]: DEBUG nova.virt.hardware [None req-50ea3e37-671c-4a4b-9c1a-d782ff1b578c tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Flavor pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 559.706371] env[63515]: DEBUG nova.virt.hardware [None req-50ea3e37-671c-4a4b-9c1a-d782ff1b578c tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Image pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 559.706512] env[63515]: DEBUG nova.virt.hardware [None req-50ea3e37-671c-4a4b-9c1a-d782ff1b578c tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 559.706712] env[63515]: DEBUG nova.virt.hardware [None req-50ea3e37-671c-4a4b-9c1a-d782ff1b578c tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 559.707683] env[63515]: DEBUG nova.virt.hardware [None req-50ea3e37-671c-4a4b-9c1a-d782ff1b578c tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 559.707683] env[63515]: DEBUG nova.virt.hardware [None req-50ea3e37-671c-4a4b-9c1a-d782ff1b578c tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Got 1 possible topologies {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 559.707683] env[63515]: DEBUG nova.virt.hardware [None req-50ea3e37-671c-4a4b-9c1a-d782ff1b578c tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 559.707820] env[63515]: DEBUG nova.virt.hardware [None req-50ea3e37-671c-4a4b-9c1a-d782ff1b578c tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 559.709172] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e01df45-8d51-4cbb-805b-b16f07ec23cd {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 559.718152] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd6ffce3-6a65-4f46-bebc-dbe29f533bce {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 559.733521] env[63515]: DEBUG nova.network.neutron [-] [instance: b641d2d1-de70-48b6-9137-fe18c3bc7511] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 560.105127] env[63515]: DEBUG oslo_concurrency.lockutils [None req-44598345-8888-488a-9c24-527daf4d66df tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.478s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 560.105663] env[63515]: DEBUG nova.compute.manager [None req-44598345-8888-488a-9c24-527daf4d66df tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] [instance: 93b28142-8454-43c0-b0a7-d61aa95c8fc8] Start building networks asynchronously for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 560.108244] env[63515]: DEBUG oslo_concurrency.lockutils [None req-c96f19aa-9abc-4b12-99f8-980442cab640 tempest-ServersAdminTestJSON-710082096 tempest-ServersAdminTestJSON-710082096-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 17.798s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 560.235974] env[63515]: DEBUG nova.network.neutron [-] [instance: b641d2d1-de70-48b6-9137-fe18c3bc7511] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 560.612385] env[63515]: DEBUG nova.compute.utils [None req-44598345-8888-488a-9c24-527daf4d66df tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] Using /dev/sd instead of None {{(pid=63515) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 560.618895] env[63515]: DEBUG nova.compute.manager [None req-44598345-8888-488a-9c24-527daf4d66df tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] [instance: 93b28142-8454-43c0-b0a7-d61aa95c8fc8] Not allocating networking since 'none' was specified. {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 560.738263] env[63515]: INFO nova.compute.manager [-] [instance: b641d2d1-de70-48b6-9137-fe18c3bc7511] Took 1.05 seconds to deallocate network for instance. [ 560.740744] env[63515]: DEBUG nova.compute.claims [None req-a5c600bc-b1c1-4a6d-9386-2b0dbd0395e4 tempest-InstanceActionsNegativeTestJSON-1061405461 tempest-InstanceActionsNegativeTestJSON-1061405461-project-member] [instance: b641d2d1-de70-48b6-9137-fe18c3bc7511] Aborting claim: {{(pid=63515) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 560.740930] env[63515]: DEBUG oslo_concurrency.lockutils [None req-a5c600bc-b1c1-4a6d-9386-2b0dbd0395e4 tempest-InstanceActionsNegativeTestJSON-1061405461 tempest-InstanceActionsNegativeTestJSON-1061405461-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 561.006734] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3c82bbd-42a9-4774-9b47-c3f23a8bb9f9 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 561.025556] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5abde000-23b3-4b44-afd1-997f4bd07d1f {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 561.059102] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-342c0900-31f7-4adc-aab0-3d50f68b1ca7 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 561.063100] env[63515]: ERROR nova.compute.manager [None req-50ea3e37-671c-4a4b-9c1a-d782ff1b578c tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 589525f2-8185-44ca-9cdc-7c1b03dd8ee3, please check neutron logs for more information. [ 561.063100] env[63515]: ERROR nova.compute.manager Traceback (most recent call last): [ 561.063100] env[63515]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 561.063100] env[63515]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 561.063100] env[63515]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 561.063100] env[63515]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 561.063100] env[63515]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 561.063100] env[63515]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 561.063100] env[63515]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 561.063100] env[63515]: ERROR nova.compute.manager self.force_reraise() [ 561.063100] env[63515]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 561.063100] env[63515]: ERROR nova.compute.manager raise self.value [ 561.063100] env[63515]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 561.063100] env[63515]: ERROR nova.compute.manager updated_port = self._update_port( [ 561.063100] env[63515]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 561.063100] env[63515]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 561.063604] env[63515]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 561.063604] env[63515]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 561.063604] env[63515]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 589525f2-8185-44ca-9cdc-7c1b03dd8ee3, please check neutron logs for more information. [ 561.063604] env[63515]: ERROR nova.compute.manager [ 561.063604] env[63515]: Traceback (most recent call last): [ 561.063604] env[63515]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 561.063604] env[63515]: listener.cb(fileno) [ 561.063604] env[63515]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 561.063604] env[63515]: result = function(*args, **kwargs) [ 561.063604] env[63515]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 561.063604] env[63515]: return func(*args, **kwargs) [ 561.063604] env[63515]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 561.063604] env[63515]: raise e [ 561.063604] env[63515]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 561.063604] env[63515]: nwinfo = self.network_api.allocate_for_instance( [ 561.063604] env[63515]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 561.063604] env[63515]: created_port_ids = self._update_ports_for_instance( [ 561.063604] env[63515]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 561.063604] env[63515]: with excutils.save_and_reraise_exception(): [ 561.063604] env[63515]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 561.063604] env[63515]: self.force_reraise() [ 561.063604] env[63515]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 561.063604] env[63515]: raise self.value [ 561.063604] env[63515]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 561.063604] env[63515]: updated_port = self._update_port( [ 561.063604] env[63515]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 561.063604] env[63515]: _ensure_no_port_binding_failure(port) [ 561.063604] env[63515]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 561.063604] env[63515]: raise exception.PortBindingFailed(port_id=port['id']) [ 561.064909] env[63515]: nova.exception.PortBindingFailed: Binding failed for port 589525f2-8185-44ca-9cdc-7c1b03dd8ee3, please check neutron logs for more information. [ 561.064909] env[63515]: Removing descriptor: 16 [ 561.064909] env[63515]: ERROR nova.compute.manager [None req-50ea3e37-671c-4a4b-9c1a-d782ff1b578c tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] [instance: cfc6c8d5-c9e8-46fb-99e7-d245664f9652] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 589525f2-8185-44ca-9cdc-7c1b03dd8ee3, please check neutron logs for more information. [ 561.064909] env[63515]: ERROR nova.compute.manager [instance: cfc6c8d5-c9e8-46fb-99e7-d245664f9652] Traceback (most recent call last): [ 561.064909] env[63515]: ERROR nova.compute.manager [instance: cfc6c8d5-c9e8-46fb-99e7-d245664f9652] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 561.064909] env[63515]: ERROR nova.compute.manager [instance: cfc6c8d5-c9e8-46fb-99e7-d245664f9652] yield resources [ 561.064909] env[63515]: ERROR nova.compute.manager [instance: cfc6c8d5-c9e8-46fb-99e7-d245664f9652] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 561.064909] env[63515]: ERROR nova.compute.manager [instance: cfc6c8d5-c9e8-46fb-99e7-d245664f9652] self.driver.spawn(context, instance, image_meta, [ 561.064909] env[63515]: ERROR nova.compute.manager [instance: cfc6c8d5-c9e8-46fb-99e7-d245664f9652] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 561.064909] env[63515]: ERROR nova.compute.manager [instance: cfc6c8d5-c9e8-46fb-99e7-d245664f9652] self._vmops.spawn(context, instance, image_meta, injected_files, [ 561.064909] env[63515]: ERROR nova.compute.manager [instance: cfc6c8d5-c9e8-46fb-99e7-d245664f9652] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 561.064909] env[63515]: ERROR nova.compute.manager [instance: cfc6c8d5-c9e8-46fb-99e7-d245664f9652] vm_ref = self.build_virtual_machine(instance, [ 561.065305] env[63515]: ERROR nova.compute.manager [instance: cfc6c8d5-c9e8-46fb-99e7-d245664f9652] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 561.065305] env[63515]: ERROR nova.compute.manager [instance: cfc6c8d5-c9e8-46fb-99e7-d245664f9652] vif_infos = vmwarevif.get_vif_info(self._session, [ 561.065305] env[63515]: ERROR nova.compute.manager [instance: cfc6c8d5-c9e8-46fb-99e7-d245664f9652] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 561.065305] env[63515]: ERROR nova.compute.manager [instance: cfc6c8d5-c9e8-46fb-99e7-d245664f9652] for vif in network_info: [ 561.065305] env[63515]: ERROR nova.compute.manager [instance: cfc6c8d5-c9e8-46fb-99e7-d245664f9652] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 561.065305] env[63515]: ERROR nova.compute.manager [instance: cfc6c8d5-c9e8-46fb-99e7-d245664f9652] return self._sync_wrapper(fn, *args, **kwargs) [ 561.065305] env[63515]: ERROR nova.compute.manager [instance: cfc6c8d5-c9e8-46fb-99e7-d245664f9652] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 561.065305] env[63515]: ERROR nova.compute.manager [instance: cfc6c8d5-c9e8-46fb-99e7-d245664f9652] self.wait() [ 561.065305] env[63515]: ERROR nova.compute.manager [instance: cfc6c8d5-c9e8-46fb-99e7-d245664f9652] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 561.065305] env[63515]: ERROR nova.compute.manager [instance: cfc6c8d5-c9e8-46fb-99e7-d245664f9652] self[:] = self._gt.wait() [ 561.065305] env[63515]: ERROR nova.compute.manager [instance: cfc6c8d5-c9e8-46fb-99e7-d245664f9652] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 561.065305] env[63515]: ERROR nova.compute.manager [instance: cfc6c8d5-c9e8-46fb-99e7-d245664f9652] return self._exit_event.wait() [ 561.065305] env[63515]: ERROR nova.compute.manager [instance: cfc6c8d5-c9e8-46fb-99e7-d245664f9652] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 561.065955] env[63515]: ERROR nova.compute.manager [instance: cfc6c8d5-c9e8-46fb-99e7-d245664f9652] result = hub.switch() [ 561.065955] env[63515]: ERROR nova.compute.manager [instance: cfc6c8d5-c9e8-46fb-99e7-d245664f9652] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 561.065955] env[63515]: ERROR nova.compute.manager [instance: cfc6c8d5-c9e8-46fb-99e7-d245664f9652] return self.greenlet.switch() [ 561.065955] env[63515]: ERROR nova.compute.manager [instance: cfc6c8d5-c9e8-46fb-99e7-d245664f9652] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 561.065955] env[63515]: ERROR nova.compute.manager [instance: cfc6c8d5-c9e8-46fb-99e7-d245664f9652] result = function(*args, **kwargs) [ 561.065955] env[63515]: ERROR nova.compute.manager [instance: cfc6c8d5-c9e8-46fb-99e7-d245664f9652] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 561.065955] env[63515]: ERROR nova.compute.manager [instance: cfc6c8d5-c9e8-46fb-99e7-d245664f9652] return func(*args, **kwargs) [ 561.065955] env[63515]: ERROR nova.compute.manager [instance: cfc6c8d5-c9e8-46fb-99e7-d245664f9652] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 561.065955] env[63515]: ERROR nova.compute.manager [instance: cfc6c8d5-c9e8-46fb-99e7-d245664f9652] raise e [ 561.065955] env[63515]: ERROR nova.compute.manager [instance: cfc6c8d5-c9e8-46fb-99e7-d245664f9652] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 561.065955] env[63515]: ERROR nova.compute.manager [instance: cfc6c8d5-c9e8-46fb-99e7-d245664f9652] nwinfo = self.network_api.allocate_for_instance( [ 561.065955] env[63515]: ERROR nova.compute.manager [instance: cfc6c8d5-c9e8-46fb-99e7-d245664f9652] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 561.065955] env[63515]: ERROR nova.compute.manager [instance: cfc6c8d5-c9e8-46fb-99e7-d245664f9652] created_port_ids = self._update_ports_for_instance( [ 561.066413] env[63515]: ERROR nova.compute.manager [instance: cfc6c8d5-c9e8-46fb-99e7-d245664f9652] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 561.066413] env[63515]: ERROR nova.compute.manager [instance: cfc6c8d5-c9e8-46fb-99e7-d245664f9652] with excutils.save_and_reraise_exception(): [ 561.066413] env[63515]: ERROR nova.compute.manager [instance: cfc6c8d5-c9e8-46fb-99e7-d245664f9652] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 561.066413] env[63515]: ERROR nova.compute.manager [instance: cfc6c8d5-c9e8-46fb-99e7-d245664f9652] self.force_reraise() [ 561.066413] env[63515]: ERROR nova.compute.manager [instance: cfc6c8d5-c9e8-46fb-99e7-d245664f9652] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 561.066413] env[63515]: ERROR nova.compute.manager [instance: cfc6c8d5-c9e8-46fb-99e7-d245664f9652] raise self.value [ 561.066413] env[63515]: ERROR nova.compute.manager [instance: cfc6c8d5-c9e8-46fb-99e7-d245664f9652] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 561.066413] env[63515]: ERROR nova.compute.manager [instance: cfc6c8d5-c9e8-46fb-99e7-d245664f9652] updated_port = self._update_port( [ 561.066413] env[63515]: ERROR nova.compute.manager [instance: cfc6c8d5-c9e8-46fb-99e7-d245664f9652] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 561.066413] env[63515]: ERROR nova.compute.manager [instance: cfc6c8d5-c9e8-46fb-99e7-d245664f9652] _ensure_no_port_binding_failure(port) [ 561.066413] env[63515]: ERROR nova.compute.manager [instance: cfc6c8d5-c9e8-46fb-99e7-d245664f9652] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 561.066413] env[63515]: ERROR nova.compute.manager [instance: cfc6c8d5-c9e8-46fb-99e7-d245664f9652] raise exception.PortBindingFailed(port_id=port['id']) [ 561.066851] env[63515]: ERROR nova.compute.manager [instance: cfc6c8d5-c9e8-46fb-99e7-d245664f9652] nova.exception.PortBindingFailed: Binding failed for port 589525f2-8185-44ca-9cdc-7c1b03dd8ee3, please check neutron logs for more information. [ 561.066851] env[63515]: ERROR nova.compute.manager [instance: cfc6c8d5-c9e8-46fb-99e7-d245664f9652] [ 561.066851] env[63515]: INFO nova.compute.manager [None req-50ea3e37-671c-4a4b-9c1a-d782ff1b578c tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] [instance: cfc6c8d5-c9e8-46fb-99e7-d245664f9652] Terminating instance [ 561.066956] env[63515]: DEBUG oslo_concurrency.lockutils [None req-50ea3e37-671c-4a4b-9c1a-d782ff1b578c tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Acquiring lock "refresh_cache-cfc6c8d5-c9e8-46fb-99e7-d245664f9652" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 561.067079] env[63515]: DEBUG oslo_concurrency.lockutils [None req-50ea3e37-671c-4a4b-9c1a-d782ff1b578c tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Acquired lock "refresh_cache-cfc6c8d5-c9e8-46fb-99e7-d245664f9652" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 561.067249] env[63515]: DEBUG nova.network.neutron [None req-50ea3e37-671c-4a4b-9c1a-d782ff1b578c tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] [instance: cfc6c8d5-c9e8-46fb-99e7-d245664f9652] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 561.071702] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32ba5399-1870-47b4-b8f4-f4102732c030 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 561.086786] env[63515]: DEBUG nova.compute.provider_tree [None req-c96f19aa-9abc-4b12-99f8-980442cab640 tempest-ServersAdminTestJSON-710082096 tempest-ServersAdminTestJSON-710082096-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 561.120520] env[63515]: DEBUG nova.compute.manager [req-32712b5b-3995-4b95-85a9-6adad4e6546f req-ce423829-7859-4dc8-9a50-376b91b4c1fd service nova] [instance: 1c5d87c7-6747-45a7-8488-6f107eeaeb78] Received event network-changed-bea3bc2e-ac36-4c4f-8398-abe757f5b66f {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 561.120742] env[63515]: DEBUG nova.compute.manager [req-32712b5b-3995-4b95-85a9-6adad4e6546f req-ce423829-7859-4dc8-9a50-376b91b4c1fd service nova] [instance: 1c5d87c7-6747-45a7-8488-6f107eeaeb78] Refreshing instance network info cache due to event network-changed-bea3bc2e-ac36-4c4f-8398-abe757f5b66f. {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 561.121127] env[63515]: DEBUG oslo_concurrency.lockutils [req-32712b5b-3995-4b95-85a9-6adad4e6546f req-ce423829-7859-4dc8-9a50-376b91b4c1fd service nova] Acquiring lock "refresh_cache-1c5d87c7-6747-45a7-8488-6f107eeaeb78" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 561.121211] env[63515]: DEBUG oslo_concurrency.lockutils [req-32712b5b-3995-4b95-85a9-6adad4e6546f req-ce423829-7859-4dc8-9a50-376b91b4c1fd service nova] Acquired lock "refresh_cache-1c5d87c7-6747-45a7-8488-6f107eeaeb78" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 561.121702] env[63515]: DEBUG nova.network.neutron [req-32712b5b-3995-4b95-85a9-6adad4e6546f req-ce423829-7859-4dc8-9a50-376b91b4c1fd service nova] [instance: 1c5d87c7-6747-45a7-8488-6f107eeaeb78] Refreshing network info cache for port bea3bc2e-ac36-4c4f-8398-abe757f5b66f {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 561.123796] env[63515]: DEBUG nova.compute.manager [None req-44598345-8888-488a-9c24-527daf4d66df tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] [instance: 93b28142-8454-43c0-b0a7-d61aa95c8fc8] Start building block device mappings for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 561.592018] env[63515]: DEBUG nova.scheduler.client.report [None req-c96f19aa-9abc-4b12-99f8-980442cab640 tempest-ServersAdminTestJSON-710082096 tempest-ServersAdminTestJSON-710082096-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 172, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 561.608573] env[63515]: DEBUG nova.network.neutron [None req-50ea3e37-671c-4a4b-9c1a-d782ff1b578c tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] [instance: cfc6c8d5-c9e8-46fb-99e7-d245664f9652] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 561.677408] env[63515]: DEBUG nova.network.neutron [req-32712b5b-3995-4b95-85a9-6adad4e6546f req-ce423829-7859-4dc8-9a50-376b91b4c1fd service nova] [instance: 1c5d87c7-6747-45a7-8488-6f107eeaeb78] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 561.798150] env[63515]: DEBUG nova.network.neutron [None req-50ea3e37-671c-4a4b-9c1a-d782ff1b578c tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] [instance: cfc6c8d5-c9e8-46fb-99e7-d245664f9652] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 561.846160] env[63515]: DEBUG nova.network.neutron [req-32712b5b-3995-4b95-85a9-6adad4e6546f req-ce423829-7859-4dc8-9a50-376b91b4c1fd service nova] [instance: 1c5d87c7-6747-45a7-8488-6f107eeaeb78] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 562.096693] env[63515]: DEBUG oslo_concurrency.lockutils [None req-c96f19aa-9abc-4b12-99f8-980442cab640 tempest-ServersAdminTestJSON-710082096 tempest-ServersAdminTestJSON-710082096-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.988s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 562.097515] env[63515]: ERROR nova.compute.manager [None req-c96f19aa-9abc-4b12-99f8-980442cab640 tempest-ServersAdminTestJSON-710082096 tempest-ServersAdminTestJSON-710082096-project-member] [instance: 29d199bc-aff3-4c62-af3a-1e6880b52446] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 75dd05f7-1fd2-44b3-858d-74b76d1a0780, please check neutron logs for more information. [ 562.097515] env[63515]: ERROR nova.compute.manager [instance: 29d199bc-aff3-4c62-af3a-1e6880b52446] Traceback (most recent call last): [ 562.097515] env[63515]: ERROR nova.compute.manager [instance: 29d199bc-aff3-4c62-af3a-1e6880b52446] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 562.097515] env[63515]: ERROR nova.compute.manager [instance: 29d199bc-aff3-4c62-af3a-1e6880b52446] self.driver.spawn(context, instance, image_meta, [ 562.097515] env[63515]: ERROR nova.compute.manager [instance: 29d199bc-aff3-4c62-af3a-1e6880b52446] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 562.097515] env[63515]: ERROR nova.compute.manager [instance: 29d199bc-aff3-4c62-af3a-1e6880b52446] self._vmops.spawn(context, instance, image_meta, injected_files, [ 562.097515] env[63515]: ERROR nova.compute.manager [instance: 29d199bc-aff3-4c62-af3a-1e6880b52446] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 562.097515] env[63515]: ERROR nova.compute.manager [instance: 29d199bc-aff3-4c62-af3a-1e6880b52446] vm_ref = self.build_virtual_machine(instance, [ 562.097515] env[63515]: ERROR nova.compute.manager [instance: 29d199bc-aff3-4c62-af3a-1e6880b52446] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 562.097515] env[63515]: ERROR nova.compute.manager [instance: 29d199bc-aff3-4c62-af3a-1e6880b52446] vif_infos = vmwarevif.get_vif_info(self._session, [ 562.097515] env[63515]: ERROR nova.compute.manager [instance: 29d199bc-aff3-4c62-af3a-1e6880b52446] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 562.097840] env[63515]: ERROR nova.compute.manager [instance: 29d199bc-aff3-4c62-af3a-1e6880b52446] for vif in network_info: [ 562.097840] env[63515]: ERROR nova.compute.manager [instance: 29d199bc-aff3-4c62-af3a-1e6880b52446] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 562.097840] env[63515]: ERROR nova.compute.manager [instance: 29d199bc-aff3-4c62-af3a-1e6880b52446] return self._sync_wrapper(fn, *args, **kwargs) [ 562.097840] env[63515]: ERROR nova.compute.manager [instance: 29d199bc-aff3-4c62-af3a-1e6880b52446] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 562.097840] env[63515]: ERROR nova.compute.manager [instance: 29d199bc-aff3-4c62-af3a-1e6880b52446] self.wait() [ 562.097840] env[63515]: ERROR nova.compute.manager [instance: 29d199bc-aff3-4c62-af3a-1e6880b52446] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 562.097840] env[63515]: ERROR nova.compute.manager [instance: 29d199bc-aff3-4c62-af3a-1e6880b52446] self[:] = self._gt.wait() [ 562.097840] env[63515]: ERROR nova.compute.manager [instance: 29d199bc-aff3-4c62-af3a-1e6880b52446] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 562.097840] env[63515]: ERROR nova.compute.manager [instance: 29d199bc-aff3-4c62-af3a-1e6880b52446] return self._exit_event.wait() [ 562.097840] env[63515]: ERROR nova.compute.manager [instance: 29d199bc-aff3-4c62-af3a-1e6880b52446] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 562.097840] env[63515]: ERROR nova.compute.manager [instance: 29d199bc-aff3-4c62-af3a-1e6880b52446] result = hub.switch() [ 562.097840] env[63515]: ERROR nova.compute.manager [instance: 29d199bc-aff3-4c62-af3a-1e6880b52446] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 562.097840] env[63515]: ERROR nova.compute.manager [instance: 29d199bc-aff3-4c62-af3a-1e6880b52446] return self.greenlet.switch() [ 562.098198] env[63515]: ERROR nova.compute.manager [instance: 29d199bc-aff3-4c62-af3a-1e6880b52446] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 562.098198] env[63515]: ERROR nova.compute.manager [instance: 29d199bc-aff3-4c62-af3a-1e6880b52446] result = function(*args, **kwargs) [ 562.098198] env[63515]: ERROR nova.compute.manager [instance: 29d199bc-aff3-4c62-af3a-1e6880b52446] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 562.098198] env[63515]: ERROR nova.compute.manager [instance: 29d199bc-aff3-4c62-af3a-1e6880b52446] return func(*args, **kwargs) [ 562.098198] env[63515]: ERROR nova.compute.manager [instance: 29d199bc-aff3-4c62-af3a-1e6880b52446] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 562.098198] env[63515]: ERROR nova.compute.manager [instance: 29d199bc-aff3-4c62-af3a-1e6880b52446] raise e [ 562.098198] env[63515]: ERROR nova.compute.manager [instance: 29d199bc-aff3-4c62-af3a-1e6880b52446] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 562.098198] env[63515]: ERROR nova.compute.manager [instance: 29d199bc-aff3-4c62-af3a-1e6880b52446] nwinfo = self.network_api.allocate_for_instance( [ 562.098198] env[63515]: ERROR nova.compute.manager [instance: 29d199bc-aff3-4c62-af3a-1e6880b52446] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 562.098198] env[63515]: ERROR nova.compute.manager [instance: 29d199bc-aff3-4c62-af3a-1e6880b52446] created_port_ids = self._update_ports_for_instance( [ 562.098198] env[63515]: ERROR nova.compute.manager [instance: 29d199bc-aff3-4c62-af3a-1e6880b52446] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 562.098198] env[63515]: ERROR nova.compute.manager [instance: 29d199bc-aff3-4c62-af3a-1e6880b52446] with excutils.save_and_reraise_exception(): [ 562.098198] env[63515]: ERROR nova.compute.manager [instance: 29d199bc-aff3-4c62-af3a-1e6880b52446] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 562.098559] env[63515]: ERROR nova.compute.manager [instance: 29d199bc-aff3-4c62-af3a-1e6880b52446] self.force_reraise() [ 562.098559] env[63515]: ERROR nova.compute.manager [instance: 29d199bc-aff3-4c62-af3a-1e6880b52446] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 562.098559] env[63515]: ERROR nova.compute.manager [instance: 29d199bc-aff3-4c62-af3a-1e6880b52446] raise self.value [ 562.098559] env[63515]: ERROR nova.compute.manager [instance: 29d199bc-aff3-4c62-af3a-1e6880b52446] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 562.098559] env[63515]: ERROR nova.compute.manager [instance: 29d199bc-aff3-4c62-af3a-1e6880b52446] updated_port = self._update_port( [ 562.098559] env[63515]: ERROR nova.compute.manager [instance: 29d199bc-aff3-4c62-af3a-1e6880b52446] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 562.098559] env[63515]: ERROR nova.compute.manager [instance: 29d199bc-aff3-4c62-af3a-1e6880b52446] _ensure_no_port_binding_failure(port) [ 562.098559] env[63515]: ERROR nova.compute.manager [instance: 29d199bc-aff3-4c62-af3a-1e6880b52446] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 562.098559] env[63515]: ERROR nova.compute.manager [instance: 29d199bc-aff3-4c62-af3a-1e6880b52446] raise exception.PortBindingFailed(port_id=port['id']) [ 562.098559] env[63515]: ERROR nova.compute.manager [instance: 29d199bc-aff3-4c62-af3a-1e6880b52446] nova.exception.PortBindingFailed: Binding failed for port 75dd05f7-1fd2-44b3-858d-74b76d1a0780, please check neutron logs for more information. [ 562.098559] env[63515]: ERROR nova.compute.manager [instance: 29d199bc-aff3-4c62-af3a-1e6880b52446] [ 562.099090] env[63515]: DEBUG nova.compute.utils [None req-c96f19aa-9abc-4b12-99f8-980442cab640 tempest-ServersAdminTestJSON-710082096 tempest-ServersAdminTestJSON-710082096-project-member] [instance: 29d199bc-aff3-4c62-af3a-1e6880b52446] Binding failed for port 75dd05f7-1fd2-44b3-858d-74b76d1a0780, please check neutron logs for more information. {{(pid=63515) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 562.102262] env[63515]: DEBUG nova.compute.manager [None req-c96f19aa-9abc-4b12-99f8-980442cab640 tempest-ServersAdminTestJSON-710082096 tempest-ServersAdminTestJSON-710082096-project-member] [instance: 29d199bc-aff3-4c62-af3a-1e6880b52446] Build of instance 29d199bc-aff3-4c62-af3a-1e6880b52446 was re-scheduled: Binding failed for port 75dd05f7-1fd2-44b3-858d-74b76d1a0780, please check neutron logs for more information. {{(pid=63515) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 562.105002] env[63515]: DEBUG nova.compute.manager [None req-c96f19aa-9abc-4b12-99f8-980442cab640 tempest-ServersAdminTestJSON-710082096 tempest-ServersAdminTestJSON-710082096-project-member] [instance: 29d199bc-aff3-4c62-af3a-1e6880b52446] Unplugging VIFs for instance {{(pid=63515) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 562.105002] env[63515]: DEBUG oslo_concurrency.lockutils [None req-c96f19aa-9abc-4b12-99f8-980442cab640 tempest-ServersAdminTestJSON-710082096 tempest-ServersAdminTestJSON-710082096-project-member] Acquiring lock "refresh_cache-29d199bc-aff3-4c62-af3a-1e6880b52446" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 562.105002] env[63515]: DEBUG oslo_concurrency.lockutils [None req-c96f19aa-9abc-4b12-99f8-980442cab640 tempest-ServersAdminTestJSON-710082096 tempest-ServersAdminTestJSON-710082096-project-member] Acquired lock "refresh_cache-29d199bc-aff3-4c62-af3a-1e6880b52446" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 562.105002] env[63515]: DEBUG nova.network.neutron [None req-c96f19aa-9abc-4b12-99f8-980442cab640 tempest-ServersAdminTestJSON-710082096 tempest-ServersAdminTestJSON-710082096-project-member] [instance: 29d199bc-aff3-4c62-af3a-1e6880b52446] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 562.105002] env[63515]: DEBUG oslo_concurrency.lockutils [None req-eaa7f189-a861-45f8-872f-29d870dc784b tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.748s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 562.109571] env[63515]: INFO nova.compute.claims [None req-eaa7f189-a861-45f8-872f-29d870dc784b tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] [instance: e39bc55e-b677-4e2d-9aa4-7bc1d5ffadfa] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 562.134056] env[63515]: DEBUG nova.compute.manager [None req-44598345-8888-488a-9c24-527daf4d66df tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] [instance: 93b28142-8454-43c0-b0a7-d61aa95c8fc8] Start spawning the instance on the hypervisor. {{(pid=63515) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 562.162995] env[63515]: DEBUG nova.virt.hardware [None req-44598345-8888-488a-9c24-527daf4d66df tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T02:52:14Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T02:51:56Z,direct_url=,disk_format='vmdk',id=8a120570-cb06-4099-b262-554ca0ad15c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b569255cc43e42e7ae3f2b4ad37c6ef4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T02:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 562.164008] env[63515]: DEBUG nova.virt.hardware [None req-44598345-8888-488a-9c24-527daf4d66df tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] Flavor limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 562.164008] env[63515]: DEBUG nova.virt.hardware [None req-44598345-8888-488a-9c24-527daf4d66df tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] Image limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 562.164008] env[63515]: DEBUG nova.virt.hardware [None req-44598345-8888-488a-9c24-527daf4d66df tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] Flavor pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 562.164008] env[63515]: DEBUG nova.virt.hardware [None req-44598345-8888-488a-9c24-527daf4d66df tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] Image pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 562.164147] env[63515]: DEBUG nova.virt.hardware [None req-44598345-8888-488a-9c24-527daf4d66df tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 562.164388] env[63515]: DEBUG nova.virt.hardware [None req-44598345-8888-488a-9c24-527daf4d66df tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 562.164594] env[63515]: DEBUG nova.virt.hardware [None req-44598345-8888-488a-9c24-527daf4d66df tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 562.164815] env[63515]: DEBUG nova.virt.hardware [None req-44598345-8888-488a-9c24-527daf4d66df tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] Got 1 possible topologies {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 562.165044] env[63515]: DEBUG nova.virt.hardware [None req-44598345-8888-488a-9c24-527daf4d66df tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 562.165276] env[63515]: DEBUG nova.virt.hardware [None req-44598345-8888-488a-9c24-527daf4d66df tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 562.166498] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a678ca6-77e7-4f1c-aeb3-ad45ade0e631 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 562.177396] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf599117-87f5-4840-a707-682023760976 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 562.192782] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-44598345-8888-488a-9c24-527daf4d66df tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] [instance: 93b28142-8454-43c0-b0a7-d61aa95c8fc8] Instance VIF info [] {{(pid=63515) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 562.205043] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-44598345-8888-488a-9c24-527daf4d66df tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] Creating folder: OpenStack. Parent ref: group-v4. {{(pid=63515) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 562.205880] env[63515]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-036556f7-35c3-42f6-9db9-f14718e79f64 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 562.219077] env[63515]: INFO nova.virt.vmwareapi.vm_util [None req-44598345-8888-488a-9c24-527daf4d66df tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] Created folder: OpenStack in parent group-v4. [ 562.219267] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-44598345-8888-488a-9c24-527daf4d66df tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] Creating folder: Project (8e2e24b32ec24467a780ec57c02a4153). Parent ref: group-v243370. {{(pid=63515) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 562.219493] env[63515]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-421660bf-9f1b-4127-9f01-97b25b9f5962 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 562.230691] env[63515]: INFO nova.virt.vmwareapi.vm_util [None req-44598345-8888-488a-9c24-527daf4d66df tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] Created folder: Project (8e2e24b32ec24467a780ec57c02a4153) in parent group-v243370. [ 562.230691] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-44598345-8888-488a-9c24-527daf4d66df tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] Creating folder: Instances. Parent ref: group-v243371. {{(pid=63515) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 562.230691] env[63515]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-6170dfaa-fa0a-463b-9678-2728c1eab219 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 562.239904] env[63515]: INFO nova.virt.vmwareapi.vm_util [None req-44598345-8888-488a-9c24-527daf4d66df tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] Created folder: Instances in parent group-v243371. [ 562.239904] env[63515]: DEBUG oslo.service.loopingcall [None req-44598345-8888-488a-9c24-527daf4d66df tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 562.239904] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 93b28142-8454-43c0-b0a7-d61aa95c8fc8] Creating VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 562.239904] env[63515]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-00f1015a-5b8e-4d13-8171-7cfd22d05f5f {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 562.257315] env[63515]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 562.257315] env[63515]: value = "task-1110846" [ 562.257315] env[63515]: _type = "Task" [ 562.257315] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 562.265098] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1110846, 'name': CreateVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 562.301763] env[63515]: DEBUG oslo_concurrency.lockutils [None req-50ea3e37-671c-4a4b-9c1a-d782ff1b578c tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Releasing lock "refresh_cache-cfc6c8d5-c9e8-46fb-99e7-d245664f9652" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 562.301763] env[63515]: DEBUG nova.compute.manager [None req-50ea3e37-671c-4a4b-9c1a-d782ff1b578c tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] [instance: cfc6c8d5-c9e8-46fb-99e7-d245664f9652] Start destroying the instance on the hypervisor. {{(pid=63515) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 562.301763] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-50ea3e37-671c-4a4b-9c1a-d782ff1b578c tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] [instance: cfc6c8d5-c9e8-46fb-99e7-d245664f9652] Destroying instance {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 562.301889] env[63515]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-91961c9a-17e0-4bc8-b608-02bd48b64d81 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 562.311547] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-378656d6-9662-432a-b598-941af0d74ad2 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 562.333530] env[63515]: WARNING nova.virt.vmwareapi.vmops [None req-50ea3e37-671c-4a4b-9c1a-d782ff1b578c tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] [instance: cfc6c8d5-c9e8-46fb-99e7-d245664f9652] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance cfc6c8d5-c9e8-46fb-99e7-d245664f9652 could not be found. [ 562.333791] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-50ea3e37-671c-4a4b-9c1a-d782ff1b578c tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] [instance: cfc6c8d5-c9e8-46fb-99e7-d245664f9652] Instance destroyed {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 562.333985] env[63515]: INFO nova.compute.manager [None req-50ea3e37-671c-4a4b-9c1a-d782ff1b578c tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] [instance: cfc6c8d5-c9e8-46fb-99e7-d245664f9652] Took 0.03 seconds to destroy the instance on the hypervisor. [ 562.334274] env[63515]: DEBUG oslo.service.loopingcall [None req-50ea3e37-671c-4a4b-9c1a-d782ff1b578c tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 562.334510] env[63515]: DEBUG nova.compute.manager [-] [instance: cfc6c8d5-c9e8-46fb-99e7-d245664f9652] Deallocating network for instance {{(pid=63515) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 562.335735] env[63515]: DEBUG nova.network.neutron [-] [instance: cfc6c8d5-c9e8-46fb-99e7-d245664f9652] deallocate_for_instance() {{(pid=63515) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 562.349848] env[63515]: DEBUG oslo_concurrency.lockutils [req-32712b5b-3995-4b95-85a9-6adad4e6546f req-ce423829-7859-4dc8-9a50-376b91b4c1fd service nova] Releasing lock "refresh_cache-1c5d87c7-6747-45a7-8488-6f107eeaeb78" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 562.349848] env[63515]: DEBUG nova.compute.manager [req-32712b5b-3995-4b95-85a9-6adad4e6546f req-ce423829-7859-4dc8-9a50-376b91b4c1fd service nova] [instance: 1c5d87c7-6747-45a7-8488-6f107eeaeb78] Received event network-vif-deleted-bea3bc2e-ac36-4c4f-8398-abe757f5b66f {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 562.349848] env[63515]: DEBUG nova.compute.manager [req-32712b5b-3995-4b95-85a9-6adad4e6546f req-ce423829-7859-4dc8-9a50-376b91b4c1fd service nova] [instance: b641d2d1-de70-48b6-9137-fe18c3bc7511] Received event network-changed-cab05064-e039-47ee-9e97-1d390704eedd {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 562.350099] env[63515]: DEBUG nova.compute.manager [req-32712b5b-3995-4b95-85a9-6adad4e6546f req-ce423829-7859-4dc8-9a50-376b91b4c1fd service nova] [instance: b641d2d1-de70-48b6-9137-fe18c3bc7511] Refreshing instance network info cache due to event network-changed-cab05064-e039-47ee-9e97-1d390704eedd. {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 562.350160] env[63515]: DEBUG oslo_concurrency.lockutils [req-32712b5b-3995-4b95-85a9-6adad4e6546f req-ce423829-7859-4dc8-9a50-376b91b4c1fd service nova] Acquiring lock "refresh_cache-b641d2d1-de70-48b6-9137-fe18c3bc7511" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 562.350336] env[63515]: DEBUG oslo_concurrency.lockutils [req-32712b5b-3995-4b95-85a9-6adad4e6546f req-ce423829-7859-4dc8-9a50-376b91b4c1fd service nova] Acquired lock "refresh_cache-b641d2d1-de70-48b6-9137-fe18c3bc7511" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 562.350530] env[63515]: DEBUG nova.network.neutron [req-32712b5b-3995-4b95-85a9-6adad4e6546f req-ce423829-7859-4dc8-9a50-376b91b4c1fd service nova] [instance: b641d2d1-de70-48b6-9137-fe18c3bc7511] Refreshing network info cache for port cab05064-e039-47ee-9e97-1d390704eedd {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 562.367056] env[63515]: DEBUG nova.network.neutron [-] [instance: cfc6c8d5-c9e8-46fb-99e7-d245664f9652] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 562.647512] env[63515]: DEBUG nova.network.neutron [None req-c96f19aa-9abc-4b12-99f8-980442cab640 tempest-ServersAdminTestJSON-710082096 tempest-ServersAdminTestJSON-710082096-project-member] [instance: 29d199bc-aff3-4c62-af3a-1e6880b52446] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 562.769282] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1110846, 'name': CreateVM_Task, 'duration_secs': 0.335327} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 562.769708] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 93b28142-8454-43c0-b0a7-d61aa95c8fc8] Created VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 562.771084] env[63515]: DEBUG oslo_vmware.service [None req-44598345-8888-488a-9c24-527daf4d66df tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8cc902f1-4d69-4792-8f54-e959e83693cb {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 562.777056] env[63515]: DEBUG oslo_concurrency.lockutils [None req-44598345-8888-488a-9c24-527daf4d66df tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 562.777253] env[63515]: DEBUG oslo_concurrency.lockutils [None req-44598345-8888-488a-9c24-527daf4d66df tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 562.777887] env[63515]: DEBUG oslo_concurrency.lockutils [None req-44598345-8888-488a-9c24-527daf4d66df tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 562.778148] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-78de2e86-77a6-4e8a-96d3-10dd588c9a8c {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 562.785310] env[63515]: DEBUG oslo_vmware.api [None req-44598345-8888-488a-9c24-527daf4d66df tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] Waiting for the task: (returnval){ [ 562.785310] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]527209a9-d177-2c5c-fb97-b86a2427873c" [ 562.785310] env[63515]: _type = "Task" [ 562.785310] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 562.797540] env[63515]: DEBUG oslo_vmware.api [None req-44598345-8888-488a-9c24-527daf4d66df tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]527209a9-d177-2c5c-fb97-b86a2427873c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 562.869225] env[63515]: DEBUG nova.network.neutron [-] [instance: cfc6c8d5-c9e8-46fb-99e7-d245664f9652] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 562.902319] env[63515]: DEBUG nova.network.neutron [req-32712b5b-3995-4b95-85a9-6adad4e6546f req-ce423829-7859-4dc8-9a50-376b91b4c1fd service nova] [instance: b641d2d1-de70-48b6-9137-fe18c3bc7511] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 562.954284] env[63515]: DEBUG nova.network.neutron [None req-c96f19aa-9abc-4b12-99f8-980442cab640 tempest-ServersAdminTestJSON-710082096 tempest-ServersAdminTestJSON-710082096-project-member] [instance: 29d199bc-aff3-4c62-af3a-1e6880b52446] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 563.195536] env[63515]: DEBUG nova.network.neutron [req-32712b5b-3995-4b95-85a9-6adad4e6546f req-ce423829-7859-4dc8-9a50-376b91b4c1fd service nova] [instance: b641d2d1-de70-48b6-9137-fe18c3bc7511] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 563.299880] env[63515]: DEBUG oslo_concurrency.lockutils [None req-44598345-8888-488a-9c24-527daf4d66df tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 563.300194] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-44598345-8888-488a-9c24-527daf4d66df tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] [instance: 93b28142-8454-43c0-b0a7-d61aa95c8fc8] Processing image 8a120570-cb06-4099-b262-554ca0ad15c5 {{(pid=63515) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 563.300448] env[63515]: DEBUG oslo_concurrency.lockutils [None req-44598345-8888-488a-9c24-527daf4d66df tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 563.300590] env[63515]: DEBUG oslo_concurrency.lockutils [None req-44598345-8888-488a-9c24-527daf4d66df tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 563.301379] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-44598345-8888-488a-9c24-527daf4d66df tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 563.301856] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-944e3eb9-b596-4c9f-beaf-38b9d5d72c1f {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 563.320358] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-44598345-8888-488a-9c24-527daf4d66df tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 563.320759] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-44598345-8888-488a-9c24-527daf4d66df tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63515) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 563.321374] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e318e784-ba96-4417-9e27-ff9e530e30e3 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 563.331932] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4e6878bd-0232-4cf0-94e6-1b0f9948700f {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 563.338606] env[63515]: DEBUG oslo_vmware.api [None req-44598345-8888-488a-9c24-527daf4d66df tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] Waiting for the task: (returnval){ [ 563.338606] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]52bd6142-05c1-94b6-bdc3-bbdb08350a1c" [ 563.338606] env[63515]: _type = "Task" [ 563.338606] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 563.346588] env[63515]: DEBUG oslo_vmware.api [None req-44598345-8888-488a-9c24-527daf4d66df tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52bd6142-05c1-94b6-bdc3-bbdb08350a1c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 563.371843] env[63515]: INFO nova.compute.manager [-] [instance: cfc6c8d5-c9e8-46fb-99e7-d245664f9652] Took 1.04 seconds to deallocate network for instance. [ 563.375034] env[63515]: DEBUG nova.compute.claims [None req-50ea3e37-671c-4a4b-9c1a-d782ff1b578c tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] [instance: cfc6c8d5-c9e8-46fb-99e7-d245664f9652] Aborting claim: {{(pid=63515) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 563.375034] env[63515]: DEBUG oslo_concurrency.lockutils [None req-50ea3e37-671c-4a4b-9c1a-d782ff1b578c tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 563.461294] env[63515]: DEBUG oslo_concurrency.lockutils [None req-c96f19aa-9abc-4b12-99f8-980442cab640 tempest-ServersAdminTestJSON-710082096 tempest-ServersAdminTestJSON-710082096-project-member] Releasing lock "refresh_cache-29d199bc-aff3-4c62-af3a-1e6880b52446" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 563.461598] env[63515]: DEBUG nova.compute.manager [None req-c96f19aa-9abc-4b12-99f8-980442cab640 tempest-ServersAdminTestJSON-710082096 tempest-ServersAdminTestJSON-710082096-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63515) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 563.461932] env[63515]: DEBUG nova.compute.manager [None req-c96f19aa-9abc-4b12-99f8-980442cab640 tempest-ServersAdminTestJSON-710082096 tempest-ServersAdminTestJSON-710082096-project-member] [instance: 29d199bc-aff3-4c62-af3a-1e6880b52446] Deallocating network for instance {{(pid=63515) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 563.462182] env[63515]: DEBUG nova.network.neutron [None req-c96f19aa-9abc-4b12-99f8-980442cab640 tempest-ServersAdminTestJSON-710082096 tempest-ServersAdminTestJSON-710082096-project-member] [instance: 29d199bc-aff3-4c62-af3a-1e6880b52446] deallocate_for_instance() {{(pid=63515) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 563.507349] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc821ec4-f1a3-4f8c-9fbe-e29e90b98138 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 563.517393] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2cbacea-5727-4228-a9bf-0306defc750a {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 563.522513] env[63515]: DEBUG nova.network.neutron [None req-c96f19aa-9abc-4b12-99f8-980442cab640 tempest-ServersAdminTestJSON-710082096 tempest-ServersAdminTestJSON-710082096-project-member] [instance: 29d199bc-aff3-4c62-af3a-1e6880b52446] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 563.560197] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82104d79-2a54-46f6-aa3f-58a306163e2b {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 563.569800] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-452ffa4a-113f-404c-9332-9ec53c030b6d {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 563.585141] env[63515]: DEBUG nova.compute.provider_tree [None req-eaa7f189-a861-45f8-872f-29d870dc784b tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 563.702304] env[63515]: DEBUG oslo_concurrency.lockutils [req-32712b5b-3995-4b95-85a9-6adad4e6546f req-ce423829-7859-4dc8-9a50-376b91b4c1fd service nova] Releasing lock "refresh_cache-b641d2d1-de70-48b6-9137-fe18c3bc7511" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 563.702582] env[63515]: DEBUG nova.compute.manager [req-32712b5b-3995-4b95-85a9-6adad4e6546f req-ce423829-7859-4dc8-9a50-376b91b4c1fd service nova] [instance: b641d2d1-de70-48b6-9137-fe18c3bc7511] Received event network-vif-deleted-cab05064-e039-47ee-9e97-1d390704eedd {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 563.752350] env[63515]: DEBUG nova.compute.manager [req-91a90eea-4da8-49c7-bc6f-244113f0e0aa req-5a17caf3-a3c5-48ad-aaec-acd1e0c8fdab service nova] [instance: cfc6c8d5-c9e8-46fb-99e7-d245664f9652] Received event network-changed-589525f2-8185-44ca-9cdc-7c1b03dd8ee3 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 563.752643] env[63515]: DEBUG nova.compute.manager [req-91a90eea-4da8-49c7-bc6f-244113f0e0aa req-5a17caf3-a3c5-48ad-aaec-acd1e0c8fdab service nova] [instance: cfc6c8d5-c9e8-46fb-99e7-d245664f9652] Refreshing instance network info cache due to event network-changed-589525f2-8185-44ca-9cdc-7c1b03dd8ee3. {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 563.752930] env[63515]: DEBUG oslo_concurrency.lockutils [req-91a90eea-4da8-49c7-bc6f-244113f0e0aa req-5a17caf3-a3c5-48ad-aaec-acd1e0c8fdab service nova] Acquiring lock "refresh_cache-cfc6c8d5-c9e8-46fb-99e7-d245664f9652" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 563.753181] env[63515]: DEBUG oslo_concurrency.lockutils [req-91a90eea-4da8-49c7-bc6f-244113f0e0aa req-5a17caf3-a3c5-48ad-aaec-acd1e0c8fdab service nova] Acquired lock "refresh_cache-cfc6c8d5-c9e8-46fb-99e7-d245664f9652" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 563.753393] env[63515]: DEBUG nova.network.neutron [req-91a90eea-4da8-49c7-bc6f-244113f0e0aa req-5a17caf3-a3c5-48ad-aaec-acd1e0c8fdab service nova] [instance: cfc6c8d5-c9e8-46fb-99e7-d245664f9652] Refreshing network info cache for port 589525f2-8185-44ca-9cdc-7c1b03dd8ee3 {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 563.851480] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-44598345-8888-488a-9c24-527daf4d66df tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] [instance: 93b28142-8454-43c0-b0a7-d61aa95c8fc8] Preparing fetch location {{(pid=63515) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 563.851954] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-44598345-8888-488a-9c24-527daf4d66df tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] Creating directory with path [datastore1] vmware_temp/e6b3038c-7dd7-4196-a714-112b26e83ee0/8a120570-cb06-4099-b262-554ca0ad15c5 {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 563.852273] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f13e805d-33b6-435c-a821-c13ca1848f90 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 563.865781] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-44598345-8888-488a-9c24-527daf4d66df tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] Created directory with path [datastore1] vmware_temp/e6b3038c-7dd7-4196-a714-112b26e83ee0/8a120570-cb06-4099-b262-554ca0ad15c5 {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 563.866014] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-44598345-8888-488a-9c24-527daf4d66df tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] [instance: 93b28142-8454-43c0-b0a7-d61aa95c8fc8] Fetch image to [datastore1] vmware_temp/e6b3038c-7dd7-4196-a714-112b26e83ee0/8a120570-cb06-4099-b262-554ca0ad15c5/tmp-sparse.vmdk {{(pid=63515) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 563.866200] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-44598345-8888-488a-9c24-527daf4d66df tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] [instance: 93b28142-8454-43c0-b0a7-d61aa95c8fc8] Downloading image file data 8a120570-cb06-4099-b262-554ca0ad15c5 to [datastore1] vmware_temp/e6b3038c-7dd7-4196-a714-112b26e83ee0/8a120570-cb06-4099-b262-554ca0ad15c5/tmp-sparse.vmdk on the data store datastore1 {{(pid=63515) _fetch_image_as_file /opt/stack/nova/nova/virt/vmwareapi/vmops.py:399}} [ 563.866973] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a48a59d-30f2-48d7-9241-e64035e6395d {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 563.879895] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e4491c2-2d26-4b90-8210-e3e17ead4e18 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 563.893141] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-799398dc-caaa-412d-827a-8d332ab7ee0c {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 563.927658] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e5deee0-824d-4052-81e0-f21f8410e4d5 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 563.933987] env[63515]: DEBUG oslo_vmware.service [-] Invoking SessionManager.AcquireGenericServiceTicket with opID=oslo.vmware-8a16458a-8242-48ed-ad81-ff96186b8b59 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 564.025953] env[63515]: DEBUG nova.network.neutron [None req-c96f19aa-9abc-4b12-99f8-980442cab640 tempest-ServersAdminTestJSON-710082096 tempest-ServersAdminTestJSON-710082096-project-member] [instance: 29d199bc-aff3-4c62-af3a-1e6880b52446] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 564.030040] env[63515]: DEBUG nova.virt.vmwareapi.images [None req-44598345-8888-488a-9c24-527daf4d66df tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] [instance: 93b28142-8454-43c0-b0a7-d61aa95c8fc8] Downloading image file data 8a120570-cb06-4099-b262-554ca0ad15c5 to the data store datastore1 {{(pid=63515) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:245}} [ 564.093132] env[63515]: DEBUG nova.scheduler.client.report [None req-eaa7f189-a861-45f8-872f-29d870dc784b tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 172, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 564.102902] env[63515]: DEBUG oslo_vmware.rw_handles [None req-44598345-8888-488a-9c24-527daf4d66df tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] Creating HTTP connection to write to file with size = 21318656 and URL = https://esx7c2n1.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/e6b3038c-7dd7-4196-a714-112b26e83ee0/8a120570-cb06-4099-b262-554ca0ad15c5/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore1. {{(pid=63515) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 564.321697] env[63515]: DEBUG nova.network.neutron [req-91a90eea-4da8-49c7-bc6f-244113f0e0aa req-5a17caf3-a3c5-48ad-aaec-acd1e0c8fdab service nova] [instance: cfc6c8d5-c9e8-46fb-99e7-d245664f9652] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 564.528035] env[63515]: DEBUG nova.network.neutron [req-91a90eea-4da8-49c7-bc6f-244113f0e0aa req-5a17caf3-a3c5-48ad-aaec-acd1e0c8fdab service nova] [instance: cfc6c8d5-c9e8-46fb-99e7-d245664f9652] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 564.535026] env[63515]: INFO nova.compute.manager [None req-c96f19aa-9abc-4b12-99f8-980442cab640 tempest-ServersAdminTestJSON-710082096 tempest-ServersAdminTestJSON-710082096-project-member] [instance: 29d199bc-aff3-4c62-af3a-1e6880b52446] Took 1.07 seconds to deallocate network for instance. [ 564.598990] env[63515]: DEBUG oslo_concurrency.lockutils [None req-eaa7f189-a861-45f8-872f-29d870dc784b tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.494s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 564.599156] env[63515]: DEBUG nova.compute.manager [None req-eaa7f189-a861-45f8-872f-29d870dc784b tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] [instance: e39bc55e-b677-4e2d-9aa4-7bc1d5ffadfa] Start building networks asynchronously for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 564.605180] env[63515]: DEBUG oslo_concurrency.lockutils [None req-ada3051f-c750-48c7-b591-0344c25b8fea tempest-ServersAdminTestJSON-710082096 tempest-ServersAdminTestJSON-710082096-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 13.562s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 564.780597] env[63515]: DEBUG oslo_vmware.rw_handles [None req-44598345-8888-488a-9c24-527daf4d66df tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] Completed reading data from the image iterator. {{(pid=63515) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 564.781022] env[63515]: DEBUG oslo_vmware.rw_handles [None req-44598345-8888-488a-9c24-527daf4d66df tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] Closing write handle for https://esx7c2n1.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/e6b3038c-7dd7-4196-a714-112b26e83ee0/8a120570-cb06-4099-b262-554ca0ad15c5/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore1. {{(pid=63515) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:281}} [ 564.921037] env[63515]: DEBUG nova.virt.vmwareapi.images [None req-44598345-8888-488a-9c24-527daf4d66df tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] [instance: 93b28142-8454-43c0-b0a7-d61aa95c8fc8] Downloaded image file data 8a120570-cb06-4099-b262-554ca0ad15c5 to vmware_temp/e6b3038c-7dd7-4196-a714-112b26e83ee0/8a120570-cb06-4099-b262-554ca0ad15c5/tmp-sparse.vmdk on the data store datastore1 {{(pid=63515) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:258}} [ 564.922829] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-44598345-8888-488a-9c24-527daf4d66df tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] [instance: 93b28142-8454-43c0-b0a7-d61aa95c8fc8] Caching image {{(pid=63515) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 564.923122] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-44598345-8888-488a-9c24-527daf4d66df tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] Copying Virtual Disk [datastore1] vmware_temp/e6b3038c-7dd7-4196-a714-112b26e83ee0/8a120570-cb06-4099-b262-554ca0ad15c5/tmp-sparse.vmdk to [datastore1] vmware_temp/e6b3038c-7dd7-4196-a714-112b26e83ee0/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 564.923443] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-015e83dd-51a2-43e4-8f2f-df245a1fe6d8 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 564.940884] env[63515]: DEBUG oslo_vmware.api [None req-44598345-8888-488a-9c24-527daf4d66df tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] Waiting for the task: (returnval){ [ 564.940884] env[63515]: value = "task-1110847" [ 564.940884] env[63515]: _type = "Task" [ 564.940884] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 564.950329] env[63515]: DEBUG oslo_vmware.api [None req-44598345-8888-488a-9c24-527daf4d66df tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] Task: {'id': task-1110847, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 565.033028] env[63515]: DEBUG oslo_concurrency.lockutils [req-91a90eea-4da8-49c7-bc6f-244113f0e0aa req-5a17caf3-a3c5-48ad-aaec-acd1e0c8fdab service nova] Releasing lock "refresh_cache-cfc6c8d5-c9e8-46fb-99e7-d245664f9652" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 565.033028] env[63515]: DEBUG nova.compute.manager [req-91a90eea-4da8-49c7-bc6f-244113f0e0aa req-5a17caf3-a3c5-48ad-aaec-acd1e0c8fdab service nova] [instance: cfc6c8d5-c9e8-46fb-99e7-d245664f9652] Received event network-vif-deleted-589525f2-8185-44ca-9cdc-7c1b03dd8ee3 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 565.124311] env[63515]: DEBUG nova.compute.utils [None req-eaa7f189-a861-45f8-872f-29d870dc784b tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] Using /dev/sd instead of None {{(pid=63515) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 565.133604] env[63515]: DEBUG nova.compute.manager [None req-eaa7f189-a861-45f8-872f-29d870dc784b tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] [instance: e39bc55e-b677-4e2d-9aa4-7bc1d5ffadfa] Allocating IP information in the background. {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 565.133825] env[63515]: DEBUG nova.network.neutron [None req-eaa7f189-a861-45f8-872f-29d870dc784b tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] [instance: e39bc55e-b677-4e2d-9aa4-7bc1d5ffadfa] allocate_for_instance() {{(pid=63515) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 565.277132] env[63515]: DEBUG nova.policy [None req-eaa7f189-a861-45f8-872f-29d870dc784b tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '21c4548852c747528a0b0530f6f42be9', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e0c07bb70a414f1787073d335c9001bf', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63515) authorize /opt/stack/nova/nova/policy.py:201}} [ 565.454497] env[63515]: DEBUG oslo_vmware.api [None req-44598345-8888-488a-9c24-527daf4d66df tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] Task: {'id': task-1110847, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 565.496620] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af5d77c5-8c80-4749-920e-6d330f5ea5f8 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 565.504148] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8bd7ad3-2eed-417c-a8dc-c7e48f6c1310 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 565.551031] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-429bf096-a7c4-44de-82b2-c175f7a792f8 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 565.562021] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5131e50c-195f-47f6-a8d2-1a54617bee7d {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 565.573783] env[63515]: DEBUG nova.compute.provider_tree [None req-ada3051f-c750-48c7-b591-0344c25b8fea tempest-ServersAdminTestJSON-710082096 tempest-ServersAdminTestJSON-710082096-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 565.581494] env[63515]: INFO nova.scheduler.client.report [None req-c96f19aa-9abc-4b12-99f8-980442cab640 tempest-ServersAdminTestJSON-710082096 tempest-ServersAdminTestJSON-710082096-project-member] Deleted allocations for instance 29d199bc-aff3-4c62-af3a-1e6880b52446 [ 565.635983] env[63515]: DEBUG nova.compute.manager [None req-eaa7f189-a861-45f8-872f-29d870dc784b tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] [instance: e39bc55e-b677-4e2d-9aa4-7bc1d5ffadfa] Start building block device mappings for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 565.953983] env[63515]: DEBUG oslo_vmware.api [None req-44598345-8888-488a-9c24-527daf4d66df tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] Task: {'id': task-1110847, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.670836} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 565.955040] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-44598345-8888-488a-9c24-527daf4d66df tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] Copied Virtual Disk [datastore1] vmware_temp/e6b3038c-7dd7-4196-a714-112b26e83ee0/8a120570-cb06-4099-b262-554ca0ad15c5/tmp-sparse.vmdk to [datastore1] vmware_temp/e6b3038c-7dd7-4196-a714-112b26e83ee0/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 565.955040] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-44598345-8888-488a-9c24-527daf4d66df tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] Deleting the datastore file [datastore1] vmware_temp/e6b3038c-7dd7-4196-a714-112b26e83ee0/8a120570-cb06-4099-b262-554ca0ad15c5/tmp-sparse.vmdk {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 565.955040] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-430e5adf-7206-474c-8c4e-f136384e48a2 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 565.964250] env[63515]: DEBUG oslo_vmware.api [None req-44598345-8888-488a-9c24-527daf4d66df tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] Waiting for the task: (returnval){ [ 565.964250] env[63515]: value = "task-1110848" [ 565.964250] env[63515]: _type = "Task" [ 565.964250] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 565.974605] env[63515]: DEBUG oslo_vmware.api [None req-44598345-8888-488a-9c24-527daf4d66df tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] Task: {'id': task-1110848, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 566.080834] env[63515]: DEBUG nova.scheduler.client.report [None req-ada3051f-c750-48c7-b591-0344c25b8fea tempest-ServersAdminTestJSON-710082096 tempest-ServersAdminTestJSON-710082096-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 172, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 566.092338] env[63515]: DEBUG oslo_concurrency.lockutils [None req-c96f19aa-9abc-4b12-99f8-980442cab640 tempest-ServersAdminTestJSON-710082096 tempest-ServersAdminTestJSON-710082096-project-member] Lock "29d199bc-aff3-4c62-af3a-1e6880b52446" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 33.469s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 566.480365] env[63515]: DEBUG oslo_vmware.api [None req-44598345-8888-488a-9c24-527daf4d66df tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] Task: {'id': task-1110848, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.023745} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 566.480655] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-44598345-8888-488a-9c24-527daf4d66df tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] Deleted the datastore file {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 566.480880] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-44598345-8888-488a-9c24-527daf4d66df tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] Moving file from [datastore1] vmware_temp/e6b3038c-7dd7-4196-a714-112b26e83ee0/8a120570-cb06-4099-b262-554ca0ad15c5 to [datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5. {{(pid=63515) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:334}} [ 566.481159] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.MoveDatastoreFile_Task with opID=oslo.vmware-e3fe8cdf-277b-4866-9687-833a32c73044 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 566.489185] env[63515]: DEBUG oslo_vmware.api [None req-44598345-8888-488a-9c24-527daf4d66df tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] Waiting for the task: (returnval){ [ 566.489185] env[63515]: value = "task-1110849" [ 566.489185] env[63515]: _type = "Task" [ 566.489185] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 566.500196] env[63515]: DEBUG oslo_concurrency.lockutils [None req-30030db8-77b3-4af4-8ae6-30653bd81e23 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Acquiring lock "02ffb723-d6df-42cb-93c5-d582705d1e03" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 566.500481] env[63515]: DEBUG oslo_concurrency.lockutils [None req-30030db8-77b3-4af4-8ae6-30653bd81e23 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Lock "02ffb723-d6df-42cb-93c5-d582705d1e03" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 566.501565] env[63515]: DEBUG nova.network.neutron [None req-eaa7f189-a861-45f8-872f-29d870dc784b tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] [instance: e39bc55e-b677-4e2d-9aa4-7bc1d5ffadfa] Successfully created port: ffab05fb-3a66-444a-9cca-115828ea7f22 {{(pid=63515) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 566.509665] env[63515]: DEBUG oslo_vmware.api [None req-44598345-8888-488a-9c24-527daf4d66df tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] Task: {'id': task-1110849, 'name': MoveDatastoreFile_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 566.595197] env[63515]: DEBUG oslo_concurrency.lockutils [None req-ada3051f-c750-48c7-b591-0344c25b8fea tempest-ServersAdminTestJSON-710082096 tempest-ServersAdminTestJSON-710082096-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.988s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 566.595197] env[63515]: ERROR nova.compute.manager [None req-ada3051f-c750-48c7-b591-0344c25b8fea tempest-ServersAdminTestJSON-710082096 tempest-ServersAdminTestJSON-710082096-project-member] [instance: d7fed938-f7cd-4013-b545-5d5350242084] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port abac7593-4174-4f4c-886d-f2653e2b4123, please check neutron logs for more information. [ 566.595197] env[63515]: ERROR nova.compute.manager [instance: d7fed938-f7cd-4013-b545-5d5350242084] Traceback (most recent call last): [ 566.595197] env[63515]: ERROR nova.compute.manager [instance: d7fed938-f7cd-4013-b545-5d5350242084] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 566.595197] env[63515]: ERROR nova.compute.manager [instance: d7fed938-f7cd-4013-b545-5d5350242084] self.driver.spawn(context, instance, image_meta, [ 566.595197] env[63515]: ERROR nova.compute.manager [instance: d7fed938-f7cd-4013-b545-5d5350242084] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 566.595197] env[63515]: ERROR nova.compute.manager [instance: d7fed938-f7cd-4013-b545-5d5350242084] self._vmops.spawn(context, instance, image_meta, injected_files, [ 566.595197] env[63515]: ERROR nova.compute.manager [instance: d7fed938-f7cd-4013-b545-5d5350242084] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 566.595197] env[63515]: ERROR nova.compute.manager [instance: d7fed938-f7cd-4013-b545-5d5350242084] vm_ref = self.build_virtual_machine(instance, [ 566.595562] env[63515]: ERROR nova.compute.manager [instance: d7fed938-f7cd-4013-b545-5d5350242084] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 566.595562] env[63515]: ERROR nova.compute.manager [instance: d7fed938-f7cd-4013-b545-5d5350242084] vif_infos = vmwarevif.get_vif_info(self._session, [ 566.595562] env[63515]: ERROR nova.compute.manager [instance: d7fed938-f7cd-4013-b545-5d5350242084] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 566.595562] env[63515]: ERROR nova.compute.manager [instance: d7fed938-f7cd-4013-b545-5d5350242084] for vif in network_info: [ 566.595562] env[63515]: ERROR nova.compute.manager [instance: d7fed938-f7cd-4013-b545-5d5350242084] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 566.595562] env[63515]: ERROR nova.compute.manager [instance: d7fed938-f7cd-4013-b545-5d5350242084] return self._sync_wrapper(fn, *args, **kwargs) [ 566.595562] env[63515]: ERROR nova.compute.manager [instance: d7fed938-f7cd-4013-b545-5d5350242084] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 566.595562] env[63515]: ERROR nova.compute.manager [instance: d7fed938-f7cd-4013-b545-5d5350242084] self.wait() [ 566.595562] env[63515]: ERROR nova.compute.manager [instance: d7fed938-f7cd-4013-b545-5d5350242084] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 566.595562] env[63515]: ERROR nova.compute.manager [instance: d7fed938-f7cd-4013-b545-5d5350242084] self[:] = self._gt.wait() [ 566.595562] env[63515]: ERROR nova.compute.manager [instance: d7fed938-f7cd-4013-b545-5d5350242084] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 566.595562] env[63515]: ERROR nova.compute.manager [instance: d7fed938-f7cd-4013-b545-5d5350242084] return self._exit_event.wait() [ 566.595562] env[63515]: ERROR nova.compute.manager [instance: d7fed938-f7cd-4013-b545-5d5350242084] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 566.595914] env[63515]: ERROR nova.compute.manager [instance: d7fed938-f7cd-4013-b545-5d5350242084] result = hub.switch() [ 566.595914] env[63515]: ERROR nova.compute.manager [instance: d7fed938-f7cd-4013-b545-5d5350242084] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 566.595914] env[63515]: ERROR nova.compute.manager [instance: d7fed938-f7cd-4013-b545-5d5350242084] return self.greenlet.switch() [ 566.595914] env[63515]: ERROR nova.compute.manager [instance: d7fed938-f7cd-4013-b545-5d5350242084] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 566.595914] env[63515]: ERROR nova.compute.manager [instance: d7fed938-f7cd-4013-b545-5d5350242084] result = function(*args, **kwargs) [ 566.595914] env[63515]: ERROR nova.compute.manager [instance: d7fed938-f7cd-4013-b545-5d5350242084] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 566.595914] env[63515]: ERROR nova.compute.manager [instance: d7fed938-f7cd-4013-b545-5d5350242084] return func(*args, **kwargs) [ 566.595914] env[63515]: ERROR nova.compute.manager [instance: d7fed938-f7cd-4013-b545-5d5350242084] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 566.595914] env[63515]: ERROR nova.compute.manager [instance: d7fed938-f7cd-4013-b545-5d5350242084] raise e [ 566.595914] env[63515]: ERROR nova.compute.manager [instance: d7fed938-f7cd-4013-b545-5d5350242084] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 566.595914] env[63515]: ERROR nova.compute.manager [instance: d7fed938-f7cd-4013-b545-5d5350242084] nwinfo = self.network_api.allocate_for_instance( [ 566.595914] env[63515]: ERROR nova.compute.manager [instance: d7fed938-f7cd-4013-b545-5d5350242084] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 566.595914] env[63515]: ERROR nova.compute.manager [instance: d7fed938-f7cd-4013-b545-5d5350242084] created_port_ids = self._update_ports_for_instance( [ 566.596306] env[63515]: ERROR nova.compute.manager [instance: d7fed938-f7cd-4013-b545-5d5350242084] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 566.596306] env[63515]: ERROR nova.compute.manager [instance: d7fed938-f7cd-4013-b545-5d5350242084] with excutils.save_and_reraise_exception(): [ 566.596306] env[63515]: ERROR nova.compute.manager [instance: d7fed938-f7cd-4013-b545-5d5350242084] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 566.596306] env[63515]: ERROR nova.compute.manager [instance: d7fed938-f7cd-4013-b545-5d5350242084] self.force_reraise() [ 566.596306] env[63515]: ERROR nova.compute.manager [instance: d7fed938-f7cd-4013-b545-5d5350242084] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 566.596306] env[63515]: ERROR nova.compute.manager [instance: d7fed938-f7cd-4013-b545-5d5350242084] raise self.value [ 566.596306] env[63515]: ERROR nova.compute.manager [instance: d7fed938-f7cd-4013-b545-5d5350242084] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 566.596306] env[63515]: ERROR nova.compute.manager [instance: d7fed938-f7cd-4013-b545-5d5350242084] updated_port = self._update_port( [ 566.596306] env[63515]: ERROR nova.compute.manager [instance: d7fed938-f7cd-4013-b545-5d5350242084] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 566.596306] env[63515]: ERROR nova.compute.manager [instance: d7fed938-f7cd-4013-b545-5d5350242084] _ensure_no_port_binding_failure(port) [ 566.596306] env[63515]: ERROR nova.compute.manager [instance: d7fed938-f7cd-4013-b545-5d5350242084] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 566.596306] env[63515]: ERROR nova.compute.manager [instance: d7fed938-f7cd-4013-b545-5d5350242084] raise exception.PortBindingFailed(port_id=port['id']) [ 566.596655] env[63515]: ERROR nova.compute.manager [instance: d7fed938-f7cd-4013-b545-5d5350242084] nova.exception.PortBindingFailed: Binding failed for port abac7593-4174-4f4c-886d-f2653e2b4123, please check neutron logs for more information. [ 566.596655] env[63515]: ERROR nova.compute.manager [instance: d7fed938-f7cd-4013-b545-5d5350242084] [ 566.596655] env[63515]: DEBUG nova.compute.utils [None req-ada3051f-c750-48c7-b591-0344c25b8fea tempest-ServersAdminTestJSON-710082096 tempest-ServersAdminTestJSON-710082096-project-member] [instance: d7fed938-f7cd-4013-b545-5d5350242084] Binding failed for port abac7593-4174-4f4c-886d-f2653e2b4123, please check neutron logs for more information. {{(pid=63515) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 566.598863] env[63515]: DEBUG nova.compute.manager [None req-ada3051f-c750-48c7-b591-0344c25b8fea tempest-ServersAdminTestJSON-710082096 tempest-ServersAdminTestJSON-710082096-project-member] [instance: d7fed938-f7cd-4013-b545-5d5350242084] Build of instance d7fed938-f7cd-4013-b545-5d5350242084 was re-scheduled: Binding failed for port abac7593-4174-4f4c-886d-f2653e2b4123, please check neutron logs for more information. {{(pid=63515) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 566.599336] env[63515]: DEBUG nova.compute.manager [None req-ada3051f-c750-48c7-b591-0344c25b8fea tempest-ServersAdminTestJSON-710082096 tempest-ServersAdminTestJSON-710082096-project-member] [instance: d7fed938-f7cd-4013-b545-5d5350242084] Unplugging VIFs for instance {{(pid=63515) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 566.600192] env[63515]: DEBUG oslo_concurrency.lockutils [None req-ada3051f-c750-48c7-b591-0344c25b8fea tempest-ServersAdminTestJSON-710082096 tempest-ServersAdminTestJSON-710082096-project-member] Acquiring lock "refresh_cache-d7fed938-f7cd-4013-b545-5d5350242084" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 566.600192] env[63515]: DEBUG oslo_concurrency.lockutils [None req-ada3051f-c750-48c7-b591-0344c25b8fea tempest-ServersAdminTestJSON-710082096 tempest-ServersAdminTestJSON-710082096-project-member] Acquired lock "refresh_cache-d7fed938-f7cd-4013-b545-5d5350242084" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 566.600192] env[63515]: DEBUG nova.network.neutron [None req-ada3051f-c750-48c7-b591-0344c25b8fea tempest-ServersAdminTestJSON-710082096 tempest-ServersAdminTestJSON-710082096-project-member] [instance: d7fed938-f7cd-4013-b545-5d5350242084] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 566.600967] env[63515]: DEBUG oslo_concurrency.lockutils [None req-7f5043f7-e6d7-46dd-88fd-99ee049c2b4b tempest-VolumesAssistedSnapshotsTest-453349191 tempest-VolumesAssistedSnapshotsTest-453349191-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.139s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 566.602469] env[63515]: INFO nova.compute.claims [None req-7f5043f7-e6d7-46dd-88fd-99ee049c2b4b tempest-VolumesAssistedSnapshotsTest-453349191 tempest-VolumesAssistedSnapshotsTest-453349191-project-member] [instance: db682708-36fa-4126-a848-bfb609df96d7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 566.605170] env[63515]: DEBUG nova.compute.manager [None req-12d73cc4-118d-457e-b393-cc4362d15b0d tempest-ListImageFiltersTestJSON-1100101045 tempest-ListImageFiltersTestJSON-1100101045-project-member] [instance: ab1539ba-33a4-4139-a2f3-1f8c507beede] Starting instance... {{(pid=63515) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 566.654248] env[63515]: DEBUG nova.compute.manager [None req-eaa7f189-a861-45f8-872f-29d870dc784b tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] [instance: e39bc55e-b677-4e2d-9aa4-7bc1d5ffadfa] Start spawning the instance on the hypervisor. {{(pid=63515) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 566.683084] env[63515]: DEBUG nova.virt.hardware [None req-eaa7f189-a861-45f8-872f-29d870dc784b tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T02:52:14Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T02:51:56Z,direct_url=,disk_format='vmdk',id=8a120570-cb06-4099-b262-554ca0ad15c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b569255cc43e42e7ae3f2b4ad37c6ef4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T02:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 566.683338] env[63515]: DEBUG nova.virt.hardware [None req-eaa7f189-a861-45f8-872f-29d870dc784b tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] Flavor limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 566.683506] env[63515]: DEBUG nova.virt.hardware [None req-eaa7f189-a861-45f8-872f-29d870dc784b tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] Image limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 566.683652] env[63515]: DEBUG nova.virt.hardware [None req-eaa7f189-a861-45f8-872f-29d870dc784b tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] Flavor pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 566.683796] env[63515]: DEBUG nova.virt.hardware [None req-eaa7f189-a861-45f8-872f-29d870dc784b tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] Image pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 566.683954] env[63515]: DEBUG nova.virt.hardware [None req-eaa7f189-a861-45f8-872f-29d870dc784b tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 566.684202] env[63515]: DEBUG nova.virt.hardware [None req-eaa7f189-a861-45f8-872f-29d870dc784b tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 566.684366] env[63515]: DEBUG nova.virt.hardware [None req-eaa7f189-a861-45f8-872f-29d870dc784b tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 566.685141] env[63515]: DEBUG nova.virt.hardware [None req-eaa7f189-a861-45f8-872f-29d870dc784b tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] Got 1 possible topologies {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 566.685141] env[63515]: DEBUG nova.virt.hardware [None req-eaa7f189-a861-45f8-872f-29d870dc784b tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 566.685141] env[63515]: DEBUG nova.virt.hardware [None req-eaa7f189-a861-45f8-872f-29d870dc784b tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 566.685754] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80ee46e4-cbf5-47a4-b42e-59289da6a1bb {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 566.694975] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5288a6b4-851a-470b-946c-d75016da9a29 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 567.000465] env[63515]: DEBUG oslo_vmware.api [None req-44598345-8888-488a-9c24-527daf4d66df tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] Task: {'id': task-1110849, 'name': MoveDatastoreFile_Task, 'duration_secs': 0.025133} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 567.000766] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-44598345-8888-488a-9c24-527daf4d66df tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] File moved {{(pid=63515) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:346}} [ 567.000928] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-44598345-8888-488a-9c24-527daf4d66df tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] [instance: 93b28142-8454-43c0-b0a7-d61aa95c8fc8] Cleaning up location [datastore1] vmware_temp/e6b3038c-7dd7-4196-a714-112b26e83ee0 {{(pid=63515) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 567.001113] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-44598345-8888-488a-9c24-527daf4d66df tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] Deleting the datastore file [datastore1] vmware_temp/e6b3038c-7dd7-4196-a714-112b26e83ee0 {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 567.001367] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3c62c565-0ca5-4a6d-b8a0-c21a53a92378 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 567.011265] env[63515]: DEBUG oslo_vmware.api [None req-44598345-8888-488a-9c24-527daf4d66df tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] Waiting for the task: (returnval){ [ 567.011265] env[63515]: value = "task-1110850" [ 567.011265] env[63515]: _type = "Task" [ 567.011265] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 567.017824] env[63515]: DEBUG oslo_vmware.api [None req-44598345-8888-488a-9c24-527daf4d66df tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] Task: {'id': task-1110850, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 567.140755] env[63515]: DEBUG oslo_concurrency.lockutils [None req-12d73cc4-118d-457e-b393-cc4362d15b0d tempest-ListImageFiltersTestJSON-1100101045 tempest-ListImageFiltersTestJSON-1100101045-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 567.151700] env[63515]: DEBUG nova.network.neutron [None req-ada3051f-c750-48c7-b591-0344c25b8fea tempest-ServersAdminTestJSON-710082096 tempest-ServersAdminTestJSON-710082096-project-member] [instance: d7fed938-f7cd-4013-b545-5d5350242084] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 567.476207] env[63515]: DEBUG nova.network.neutron [None req-ada3051f-c750-48c7-b591-0344c25b8fea tempest-ServersAdminTestJSON-710082096 tempest-ServersAdminTestJSON-710082096-project-member] [instance: d7fed938-f7cd-4013-b545-5d5350242084] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 567.525395] env[63515]: DEBUG oslo_vmware.api [None req-44598345-8888-488a-9c24-527daf4d66df tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] Task: {'id': task-1110850, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.030256} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 567.525395] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-44598345-8888-488a-9c24-527daf4d66df tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] Deleted the datastore file {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 567.526134] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-70b2fd68-1322-49c4-8005-9f763d1a43b8 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 567.533995] env[63515]: DEBUG oslo_vmware.api [None req-44598345-8888-488a-9c24-527daf4d66df tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] Waiting for the task: (returnval){ [ 567.533995] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]5271ed43-61b1-0184-6b72-884cd3287851" [ 567.533995] env[63515]: _type = "Task" [ 567.533995] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 567.545802] env[63515]: DEBUG oslo_vmware.api [None req-44598345-8888-488a-9c24-527daf4d66df tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]5271ed43-61b1-0184-6b72-884cd3287851, 'name': SearchDatastore_Task, 'duration_secs': 0.009624} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 567.546581] env[63515]: DEBUG oslo_concurrency.lockutils [None req-44598345-8888-488a-9c24-527daf4d66df tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 567.546581] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-44598345-8888-488a-9c24-527daf4d66df tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk to [datastore1] 93b28142-8454-43c0-b0a7-d61aa95c8fc8/93b28142-8454-43c0-b0a7-d61aa95c8fc8.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 567.546581] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-59eb17d4-6d0b-4f6b-b714-0d2599ec44d7 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 567.553542] env[63515]: DEBUG oslo_vmware.api [None req-44598345-8888-488a-9c24-527daf4d66df tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] Waiting for the task: (returnval){ [ 567.553542] env[63515]: value = "task-1110851" [ 567.553542] env[63515]: _type = "Task" [ 567.553542] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 567.569552] env[63515]: DEBUG oslo_vmware.api [None req-44598345-8888-488a-9c24-527daf4d66df tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] Task: {'id': task-1110851, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 567.979992] env[63515]: DEBUG oslo_concurrency.lockutils [None req-ada3051f-c750-48c7-b591-0344c25b8fea tempest-ServersAdminTestJSON-710082096 tempest-ServersAdminTestJSON-710082096-project-member] Releasing lock "refresh_cache-d7fed938-f7cd-4013-b545-5d5350242084" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 567.980291] env[63515]: DEBUG nova.compute.manager [None req-ada3051f-c750-48c7-b591-0344c25b8fea tempest-ServersAdminTestJSON-710082096 tempest-ServersAdminTestJSON-710082096-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63515) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 567.980483] env[63515]: DEBUG nova.compute.manager [None req-ada3051f-c750-48c7-b591-0344c25b8fea tempest-ServersAdminTestJSON-710082096 tempest-ServersAdminTestJSON-710082096-project-member] [instance: d7fed938-f7cd-4013-b545-5d5350242084] Deallocating network for instance {{(pid=63515) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 567.980613] env[63515]: DEBUG nova.network.neutron [None req-ada3051f-c750-48c7-b591-0344c25b8fea tempest-ServersAdminTestJSON-710082096 tempest-ServersAdminTestJSON-710082096-project-member] [instance: d7fed938-f7cd-4013-b545-5d5350242084] deallocate_for_instance() {{(pid=63515) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 568.023030] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-127849f7-7cd0-4e5b-ab1e-7c8515df952b {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 568.030506] env[63515]: DEBUG nova.network.neutron [None req-ada3051f-c750-48c7-b591-0344c25b8fea tempest-ServersAdminTestJSON-710082096 tempest-ServersAdminTestJSON-710082096-project-member] [instance: d7fed938-f7cd-4013-b545-5d5350242084] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 568.040637] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2c78e64-c724-4b5d-a6b2-306978b7ff6c {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 568.081068] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-024b6aa6-b995-492b-a2cc-d8b6a28de694 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 568.090931] env[63515]: DEBUG oslo_vmware.api [None req-44598345-8888-488a-9c24-527daf4d66df tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] Task: {'id': task-1110851, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.505162} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 568.092789] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-44598345-8888-488a-9c24-527daf4d66df tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk to [datastore1] 93b28142-8454-43c0-b0a7-d61aa95c8fc8/93b28142-8454-43c0-b0a7-d61aa95c8fc8.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 568.093090] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-44598345-8888-488a-9c24-527daf4d66df tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] [instance: 93b28142-8454-43c0-b0a7-d61aa95c8fc8] Extending root virtual disk to 1048576 {{(pid=63515) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 568.093646] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ed846aa3-6ac1-45cd-a350-bff8014fa3bf {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 568.105171] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6ba2c42-a833-4404-a938-192352192839 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 568.110345] env[63515]: DEBUG oslo_vmware.api [None req-44598345-8888-488a-9c24-527daf4d66df tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] Waiting for the task: (returnval){ [ 568.110345] env[63515]: value = "task-1110852" [ 568.110345] env[63515]: _type = "Task" [ 568.110345] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 568.122802] env[63515]: DEBUG nova.compute.provider_tree [None req-7f5043f7-e6d7-46dd-88fd-99ee049c2b4b tempest-VolumesAssistedSnapshotsTest-453349191 tempest-VolumesAssistedSnapshotsTest-453349191-project-member] Updating inventory in ProviderTree for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 568.130569] env[63515]: DEBUG oslo_vmware.api [None req-44598345-8888-488a-9c24-527daf4d66df tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] Task: {'id': task-1110852, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 568.534314] env[63515]: DEBUG nova.network.neutron [None req-ada3051f-c750-48c7-b591-0344c25b8fea tempest-ServersAdminTestJSON-710082096 tempest-ServersAdminTestJSON-710082096-project-member] [instance: d7fed938-f7cd-4013-b545-5d5350242084] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 568.624863] env[63515]: DEBUG oslo_vmware.api [None req-44598345-8888-488a-9c24-527daf4d66df tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] Task: {'id': task-1110852, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.066946} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 568.625262] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-44598345-8888-488a-9c24-527daf4d66df tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] [instance: 93b28142-8454-43c0-b0a7-d61aa95c8fc8] Extended root virtual disk {{(pid=63515) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 568.628015] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a3a6f88-fff9-41d2-8984-c021074e7abd {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 568.658043] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-44598345-8888-488a-9c24-527daf4d66df tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] [instance: 93b28142-8454-43c0-b0a7-d61aa95c8fc8] Reconfiguring VM instance instance-0000000a to attach disk [datastore1] 93b28142-8454-43c0-b0a7-d61aa95c8fc8/93b28142-8454-43c0-b0a7-d61aa95c8fc8.vmdk or device None with type sparse {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 568.660326] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-81e1352f-b17b-4669-ac0f-d6723fed5f87 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 568.676623] env[63515]: ERROR nova.scheduler.client.report [None req-7f5043f7-e6d7-46dd-88fd-99ee049c2b4b tempest-VolumesAssistedSnapshotsTest-453349191 tempest-VolumesAssistedSnapshotsTest-453349191-project-member] [req-570307d7-6fd0-4cac-b5e8-72504cc84995] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 2bd86232-2b6f-44d5-9057-1a3a6b27185a. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-570307d7-6fd0-4cac-b5e8-72504cc84995"}]} [ 568.685945] env[63515]: DEBUG oslo_vmware.api [None req-44598345-8888-488a-9c24-527daf4d66df tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] Waiting for the task: (returnval){ [ 568.685945] env[63515]: value = "task-1110853" [ 568.685945] env[63515]: _type = "Task" [ 568.685945] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 568.696682] env[63515]: DEBUG oslo_vmware.api [None req-44598345-8888-488a-9c24-527daf4d66df tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] Task: {'id': task-1110853, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 568.697798] env[63515]: DEBUG nova.scheduler.client.report [None req-7f5043f7-e6d7-46dd-88fd-99ee049c2b4b tempest-VolumesAssistedSnapshotsTest-453349191 tempest-VolumesAssistedSnapshotsTest-453349191-project-member] Refreshing inventories for resource provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 568.720307] env[63515]: DEBUG nova.scheduler.client.report [None req-7f5043f7-e6d7-46dd-88fd-99ee049c2b4b tempest-VolumesAssistedSnapshotsTest-453349191 tempest-VolumesAssistedSnapshotsTest-453349191-project-member] Updating ProviderTree inventory for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 172, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 568.720307] env[63515]: DEBUG nova.compute.provider_tree [None req-7f5043f7-e6d7-46dd-88fd-99ee049c2b4b tempest-VolumesAssistedSnapshotsTest-453349191 tempest-VolumesAssistedSnapshotsTest-453349191-project-member] Updating inventory in ProviderTree for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 172, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 568.736135] env[63515]: DEBUG nova.scheduler.client.report [None req-7f5043f7-e6d7-46dd-88fd-99ee049c2b4b tempest-VolumesAssistedSnapshotsTest-453349191 tempest-VolumesAssistedSnapshotsTest-453349191-project-member] Refreshing aggregate associations for resource provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a, aggregates: None {{(pid=63515) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 568.758664] env[63515]: DEBUG nova.scheduler.client.report [None req-7f5043f7-e6d7-46dd-88fd-99ee049c2b4b tempest-VolumesAssistedSnapshotsTest-453349191 tempest-VolumesAssistedSnapshotsTest-453349191-project-member] Refreshing trait associations for resource provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a, traits: COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=63515) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 569.039027] env[63515]: INFO nova.compute.manager [None req-ada3051f-c750-48c7-b591-0344c25b8fea tempest-ServersAdminTestJSON-710082096 tempest-ServersAdminTestJSON-710082096-project-member] [instance: d7fed938-f7cd-4013-b545-5d5350242084] Took 1.06 seconds to deallocate network for instance. [ 569.093343] env[63515]: ERROR nova.compute.manager [None req-eaa7f189-a861-45f8-872f-29d870dc784b tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port ffab05fb-3a66-444a-9cca-115828ea7f22, please check neutron logs for more information. [ 569.093343] env[63515]: ERROR nova.compute.manager Traceback (most recent call last): [ 569.093343] env[63515]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 569.093343] env[63515]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 569.093343] env[63515]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 569.093343] env[63515]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 569.093343] env[63515]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 569.093343] env[63515]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 569.093343] env[63515]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 569.093343] env[63515]: ERROR nova.compute.manager self.force_reraise() [ 569.093343] env[63515]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 569.093343] env[63515]: ERROR nova.compute.manager raise self.value [ 569.093343] env[63515]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 569.093343] env[63515]: ERROR nova.compute.manager updated_port = self._update_port( [ 569.093343] env[63515]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 569.093343] env[63515]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 569.093878] env[63515]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 569.093878] env[63515]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 569.093878] env[63515]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port ffab05fb-3a66-444a-9cca-115828ea7f22, please check neutron logs for more information. [ 569.093878] env[63515]: ERROR nova.compute.manager [ 569.093878] env[63515]: Traceback (most recent call last): [ 569.093878] env[63515]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 569.093878] env[63515]: listener.cb(fileno) [ 569.093878] env[63515]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 569.093878] env[63515]: result = function(*args, **kwargs) [ 569.093878] env[63515]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 569.093878] env[63515]: return func(*args, **kwargs) [ 569.093878] env[63515]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 569.093878] env[63515]: raise e [ 569.093878] env[63515]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 569.093878] env[63515]: nwinfo = self.network_api.allocate_for_instance( [ 569.093878] env[63515]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 569.093878] env[63515]: created_port_ids = self._update_ports_for_instance( [ 569.093878] env[63515]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 569.093878] env[63515]: with excutils.save_and_reraise_exception(): [ 569.093878] env[63515]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 569.093878] env[63515]: self.force_reraise() [ 569.093878] env[63515]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 569.093878] env[63515]: raise self.value [ 569.093878] env[63515]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 569.093878] env[63515]: updated_port = self._update_port( [ 569.093878] env[63515]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 569.093878] env[63515]: _ensure_no_port_binding_failure(port) [ 569.093878] env[63515]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 569.093878] env[63515]: raise exception.PortBindingFailed(port_id=port['id']) [ 569.094788] env[63515]: nova.exception.PortBindingFailed: Binding failed for port ffab05fb-3a66-444a-9cca-115828ea7f22, please check neutron logs for more information. [ 569.094788] env[63515]: Removing descriptor: 16 [ 569.094788] env[63515]: ERROR nova.compute.manager [None req-eaa7f189-a861-45f8-872f-29d870dc784b tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] [instance: e39bc55e-b677-4e2d-9aa4-7bc1d5ffadfa] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port ffab05fb-3a66-444a-9cca-115828ea7f22, please check neutron logs for more information. [ 569.094788] env[63515]: ERROR nova.compute.manager [instance: e39bc55e-b677-4e2d-9aa4-7bc1d5ffadfa] Traceback (most recent call last): [ 569.094788] env[63515]: ERROR nova.compute.manager [instance: e39bc55e-b677-4e2d-9aa4-7bc1d5ffadfa] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 569.094788] env[63515]: ERROR nova.compute.manager [instance: e39bc55e-b677-4e2d-9aa4-7bc1d5ffadfa] yield resources [ 569.094788] env[63515]: ERROR nova.compute.manager [instance: e39bc55e-b677-4e2d-9aa4-7bc1d5ffadfa] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 569.094788] env[63515]: ERROR nova.compute.manager [instance: e39bc55e-b677-4e2d-9aa4-7bc1d5ffadfa] self.driver.spawn(context, instance, image_meta, [ 569.094788] env[63515]: ERROR nova.compute.manager [instance: e39bc55e-b677-4e2d-9aa4-7bc1d5ffadfa] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 569.094788] env[63515]: ERROR nova.compute.manager [instance: e39bc55e-b677-4e2d-9aa4-7bc1d5ffadfa] self._vmops.spawn(context, instance, image_meta, injected_files, [ 569.094788] env[63515]: ERROR nova.compute.manager [instance: e39bc55e-b677-4e2d-9aa4-7bc1d5ffadfa] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 569.094788] env[63515]: ERROR nova.compute.manager [instance: e39bc55e-b677-4e2d-9aa4-7bc1d5ffadfa] vm_ref = self.build_virtual_machine(instance, [ 569.095176] env[63515]: ERROR nova.compute.manager [instance: e39bc55e-b677-4e2d-9aa4-7bc1d5ffadfa] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 569.095176] env[63515]: ERROR nova.compute.manager [instance: e39bc55e-b677-4e2d-9aa4-7bc1d5ffadfa] vif_infos = vmwarevif.get_vif_info(self._session, [ 569.095176] env[63515]: ERROR nova.compute.manager [instance: e39bc55e-b677-4e2d-9aa4-7bc1d5ffadfa] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 569.095176] env[63515]: ERROR nova.compute.manager [instance: e39bc55e-b677-4e2d-9aa4-7bc1d5ffadfa] for vif in network_info: [ 569.095176] env[63515]: ERROR nova.compute.manager [instance: e39bc55e-b677-4e2d-9aa4-7bc1d5ffadfa] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 569.095176] env[63515]: ERROR nova.compute.manager [instance: e39bc55e-b677-4e2d-9aa4-7bc1d5ffadfa] return self._sync_wrapper(fn, *args, **kwargs) [ 569.095176] env[63515]: ERROR nova.compute.manager [instance: e39bc55e-b677-4e2d-9aa4-7bc1d5ffadfa] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 569.095176] env[63515]: ERROR nova.compute.manager [instance: e39bc55e-b677-4e2d-9aa4-7bc1d5ffadfa] self.wait() [ 569.095176] env[63515]: ERROR nova.compute.manager [instance: e39bc55e-b677-4e2d-9aa4-7bc1d5ffadfa] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 569.095176] env[63515]: ERROR nova.compute.manager [instance: e39bc55e-b677-4e2d-9aa4-7bc1d5ffadfa] self[:] = self._gt.wait() [ 569.095176] env[63515]: ERROR nova.compute.manager [instance: e39bc55e-b677-4e2d-9aa4-7bc1d5ffadfa] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 569.095176] env[63515]: ERROR nova.compute.manager [instance: e39bc55e-b677-4e2d-9aa4-7bc1d5ffadfa] return self._exit_event.wait() [ 569.095176] env[63515]: ERROR nova.compute.manager [instance: e39bc55e-b677-4e2d-9aa4-7bc1d5ffadfa] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 569.095544] env[63515]: ERROR nova.compute.manager [instance: e39bc55e-b677-4e2d-9aa4-7bc1d5ffadfa] result = hub.switch() [ 569.095544] env[63515]: ERROR nova.compute.manager [instance: e39bc55e-b677-4e2d-9aa4-7bc1d5ffadfa] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 569.095544] env[63515]: ERROR nova.compute.manager [instance: e39bc55e-b677-4e2d-9aa4-7bc1d5ffadfa] return self.greenlet.switch() [ 569.095544] env[63515]: ERROR nova.compute.manager [instance: e39bc55e-b677-4e2d-9aa4-7bc1d5ffadfa] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 569.095544] env[63515]: ERROR nova.compute.manager [instance: e39bc55e-b677-4e2d-9aa4-7bc1d5ffadfa] result = function(*args, **kwargs) [ 569.095544] env[63515]: ERROR nova.compute.manager [instance: e39bc55e-b677-4e2d-9aa4-7bc1d5ffadfa] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 569.095544] env[63515]: ERROR nova.compute.manager [instance: e39bc55e-b677-4e2d-9aa4-7bc1d5ffadfa] return func(*args, **kwargs) [ 569.095544] env[63515]: ERROR nova.compute.manager [instance: e39bc55e-b677-4e2d-9aa4-7bc1d5ffadfa] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 569.095544] env[63515]: ERROR nova.compute.manager [instance: e39bc55e-b677-4e2d-9aa4-7bc1d5ffadfa] raise e [ 569.095544] env[63515]: ERROR nova.compute.manager [instance: e39bc55e-b677-4e2d-9aa4-7bc1d5ffadfa] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 569.095544] env[63515]: ERROR nova.compute.manager [instance: e39bc55e-b677-4e2d-9aa4-7bc1d5ffadfa] nwinfo = self.network_api.allocate_for_instance( [ 569.095544] env[63515]: ERROR nova.compute.manager [instance: e39bc55e-b677-4e2d-9aa4-7bc1d5ffadfa] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 569.095544] env[63515]: ERROR nova.compute.manager [instance: e39bc55e-b677-4e2d-9aa4-7bc1d5ffadfa] created_port_ids = self._update_ports_for_instance( [ 569.096629] env[63515]: ERROR nova.compute.manager [instance: e39bc55e-b677-4e2d-9aa4-7bc1d5ffadfa] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 569.096629] env[63515]: ERROR nova.compute.manager [instance: e39bc55e-b677-4e2d-9aa4-7bc1d5ffadfa] with excutils.save_and_reraise_exception(): [ 569.096629] env[63515]: ERROR nova.compute.manager [instance: e39bc55e-b677-4e2d-9aa4-7bc1d5ffadfa] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 569.096629] env[63515]: ERROR nova.compute.manager [instance: e39bc55e-b677-4e2d-9aa4-7bc1d5ffadfa] self.force_reraise() [ 569.096629] env[63515]: ERROR nova.compute.manager [instance: e39bc55e-b677-4e2d-9aa4-7bc1d5ffadfa] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 569.096629] env[63515]: ERROR nova.compute.manager [instance: e39bc55e-b677-4e2d-9aa4-7bc1d5ffadfa] raise self.value [ 569.096629] env[63515]: ERROR nova.compute.manager [instance: e39bc55e-b677-4e2d-9aa4-7bc1d5ffadfa] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 569.096629] env[63515]: ERROR nova.compute.manager [instance: e39bc55e-b677-4e2d-9aa4-7bc1d5ffadfa] updated_port = self._update_port( [ 569.096629] env[63515]: ERROR nova.compute.manager [instance: e39bc55e-b677-4e2d-9aa4-7bc1d5ffadfa] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 569.096629] env[63515]: ERROR nova.compute.manager [instance: e39bc55e-b677-4e2d-9aa4-7bc1d5ffadfa] _ensure_no_port_binding_failure(port) [ 569.096629] env[63515]: ERROR nova.compute.manager [instance: e39bc55e-b677-4e2d-9aa4-7bc1d5ffadfa] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 569.096629] env[63515]: ERROR nova.compute.manager [instance: e39bc55e-b677-4e2d-9aa4-7bc1d5ffadfa] raise exception.PortBindingFailed(port_id=port['id']) [ 569.097317] env[63515]: ERROR nova.compute.manager [instance: e39bc55e-b677-4e2d-9aa4-7bc1d5ffadfa] nova.exception.PortBindingFailed: Binding failed for port ffab05fb-3a66-444a-9cca-115828ea7f22, please check neutron logs for more information. [ 569.097317] env[63515]: ERROR nova.compute.manager [instance: e39bc55e-b677-4e2d-9aa4-7bc1d5ffadfa] [ 569.097317] env[63515]: INFO nova.compute.manager [None req-eaa7f189-a861-45f8-872f-29d870dc784b tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] [instance: e39bc55e-b677-4e2d-9aa4-7bc1d5ffadfa] Terminating instance [ 569.099214] env[63515]: DEBUG oslo_concurrency.lockutils [None req-eaa7f189-a861-45f8-872f-29d870dc784b tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] Acquiring lock "refresh_cache-e39bc55e-b677-4e2d-9aa4-7bc1d5ffadfa" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 569.099214] env[63515]: DEBUG oslo_concurrency.lockutils [None req-eaa7f189-a861-45f8-872f-29d870dc784b tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] Acquired lock "refresh_cache-e39bc55e-b677-4e2d-9aa4-7bc1d5ffadfa" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 569.099214] env[63515]: DEBUG nova.network.neutron [None req-eaa7f189-a861-45f8-872f-29d870dc784b tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] [instance: e39bc55e-b677-4e2d-9aa4-7bc1d5ffadfa] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 569.141126] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6884c415-2add-4baf-8b1c-e45b7b125f94 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.154175] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4942a33e-9578-41ba-bb80-e19828313bca {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.193055] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33f1f3c1-cb4e-494d-8751-1d84883818e3 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.201719] env[63515]: DEBUG oslo_vmware.api [None req-44598345-8888-488a-9c24-527daf4d66df tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] Task: {'id': task-1110853, 'name': ReconfigVM_Task, 'duration_secs': 0.279645} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 569.204310] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-44598345-8888-488a-9c24-527daf4d66df tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] [instance: 93b28142-8454-43c0-b0a7-d61aa95c8fc8] Reconfigured VM instance instance-0000000a to attach disk [datastore1] 93b28142-8454-43c0-b0a7-d61aa95c8fc8/93b28142-8454-43c0-b0a7-d61aa95c8fc8.vmdk or device None with type sparse {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 569.205891] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-76988d91-fd4a-41b9-869c-54aa9918e6e2 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.208490] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85b6652b-19a5-4538-bacf-d183100eb0cf {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.225596] env[63515]: DEBUG oslo_vmware.api [None req-44598345-8888-488a-9c24-527daf4d66df tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] Waiting for the task: (returnval){ [ 569.225596] env[63515]: value = "task-1110854" [ 569.225596] env[63515]: _type = "Task" [ 569.225596] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 569.226112] env[63515]: DEBUG nova.compute.provider_tree [None req-7f5043f7-e6d7-46dd-88fd-99ee049c2b4b tempest-VolumesAssistedSnapshotsTest-453349191 tempest-VolumesAssistedSnapshotsTest-453349191-project-member] Updating inventory in ProviderTree for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 569.239155] env[63515]: DEBUG oslo_vmware.api [None req-44598345-8888-488a-9c24-527daf4d66df tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] Task: {'id': task-1110854, 'name': Rename_Task} progress is 14%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 569.452035] env[63515]: DEBUG nova.compute.manager [req-70c40b80-8c24-4cb3-8d36-23f8abd79a05 req-bdea7a46-aaf5-4080-887f-01e6a8d62e41 service nova] [instance: e39bc55e-b677-4e2d-9aa4-7bc1d5ffadfa] Received event network-changed-ffab05fb-3a66-444a-9cca-115828ea7f22 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 569.452271] env[63515]: DEBUG nova.compute.manager [req-70c40b80-8c24-4cb3-8d36-23f8abd79a05 req-bdea7a46-aaf5-4080-887f-01e6a8d62e41 service nova] [instance: e39bc55e-b677-4e2d-9aa4-7bc1d5ffadfa] Refreshing instance network info cache due to event network-changed-ffab05fb-3a66-444a-9cca-115828ea7f22. {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 569.452466] env[63515]: DEBUG oslo_concurrency.lockutils [req-70c40b80-8c24-4cb3-8d36-23f8abd79a05 req-bdea7a46-aaf5-4080-887f-01e6a8d62e41 service nova] Acquiring lock "refresh_cache-e39bc55e-b677-4e2d-9aa4-7bc1d5ffadfa" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 569.630733] env[63515]: DEBUG nova.network.neutron [None req-eaa7f189-a861-45f8-872f-29d870dc784b tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] [instance: e39bc55e-b677-4e2d-9aa4-7bc1d5ffadfa] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 569.707192] env[63515]: DEBUG oslo_concurrency.lockutils [None req-30987b61-ce89-4676-89be-fee7786462c9 tempest-ImagesOneServerNegativeTestJSON-52225474 tempest-ImagesOneServerNegativeTestJSON-52225474-project-member] Acquiring lock "bbce1137-d691-4633-87ff-f9f4ea257ed7" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 569.707445] env[63515]: DEBUG oslo_concurrency.lockutils [None req-30987b61-ce89-4676-89be-fee7786462c9 tempest-ImagesOneServerNegativeTestJSON-52225474 tempest-ImagesOneServerNegativeTestJSON-52225474-project-member] Lock "bbce1137-d691-4633-87ff-f9f4ea257ed7" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 569.750191] env[63515]: DEBUG oslo_vmware.api [None req-44598345-8888-488a-9c24-527daf4d66df tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] Task: {'id': task-1110854, 'name': Rename_Task, 'duration_secs': 0.129236} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 569.750896] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-44598345-8888-488a-9c24-527daf4d66df tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] [instance: 93b28142-8454-43c0-b0a7-d61aa95c8fc8] Powering on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 569.751694] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5f88ca38-df54-4bf7-92e2-54cb0a11f248 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.761021] env[63515]: DEBUG oslo_vmware.api [None req-44598345-8888-488a-9c24-527daf4d66df tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] Waiting for the task: (returnval){ [ 569.761021] env[63515]: value = "task-1110855" [ 569.761021] env[63515]: _type = "Task" [ 569.761021] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 569.769301] env[63515]: ERROR nova.scheduler.client.report [None req-7f5043f7-e6d7-46dd-88fd-99ee049c2b4b tempest-VolumesAssistedSnapshotsTest-453349191 tempest-VolumesAssistedSnapshotsTest-453349191-project-member] [req-290df22c-cd32-46f7-9350-d4ffeebfa115] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 2bd86232-2b6f-44d5-9057-1a3a6b27185a. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-290df22c-cd32-46f7-9350-d4ffeebfa115"}]} [ 569.769773] env[63515]: DEBUG oslo_vmware.api [None req-44598345-8888-488a-9c24-527daf4d66df tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] Task: {'id': task-1110855, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 569.790169] env[63515]: DEBUG nova.network.neutron [None req-eaa7f189-a861-45f8-872f-29d870dc784b tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] [instance: e39bc55e-b677-4e2d-9aa4-7bc1d5ffadfa] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 569.799430] env[63515]: DEBUG nova.scheduler.client.report [None req-7f5043f7-e6d7-46dd-88fd-99ee049c2b4b tempest-VolumesAssistedSnapshotsTest-453349191 tempest-VolumesAssistedSnapshotsTest-453349191-project-member] Refreshing inventories for resource provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 569.822128] env[63515]: DEBUG nova.scheduler.client.report [None req-7f5043f7-e6d7-46dd-88fd-99ee049c2b4b tempest-VolumesAssistedSnapshotsTest-453349191 tempest-VolumesAssistedSnapshotsTest-453349191-project-member] Updating ProviderTree inventory for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 172, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 569.822361] env[63515]: DEBUG nova.compute.provider_tree [None req-7f5043f7-e6d7-46dd-88fd-99ee049c2b4b tempest-VolumesAssistedSnapshotsTest-453349191 tempest-VolumesAssistedSnapshotsTest-453349191-project-member] Updating inventory in ProviderTree for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 172, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 569.847981] env[63515]: DEBUG nova.scheduler.client.report [None req-7f5043f7-e6d7-46dd-88fd-99ee049c2b4b tempest-VolumesAssistedSnapshotsTest-453349191 tempest-VolumesAssistedSnapshotsTest-453349191-project-member] Refreshing aggregate associations for resource provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a, aggregates: None {{(pid=63515) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 569.870197] env[63515]: DEBUG nova.scheduler.client.report [None req-7f5043f7-e6d7-46dd-88fd-99ee049c2b4b tempest-VolumesAssistedSnapshotsTest-453349191 tempest-VolumesAssistedSnapshotsTest-453349191-project-member] Refreshing trait associations for resource provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a, traits: COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=63515) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 570.099176] env[63515]: INFO nova.scheduler.client.report [None req-ada3051f-c750-48c7-b591-0344c25b8fea tempest-ServersAdminTestJSON-710082096 tempest-ServersAdminTestJSON-710082096-project-member] Deleted allocations for instance d7fed938-f7cd-4013-b545-5d5350242084 [ 570.269620] env[63515]: DEBUG oslo_vmware.api [None req-44598345-8888-488a-9c24-527daf4d66df tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] Task: {'id': task-1110855, 'name': PowerOnVM_Task, 'duration_secs': 0.426054} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 570.269978] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-44598345-8888-488a-9c24-527daf4d66df tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] [instance: 93b28142-8454-43c0-b0a7-d61aa95c8fc8] Powered on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 570.270188] env[63515]: INFO nova.compute.manager [None req-44598345-8888-488a-9c24-527daf4d66df tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] [instance: 93b28142-8454-43c0-b0a7-d61aa95c8fc8] Took 8.14 seconds to spawn the instance on the hypervisor. [ 570.272250] env[63515]: DEBUG nova.compute.manager [None req-44598345-8888-488a-9c24-527daf4d66df tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] [instance: 93b28142-8454-43c0-b0a7-d61aa95c8fc8] Checking state {{(pid=63515) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 570.273016] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb9d79d6-44ce-4f5d-a34e-eae6b3b61b76 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 570.302831] env[63515]: DEBUG oslo_concurrency.lockutils [None req-eaa7f189-a861-45f8-872f-29d870dc784b tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] Releasing lock "refresh_cache-e39bc55e-b677-4e2d-9aa4-7bc1d5ffadfa" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 570.302831] env[63515]: DEBUG nova.compute.manager [None req-eaa7f189-a861-45f8-872f-29d870dc784b tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] [instance: e39bc55e-b677-4e2d-9aa4-7bc1d5ffadfa] Start destroying the instance on the hypervisor. {{(pid=63515) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 570.302831] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-eaa7f189-a861-45f8-872f-29d870dc784b tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] [instance: e39bc55e-b677-4e2d-9aa4-7bc1d5ffadfa] Destroying instance {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 570.302831] env[63515]: DEBUG oslo_concurrency.lockutils [req-70c40b80-8c24-4cb3-8d36-23f8abd79a05 req-bdea7a46-aaf5-4080-887f-01e6a8d62e41 service nova] Acquired lock "refresh_cache-e39bc55e-b677-4e2d-9aa4-7bc1d5ffadfa" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 570.302831] env[63515]: DEBUG nova.network.neutron [req-70c40b80-8c24-4cb3-8d36-23f8abd79a05 req-bdea7a46-aaf5-4080-887f-01e6a8d62e41 service nova] [instance: e39bc55e-b677-4e2d-9aa4-7bc1d5ffadfa] Refreshing network info cache for port ffab05fb-3a66-444a-9cca-115828ea7f22 {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 570.303221] env[63515]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e3c6248a-a2db-417c-b9e8-61adee9398fa {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 570.317141] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe36792d-f853-451d-b616-b5686db68982 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 570.346168] env[63515]: WARNING nova.virt.vmwareapi.vmops [None req-eaa7f189-a861-45f8-872f-29d870dc784b tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] [instance: e39bc55e-b677-4e2d-9aa4-7bc1d5ffadfa] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance e39bc55e-b677-4e2d-9aa4-7bc1d5ffadfa could not be found. [ 570.346493] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-eaa7f189-a861-45f8-872f-29d870dc784b tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] [instance: e39bc55e-b677-4e2d-9aa4-7bc1d5ffadfa] Instance destroyed {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 570.346601] env[63515]: INFO nova.compute.manager [None req-eaa7f189-a861-45f8-872f-29d870dc784b tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] [instance: e39bc55e-b677-4e2d-9aa4-7bc1d5ffadfa] Took 0.05 seconds to destroy the instance on the hypervisor. [ 570.347255] env[63515]: DEBUG oslo.service.loopingcall [None req-eaa7f189-a861-45f8-872f-29d870dc784b tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 570.347770] env[63515]: DEBUG nova.compute.manager [-] [instance: e39bc55e-b677-4e2d-9aa4-7bc1d5ffadfa] Deallocating network for instance {{(pid=63515) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 570.347770] env[63515]: DEBUG nova.network.neutron [-] [instance: e39bc55e-b677-4e2d-9aa4-7bc1d5ffadfa] deallocate_for_instance() {{(pid=63515) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 570.375066] env[63515]: DEBUG nova.network.neutron [-] [instance: e39bc55e-b677-4e2d-9aa4-7bc1d5ffadfa] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 570.409384] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a42f8c1e-4071-47ff-b154-10d4e9d74542 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 570.418801] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3feaa047-6174-47e7-aae9-6f999810badd {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 570.455898] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-272ac84e-f3af-4aca-8302-6715c8138736 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 570.463773] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb33401c-4b3f-40dd-9312-2472470af40d {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 570.482728] env[63515]: DEBUG nova.compute.provider_tree [None req-7f5043f7-e6d7-46dd-88fd-99ee049c2b4b tempest-VolumesAssistedSnapshotsTest-453349191 tempest-VolumesAssistedSnapshotsTest-453349191-project-member] Updating inventory in ProviderTree for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 570.619852] env[63515]: DEBUG oslo_concurrency.lockutils [None req-ada3051f-c750-48c7-b591-0344c25b8fea tempest-ServersAdminTestJSON-710082096 tempest-ServersAdminTestJSON-710082096-project-member] Lock "d7fed938-f7cd-4013-b545-5d5350242084" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 35.195s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 570.803618] env[63515]: INFO nova.compute.manager [None req-44598345-8888-488a-9c24-527daf4d66df tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] [instance: 93b28142-8454-43c0-b0a7-d61aa95c8fc8] Took 28.77 seconds to build instance. [ 570.836538] env[63515]: DEBUG nova.network.neutron [req-70c40b80-8c24-4cb3-8d36-23f8abd79a05 req-bdea7a46-aaf5-4080-887f-01e6a8d62e41 service nova] [instance: e39bc55e-b677-4e2d-9aa4-7bc1d5ffadfa] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 570.879223] env[63515]: DEBUG nova.network.neutron [-] [instance: e39bc55e-b677-4e2d-9aa4-7bc1d5ffadfa] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 571.029957] env[63515]: DEBUG nova.scheduler.client.report [None req-7f5043f7-e6d7-46dd-88fd-99ee049c2b4b tempest-VolumesAssistedSnapshotsTest-453349191 tempest-VolumesAssistedSnapshotsTest-453349191-project-member] Updated inventory for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a with generation 32 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 571.030368] env[63515]: DEBUG nova.compute.provider_tree [None req-7f5043f7-e6d7-46dd-88fd-99ee049c2b4b tempest-VolumesAssistedSnapshotsTest-453349191 tempest-VolumesAssistedSnapshotsTest-453349191-project-member] Updating resource provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a generation from 32 to 33 during operation: update_inventory {{(pid=63515) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 571.030661] env[63515]: DEBUG nova.compute.provider_tree [None req-7f5043f7-e6d7-46dd-88fd-99ee049c2b4b tempest-VolumesAssistedSnapshotsTest-453349191 tempest-VolumesAssistedSnapshotsTest-453349191-project-member] Updating inventory in ProviderTree for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 571.121844] env[63515]: DEBUG nova.compute.manager [None req-c80cd50f-4127-4451-9298-83a67d8f0ab6 tempest-ServersTestFqdnHostnames-1452522528 tempest-ServersTestFqdnHostnames-1452522528-project-member] [instance: b84c6c8d-e358-4a14-8425-c4f1d8adf4ee] Starting instance... {{(pid=63515) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 571.131061] env[63515]: DEBUG nova.network.neutron [req-70c40b80-8c24-4cb3-8d36-23f8abd79a05 req-bdea7a46-aaf5-4080-887f-01e6a8d62e41 service nova] [instance: e39bc55e-b677-4e2d-9aa4-7bc1d5ffadfa] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 571.312207] env[63515]: DEBUG oslo_concurrency.lockutils [None req-44598345-8888-488a-9c24-527daf4d66df tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] Lock "93b28142-8454-43c0-b0a7-d61aa95c8fc8" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 30.287s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 571.380841] env[63515]: INFO nova.compute.manager [-] [instance: e39bc55e-b677-4e2d-9aa4-7bc1d5ffadfa] Took 1.03 seconds to deallocate network for instance. [ 571.390472] env[63515]: DEBUG nova.compute.claims [None req-eaa7f189-a861-45f8-872f-29d870dc784b tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] [instance: e39bc55e-b677-4e2d-9aa4-7bc1d5ffadfa] Aborting claim: {{(pid=63515) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 571.390690] env[63515]: DEBUG oslo_concurrency.lockutils [None req-eaa7f189-a861-45f8-872f-29d870dc784b tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 571.536184] env[63515]: DEBUG oslo_concurrency.lockutils [None req-7f5043f7-e6d7-46dd-88fd-99ee049c2b4b tempest-VolumesAssistedSnapshotsTest-453349191 tempest-VolumesAssistedSnapshotsTest-453349191-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 4.935s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 571.536444] env[63515]: DEBUG nova.compute.manager [None req-7f5043f7-e6d7-46dd-88fd-99ee049c2b4b tempest-VolumesAssistedSnapshotsTest-453349191 tempest-VolumesAssistedSnapshotsTest-453349191-project-member] [instance: db682708-36fa-4126-a848-bfb609df96d7] Start building networks asynchronously for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 571.543323] env[63515]: DEBUG oslo_concurrency.lockutils [None req-222754a5-7546-43c7-a99a-1c936b9e5035 tempest-ImagesOneServerTestJSON-1950372435 tempest-ImagesOneServerTestJSON-1950372435-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 17.623s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 571.631343] env[63515]: DEBUG oslo_concurrency.lockutils [None req-480b7d6d-2f7f-4da3-9eb1-818f0711a940 tempest-ServerRescueTestJSON-1961208277 tempest-ServerRescueTestJSON-1961208277-project-member] Acquiring lock "7f49d79e-3bda-4949-a976-7e3e6513b2c7" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 571.631343] env[63515]: DEBUG oslo_concurrency.lockutils [None req-480b7d6d-2f7f-4da3-9eb1-818f0711a940 tempest-ServerRescueTestJSON-1961208277 tempest-ServerRescueTestJSON-1961208277-project-member] Lock "7f49d79e-3bda-4949-a976-7e3e6513b2c7" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 571.633540] env[63515]: DEBUG oslo_concurrency.lockutils [req-70c40b80-8c24-4cb3-8d36-23f8abd79a05 req-bdea7a46-aaf5-4080-887f-01e6a8d62e41 service nova] Releasing lock "refresh_cache-e39bc55e-b677-4e2d-9aa4-7bc1d5ffadfa" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 571.652931] env[63515]: DEBUG oslo_concurrency.lockutils [None req-c80cd50f-4127-4451-9298-83a67d8f0ab6 tempest-ServersTestFqdnHostnames-1452522528 tempest-ServersTestFqdnHostnames-1452522528-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 571.793599] env[63515]: DEBUG nova.compute.manager [req-398848a2-67fa-47b3-8981-8430b6af4060 req-430eea61-e3a3-4bf1-9a6c-1f69ed21751c service nova] [instance: e39bc55e-b677-4e2d-9aa4-7bc1d5ffadfa] Received event network-vif-deleted-ffab05fb-3a66-444a-9cca-115828ea7f22 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 571.820342] env[63515]: DEBUG nova.compute.manager [None req-22f02712-96d1-4031-8b8f-5bc880f34c33 tempest-ImagesNegativeTestJSON-2041328384 tempest-ImagesNegativeTestJSON-2041328384-project-member] [instance: d5f988f2-765b-4ecf-bad5-4a70fc95d74c] Starting instance... {{(pid=63515) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 572.051527] env[63515]: DEBUG nova.compute.utils [None req-7f5043f7-e6d7-46dd-88fd-99ee049c2b4b tempest-VolumesAssistedSnapshotsTest-453349191 tempest-VolumesAssistedSnapshotsTest-453349191-project-member] Using /dev/sd instead of None {{(pid=63515) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 572.058297] env[63515]: DEBUG nova.compute.manager [None req-7f5043f7-e6d7-46dd-88fd-99ee049c2b4b tempest-VolumesAssistedSnapshotsTest-453349191 tempest-VolumesAssistedSnapshotsTest-453349191-project-member] [instance: db682708-36fa-4126-a848-bfb609df96d7] Allocating IP information in the background. {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 572.058363] env[63515]: DEBUG nova.network.neutron [None req-7f5043f7-e6d7-46dd-88fd-99ee049c2b4b tempest-VolumesAssistedSnapshotsTest-453349191 tempest-VolumesAssistedSnapshotsTest-453349191-project-member] [instance: db682708-36fa-4126-a848-bfb609df96d7] allocate_for_instance() {{(pid=63515) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 572.123975] env[63515]: DEBUG nova.policy [None req-7f5043f7-e6d7-46dd-88fd-99ee049c2b4b tempest-VolumesAssistedSnapshotsTest-453349191 tempest-VolumesAssistedSnapshotsTest-453349191-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '3f076858b31c46aab0b53821eaa4372c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0fa008326cd343b98048b8b874e57b35', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63515) authorize /opt/stack/nova/nova/policy.py:201}} [ 572.343465] env[63515]: DEBUG oslo_concurrency.lockutils [None req-22f02712-96d1-4031-8b8f-5bc880f34c33 tempest-ImagesNegativeTestJSON-2041328384 tempest-ImagesNegativeTestJSON-2041328384-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 572.416786] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23c0abd3-98ff-4b15-94fc-5d58fbaaff74 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 572.425037] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c22834d3-29da-4092-9ddd-1d5d940f4289 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 572.458137] env[63515]: INFO nova.compute.manager [None req-b1f9566f-fcad-4d5d-9d53-b04b591d6c49 tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] [instance: 93b28142-8454-43c0-b0a7-d61aa95c8fc8] Rebuilding instance [ 572.461465] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad0f8440-1c6e-4518-bfe0-ccca23016249 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 572.470716] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a671b233-fada-4f66-9dd6-38c1475f0240 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 572.487371] env[63515]: DEBUG nova.compute.provider_tree [None req-222754a5-7546-43c7-a99a-1c936b9e5035 tempest-ImagesOneServerTestJSON-1950372435 tempest-ImagesOneServerTestJSON-1950372435-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 572.520157] env[63515]: DEBUG nova.compute.manager [None req-b1f9566f-fcad-4d5d-9d53-b04b591d6c49 tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] [instance: 93b28142-8454-43c0-b0a7-d61aa95c8fc8] Checking state {{(pid=63515) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 572.521179] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9b76ccd-f5fc-4256-9fac-25b50ad819a4 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 572.559367] env[63515]: DEBUG nova.compute.manager [None req-7f5043f7-e6d7-46dd-88fd-99ee049c2b4b tempest-VolumesAssistedSnapshotsTest-453349191 tempest-VolumesAssistedSnapshotsTest-453349191-project-member] [instance: db682708-36fa-4126-a848-bfb609df96d7] Start building block device mappings for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 572.968660] env[63515]: DEBUG nova.network.neutron [None req-7f5043f7-e6d7-46dd-88fd-99ee049c2b4b tempest-VolumesAssistedSnapshotsTest-453349191 tempest-VolumesAssistedSnapshotsTest-453349191-project-member] [instance: db682708-36fa-4126-a848-bfb609df96d7] Successfully created port: ee30f5c5-4733-4b41-ad1e-b45f0ae502cb {{(pid=63515) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 572.993495] env[63515]: DEBUG nova.scheduler.client.report [None req-222754a5-7546-43c7-a99a-1c936b9e5035 tempest-ImagesOneServerTestJSON-1950372435 tempest-ImagesOneServerTestJSON-1950372435-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 573.038538] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-b1f9566f-fcad-4d5d-9d53-b04b591d6c49 tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] [instance: 93b28142-8454-43c0-b0a7-d61aa95c8fc8] Powering off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 573.039158] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-20765c23-c3d2-445e-afcf-086dc5279d6a {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 573.048118] env[63515]: DEBUG oslo_vmware.api [None req-b1f9566f-fcad-4d5d-9d53-b04b591d6c49 tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] Waiting for the task: (returnval){ [ 573.048118] env[63515]: value = "task-1110860" [ 573.048118] env[63515]: _type = "Task" [ 573.048118] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 573.060505] env[63515]: DEBUG oslo_vmware.api [None req-b1f9566f-fcad-4d5d-9d53-b04b591d6c49 tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] Task: {'id': task-1110860, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 573.166170] env[63515]: DEBUG oslo_concurrency.lockutils [None req-b9c961a0-a559-4690-abb6-579359ed447c tempest-ServerPasswordTestJSON-433685770 tempest-ServerPasswordTestJSON-433685770-project-member] Acquiring lock "07499601-62bc-4c31-b295-23f34a6e2e91" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 573.166170] env[63515]: DEBUG oslo_concurrency.lockutils [None req-b9c961a0-a559-4690-abb6-579359ed447c tempest-ServerPasswordTestJSON-433685770 tempest-ServerPasswordTestJSON-433685770-project-member] Lock "07499601-62bc-4c31-b295-23f34a6e2e91" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 573.503939] env[63515]: DEBUG oslo_concurrency.lockutils [None req-222754a5-7546-43c7-a99a-1c936b9e5035 tempest-ImagesOneServerTestJSON-1950372435 tempest-ImagesOneServerTestJSON-1950372435-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.963s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 573.505332] env[63515]: ERROR nova.compute.manager [None req-222754a5-7546-43c7-a99a-1c936b9e5035 tempest-ImagesOneServerTestJSON-1950372435 tempest-ImagesOneServerTestJSON-1950372435-project-member] [instance: aec66939-7ebc-4bfc-bd8b-bae22e011239] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port b5d340e0-18c9-43c1-bcc4-7b682823c4c4, please check neutron logs for more information. [ 573.505332] env[63515]: ERROR nova.compute.manager [instance: aec66939-7ebc-4bfc-bd8b-bae22e011239] Traceback (most recent call last): [ 573.505332] env[63515]: ERROR nova.compute.manager [instance: aec66939-7ebc-4bfc-bd8b-bae22e011239] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 573.505332] env[63515]: ERROR nova.compute.manager [instance: aec66939-7ebc-4bfc-bd8b-bae22e011239] self.driver.spawn(context, instance, image_meta, [ 573.505332] env[63515]: ERROR nova.compute.manager [instance: aec66939-7ebc-4bfc-bd8b-bae22e011239] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 573.505332] env[63515]: ERROR nova.compute.manager [instance: aec66939-7ebc-4bfc-bd8b-bae22e011239] self._vmops.spawn(context, instance, image_meta, injected_files, [ 573.505332] env[63515]: ERROR nova.compute.manager [instance: aec66939-7ebc-4bfc-bd8b-bae22e011239] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 573.505332] env[63515]: ERROR nova.compute.manager [instance: aec66939-7ebc-4bfc-bd8b-bae22e011239] vm_ref = self.build_virtual_machine(instance, [ 573.505332] env[63515]: ERROR nova.compute.manager [instance: aec66939-7ebc-4bfc-bd8b-bae22e011239] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 573.505332] env[63515]: ERROR nova.compute.manager [instance: aec66939-7ebc-4bfc-bd8b-bae22e011239] vif_infos = vmwarevif.get_vif_info(self._session, [ 573.505332] env[63515]: ERROR nova.compute.manager [instance: aec66939-7ebc-4bfc-bd8b-bae22e011239] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 573.505938] env[63515]: ERROR nova.compute.manager [instance: aec66939-7ebc-4bfc-bd8b-bae22e011239] for vif in network_info: [ 573.505938] env[63515]: ERROR nova.compute.manager [instance: aec66939-7ebc-4bfc-bd8b-bae22e011239] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 573.505938] env[63515]: ERROR nova.compute.manager [instance: aec66939-7ebc-4bfc-bd8b-bae22e011239] return self._sync_wrapper(fn, *args, **kwargs) [ 573.505938] env[63515]: ERROR nova.compute.manager [instance: aec66939-7ebc-4bfc-bd8b-bae22e011239] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 573.505938] env[63515]: ERROR nova.compute.manager [instance: aec66939-7ebc-4bfc-bd8b-bae22e011239] self.wait() [ 573.505938] env[63515]: ERROR nova.compute.manager [instance: aec66939-7ebc-4bfc-bd8b-bae22e011239] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 573.505938] env[63515]: ERROR nova.compute.manager [instance: aec66939-7ebc-4bfc-bd8b-bae22e011239] self[:] = self._gt.wait() [ 573.505938] env[63515]: ERROR nova.compute.manager [instance: aec66939-7ebc-4bfc-bd8b-bae22e011239] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 573.505938] env[63515]: ERROR nova.compute.manager [instance: aec66939-7ebc-4bfc-bd8b-bae22e011239] return self._exit_event.wait() [ 573.505938] env[63515]: ERROR nova.compute.manager [instance: aec66939-7ebc-4bfc-bd8b-bae22e011239] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 573.505938] env[63515]: ERROR nova.compute.manager [instance: aec66939-7ebc-4bfc-bd8b-bae22e011239] result = hub.switch() [ 573.505938] env[63515]: ERROR nova.compute.manager [instance: aec66939-7ebc-4bfc-bd8b-bae22e011239] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 573.505938] env[63515]: ERROR nova.compute.manager [instance: aec66939-7ebc-4bfc-bd8b-bae22e011239] return self.greenlet.switch() [ 573.506473] env[63515]: ERROR nova.compute.manager [instance: aec66939-7ebc-4bfc-bd8b-bae22e011239] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 573.506473] env[63515]: ERROR nova.compute.manager [instance: aec66939-7ebc-4bfc-bd8b-bae22e011239] result = function(*args, **kwargs) [ 573.506473] env[63515]: ERROR nova.compute.manager [instance: aec66939-7ebc-4bfc-bd8b-bae22e011239] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 573.506473] env[63515]: ERROR nova.compute.manager [instance: aec66939-7ebc-4bfc-bd8b-bae22e011239] return func(*args, **kwargs) [ 573.506473] env[63515]: ERROR nova.compute.manager [instance: aec66939-7ebc-4bfc-bd8b-bae22e011239] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 573.506473] env[63515]: ERROR nova.compute.manager [instance: aec66939-7ebc-4bfc-bd8b-bae22e011239] raise e [ 573.506473] env[63515]: ERROR nova.compute.manager [instance: aec66939-7ebc-4bfc-bd8b-bae22e011239] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 573.506473] env[63515]: ERROR nova.compute.manager [instance: aec66939-7ebc-4bfc-bd8b-bae22e011239] nwinfo = self.network_api.allocate_for_instance( [ 573.506473] env[63515]: ERROR nova.compute.manager [instance: aec66939-7ebc-4bfc-bd8b-bae22e011239] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 573.506473] env[63515]: ERROR nova.compute.manager [instance: aec66939-7ebc-4bfc-bd8b-bae22e011239] created_port_ids = self._update_ports_for_instance( [ 573.506473] env[63515]: ERROR nova.compute.manager [instance: aec66939-7ebc-4bfc-bd8b-bae22e011239] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 573.506473] env[63515]: ERROR nova.compute.manager [instance: aec66939-7ebc-4bfc-bd8b-bae22e011239] with excutils.save_and_reraise_exception(): [ 573.506473] env[63515]: ERROR nova.compute.manager [instance: aec66939-7ebc-4bfc-bd8b-bae22e011239] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 573.506852] env[63515]: ERROR nova.compute.manager [instance: aec66939-7ebc-4bfc-bd8b-bae22e011239] self.force_reraise() [ 573.506852] env[63515]: ERROR nova.compute.manager [instance: aec66939-7ebc-4bfc-bd8b-bae22e011239] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 573.506852] env[63515]: ERROR nova.compute.manager [instance: aec66939-7ebc-4bfc-bd8b-bae22e011239] raise self.value [ 573.506852] env[63515]: ERROR nova.compute.manager [instance: aec66939-7ebc-4bfc-bd8b-bae22e011239] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 573.506852] env[63515]: ERROR nova.compute.manager [instance: aec66939-7ebc-4bfc-bd8b-bae22e011239] updated_port = self._update_port( [ 573.506852] env[63515]: ERROR nova.compute.manager [instance: aec66939-7ebc-4bfc-bd8b-bae22e011239] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 573.506852] env[63515]: ERROR nova.compute.manager [instance: aec66939-7ebc-4bfc-bd8b-bae22e011239] _ensure_no_port_binding_failure(port) [ 573.506852] env[63515]: ERROR nova.compute.manager [instance: aec66939-7ebc-4bfc-bd8b-bae22e011239] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 573.506852] env[63515]: ERROR nova.compute.manager [instance: aec66939-7ebc-4bfc-bd8b-bae22e011239] raise exception.PortBindingFailed(port_id=port['id']) [ 573.506852] env[63515]: ERROR nova.compute.manager [instance: aec66939-7ebc-4bfc-bd8b-bae22e011239] nova.exception.PortBindingFailed: Binding failed for port b5d340e0-18c9-43c1-bcc4-7b682823c4c4, please check neutron logs for more information. [ 573.506852] env[63515]: ERROR nova.compute.manager [instance: aec66939-7ebc-4bfc-bd8b-bae22e011239] [ 573.507244] env[63515]: DEBUG nova.compute.utils [None req-222754a5-7546-43c7-a99a-1c936b9e5035 tempest-ImagesOneServerTestJSON-1950372435 tempest-ImagesOneServerTestJSON-1950372435-project-member] [instance: aec66939-7ebc-4bfc-bd8b-bae22e011239] Binding failed for port b5d340e0-18c9-43c1-bcc4-7b682823c4c4, please check neutron logs for more information. {{(pid=63515) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 573.507278] env[63515]: DEBUG oslo_concurrency.lockutils [None req-5e279136-7e72-43a2-85c4-9cd11158a72b tempest-ListImageFiltersTestJSON-1100101045 tempest-ListImageFiltersTestJSON-1100101045-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.655s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 573.510215] env[63515]: INFO nova.compute.claims [None req-5e279136-7e72-43a2-85c4-9cd11158a72b tempest-ListImageFiltersTestJSON-1100101045 tempest-ListImageFiltersTestJSON-1100101045-project-member] [instance: 48908cb0-a989-4bad-84a4-25a4bdd3baeb] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 573.511796] env[63515]: DEBUG nova.compute.manager [None req-222754a5-7546-43c7-a99a-1c936b9e5035 tempest-ImagesOneServerTestJSON-1950372435 tempest-ImagesOneServerTestJSON-1950372435-project-member] [instance: aec66939-7ebc-4bfc-bd8b-bae22e011239] Build of instance aec66939-7ebc-4bfc-bd8b-bae22e011239 was re-scheduled: Binding failed for port b5d340e0-18c9-43c1-bcc4-7b682823c4c4, please check neutron logs for more information. {{(pid=63515) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 573.512128] env[63515]: DEBUG nova.compute.manager [None req-222754a5-7546-43c7-a99a-1c936b9e5035 tempest-ImagesOneServerTestJSON-1950372435 tempest-ImagesOneServerTestJSON-1950372435-project-member] [instance: aec66939-7ebc-4bfc-bd8b-bae22e011239] Unplugging VIFs for instance {{(pid=63515) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 573.512128] env[63515]: DEBUG oslo_concurrency.lockutils [None req-222754a5-7546-43c7-a99a-1c936b9e5035 tempest-ImagesOneServerTestJSON-1950372435 tempest-ImagesOneServerTestJSON-1950372435-project-member] Acquiring lock "refresh_cache-aec66939-7ebc-4bfc-bd8b-bae22e011239" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 573.513082] env[63515]: DEBUG oslo_concurrency.lockutils [None req-222754a5-7546-43c7-a99a-1c936b9e5035 tempest-ImagesOneServerTestJSON-1950372435 tempest-ImagesOneServerTestJSON-1950372435-project-member] Acquired lock "refresh_cache-aec66939-7ebc-4bfc-bd8b-bae22e011239" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 573.513082] env[63515]: DEBUG nova.network.neutron [None req-222754a5-7546-43c7-a99a-1c936b9e5035 tempest-ImagesOneServerTestJSON-1950372435 tempest-ImagesOneServerTestJSON-1950372435-project-member] [instance: aec66939-7ebc-4bfc-bd8b-bae22e011239] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 573.559560] env[63515]: DEBUG oslo_vmware.api [None req-b1f9566f-fcad-4d5d-9d53-b04b591d6c49 tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] Task: {'id': task-1110860, 'name': PowerOffVM_Task, 'duration_secs': 0.153514} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 573.561088] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-b1f9566f-fcad-4d5d-9d53-b04b591d6c49 tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] [instance: 93b28142-8454-43c0-b0a7-d61aa95c8fc8] Powered off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 573.561088] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-b1f9566f-fcad-4d5d-9d53-b04b591d6c49 tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] [instance: 93b28142-8454-43c0-b0a7-d61aa95c8fc8] Destroying instance {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 573.561445] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-228117b6-8083-4b23-9e2f-562695f9c23f {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 573.569924] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-b1f9566f-fcad-4d5d-9d53-b04b591d6c49 tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] [instance: 93b28142-8454-43c0-b0a7-d61aa95c8fc8] Unregistering the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 573.570233] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-29eeacfc-bdf6-4c92-aea7-97a6688feb96 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 573.573633] env[63515]: DEBUG nova.compute.manager [None req-7f5043f7-e6d7-46dd-88fd-99ee049c2b4b tempest-VolumesAssistedSnapshotsTest-453349191 tempest-VolumesAssistedSnapshotsTest-453349191-project-member] [instance: db682708-36fa-4126-a848-bfb609df96d7] Start spawning the instance on the hypervisor. {{(pid=63515) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 573.598607] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-b1f9566f-fcad-4d5d-9d53-b04b591d6c49 tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] [instance: 93b28142-8454-43c0-b0a7-d61aa95c8fc8] Unregistered the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 573.598869] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-b1f9566f-fcad-4d5d-9d53-b04b591d6c49 tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] [instance: 93b28142-8454-43c0-b0a7-d61aa95c8fc8] Deleting contents of the VM from datastore datastore1 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 573.599074] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-b1f9566f-fcad-4d5d-9d53-b04b591d6c49 tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] Deleting the datastore file [datastore1] 93b28142-8454-43c0-b0a7-d61aa95c8fc8 {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 573.599340] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4b7ed31f-6f3c-44e7-a531-9a20fe79cee4 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 573.607448] env[63515]: DEBUG nova.virt.hardware [None req-7f5043f7-e6d7-46dd-88fd-99ee049c2b4b tempest-VolumesAssistedSnapshotsTest-453349191 tempest-VolumesAssistedSnapshotsTest-453349191-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T02:52:14Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T02:51:56Z,direct_url=,disk_format='vmdk',id=8a120570-cb06-4099-b262-554ca0ad15c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b569255cc43e42e7ae3f2b4ad37c6ef4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T02:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 573.607608] env[63515]: DEBUG nova.virt.hardware [None req-7f5043f7-e6d7-46dd-88fd-99ee049c2b4b tempest-VolumesAssistedSnapshotsTest-453349191 tempest-VolumesAssistedSnapshotsTest-453349191-project-member] Flavor limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 573.607608] env[63515]: DEBUG nova.virt.hardware [None req-7f5043f7-e6d7-46dd-88fd-99ee049c2b4b tempest-VolumesAssistedSnapshotsTest-453349191 tempest-VolumesAssistedSnapshotsTest-453349191-project-member] Image limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 573.607876] env[63515]: DEBUG nova.virt.hardware [None req-7f5043f7-e6d7-46dd-88fd-99ee049c2b4b tempest-VolumesAssistedSnapshotsTest-453349191 tempest-VolumesAssistedSnapshotsTest-453349191-project-member] Flavor pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 573.607990] env[63515]: DEBUG nova.virt.hardware [None req-7f5043f7-e6d7-46dd-88fd-99ee049c2b4b tempest-VolumesAssistedSnapshotsTest-453349191 tempest-VolumesAssistedSnapshotsTest-453349191-project-member] Image pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 573.608086] env[63515]: DEBUG nova.virt.hardware [None req-7f5043f7-e6d7-46dd-88fd-99ee049c2b4b tempest-VolumesAssistedSnapshotsTest-453349191 tempest-VolumesAssistedSnapshotsTest-453349191-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 573.608293] env[63515]: DEBUG nova.virt.hardware [None req-7f5043f7-e6d7-46dd-88fd-99ee049c2b4b tempest-VolumesAssistedSnapshotsTest-453349191 tempest-VolumesAssistedSnapshotsTest-453349191-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 573.608446] env[63515]: DEBUG nova.virt.hardware [None req-7f5043f7-e6d7-46dd-88fd-99ee049c2b4b tempest-VolumesAssistedSnapshotsTest-453349191 tempest-VolumesAssistedSnapshotsTest-453349191-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 573.608809] env[63515]: DEBUG nova.virt.hardware [None req-7f5043f7-e6d7-46dd-88fd-99ee049c2b4b tempest-VolumesAssistedSnapshotsTest-453349191 tempest-VolumesAssistedSnapshotsTest-453349191-project-member] Got 1 possible topologies {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 573.608809] env[63515]: DEBUG nova.virt.hardware [None req-7f5043f7-e6d7-46dd-88fd-99ee049c2b4b tempest-VolumesAssistedSnapshotsTest-453349191 tempest-VolumesAssistedSnapshotsTest-453349191-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 573.608914] env[63515]: DEBUG nova.virt.hardware [None req-7f5043f7-e6d7-46dd-88fd-99ee049c2b4b tempest-VolumesAssistedSnapshotsTest-453349191 tempest-VolumesAssistedSnapshotsTest-453349191-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 573.611780] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e3556e0-c839-43ef-96f0-27287f05fc7c {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 573.620231] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d401a7a-e06b-4f71-9ab7-6bb6b863eb5a {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 573.625302] env[63515]: DEBUG oslo_vmware.api [None req-b1f9566f-fcad-4d5d-9d53-b04b591d6c49 tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] Waiting for the task: (returnval){ [ 573.625302] env[63515]: value = "task-1110862" [ 573.625302] env[63515]: _type = "Task" [ 573.625302] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 573.642273] env[63515]: DEBUG oslo_vmware.api [None req-b1f9566f-fcad-4d5d-9d53-b04b591d6c49 tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] Task: {'id': task-1110862, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 574.052046] env[63515]: DEBUG nova.network.neutron [None req-222754a5-7546-43c7-a99a-1c936b9e5035 tempest-ImagesOneServerTestJSON-1950372435 tempest-ImagesOneServerTestJSON-1950372435-project-member] [instance: aec66939-7ebc-4bfc-bd8b-bae22e011239] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 574.144104] env[63515]: DEBUG oslo_vmware.api [None req-b1f9566f-fcad-4d5d-9d53-b04b591d6c49 tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] Task: {'id': task-1110862, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.121399} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 574.144832] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-b1f9566f-fcad-4d5d-9d53-b04b591d6c49 tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] Deleted the datastore file {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 574.144832] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-b1f9566f-fcad-4d5d-9d53-b04b591d6c49 tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] [instance: 93b28142-8454-43c0-b0a7-d61aa95c8fc8] Deleted contents of the VM from datastore datastore1 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 574.144832] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-b1f9566f-fcad-4d5d-9d53-b04b591d6c49 tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] [instance: 93b28142-8454-43c0-b0a7-d61aa95c8fc8] Instance destroyed {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 574.274908] env[63515]: DEBUG nova.network.neutron [None req-222754a5-7546-43c7-a99a-1c936b9e5035 tempest-ImagesOneServerTestJSON-1950372435 tempest-ImagesOneServerTestJSON-1950372435-project-member] [instance: aec66939-7ebc-4bfc-bd8b-bae22e011239] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 574.781350] env[63515]: DEBUG oslo_concurrency.lockutils [None req-222754a5-7546-43c7-a99a-1c936b9e5035 tempest-ImagesOneServerTestJSON-1950372435 tempest-ImagesOneServerTestJSON-1950372435-project-member] Releasing lock "refresh_cache-aec66939-7ebc-4bfc-bd8b-bae22e011239" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 574.781582] env[63515]: DEBUG nova.compute.manager [None req-222754a5-7546-43c7-a99a-1c936b9e5035 tempest-ImagesOneServerTestJSON-1950372435 tempest-ImagesOneServerTestJSON-1950372435-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63515) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 574.781761] env[63515]: DEBUG nova.compute.manager [None req-222754a5-7546-43c7-a99a-1c936b9e5035 tempest-ImagesOneServerTestJSON-1950372435 tempest-ImagesOneServerTestJSON-1950372435-project-member] [instance: aec66939-7ebc-4bfc-bd8b-bae22e011239] Deallocating network for instance {{(pid=63515) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 574.781923] env[63515]: DEBUG nova.network.neutron [None req-222754a5-7546-43c7-a99a-1c936b9e5035 tempest-ImagesOneServerTestJSON-1950372435 tempest-ImagesOneServerTestJSON-1950372435-project-member] [instance: aec66939-7ebc-4bfc-bd8b-bae22e011239] deallocate_for_instance() {{(pid=63515) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 574.834982] env[63515]: DEBUG nova.network.neutron [None req-222754a5-7546-43c7-a99a-1c936b9e5035 tempest-ImagesOneServerTestJSON-1950372435 tempest-ImagesOneServerTestJSON-1950372435-project-member] [instance: aec66939-7ebc-4bfc-bd8b-bae22e011239] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 574.890680] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f20b76a-5094-49ae-b1c7-7384854af781 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 574.899975] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d03b660d-4f12-4710-b062-38b906832e23 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 574.933844] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d7b16d2-e687-438b-aaa7-41bcad49a168 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 574.941798] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0d91737-6264-4f6d-b932-35be17d6f1f9 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 574.957610] env[63515]: DEBUG nova.compute.provider_tree [None req-5e279136-7e72-43a2-85c4-9cd11158a72b tempest-ListImageFiltersTestJSON-1100101045 tempest-ListImageFiltersTestJSON-1100101045-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 575.007738] env[63515]: DEBUG nova.compute.manager [req-065afdda-20bd-4e5b-b5d5-7bc8aa98573c req-66cf8c43-7a24-4fd9-b548-d7386bf078e9 service nova] [instance: db682708-36fa-4126-a848-bfb609df96d7] Received event network-changed-ee30f5c5-4733-4b41-ad1e-b45f0ae502cb {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 575.007738] env[63515]: DEBUG nova.compute.manager [req-065afdda-20bd-4e5b-b5d5-7bc8aa98573c req-66cf8c43-7a24-4fd9-b548-d7386bf078e9 service nova] [instance: db682708-36fa-4126-a848-bfb609df96d7] Refreshing instance network info cache due to event network-changed-ee30f5c5-4733-4b41-ad1e-b45f0ae502cb. {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 575.007738] env[63515]: DEBUG oslo_concurrency.lockutils [req-065afdda-20bd-4e5b-b5d5-7bc8aa98573c req-66cf8c43-7a24-4fd9-b548-d7386bf078e9 service nova] Acquiring lock "refresh_cache-db682708-36fa-4126-a848-bfb609df96d7" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 575.007738] env[63515]: DEBUG oslo_concurrency.lockutils [req-065afdda-20bd-4e5b-b5d5-7bc8aa98573c req-66cf8c43-7a24-4fd9-b548-d7386bf078e9 service nova] Acquired lock "refresh_cache-db682708-36fa-4126-a848-bfb609df96d7" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 575.007738] env[63515]: DEBUG nova.network.neutron [req-065afdda-20bd-4e5b-b5d5-7bc8aa98573c req-66cf8c43-7a24-4fd9-b548-d7386bf078e9 service nova] [instance: db682708-36fa-4126-a848-bfb609df96d7] Refreshing network info cache for port ee30f5c5-4733-4b41-ad1e-b45f0ae502cb {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 575.185585] env[63515]: DEBUG nova.virt.hardware [None req-b1f9566f-fcad-4d5d-9d53-b04b591d6c49 tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T02:52:14Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T02:51:56Z,direct_url=,disk_format='vmdk',id=8a120570-cb06-4099-b262-554ca0ad15c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b569255cc43e42e7ae3f2b4ad37c6ef4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T02:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 575.185866] env[63515]: DEBUG nova.virt.hardware [None req-b1f9566f-fcad-4d5d-9d53-b04b591d6c49 tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] Flavor limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 575.186035] env[63515]: DEBUG nova.virt.hardware [None req-b1f9566f-fcad-4d5d-9d53-b04b591d6c49 tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] Image limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 575.186221] env[63515]: DEBUG nova.virt.hardware [None req-b1f9566f-fcad-4d5d-9d53-b04b591d6c49 tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] Flavor pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 575.186367] env[63515]: DEBUG nova.virt.hardware [None req-b1f9566f-fcad-4d5d-9d53-b04b591d6c49 tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] Image pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 575.188060] env[63515]: DEBUG nova.virt.hardware [None req-b1f9566f-fcad-4d5d-9d53-b04b591d6c49 tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 575.188060] env[63515]: DEBUG nova.virt.hardware [None req-b1f9566f-fcad-4d5d-9d53-b04b591d6c49 tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 575.188060] env[63515]: DEBUG nova.virt.hardware [None req-b1f9566f-fcad-4d5d-9d53-b04b591d6c49 tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 575.188060] env[63515]: DEBUG nova.virt.hardware [None req-b1f9566f-fcad-4d5d-9d53-b04b591d6c49 tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] Got 1 possible topologies {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 575.188060] env[63515]: DEBUG nova.virt.hardware [None req-b1f9566f-fcad-4d5d-9d53-b04b591d6c49 tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 575.188263] env[63515]: DEBUG nova.virt.hardware [None req-b1f9566f-fcad-4d5d-9d53-b04b591d6c49 tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 575.188263] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b4b829a-b2e6-44f6-808b-8c479c870027 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 575.197584] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62c6d0c5-5c74-4348-a3e3-3e125e1003fd {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 575.211491] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-b1f9566f-fcad-4d5d-9d53-b04b591d6c49 tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] [instance: 93b28142-8454-43c0-b0a7-d61aa95c8fc8] Instance VIF info [] {{(pid=63515) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 575.218702] env[63515]: DEBUG oslo.service.loopingcall [None req-b1f9566f-fcad-4d5d-9d53-b04b591d6c49 tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 575.218702] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 93b28142-8454-43c0-b0a7-d61aa95c8fc8] Creating VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 575.218702] env[63515]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-443108de-8e67-4575-99d1-8f7d37a75197 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 575.236817] env[63515]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 575.236817] env[63515]: value = "task-1110864" [ 575.236817] env[63515]: _type = "Task" [ 575.236817] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 575.244869] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1110864, 'name': CreateVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 575.331967] env[63515]: ERROR nova.compute.manager [None req-7f5043f7-e6d7-46dd-88fd-99ee049c2b4b tempest-VolumesAssistedSnapshotsTest-453349191 tempest-VolumesAssistedSnapshotsTest-453349191-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port ee30f5c5-4733-4b41-ad1e-b45f0ae502cb, please check neutron logs for more information. [ 575.331967] env[63515]: ERROR nova.compute.manager Traceback (most recent call last): [ 575.331967] env[63515]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 575.331967] env[63515]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 575.331967] env[63515]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 575.331967] env[63515]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 575.331967] env[63515]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 575.331967] env[63515]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 575.331967] env[63515]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 575.331967] env[63515]: ERROR nova.compute.manager self.force_reraise() [ 575.331967] env[63515]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 575.331967] env[63515]: ERROR nova.compute.manager raise self.value [ 575.331967] env[63515]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 575.331967] env[63515]: ERROR nova.compute.manager updated_port = self._update_port( [ 575.331967] env[63515]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 575.331967] env[63515]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 575.332689] env[63515]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 575.332689] env[63515]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 575.332689] env[63515]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port ee30f5c5-4733-4b41-ad1e-b45f0ae502cb, please check neutron logs for more information. [ 575.332689] env[63515]: ERROR nova.compute.manager [ 575.332689] env[63515]: Traceback (most recent call last): [ 575.332689] env[63515]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 575.332689] env[63515]: listener.cb(fileno) [ 575.332689] env[63515]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 575.332689] env[63515]: result = function(*args, **kwargs) [ 575.332689] env[63515]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 575.332689] env[63515]: return func(*args, **kwargs) [ 575.332689] env[63515]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 575.332689] env[63515]: raise e [ 575.332689] env[63515]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 575.332689] env[63515]: nwinfo = self.network_api.allocate_for_instance( [ 575.332689] env[63515]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 575.332689] env[63515]: created_port_ids = self._update_ports_for_instance( [ 575.332689] env[63515]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 575.332689] env[63515]: with excutils.save_and_reraise_exception(): [ 575.332689] env[63515]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 575.332689] env[63515]: self.force_reraise() [ 575.332689] env[63515]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 575.332689] env[63515]: raise self.value [ 575.332689] env[63515]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 575.332689] env[63515]: updated_port = self._update_port( [ 575.332689] env[63515]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 575.332689] env[63515]: _ensure_no_port_binding_failure(port) [ 575.332689] env[63515]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 575.332689] env[63515]: raise exception.PortBindingFailed(port_id=port['id']) [ 575.333689] env[63515]: nova.exception.PortBindingFailed: Binding failed for port ee30f5c5-4733-4b41-ad1e-b45f0ae502cb, please check neutron logs for more information. [ 575.333689] env[63515]: Removing descriptor: 16 [ 575.333689] env[63515]: ERROR nova.compute.manager [None req-7f5043f7-e6d7-46dd-88fd-99ee049c2b4b tempest-VolumesAssistedSnapshotsTest-453349191 tempest-VolumesAssistedSnapshotsTest-453349191-project-member] [instance: db682708-36fa-4126-a848-bfb609df96d7] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port ee30f5c5-4733-4b41-ad1e-b45f0ae502cb, please check neutron logs for more information. [ 575.333689] env[63515]: ERROR nova.compute.manager [instance: db682708-36fa-4126-a848-bfb609df96d7] Traceback (most recent call last): [ 575.333689] env[63515]: ERROR nova.compute.manager [instance: db682708-36fa-4126-a848-bfb609df96d7] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 575.333689] env[63515]: ERROR nova.compute.manager [instance: db682708-36fa-4126-a848-bfb609df96d7] yield resources [ 575.333689] env[63515]: ERROR nova.compute.manager [instance: db682708-36fa-4126-a848-bfb609df96d7] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 575.333689] env[63515]: ERROR nova.compute.manager [instance: db682708-36fa-4126-a848-bfb609df96d7] self.driver.spawn(context, instance, image_meta, [ 575.333689] env[63515]: ERROR nova.compute.manager [instance: db682708-36fa-4126-a848-bfb609df96d7] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 575.333689] env[63515]: ERROR nova.compute.manager [instance: db682708-36fa-4126-a848-bfb609df96d7] self._vmops.spawn(context, instance, image_meta, injected_files, [ 575.333689] env[63515]: ERROR nova.compute.manager [instance: db682708-36fa-4126-a848-bfb609df96d7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 575.333689] env[63515]: ERROR nova.compute.manager [instance: db682708-36fa-4126-a848-bfb609df96d7] vm_ref = self.build_virtual_machine(instance, [ 575.335639] env[63515]: ERROR nova.compute.manager [instance: db682708-36fa-4126-a848-bfb609df96d7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 575.335639] env[63515]: ERROR nova.compute.manager [instance: db682708-36fa-4126-a848-bfb609df96d7] vif_infos = vmwarevif.get_vif_info(self._session, [ 575.335639] env[63515]: ERROR nova.compute.manager [instance: db682708-36fa-4126-a848-bfb609df96d7] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 575.335639] env[63515]: ERROR nova.compute.manager [instance: db682708-36fa-4126-a848-bfb609df96d7] for vif in network_info: [ 575.335639] env[63515]: ERROR nova.compute.manager [instance: db682708-36fa-4126-a848-bfb609df96d7] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 575.335639] env[63515]: ERROR nova.compute.manager [instance: db682708-36fa-4126-a848-bfb609df96d7] return self._sync_wrapper(fn, *args, **kwargs) [ 575.335639] env[63515]: ERROR nova.compute.manager [instance: db682708-36fa-4126-a848-bfb609df96d7] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 575.335639] env[63515]: ERROR nova.compute.manager [instance: db682708-36fa-4126-a848-bfb609df96d7] self.wait() [ 575.335639] env[63515]: ERROR nova.compute.manager [instance: db682708-36fa-4126-a848-bfb609df96d7] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 575.335639] env[63515]: ERROR nova.compute.manager [instance: db682708-36fa-4126-a848-bfb609df96d7] self[:] = self._gt.wait() [ 575.335639] env[63515]: ERROR nova.compute.manager [instance: db682708-36fa-4126-a848-bfb609df96d7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 575.335639] env[63515]: ERROR nova.compute.manager [instance: db682708-36fa-4126-a848-bfb609df96d7] return self._exit_event.wait() [ 575.335639] env[63515]: ERROR nova.compute.manager [instance: db682708-36fa-4126-a848-bfb609df96d7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 575.336041] env[63515]: ERROR nova.compute.manager [instance: db682708-36fa-4126-a848-bfb609df96d7] result = hub.switch() [ 575.336041] env[63515]: ERROR nova.compute.manager [instance: db682708-36fa-4126-a848-bfb609df96d7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 575.336041] env[63515]: ERROR nova.compute.manager [instance: db682708-36fa-4126-a848-bfb609df96d7] return self.greenlet.switch() [ 575.336041] env[63515]: ERROR nova.compute.manager [instance: db682708-36fa-4126-a848-bfb609df96d7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 575.336041] env[63515]: ERROR nova.compute.manager [instance: db682708-36fa-4126-a848-bfb609df96d7] result = function(*args, **kwargs) [ 575.336041] env[63515]: ERROR nova.compute.manager [instance: db682708-36fa-4126-a848-bfb609df96d7] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 575.336041] env[63515]: ERROR nova.compute.manager [instance: db682708-36fa-4126-a848-bfb609df96d7] return func(*args, **kwargs) [ 575.336041] env[63515]: ERROR nova.compute.manager [instance: db682708-36fa-4126-a848-bfb609df96d7] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 575.336041] env[63515]: ERROR nova.compute.manager [instance: db682708-36fa-4126-a848-bfb609df96d7] raise e [ 575.336041] env[63515]: ERROR nova.compute.manager [instance: db682708-36fa-4126-a848-bfb609df96d7] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 575.336041] env[63515]: ERROR nova.compute.manager [instance: db682708-36fa-4126-a848-bfb609df96d7] nwinfo = self.network_api.allocate_for_instance( [ 575.336041] env[63515]: ERROR nova.compute.manager [instance: db682708-36fa-4126-a848-bfb609df96d7] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 575.336041] env[63515]: ERROR nova.compute.manager [instance: db682708-36fa-4126-a848-bfb609df96d7] created_port_ids = self._update_ports_for_instance( [ 575.336484] env[63515]: ERROR nova.compute.manager [instance: db682708-36fa-4126-a848-bfb609df96d7] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 575.336484] env[63515]: ERROR nova.compute.manager [instance: db682708-36fa-4126-a848-bfb609df96d7] with excutils.save_and_reraise_exception(): [ 575.336484] env[63515]: ERROR nova.compute.manager [instance: db682708-36fa-4126-a848-bfb609df96d7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 575.336484] env[63515]: ERROR nova.compute.manager [instance: db682708-36fa-4126-a848-bfb609df96d7] self.force_reraise() [ 575.336484] env[63515]: ERROR nova.compute.manager [instance: db682708-36fa-4126-a848-bfb609df96d7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 575.336484] env[63515]: ERROR nova.compute.manager [instance: db682708-36fa-4126-a848-bfb609df96d7] raise self.value [ 575.336484] env[63515]: ERROR nova.compute.manager [instance: db682708-36fa-4126-a848-bfb609df96d7] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 575.336484] env[63515]: ERROR nova.compute.manager [instance: db682708-36fa-4126-a848-bfb609df96d7] updated_port = self._update_port( [ 575.336484] env[63515]: ERROR nova.compute.manager [instance: db682708-36fa-4126-a848-bfb609df96d7] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 575.336484] env[63515]: ERROR nova.compute.manager [instance: db682708-36fa-4126-a848-bfb609df96d7] _ensure_no_port_binding_failure(port) [ 575.336484] env[63515]: ERROR nova.compute.manager [instance: db682708-36fa-4126-a848-bfb609df96d7] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 575.336484] env[63515]: ERROR nova.compute.manager [instance: db682708-36fa-4126-a848-bfb609df96d7] raise exception.PortBindingFailed(port_id=port['id']) [ 575.336839] env[63515]: ERROR nova.compute.manager [instance: db682708-36fa-4126-a848-bfb609df96d7] nova.exception.PortBindingFailed: Binding failed for port ee30f5c5-4733-4b41-ad1e-b45f0ae502cb, please check neutron logs for more information. [ 575.336839] env[63515]: ERROR nova.compute.manager [instance: db682708-36fa-4126-a848-bfb609df96d7] [ 575.336839] env[63515]: INFO nova.compute.manager [None req-7f5043f7-e6d7-46dd-88fd-99ee049c2b4b tempest-VolumesAssistedSnapshotsTest-453349191 tempest-VolumesAssistedSnapshotsTest-453349191-project-member] [instance: db682708-36fa-4126-a848-bfb609df96d7] Terminating instance [ 575.336839] env[63515]: DEBUG oslo_concurrency.lockutils [None req-7f5043f7-e6d7-46dd-88fd-99ee049c2b4b tempest-VolumesAssistedSnapshotsTest-453349191 tempest-VolumesAssistedSnapshotsTest-453349191-project-member] Acquiring lock "refresh_cache-db682708-36fa-4126-a848-bfb609df96d7" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 575.341376] env[63515]: DEBUG nova.network.neutron [None req-222754a5-7546-43c7-a99a-1c936b9e5035 tempest-ImagesOneServerTestJSON-1950372435 tempest-ImagesOneServerTestJSON-1950372435-project-member] [instance: aec66939-7ebc-4bfc-bd8b-bae22e011239] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 575.462655] env[63515]: DEBUG nova.scheduler.client.report [None req-5e279136-7e72-43a2-85c4-9cd11158a72b tempest-ListImageFiltersTestJSON-1100101045 tempest-ListImageFiltersTestJSON-1100101045-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 575.531421] env[63515]: DEBUG nova.network.neutron [req-065afdda-20bd-4e5b-b5d5-7bc8aa98573c req-66cf8c43-7a24-4fd9-b548-d7386bf078e9 service nova] [instance: db682708-36fa-4126-a848-bfb609df96d7] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 575.643140] env[63515]: DEBUG nova.network.neutron [req-065afdda-20bd-4e5b-b5d5-7bc8aa98573c req-66cf8c43-7a24-4fd9-b548-d7386bf078e9 service nova] [instance: db682708-36fa-4126-a848-bfb609df96d7] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 575.753227] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1110864, 'name': CreateVM_Task, 'duration_secs': 0.32603} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 575.753414] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 93b28142-8454-43c0-b0a7-d61aa95c8fc8] Created VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 575.753927] env[63515]: DEBUG oslo_concurrency.lockutils [None req-b1f9566f-fcad-4d5d-9d53-b04b591d6c49 tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 575.754051] env[63515]: DEBUG oslo_concurrency.lockutils [None req-b1f9566f-fcad-4d5d-9d53-b04b591d6c49 tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 575.754414] env[63515]: DEBUG oslo_concurrency.lockutils [None req-b1f9566f-fcad-4d5d-9d53-b04b591d6c49 tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 575.754666] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b4620cd9-c06c-4d94-8d43-5d2a4794dce6 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 575.761337] env[63515]: DEBUG oslo_vmware.api [None req-b1f9566f-fcad-4d5d-9d53-b04b591d6c49 tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] Waiting for the task: (returnval){ [ 575.761337] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]52d71ed3-ea5c-2eb4-c843-a15834e4c848" [ 575.761337] env[63515]: _type = "Task" [ 575.761337] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 575.775667] env[63515]: DEBUG oslo_vmware.api [None req-b1f9566f-fcad-4d5d-9d53-b04b591d6c49 tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52d71ed3-ea5c-2eb4-c843-a15834e4c848, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 575.845888] env[63515]: INFO nova.compute.manager [None req-222754a5-7546-43c7-a99a-1c936b9e5035 tempest-ImagesOneServerTestJSON-1950372435 tempest-ImagesOneServerTestJSON-1950372435-project-member] [instance: aec66939-7ebc-4bfc-bd8b-bae22e011239] Took 1.06 seconds to deallocate network for instance. [ 575.974157] env[63515]: DEBUG oslo_concurrency.lockutils [None req-5e279136-7e72-43a2-85c4-9cd11158a72b tempest-ListImageFiltersTestJSON-1100101045 tempest-ListImageFiltersTestJSON-1100101045-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.465s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 575.974157] env[63515]: DEBUG nova.compute.manager [None req-5e279136-7e72-43a2-85c4-9cd11158a72b tempest-ListImageFiltersTestJSON-1100101045 tempest-ListImageFiltersTestJSON-1100101045-project-member] [instance: 48908cb0-a989-4bad-84a4-25a4bdd3baeb] Start building networks asynchronously for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 575.975378] env[63515]: DEBUG oslo_concurrency.lockutils [None req-757a72d0-3dae-4296-830e-18ff5e948610 tempest-ServerDiagnosticsNegativeTest-1799249066 tempest-ServerDiagnosticsNegativeTest-1799249066-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 17.163s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 576.144156] env[63515]: DEBUG oslo_concurrency.lockutils [req-065afdda-20bd-4e5b-b5d5-7bc8aa98573c req-66cf8c43-7a24-4fd9-b548-d7386bf078e9 service nova] Releasing lock "refresh_cache-db682708-36fa-4126-a848-bfb609df96d7" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 576.144605] env[63515]: DEBUG oslo_concurrency.lockutils [None req-7f5043f7-e6d7-46dd-88fd-99ee049c2b4b tempest-VolumesAssistedSnapshotsTest-453349191 tempest-VolumesAssistedSnapshotsTest-453349191-project-member] Acquired lock "refresh_cache-db682708-36fa-4126-a848-bfb609df96d7" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 576.144790] env[63515]: DEBUG nova.network.neutron [None req-7f5043f7-e6d7-46dd-88fd-99ee049c2b4b tempest-VolumesAssistedSnapshotsTest-453349191 tempest-VolumesAssistedSnapshotsTest-453349191-project-member] [instance: db682708-36fa-4126-a848-bfb609df96d7] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 576.275557] env[63515]: DEBUG oslo_vmware.api [None req-b1f9566f-fcad-4d5d-9d53-b04b591d6c49 tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52d71ed3-ea5c-2eb4-c843-a15834e4c848, 'name': SearchDatastore_Task, 'duration_secs': 0.009846} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 576.275557] env[63515]: DEBUG oslo_concurrency.lockutils [None req-b1f9566f-fcad-4d5d-9d53-b04b591d6c49 tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 576.275557] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-b1f9566f-fcad-4d5d-9d53-b04b591d6c49 tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] [instance: 93b28142-8454-43c0-b0a7-d61aa95c8fc8] Processing image 8a120570-cb06-4099-b262-554ca0ad15c5 {{(pid=63515) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 576.275557] env[63515]: DEBUG oslo_concurrency.lockutils [None req-b1f9566f-fcad-4d5d-9d53-b04b591d6c49 tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 576.276031] env[63515]: DEBUG oslo_concurrency.lockutils [None req-b1f9566f-fcad-4d5d-9d53-b04b591d6c49 tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 576.276031] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-b1f9566f-fcad-4d5d-9d53-b04b591d6c49 tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 576.276031] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-13a098a0-07fa-462b-9be5-1beca621f718 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 576.284046] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-b1f9566f-fcad-4d5d-9d53-b04b591d6c49 tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 576.284046] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-b1f9566f-fcad-4d5d-9d53-b04b591d6c49 tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63515) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 576.285097] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-10180587-eeb4-4309-9b66-0e70dbb75d0f {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 576.291015] env[63515]: DEBUG oslo_vmware.api [None req-b1f9566f-fcad-4d5d-9d53-b04b591d6c49 tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] Waiting for the task: (returnval){ [ 576.291015] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]524a63ae-0f45-cb5c-8a9f-f4497527b15d" [ 576.291015] env[63515]: _type = "Task" [ 576.291015] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 576.301601] env[63515]: DEBUG oslo_vmware.api [None req-b1f9566f-fcad-4d5d-9d53-b04b591d6c49 tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]524a63ae-0f45-cb5c-8a9f-f4497527b15d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 576.479954] env[63515]: DEBUG nova.compute.utils [None req-5e279136-7e72-43a2-85c4-9cd11158a72b tempest-ListImageFiltersTestJSON-1100101045 tempest-ListImageFiltersTestJSON-1100101045-project-member] Using /dev/sd instead of None {{(pid=63515) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 576.481760] env[63515]: DEBUG nova.compute.manager [None req-5e279136-7e72-43a2-85c4-9cd11158a72b tempest-ListImageFiltersTestJSON-1100101045 tempest-ListImageFiltersTestJSON-1100101045-project-member] [instance: 48908cb0-a989-4bad-84a4-25a4bdd3baeb] Allocating IP information in the background. {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 576.481760] env[63515]: DEBUG nova.network.neutron [None req-5e279136-7e72-43a2-85c4-9cd11158a72b tempest-ListImageFiltersTestJSON-1100101045 tempest-ListImageFiltersTestJSON-1100101045-project-member] [instance: 48908cb0-a989-4bad-84a4-25a4bdd3baeb] allocate_for_instance() {{(pid=63515) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 576.692461] env[63515]: DEBUG nova.network.neutron [None req-7f5043f7-e6d7-46dd-88fd-99ee049c2b4b tempest-VolumesAssistedSnapshotsTest-453349191 tempest-VolumesAssistedSnapshotsTest-453349191-project-member] [instance: db682708-36fa-4126-a848-bfb609df96d7] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 576.739723] env[63515]: DEBUG nova.policy [None req-5e279136-7e72-43a2-85c4-9cd11158a72b tempest-ListImageFiltersTestJSON-1100101045 tempest-ListImageFiltersTestJSON-1100101045-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'fb6aad1c2ef9400fb9e7cfe4572e5cac', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '09ce0d978c1b4ecfb2b91fa1b1c20622', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63515) authorize /opt/stack/nova/nova/policy.py:201}} [ 576.805857] env[63515]: DEBUG oslo_vmware.api [None req-b1f9566f-fcad-4d5d-9d53-b04b591d6c49 tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]524a63ae-0f45-cb5c-8a9f-f4497527b15d, 'name': SearchDatastore_Task, 'duration_secs': 0.008338} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 576.809960] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8ea102a1-1122-47dc-abc3-0fda1497ce54 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 576.817073] env[63515]: DEBUG oslo_vmware.api [None req-b1f9566f-fcad-4d5d-9d53-b04b591d6c49 tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] Waiting for the task: (returnval){ [ 576.817073] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]52cba51e-5eeb-a8a1-842f-94a44e4835b0" [ 576.817073] env[63515]: _type = "Task" [ 576.817073] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 576.829881] env[63515]: DEBUG oslo_vmware.api [None req-b1f9566f-fcad-4d5d-9d53-b04b591d6c49 tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52cba51e-5eeb-a8a1-842f-94a44e4835b0, 'name': SearchDatastore_Task, 'duration_secs': 0.009172} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 576.830100] env[63515]: DEBUG oslo_concurrency.lockutils [None req-b1f9566f-fcad-4d5d-9d53-b04b591d6c49 tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 576.830355] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-b1f9566f-fcad-4d5d-9d53-b04b591d6c49 tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk to [datastore1] 93b28142-8454-43c0-b0a7-d61aa95c8fc8/93b28142-8454-43c0-b0a7-d61aa95c8fc8.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 576.830857] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-01dc8e3d-a3ee-45e8-b897-0f3b5aeb7816 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 576.838386] env[63515]: DEBUG oslo_vmware.api [None req-b1f9566f-fcad-4d5d-9d53-b04b591d6c49 tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] Waiting for the task: (returnval){ [ 576.838386] env[63515]: value = "task-1110866" [ 576.838386] env[63515]: _type = "Task" [ 576.838386] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 576.846392] env[63515]: DEBUG oslo_vmware.api [None req-b1f9566f-fcad-4d5d-9d53-b04b591d6c49 tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] Task: {'id': task-1110866, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 576.854322] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c3bd3c3-4e1f-40b6-a9e9-19ee219e15f0 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 576.859675] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b34ececf-2bb9-40ad-8841-67ddbf1770c8 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 576.904233] env[63515]: INFO nova.scheduler.client.report [None req-222754a5-7546-43c7-a99a-1c936b9e5035 tempest-ImagesOneServerTestJSON-1950372435 tempest-ImagesOneServerTestJSON-1950372435-project-member] Deleted allocations for instance aec66939-7ebc-4bfc-bd8b-bae22e011239 [ 576.911793] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a048a70-b60a-424d-b104-f6b628ba09c7 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 576.922719] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-638a1bed-c12e-463d-8eb8-2ad9c194772c {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 576.938188] env[63515]: DEBUG nova.compute.provider_tree [None req-757a72d0-3dae-4296-830e-18ff5e948610 tempest-ServerDiagnosticsNegativeTest-1799249066 tempest-ServerDiagnosticsNegativeTest-1799249066-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 576.987772] env[63515]: DEBUG nova.compute.manager [None req-5e279136-7e72-43a2-85c4-9cd11158a72b tempest-ListImageFiltersTestJSON-1100101045 tempest-ListImageFiltersTestJSON-1100101045-project-member] [instance: 48908cb0-a989-4bad-84a4-25a4bdd3baeb] Start building block device mappings for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 577.118240] env[63515]: DEBUG nova.network.neutron [None req-7f5043f7-e6d7-46dd-88fd-99ee049c2b4b tempest-VolumesAssistedSnapshotsTest-453349191 tempest-VolumesAssistedSnapshotsTest-453349191-project-member] [instance: db682708-36fa-4126-a848-bfb609df96d7] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 577.286548] env[63515]: DEBUG nova.compute.manager [req-1660bbf0-8df7-4691-804e-cbe393cafd49 req-1b49b0d5-0c4b-4580-aa38-9709fa9d6ff1 service nova] [instance: db682708-36fa-4126-a848-bfb609df96d7] Received event network-vif-deleted-ee30f5c5-4733-4b41-ad1e-b45f0ae502cb {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 577.351577] env[63515]: DEBUG oslo_vmware.api [None req-b1f9566f-fcad-4d5d-9d53-b04b591d6c49 tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] Task: {'id': task-1110866, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 577.416773] env[63515]: DEBUG oslo_concurrency.lockutils [None req-222754a5-7546-43c7-a99a-1c936b9e5035 tempest-ImagesOneServerTestJSON-1950372435 tempest-ImagesOneServerTestJSON-1950372435-project-member] Lock "aec66939-7ebc-4bfc-bd8b-bae22e011239" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 41.750s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 577.440383] env[63515]: DEBUG nova.scheduler.client.report [None req-757a72d0-3dae-4296-830e-18ff5e948610 tempest-ServerDiagnosticsNegativeTest-1799249066 tempest-ServerDiagnosticsNegativeTest-1799249066-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 577.622156] env[63515]: DEBUG oslo_concurrency.lockutils [None req-7f5043f7-e6d7-46dd-88fd-99ee049c2b4b tempest-VolumesAssistedSnapshotsTest-453349191 tempest-VolumesAssistedSnapshotsTest-453349191-project-member] Releasing lock "refresh_cache-db682708-36fa-4126-a848-bfb609df96d7" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 577.622156] env[63515]: DEBUG nova.compute.manager [None req-7f5043f7-e6d7-46dd-88fd-99ee049c2b4b tempest-VolumesAssistedSnapshotsTest-453349191 tempest-VolumesAssistedSnapshotsTest-453349191-project-member] [instance: db682708-36fa-4126-a848-bfb609df96d7] Start destroying the instance on the hypervisor. {{(pid=63515) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 577.622430] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-7f5043f7-e6d7-46dd-88fd-99ee049c2b4b tempest-VolumesAssistedSnapshotsTest-453349191 tempest-VolumesAssistedSnapshotsTest-453349191-project-member] [instance: db682708-36fa-4126-a848-bfb609df96d7] Destroying instance {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 577.622649] env[63515]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-7ac36ce2-75b0-4f00-9b51-27a2a5045409 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 577.634838] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a385df5-374b-4dcb-bdef-2855ca7116b3 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 577.660235] env[63515]: WARNING nova.virt.vmwareapi.vmops [None req-7f5043f7-e6d7-46dd-88fd-99ee049c2b4b tempest-VolumesAssistedSnapshotsTest-453349191 tempest-VolumesAssistedSnapshotsTest-453349191-project-member] [instance: db682708-36fa-4126-a848-bfb609df96d7] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance db682708-36fa-4126-a848-bfb609df96d7 could not be found. [ 577.660235] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-7f5043f7-e6d7-46dd-88fd-99ee049c2b4b tempest-VolumesAssistedSnapshotsTest-453349191 tempest-VolumesAssistedSnapshotsTest-453349191-project-member] [instance: db682708-36fa-4126-a848-bfb609df96d7] Instance destroyed {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 577.660235] env[63515]: INFO nova.compute.manager [None req-7f5043f7-e6d7-46dd-88fd-99ee049c2b4b tempest-VolumesAssistedSnapshotsTest-453349191 tempest-VolumesAssistedSnapshotsTest-453349191-project-member] [instance: db682708-36fa-4126-a848-bfb609df96d7] Took 0.04 seconds to destroy the instance on the hypervisor. [ 577.660235] env[63515]: DEBUG oslo.service.loopingcall [None req-7f5043f7-e6d7-46dd-88fd-99ee049c2b4b tempest-VolumesAssistedSnapshotsTest-453349191 tempest-VolumesAssistedSnapshotsTest-453349191-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 577.660235] env[63515]: DEBUG nova.compute.manager [-] [instance: db682708-36fa-4126-a848-bfb609df96d7] Deallocating network for instance {{(pid=63515) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 577.660235] env[63515]: DEBUG nova.network.neutron [-] [instance: db682708-36fa-4126-a848-bfb609df96d7] deallocate_for_instance() {{(pid=63515) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 577.701225] env[63515]: DEBUG nova.network.neutron [-] [instance: db682708-36fa-4126-a848-bfb609df96d7] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 577.850396] env[63515]: DEBUG oslo_vmware.api [None req-b1f9566f-fcad-4d5d-9d53-b04b591d6c49 tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] Task: {'id': task-1110866, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.542484} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 577.850666] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-b1f9566f-fcad-4d5d-9d53-b04b591d6c49 tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk to [datastore1] 93b28142-8454-43c0-b0a7-d61aa95c8fc8/93b28142-8454-43c0-b0a7-d61aa95c8fc8.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 577.850871] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-b1f9566f-fcad-4d5d-9d53-b04b591d6c49 tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] [instance: 93b28142-8454-43c0-b0a7-d61aa95c8fc8] Extending root virtual disk to 1048576 {{(pid=63515) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 577.851129] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-cb960b41-fc87-4123-a6bf-6f15e9a06843 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 577.857720] env[63515]: DEBUG oslo_vmware.api [None req-b1f9566f-fcad-4d5d-9d53-b04b591d6c49 tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] Waiting for the task: (returnval){ [ 577.857720] env[63515]: value = "task-1110867" [ 577.857720] env[63515]: _type = "Task" [ 577.857720] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 577.865703] env[63515]: DEBUG oslo_vmware.api [None req-b1f9566f-fcad-4d5d-9d53-b04b591d6c49 tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] Task: {'id': task-1110867, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 577.921832] env[63515]: DEBUG nova.compute.manager [None req-9e973647-d524-4c71-b05a-f1a253758bab tempest-ServerAddressesTestJSON-9863250 tempest-ServerAddressesTestJSON-9863250-project-member] [instance: 7fee0afb-88c9-4e93-a4d9-94312394f927] Starting instance... {{(pid=63515) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 577.946126] env[63515]: DEBUG oslo_concurrency.lockutils [None req-757a72d0-3dae-4296-830e-18ff5e948610 tempest-ServerDiagnosticsNegativeTest-1799249066 tempest-ServerDiagnosticsNegativeTest-1799249066-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.971s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 577.946924] env[63515]: ERROR nova.compute.manager [None req-757a72d0-3dae-4296-830e-18ff5e948610 tempest-ServerDiagnosticsNegativeTest-1799249066 tempest-ServerDiagnosticsNegativeTest-1799249066-project-member] [instance: 1c5d87c7-6747-45a7-8488-6f107eeaeb78] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port bea3bc2e-ac36-4c4f-8398-abe757f5b66f, please check neutron logs for more information. [ 577.946924] env[63515]: ERROR nova.compute.manager [instance: 1c5d87c7-6747-45a7-8488-6f107eeaeb78] Traceback (most recent call last): [ 577.946924] env[63515]: ERROR nova.compute.manager [instance: 1c5d87c7-6747-45a7-8488-6f107eeaeb78] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 577.946924] env[63515]: ERROR nova.compute.manager [instance: 1c5d87c7-6747-45a7-8488-6f107eeaeb78] self.driver.spawn(context, instance, image_meta, [ 577.946924] env[63515]: ERROR nova.compute.manager [instance: 1c5d87c7-6747-45a7-8488-6f107eeaeb78] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 577.946924] env[63515]: ERROR nova.compute.manager [instance: 1c5d87c7-6747-45a7-8488-6f107eeaeb78] self._vmops.spawn(context, instance, image_meta, injected_files, [ 577.946924] env[63515]: ERROR nova.compute.manager [instance: 1c5d87c7-6747-45a7-8488-6f107eeaeb78] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 577.946924] env[63515]: ERROR nova.compute.manager [instance: 1c5d87c7-6747-45a7-8488-6f107eeaeb78] vm_ref = self.build_virtual_machine(instance, [ 577.946924] env[63515]: ERROR nova.compute.manager [instance: 1c5d87c7-6747-45a7-8488-6f107eeaeb78] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 577.946924] env[63515]: ERROR nova.compute.manager [instance: 1c5d87c7-6747-45a7-8488-6f107eeaeb78] vif_infos = vmwarevif.get_vif_info(self._session, [ 577.946924] env[63515]: ERROR nova.compute.manager [instance: 1c5d87c7-6747-45a7-8488-6f107eeaeb78] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 577.947336] env[63515]: ERROR nova.compute.manager [instance: 1c5d87c7-6747-45a7-8488-6f107eeaeb78] for vif in network_info: [ 577.947336] env[63515]: ERROR nova.compute.manager [instance: 1c5d87c7-6747-45a7-8488-6f107eeaeb78] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 577.947336] env[63515]: ERROR nova.compute.manager [instance: 1c5d87c7-6747-45a7-8488-6f107eeaeb78] return self._sync_wrapper(fn, *args, **kwargs) [ 577.947336] env[63515]: ERROR nova.compute.manager [instance: 1c5d87c7-6747-45a7-8488-6f107eeaeb78] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 577.947336] env[63515]: ERROR nova.compute.manager [instance: 1c5d87c7-6747-45a7-8488-6f107eeaeb78] self.wait() [ 577.947336] env[63515]: ERROR nova.compute.manager [instance: 1c5d87c7-6747-45a7-8488-6f107eeaeb78] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 577.947336] env[63515]: ERROR nova.compute.manager [instance: 1c5d87c7-6747-45a7-8488-6f107eeaeb78] self[:] = self._gt.wait() [ 577.947336] env[63515]: ERROR nova.compute.manager [instance: 1c5d87c7-6747-45a7-8488-6f107eeaeb78] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 577.947336] env[63515]: ERROR nova.compute.manager [instance: 1c5d87c7-6747-45a7-8488-6f107eeaeb78] return self._exit_event.wait() [ 577.947336] env[63515]: ERROR nova.compute.manager [instance: 1c5d87c7-6747-45a7-8488-6f107eeaeb78] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 577.947336] env[63515]: ERROR nova.compute.manager [instance: 1c5d87c7-6747-45a7-8488-6f107eeaeb78] result = hub.switch() [ 577.947336] env[63515]: ERROR nova.compute.manager [instance: 1c5d87c7-6747-45a7-8488-6f107eeaeb78] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 577.947336] env[63515]: ERROR nova.compute.manager [instance: 1c5d87c7-6747-45a7-8488-6f107eeaeb78] return self.greenlet.switch() [ 577.947807] env[63515]: ERROR nova.compute.manager [instance: 1c5d87c7-6747-45a7-8488-6f107eeaeb78] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 577.947807] env[63515]: ERROR nova.compute.manager [instance: 1c5d87c7-6747-45a7-8488-6f107eeaeb78] result = function(*args, **kwargs) [ 577.947807] env[63515]: ERROR nova.compute.manager [instance: 1c5d87c7-6747-45a7-8488-6f107eeaeb78] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 577.947807] env[63515]: ERROR nova.compute.manager [instance: 1c5d87c7-6747-45a7-8488-6f107eeaeb78] return func(*args, **kwargs) [ 577.947807] env[63515]: ERROR nova.compute.manager [instance: 1c5d87c7-6747-45a7-8488-6f107eeaeb78] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 577.947807] env[63515]: ERROR nova.compute.manager [instance: 1c5d87c7-6747-45a7-8488-6f107eeaeb78] raise e [ 577.947807] env[63515]: ERROR nova.compute.manager [instance: 1c5d87c7-6747-45a7-8488-6f107eeaeb78] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 577.947807] env[63515]: ERROR nova.compute.manager [instance: 1c5d87c7-6747-45a7-8488-6f107eeaeb78] nwinfo = self.network_api.allocate_for_instance( [ 577.947807] env[63515]: ERROR nova.compute.manager [instance: 1c5d87c7-6747-45a7-8488-6f107eeaeb78] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 577.947807] env[63515]: ERROR nova.compute.manager [instance: 1c5d87c7-6747-45a7-8488-6f107eeaeb78] created_port_ids = self._update_ports_for_instance( [ 577.947807] env[63515]: ERROR nova.compute.manager [instance: 1c5d87c7-6747-45a7-8488-6f107eeaeb78] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 577.947807] env[63515]: ERROR nova.compute.manager [instance: 1c5d87c7-6747-45a7-8488-6f107eeaeb78] with excutils.save_and_reraise_exception(): [ 577.947807] env[63515]: ERROR nova.compute.manager [instance: 1c5d87c7-6747-45a7-8488-6f107eeaeb78] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 577.948178] env[63515]: ERROR nova.compute.manager [instance: 1c5d87c7-6747-45a7-8488-6f107eeaeb78] self.force_reraise() [ 577.948178] env[63515]: ERROR nova.compute.manager [instance: 1c5d87c7-6747-45a7-8488-6f107eeaeb78] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 577.948178] env[63515]: ERROR nova.compute.manager [instance: 1c5d87c7-6747-45a7-8488-6f107eeaeb78] raise self.value [ 577.948178] env[63515]: ERROR nova.compute.manager [instance: 1c5d87c7-6747-45a7-8488-6f107eeaeb78] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 577.948178] env[63515]: ERROR nova.compute.manager [instance: 1c5d87c7-6747-45a7-8488-6f107eeaeb78] updated_port = self._update_port( [ 577.948178] env[63515]: ERROR nova.compute.manager [instance: 1c5d87c7-6747-45a7-8488-6f107eeaeb78] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 577.948178] env[63515]: ERROR nova.compute.manager [instance: 1c5d87c7-6747-45a7-8488-6f107eeaeb78] _ensure_no_port_binding_failure(port) [ 577.948178] env[63515]: ERROR nova.compute.manager [instance: 1c5d87c7-6747-45a7-8488-6f107eeaeb78] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 577.948178] env[63515]: ERROR nova.compute.manager [instance: 1c5d87c7-6747-45a7-8488-6f107eeaeb78] raise exception.PortBindingFailed(port_id=port['id']) [ 577.948178] env[63515]: ERROR nova.compute.manager [instance: 1c5d87c7-6747-45a7-8488-6f107eeaeb78] nova.exception.PortBindingFailed: Binding failed for port bea3bc2e-ac36-4c4f-8398-abe757f5b66f, please check neutron logs for more information. [ 577.948178] env[63515]: ERROR nova.compute.manager [instance: 1c5d87c7-6747-45a7-8488-6f107eeaeb78] [ 577.948485] env[63515]: DEBUG nova.compute.utils [None req-757a72d0-3dae-4296-830e-18ff5e948610 tempest-ServerDiagnosticsNegativeTest-1799249066 tempest-ServerDiagnosticsNegativeTest-1799249066-project-member] [instance: 1c5d87c7-6747-45a7-8488-6f107eeaeb78] Binding failed for port bea3bc2e-ac36-4c4f-8398-abe757f5b66f, please check neutron logs for more information. {{(pid=63515) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 577.949116] env[63515]: DEBUG oslo_concurrency.lockutils [None req-a5c600bc-b1c1-4a6d-9386-2b0dbd0395e4 tempest-InstanceActionsNegativeTestJSON-1061405461 tempest-InstanceActionsNegativeTestJSON-1061405461-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 17.208s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 577.962551] env[63515]: DEBUG nova.compute.manager [None req-757a72d0-3dae-4296-830e-18ff5e948610 tempest-ServerDiagnosticsNegativeTest-1799249066 tempest-ServerDiagnosticsNegativeTest-1799249066-project-member] [instance: 1c5d87c7-6747-45a7-8488-6f107eeaeb78] Build of instance 1c5d87c7-6747-45a7-8488-6f107eeaeb78 was re-scheduled: Binding failed for port bea3bc2e-ac36-4c4f-8398-abe757f5b66f, please check neutron logs for more information. {{(pid=63515) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 577.962551] env[63515]: DEBUG nova.compute.manager [None req-757a72d0-3dae-4296-830e-18ff5e948610 tempest-ServerDiagnosticsNegativeTest-1799249066 tempest-ServerDiagnosticsNegativeTest-1799249066-project-member] [instance: 1c5d87c7-6747-45a7-8488-6f107eeaeb78] Unplugging VIFs for instance {{(pid=63515) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 577.962551] env[63515]: DEBUG oslo_concurrency.lockutils [None req-757a72d0-3dae-4296-830e-18ff5e948610 tempest-ServerDiagnosticsNegativeTest-1799249066 tempest-ServerDiagnosticsNegativeTest-1799249066-project-member] Acquiring lock "refresh_cache-1c5d87c7-6747-45a7-8488-6f107eeaeb78" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 577.962551] env[63515]: DEBUG oslo_concurrency.lockutils [None req-757a72d0-3dae-4296-830e-18ff5e948610 tempest-ServerDiagnosticsNegativeTest-1799249066 tempest-ServerDiagnosticsNegativeTest-1799249066-project-member] Acquired lock "refresh_cache-1c5d87c7-6747-45a7-8488-6f107eeaeb78" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 577.962766] env[63515]: DEBUG nova.network.neutron [None req-757a72d0-3dae-4296-830e-18ff5e948610 tempest-ServerDiagnosticsNegativeTest-1799249066 tempest-ServerDiagnosticsNegativeTest-1799249066-project-member] [instance: 1c5d87c7-6747-45a7-8488-6f107eeaeb78] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 578.003812] env[63515]: DEBUG nova.compute.manager [None req-5e279136-7e72-43a2-85c4-9cd11158a72b tempest-ListImageFiltersTestJSON-1100101045 tempest-ListImageFiltersTestJSON-1100101045-project-member] [instance: 48908cb0-a989-4bad-84a4-25a4bdd3baeb] Start spawning the instance on the hypervisor. {{(pid=63515) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 578.036260] env[63515]: DEBUG nova.virt.hardware [None req-5e279136-7e72-43a2-85c4-9cd11158a72b tempest-ListImageFiltersTestJSON-1100101045 tempest-ListImageFiltersTestJSON-1100101045-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T02:52:14Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T02:51:56Z,direct_url=,disk_format='vmdk',id=8a120570-cb06-4099-b262-554ca0ad15c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b569255cc43e42e7ae3f2b4ad37c6ef4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T02:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 578.036260] env[63515]: DEBUG nova.virt.hardware [None req-5e279136-7e72-43a2-85c4-9cd11158a72b tempest-ListImageFiltersTestJSON-1100101045 tempest-ListImageFiltersTestJSON-1100101045-project-member] Flavor limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 578.036260] env[63515]: DEBUG nova.virt.hardware [None req-5e279136-7e72-43a2-85c4-9cd11158a72b tempest-ListImageFiltersTestJSON-1100101045 tempest-ListImageFiltersTestJSON-1100101045-project-member] Image limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 578.036567] env[63515]: DEBUG nova.virt.hardware [None req-5e279136-7e72-43a2-85c4-9cd11158a72b tempest-ListImageFiltersTestJSON-1100101045 tempest-ListImageFiltersTestJSON-1100101045-project-member] Flavor pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 578.036567] env[63515]: DEBUG nova.virt.hardware [None req-5e279136-7e72-43a2-85c4-9cd11158a72b tempest-ListImageFiltersTestJSON-1100101045 tempest-ListImageFiltersTestJSON-1100101045-project-member] Image pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 578.036567] env[63515]: DEBUG nova.virt.hardware [None req-5e279136-7e72-43a2-85c4-9cd11158a72b tempest-ListImageFiltersTestJSON-1100101045 tempest-ListImageFiltersTestJSON-1100101045-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 578.039185] env[63515]: DEBUG nova.virt.hardware [None req-5e279136-7e72-43a2-85c4-9cd11158a72b tempest-ListImageFiltersTestJSON-1100101045 tempest-ListImageFiltersTestJSON-1100101045-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 578.039185] env[63515]: DEBUG nova.virt.hardware [None req-5e279136-7e72-43a2-85c4-9cd11158a72b tempest-ListImageFiltersTestJSON-1100101045 tempest-ListImageFiltersTestJSON-1100101045-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 578.039185] env[63515]: DEBUG nova.virt.hardware [None req-5e279136-7e72-43a2-85c4-9cd11158a72b tempest-ListImageFiltersTestJSON-1100101045 tempest-ListImageFiltersTestJSON-1100101045-project-member] Got 1 possible topologies {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 578.039185] env[63515]: DEBUG nova.virt.hardware [None req-5e279136-7e72-43a2-85c4-9cd11158a72b tempest-ListImageFiltersTestJSON-1100101045 tempest-ListImageFiltersTestJSON-1100101045-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 578.039185] env[63515]: DEBUG nova.virt.hardware [None req-5e279136-7e72-43a2-85c4-9cd11158a72b tempest-ListImageFiltersTestJSON-1100101045 tempest-ListImageFiltersTestJSON-1100101045-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 578.040023] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-048c7ddd-342b-42d4-aba3-80bc97b4a726 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 578.048956] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7466e15f-2f26-414e-b503-9a183f886dc8 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 578.155251] env[63515]: DEBUG nova.network.neutron [None req-5e279136-7e72-43a2-85c4-9cd11158a72b tempest-ListImageFiltersTestJSON-1100101045 tempest-ListImageFiltersTestJSON-1100101045-project-member] [instance: 48908cb0-a989-4bad-84a4-25a4bdd3baeb] Successfully created port: e7173eb8-86a4-4d43-8391-1d565f1dfaa9 {{(pid=63515) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 578.202856] env[63515]: DEBUG nova.network.neutron [-] [instance: db682708-36fa-4126-a848-bfb609df96d7] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 578.371097] env[63515]: DEBUG oslo_vmware.api [None req-b1f9566f-fcad-4d5d-9d53-b04b591d6c49 tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] Task: {'id': task-1110867, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.066256} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 578.371359] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-b1f9566f-fcad-4d5d-9d53-b04b591d6c49 tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] [instance: 93b28142-8454-43c0-b0a7-d61aa95c8fc8] Extended root virtual disk {{(pid=63515) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 578.372300] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8479384e-6bb6-4ede-bea0-ca04a40dff3e {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 578.393884] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-b1f9566f-fcad-4d5d-9d53-b04b591d6c49 tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] [instance: 93b28142-8454-43c0-b0a7-d61aa95c8fc8] Reconfiguring VM instance instance-0000000a to attach disk [datastore1] 93b28142-8454-43c0-b0a7-d61aa95c8fc8/93b28142-8454-43c0-b0a7-d61aa95c8fc8.vmdk or device None with type sparse {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 578.393884] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c3016933-efe1-4c22-b8c6-15d7fcd5e7ec {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 578.414201] env[63515]: DEBUG oslo_vmware.api [None req-b1f9566f-fcad-4d5d-9d53-b04b591d6c49 tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] Waiting for the task: (returnval){ [ 578.414201] env[63515]: value = "task-1110868" [ 578.414201] env[63515]: _type = "Task" [ 578.414201] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 578.435454] env[63515]: DEBUG oslo_vmware.api [None req-b1f9566f-fcad-4d5d-9d53-b04b591d6c49 tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] Task: {'id': task-1110868, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 578.457847] env[63515]: DEBUG oslo_concurrency.lockutils [None req-9e973647-d524-4c71-b05a-f1a253758bab tempest-ServerAddressesTestJSON-9863250 tempest-ServerAddressesTestJSON-9863250-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 578.548084] env[63515]: DEBUG nova.network.neutron [None req-757a72d0-3dae-4296-830e-18ff5e948610 tempest-ServerDiagnosticsNegativeTest-1799249066 tempest-ServerDiagnosticsNegativeTest-1799249066-project-member] [instance: 1c5d87c7-6747-45a7-8488-6f107eeaeb78] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 578.707123] env[63515]: INFO nova.compute.manager [-] [instance: db682708-36fa-4126-a848-bfb609df96d7] Took 1.05 seconds to deallocate network for instance. [ 578.715102] env[63515]: DEBUG nova.compute.claims [None req-7f5043f7-e6d7-46dd-88fd-99ee049c2b4b tempest-VolumesAssistedSnapshotsTest-453349191 tempest-VolumesAssistedSnapshotsTest-453349191-project-member] [instance: db682708-36fa-4126-a848-bfb609df96d7] Aborting claim: {{(pid=63515) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 578.715697] env[63515]: DEBUG oslo_concurrency.lockutils [None req-7f5043f7-e6d7-46dd-88fd-99ee049c2b4b tempest-VolumesAssistedSnapshotsTest-453349191 tempest-VolumesAssistedSnapshotsTest-453349191-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 578.719604] env[63515]: DEBUG nova.network.neutron [None req-757a72d0-3dae-4296-830e-18ff5e948610 tempest-ServerDiagnosticsNegativeTest-1799249066 tempest-ServerDiagnosticsNegativeTest-1799249066-project-member] [instance: 1c5d87c7-6747-45a7-8488-6f107eeaeb78] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 578.871635] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3029a926-9cd6-4cdf-aeae-315943f8c0e3 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 578.883413] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad204351-b4aa-424d-be45-a5f3f0566793 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 578.936634] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5564d258-1e20-49af-a33b-28f51b483961 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 578.947586] env[63515]: DEBUG oslo_vmware.api [None req-b1f9566f-fcad-4d5d-9d53-b04b591d6c49 tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] Task: {'id': task-1110868, 'name': ReconfigVM_Task} progress is 99%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 578.952128] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f0f0c75-e442-4ee4-9833-4e431b428b78 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 578.966553] env[63515]: DEBUG nova.compute.provider_tree [None req-a5c600bc-b1c1-4a6d-9386-2b0dbd0395e4 tempest-InstanceActionsNegativeTestJSON-1061405461 tempest-InstanceActionsNegativeTestJSON-1061405461-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 579.222998] env[63515]: DEBUG oslo_concurrency.lockutils [None req-757a72d0-3dae-4296-830e-18ff5e948610 tempest-ServerDiagnosticsNegativeTest-1799249066 tempest-ServerDiagnosticsNegativeTest-1799249066-project-member] Releasing lock "refresh_cache-1c5d87c7-6747-45a7-8488-6f107eeaeb78" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 579.222998] env[63515]: DEBUG nova.compute.manager [None req-757a72d0-3dae-4296-830e-18ff5e948610 tempest-ServerDiagnosticsNegativeTest-1799249066 tempest-ServerDiagnosticsNegativeTest-1799249066-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63515) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 579.223127] env[63515]: DEBUG nova.compute.manager [None req-757a72d0-3dae-4296-830e-18ff5e948610 tempest-ServerDiagnosticsNegativeTest-1799249066 tempest-ServerDiagnosticsNegativeTest-1799249066-project-member] [instance: 1c5d87c7-6747-45a7-8488-6f107eeaeb78] Deallocating network for instance {{(pid=63515) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 579.223305] env[63515]: DEBUG nova.network.neutron [None req-757a72d0-3dae-4296-830e-18ff5e948610 tempest-ServerDiagnosticsNegativeTest-1799249066 tempest-ServerDiagnosticsNegativeTest-1799249066-project-member] [instance: 1c5d87c7-6747-45a7-8488-6f107eeaeb78] deallocate_for_instance() {{(pid=63515) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 579.258062] env[63515]: DEBUG nova.network.neutron [None req-757a72d0-3dae-4296-830e-18ff5e948610 tempest-ServerDiagnosticsNegativeTest-1799249066 tempest-ServerDiagnosticsNegativeTest-1799249066-project-member] [instance: 1c5d87c7-6747-45a7-8488-6f107eeaeb78] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 579.441605] env[63515]: DEBUG oslo_vmware.api [None req-b1f9566f-fcad-4d5d-9d53-b04b591d6c49 tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] Task: {'id': task-1110868, 'name': ReconfigVM_Task} progress is 99%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 579.471232] env[63515]: DEBUG nova.scheduler.client.report [None req-a5c600bc-b1c1-4a6d-9386-2b0dbd0395e4 tempest-InstanceActionsNegativeTestJSON-1061405461 tempest-InstanceActionsNegativeTestJSON-1061405461-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 579.555949] env[63515]: DEBUG oslo_concurrency.lockutils [None req-fa933c74-4a5b-4cc9-a499-9ea16a38d8b7 tempest-ServerActionsTestJSON-1591637270 tempest-ServerActionsTestJSON-1591637270-project-member] Acquiring lock "005f45b3-630d-400f-9605-100f6aad88e6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 579.555949] env[63515]: DEBUG oslo_concurrency.lockutils [None req-fa933c74-4a5b-4cc9-a499-9ea16a38d8b7 tempest-ServerActionsTestJSON-1591637270 tempest-ServerActionsTestJSON-1591637270-project-member] Lock "005f45b3-630d-400f-9605-100f6aad88e6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 579.761794] env[63515]: DEBUG nova.network.neutron [None req-757a72d0-3dae-4296-830e-18ff5e948610 tempest-ServerDiagnosticsNegativeTest-1799249066 tempest-ServerDiagnosticsNegativeTest-1799249066-project-member] [instance: 1c5d87c7-6747-45a7-8488-6f107eeaeb78] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 579.941138] env[63515]: DEBUG oslo_concurrency.lockutils [None req-45f6ec8d-6d74-4e46-8fdf-6b0f7c107bf4 tempest-ServerDiagnosticsTest-1160321831 tempest-ServerDiagnosticsTest-1160321831-project-member] Acquiring lock "4ad5440f-cd18-44c5-8836-0aa39824cf03" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 579.942349] env[63515]: DEBUG oslo_concurrency.lockutils [None req-45f6ec8d-6d74-4e46-8fdf-6b0f7c107bf4 tempest-ServerDiagnosticsTest-1160321831 tempest-ServerDiagnosticsTest-1160321831-project-member] Lock "4ad5440f-cd18-44c5-8836-0aa39824cf03" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 579.949278] env[63515]: DEBUG oslo_vmware.api [None req-b1f9566f-fcad-4d5d-9d53-b04b591d6c49 tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] Task: {'id': task-1110868, 'name': ReconfigVM_Task, 'duration_secs': 1.269941} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 579.952574] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-b1f9566f-fcad-4d5d-9d53-b04b591d6c49 tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] [instance: 93b28142-8454-43c0-b0a7-d61aa95c8fc8] Reconfigured VM instance instance-0000000a to attach disk [datastore1] 93b28142-8454-43c0-b0a7-d61aa95c8fc8/93b28142-8454-43c0-b0a7-d61aa95c8fc8.vmdk or device None with type sparse {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 579.952574] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f676e24f-94ff-4b97-b8ba-6a2ce4ef7d72 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 579.960090] env[63515]: DEBUG oslo_vmware.api [None req-b1f9566f-fcad-4d5d-9d53-b04b591d6c49 tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] Waiting for the task: (returnval){ [ 579.960090] env[63515]: value = "task-1110870" [ 579.960090] env[63515]: _type = "Task" [ 579.960090] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 579.972219] env[63515]: DEBUG oslo_vmware.api [None req-b1f9566f-fcad-4d5d-9d53-b04b591d6c49 tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] Task: {'id': task-1110870, 'name': Rename_Task} progress is 5%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 579.977585] env[63515]: DEBUG oslo_concurrency.lockutils [None req-a5c600bc-b1c1-4a6d-9386-2b0dbd0395e4 tempest-InstanceActionsNegativeTestJSON-1061405461 tempest-InstanceActionsNegativeTestJSON-1061405461-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.028s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 579.978159] env[63515]: ERROR nova.compute.manager [None req-a5c600bc-b1c1-4a6d-9386-2b0dbd0395e4 tempest-InstanceActionsNegativeTestJSON-1061405461 tempest-InstanceActionsNegativeTestJSON-1061405461-project-member] [instance: b641d2d1-de70-48b6-9137-fe18c3bc7511] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port cab05064-e039-47ee-9e97-1d390704eedd, please check neutron logs for more information. [ 579.978159] env[63515]: ERROR nova.compute.manager [instance: b641d2d1-de70-48b6-9137-fe18c3bc7511] Traceback (most recent call last): [ 579.978159] env[63515]: ERROR nova.compute.manager [instance: b641d2d1-de70-48b6-9137-fe18c3bc7511] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 579.978159] env[63515]: ERROR nova.compute.manager [instance: b641d2d1-de70-48b6-9137-fe18c3bc7511] self.driver.spawn(context, instance, image_meta, [ 579.978159] env[63515]: ERROR nova.compute.manager [instance: b641d2d1-de70-48b6-9137-fe18c3bc7511] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 579.978159] env[63515]: ERROR nova.compute.manager [instance: b641d2d1-de70-48b6-9137-fe18c3bc7511] self._vmops.spawn(context, instance, image_meta, injected_files, [ 579.978159] env[63515]: ERROR nova.compute.manager [instance: b641d2d1-de70-48b6-9137-fe18c3bc7511] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 579.978159] env[63515]: ERROR nova.compute.manager [instance: b641d2d1-de70-48b6-9137-fe18c3bc7511] vm_ref = self.build_virtual_machine(instance, [ 579.978159] env[63515]: ERROR nova.compute.manager [instance: b641d2d1-de70-48b6-9137-fe18c3bc7511] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 579.978159] env[63515]: ERROR nova.compute.manager [instance: b641d2d1-de70-48b6-9137-fe18c3bc7511] vif_infos = vmwarevif.get_vif_info(self._session, [ 579.978159] env[63515]: ERROR nova.compute.manager [instance: b641d2d1-de70-48b6-9137-fe18c3bc7511] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 579.978882] env[63515]: ERROR nova.compute.manager [instance: b641d2d1-de70-48b6-9137-fe18c3bc7511] for vif in network_info: [ 579.978882] env[63515]: ERROR nova.compute.manager [instance: b641d2d1-de70-48b6-9137-fe18c3bc7511] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 579.978882] env[63515]: ERROR nova.compute.manager [instance: b641d2d1-de70-48b6-9137-fe18c3bc7511] return self._sync_wrapper(fn, *args, **kwargs) [ 579.978882] env[63515]: ERROR nova.compute.manager [instance: b641d2d1-de70-48b6-9137-fe18c3bc7511] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 579.978882] env[63515]: ERROR nova.compute.manager [instance: b641d2d1-de70-48b6-9137-fe18c3bc7511] self.wait() [ 579.978882] env[63515]: ERROR nova.compute.manager [instance: b641d2d1-de70-48b6-9137-fe18c3bc7511] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 579.978882] env[63515]: ERROR nova.compute.manager [instance: b641d2d1-de70-48b6-9137-fe18c3bc7511] self[:] = self._gt.wait() [ 579.978882] env[63515]: ERROR nova.compute.manager [instance: b641d2d1-de70-48b6-9137-fe18c3bc7511] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 579.978882] env[63515]: ERROR nova.compute.manager [instance: b641d2d1-de70-48b6-9137-fe18c3bc7511] return self._exit_event.wait() [ 579.978882] env[63515]: ERROR nova.compute.manager [instance: b641d2d1-de70-48b6-9137-fe18c3bc7511] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 579.978882] env[63515]: ERROR nova.compute.manager [instance: b641d2d1-de70-48b6-9137-fe18c3bc7511] result = hub.switch() [ 579.978882] env[63515]: ERROR nova.compute.manager [instance: b641d2d1-de70-48b6-9137-fe18c3bc7511] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 579.978882] env[63515]: ERROR nova.compute.manager [instance: b641d2d1-de70-48b6-9137-fe18c3bc7511] return self.greenlet.switch() [ 579.980211] env[63515]: ERROR nova.compute.manager [instance: b641d2d1-de70-48b6-9137-fe18c3bc7511] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 579.980211] env[63515]: ERROR nova.compute.manager [instance: b641d2d1-de70-48b6-9137-fe18c3bc7511] result = function(*args, **kwargs) [ 579.980211] env[63515]: ERROR nova.compute.manager [instance: b641d2d1-de70-48b6-9137-fe18c3bc7511] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 579.980211] env[63515]: ERROR nova.compute.manager [instance: b641d2d1-de70-48b6-9137-fe18c3bc7511] return func(*args, **kwargs) [ 579.980211] env[63515]: ERROR nova.compute.manager [instance: b641d2d1-de70-48b6-9137-fe18c3bc7511] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 579.980211] env[63515]: ERROR nova.compute.manager [instance: b641d2d1-de70-48b6-9137-fe18c3bc7511] raise e [ 579.980211] env[63515]: ERROR nova.compute.manager [instance: b641d2d1-de70-48b6-9137-fe18c3bc7511] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 579.980211] env[63515]: ERROR nova.compute.manager [instance: b641d2d1-de70-48b6-9137-fe18c3bc7511] nwinfo = self.network_api.allocate_for_instance( [ 579.980211] env[63515]: ERROR nova.compute.manager [instance: b641d2d1-de70-48b6-9137-fe18c3bc7511] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 579.980211] env[63515]: ERROR nova.compute.manager [instance: b641d2d1-de70-48b6-9137-fe18c3bc7511] created_port_ids = self._update_ports_for_instance( [ 579.980211] env[63515]: ERROR nova.compute.manager [instance: b641d2d1-de70-48b6-9137-fe18c3bc7511] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 579.980211] env[63515]: ERROR nova.compute.manager [instance: b641d2d1-de70-48b6-9137-fe18c3bc7511] with excutils.save_and_reraise_exception(): [ 579.980211] env[63515]: ERROR nova.compute.manager [instance: b641d2d1-de70-48b6-9137-fe18c3bc7511] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 579.981095] env[63515]: ERROR nova.compute.manager [instance: b641d2d1-de70-48b6-9137-fe18c3bc7511] self.force_reraise() [ 579.981095] env[63515]: ERROR nova.compute.manager [instance: b641d2d1-de70-48b6-9137-fe18c3bc7511] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 579.981095] env[63515]: ERROR nova.compute.manager [instance: b641d2d1-de70-48b6-9137-fe18c3bc7511] raise self.value [ 579.981095] env[63515]: ERROR nova.compute.manager [instance: b641d2d1-de70-48b6-9137-fe18c3bc7511] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 579.981095] env[63515]: ERROR nova.compute.manager [instance: b641d2d1-de70-48b6-9137-fe18c3bc7511] updated_port = self._update_port( [ 579.981095] env[63515]: ERROR nova.compute.manager [instance: b641d2d1-de70-48b6-9137-fe18c3bc7511] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 579.981095] env[63515]: ERROR nova.compute.manager [instance: b641d2d1-de70-48b6-9137-fe18c3bc7511] _ensure_no_port_binding_failure(port) [ 579.981095] env[63515]: ERROR nova.compute.manager [instance: b641d2d1-de70-48b6-9137-fe18c3bc7511] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 579.981095] env[63515]: ERROR nova.compute.manager [instance: b641d2d1-de70-48b6-9137-fe18c3bc7511] raise exception.PortBindingFailed(port_id=port['id']) [ 579.981095] env[63515]: ERROR nova.compute.manager [instance: b641d2d1-de70-48b6-9137-fe18c3bc7511] nova.exception.PortBindingFailed: Binding failed for port cab05064-e039-47ee-9e97-1d390704eedd, please check neutron logs for more information. [ 579.981095] env[63515]: ERROR nova.compute.manager [instance: b641d2d1-de70-48b6-9137-fe18c3bc7511] [ 579.981633] env[63515]: DEBUG nova.compute.utils [None req-a5c600bc-b1c1-4a6d-9386-2b0dbd0395e4 tempest-InstanceActionsNegativeTestJSON-1061405461 tempest-InstanceActionsNegativeTestJSON-1061405461-project-member] [instance: b641d2d1-de70-48b6-9137-fe18c3bc7511] Binding failed for port cab05064-e039-47ee-9e97-1d390704eedd, please check neutron logs for more information. {{(pid=63515) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 579.981633] env[63515]: DEBUG oslo_concurrency.lockutils [None req-50ea3e37-671c-4a4b-9c1a-d782ff1b578c tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.605s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 579.984036] env[63515]: DEBUG nova.compute.manager [None req-a5c600bc-b1c1-4a6d-9386-2b0dbd0395e4 tempest-InstanceActionsNegativeTestJSON-1061405461 tempest-InstanceActionsNegativeTestJSON-1061405461-project-member] [instance: b641d2d1-de70-48b6-9137-fe18c3bc7511] Build of instance b641d2d1-de70-48b6-9137-fe18c3bc7511 was re-scheduled: Binding failed for port cab05064-e039-47ee-9e97-1d390704eedd, please check neutron logs for more information. {{(pid=63515) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 579.984382] env[63515]: DEBUG nova.compute.manager [None req-a5c600bc-b1c1-4a6d-9386-2b0dbd0395e4 tempest-InstanceActionsNegativeTestJSON-1061405461 tempest-InstanceActionsNegativeTestJSON-1061405461-project-member] [instance: b641d2d1-de70-48b6-9137-fe18c3bc7511] Unplugging VIFs for instance {{(pid=63515) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 579.984855] env[63515]: DEBUG oslo_concurrency.lockutils [None req-a5c600bc-b1c1-4a6d-9386-2b0dbd0395e4 tempest-InstanceActionsNegativeTestJSON-1061405461 tempest-InstanceActionsNegativeTestJSON-1061405461-project-member] Acquiring lock "refresh_cache-b641d2d1-de70-48b6-9137-fe18c3bc7511" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 579.984963] env[63515]: DEBUG oslo_concurrency.lockutils [None req-a5c600bc-b1c1-4a6d-9386-2b0dbd0395e4 tempest-InstanceActionsNegativeTestJSON-1061405461 tempest-InstanceActionsNegativeTestJSON-1061405461-project-member] Acquired lock "refresh_cache-b641d2d1-de70-48b6-9137-fe18c3bc7511" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 579.985123] env[63515]: DEBUG nova.network.neutron [None req-a5c600bc-b1c1-4a6d-9386-2b0dbd0395e4 tempest-InstanceActionsNegativeTestJSON-1061405461 tempest-InstanceActionsNegativeTestJSON-1061405461-project-member] [instance: b641d2d1-de70-48b6-9137-fe18c3bc7511] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 580.264934] env[63515]: INFO nova.compute.manager [None req-757a72d0-3dae-4296-830e-18ff5e948610 tempest-ServerDiagnosticsNegativeTest-1799249066 tempest-ServerDiagnosticsNegativeTest-1799249066-project-member] [instance: 1c5d87c7-6747-45a7-8488-6f107eeaeb78] Took 1.04 seconds to deallocate network for instance. [ 580.298969] env[63515]: DEBUG nova.compute.manager [req-6809d388-0d7d-46fe-adf3-c1d10bbb9a98 req-1f7249e4-7a80-4d0b-9cce-e7e10f725265 service nova] [instance: 48908cb0-a989-4bad-84a4-25a4bdd3baeb] Received event network-changed-e7173eb8-86a4-4d43-8391-1d565f1dfaa9 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 580.299775] env[63515]: DEBUG nova.compute.manager [req-6809d388-0d7d-46fe-adf3-c1d10bbb9a98 req-1f7249e4-7a80-4d0b-9cce-e7e10f725265 service nova] [instance: 48908cb0-a989-4bad-84a4-25a4bdd3baeb] Refreshing instance network info cache due to event network-changed-e7173eb8-86a4-4d43-8391-1d565f1dfaa9. {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 580.299925] env[63515]: DEBUG oslo_concurrency.lockutils [req-6809d388-0d7d-46fe-adf3-c1d10bbb9a98 req-1f7249e4-7a80-4d0b-9cce-e7e10f725265 service nova] Acquiring lock "refresh_cache-48908cb0-a989-4bad-84a4-25a4bdd3baeb" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 580.303378] env[63515]: DEBUG oslo_concurrency.lockutils [req-6809d388-0d7d-46fe-adf3-c1d10bbb9a98 req-1f7249e4-7a80-4d0b-9cce-e7e10f725265 service nova] Acquired lock "refresh_cache-48908cb0-a989-4bad-84a4-25a4bdd3baeb" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 580.303596] env[63515]: DEBUG nova.network.neutron [req-6809d388-0d7d-46fe-adf3-c1d10bbb9a98 req-1f7249e4-7a80-4d0b-9cce-e7e10f725265 service nova] [instance: 48908cb0-a989-4bad-84a4-25a4bdd3baeb] Refreshing network info cache for port e7173eb8-86a4-4d43-8391-1d565f1dfaa9 {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 580.472856] env[63515]: DEBUG oslo_vmware.api [None req-b1f9566f-fcad-4d5d-9d53-b04b591d6c49 tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] Task: {'id': task-1110870, 'name': Rename_Task, 'duration_secs': 0.136052} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 580.476701] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-b1f9566f-fcad-4d5d-9d53-b04b591d6c49 tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] [instance: 93b28142-8454-43c0-b0a7-d61aa95c8fc8] Powering on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 580.476806] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7a9b0d04-655b-4400-a517-a7778f70196c {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 580.484585] env[63515]: DEBUG oslo_vmware.api [None req-b1f9566f-fcad-4d5d-9d53-b04b591d6c49 tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] Waiting for the task: (returnval){ [ 580.484585] env[63515]: value = "task-1110871" [ 580.484585] env[63515]: _type = "Task" [ 580.484585] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 580.498914] env[63515]: DEBUG oslo_vmware.api [None req-b1f9566f-fcad-4d5d-9d53-b04b591d6c49 tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] Task: {'id': task-1110871, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 580.560523] env[63515]: DEBUG nova.network.neutron [None req-a5c600bc-b1c1-4a6d-9386-2b0dbd0395e4 tempest-InstanceActionsNegativeTestJSON-1061405461 tempest-InstanceActionsNegativeTestJSON-1061405461-project-member] [instance: b641d2d1-de70-48b6-9137-fe18c3bc7511] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 580.888999] env[63515]: DEBUG nova.network.neutron [None req-a5c600bc-b1c1-4a6d-9386-2b0dbd0395e4 tempest-InstanceActionsNegativeTestJSON-1061405461 tempest-InstanceActionsNegativeTestJSON-1061405461-project-member] [instance: b641d2d1-de70-48b6-9137-fe18c3bc7511] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 580.948305] env[63515]: DEBUG nova.network.neutron [req-6809d388-0d7d-46fe-adf3-c1d10bbb9a98 req-1f7249e4-7a80-4d0b-9cce-e7e10f725265 service nova] [instance: 48908cb0-a989-4bad-84a4-25a4bdd3baeb] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 580.993186] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-672d5dce-0c07-4909-8590-2a92020330cb {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 581.000534] env[63515]: DEBUG oslo_vmware.api [None req-b1f9566f-fcad-4d5d-9d53-b04b591d6c49 tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] Task: {'id': task-1110871, 'name': PowerOnVM_Task, 'duration_secs': 0.447166} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 581.002878] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-b1f9566f-fcad-4d5d-9d53-b04b591d6c49 tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] [instance: 93b28142-8454-43c0-b0a7-d61aa95c8fc8] Powered on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 581.002878] env[63515]: DEBUG nova.compute.manager [None req-b1f9566f-fcad-4d5d-9d53-b04b591d6c49 tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] [instance: 93b28142-8454-43c0-b0a7-d61aa95c8fc8] Checking state {{(pid=63515) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 581.002878] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d0aed95-2a54-4711-aa52-ae84bef57db5 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 581.009770] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e6c1315-ff59-4a6a-84d7-3739f301ee8d {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 581.054779] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca1d31d4-67cc-4ced-be29-6fcf43e433fe {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 581.063935] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-420eec38-fa64-4fce-8088-0d685bad7d4f {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 581.080224] env[63515]: DEBUG nova.compute.provider_tree [None req-50ea3e37-671c-4a4b-9c1a-d782ff1b578c tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 581.083061] env[63515]: ERROR nova.compute.manager [None req-5e279136-7e72-43a2-85c4-9cd11158a72b tempest-ListImageFiltersTestJSON-1100101045 tempest-ListImageFiltersTestJSON-1100101045-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port e7173eb8-86a4-4d43-8391-1d565f1dfaa9, please check neutron logs for more information. [ 581.083061] env[63515]: ERROR nova.compute.manager Traceback (most recent call last): [ 581.083061] env[63515]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 581.083061] env[63515]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 581.083061] env[63515]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 581.083061] env[63515]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 581.083061] env[63515]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 581.083061] env[63515]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 581.083061] env[63515]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 581.083061] env[63515]: ERROR nova.compute.manager self.force_reraise() [ 581.083061] env[63515]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 581.083061] env[63515]: ERROR nova.compute.manager raise self.value [ 581.083061] env[63515]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 581.083061] env[63515]: ERROR nova.compute.manager updated_port = self._update_port( [ 581.083061] env[63515]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 581.083061] env[63515]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 581.083702] env[63515]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 581.083702] env[63515]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 581.083702] env[63515]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port e7173eb8-86a4-4d43-8391-1d565f1dfaa9, please check neutron logs for more information. [ 581.083702] env[63515]: ERROR nova.compute.manager [ 581.083702] env[63515]: Traceback (most recent call last): [ 581.083702] env[63515]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 581.083702] env[63515]: listener.cb(fileno) [ 581.083702] env[63515]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 581.083702] env[63515]: result = function(*args, **kwargs) [ 581.083702] env[63515]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 581.083702] env[63515]: return func(*args, **kwargs) [ 581.083702] env[63515]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 581.083702] env[63515]: raise e [ 581.083702] env[63515]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 581.083702] env[63515]: nwinfo = self.network_api.allocate_for_instance( [ 581.083702] env[63515]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 581.083702] env[63515]: created_port_ids = self._update_ports_for_instance( [ 581.083702] env[63515]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 581.083702] env[63515]: with excutils.save_and_reraise_exception(): [ 581.083702] env[63515]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 581.083702] env[63515]: self.force_reraise() [ 581.083702] env[63515]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 581.083702] env[63515]: raise self.value [ 581.083702] env[63515]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 581.083702] env[63515]: updated_port = self._update_port( [ 581.083702] env[63515]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 581.083702] env[63515]: _ensure_no_port_binding_failure(port) [ 581.083702] env[63515]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 581.083702] env[63515]: raise exception.PortBindingFailed(port_id=port['id']) [ 581.084727] env[63515]: nova.exception.PortBindingFailed: Binding failed for port e7173eb8-86a4-4d43-8391-1d565f1dfaa9, please check neutron logs for more information. [ 581.084727] env[63515]: Removing descriptor: 16 [ 581.084727] env[63515]: ERROR nova.compute.manager [None req-5e279136-7e72-43a2-85c4-9cd11158a72b tempest-ListImageFiltersTestJSON-1100101045 tempest-ListImageFiltersTestJSON-1100101045-project-member] [instance: 48908cb0-a989-4bad-84a4-25a4bdd3baeb] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port e7173eb8-86a4-4d43-8391-1d565f1dfaa9, please check neutron logs for more information. [ 581.084727] env[63515]: ERROR nova.compute.manager [instance: 48908cb0-a989-4bad-84a4-25a4bdd3baeb] Traceback (most recent call last): [ 581.084727] env[63515]: ERROR nova.compute.manager [instance: 48908cb0-a989-4bad-84a4-25a4bdd3baeb] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 581.084727] env[63515]: ERROR nova.compute.manager [instance: 48908cb0-a989-4bad-84a4-25a4bdd3baeb] yield resources [ 581.084727] env[63515]: ERROR nova.compute.manager [instance: 48908cb0-a989-4bad-84a4-25a4bdd3baeb] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 581.084727] env[63515]: ERROR nova.compute.manager [instance: 48908cb0-a989-4bad-84a4-25a4bdd3baeb] self.driver.spawn(context, instance, image_meta, [ 581.084727] env[63515]: ERROR nova.compute.manager [instance: 48908cb0-a989-4bad-84a4-25a4bdd3baeb] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 581.084727] env[63515]: ERROR nova.compute.manager [instance: 48908cb0-a989-4bad-84a4-25a4bdd3baeb] self._vmops.spawn(context, instance, image_meta, injected_files, [ 581.084727] env[63515]: ERROR nova.compute.manager [instance: 48908cb0-a989-4bad-84a4-25a4bdd3baeb] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 581.084727] env[63515]: ERROR nova.compute.manager [instance: 48908cb0-a989-4bad-84a4-25a4bdd3baeb] vm_ref = self.build_virtual_machine(instance, [ 581.085163] env[63515]: ERROR nova.compute.manager [instance: 48908cb0-a989-4bad-84a4-25a4bdd3baeb] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 581.085163] env[63515]: ERROR nova.compute.manager [instance: 48908cb0-a989-4bad-84a4-25a4bdd3baeb] vif_infos = vmwarevif.get_vif_info(self._session, [ 581.085163] env[63515]: ERROR nova.compute.manager [instance: 48908cb0-a989-4bad-84a4-25a4bdd3baeb] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 581.085163] env[63515]: ERROR nova.compute.manager [instance: 48908cb0-a989-4bad-84a4-25a4bdd3baeb] for vif in network_info: [ 581.085163] env[63515]: ERROR nova.compute.manager [instance: 48908cb0-a989-4bad-84a4-25a4bdd3baeb] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 581.085163] env[63515]: ERROR nova.compute.manager [instance: 48908cb0-a989-4bad-84a4-25a4bdd3baeb] return self._sync_wrapper(fn, *args, **kwargs) [ 581.085163] env[63515]: ERROR nova.compute.manager [instance: 48908cb0-a989-4bad-84a4-25a4bdd3baeb] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 581.085163] env[63515]: ERROR nova.compute.manager [instance: 48908cb0-a989-4bad-84a4-25a4bdd3baeb] self.wait() [ 581.085163] env[63515]: ERROR nova.compute.manager [instance: 48908cb0-a989-4bad-84a4-25a4bdd3baeb] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 581.085163] env[63515]: ERROR nova.compute.manager [instance: 48908cb0-a989-4bad-84a4-25a4bdd3baeb] self[:] = self._gt.wait() [ 581.085163] env[63515]: ERROR nova.compute.manager [instance: 48908cb0-a989-4bad-84a4-25a4bdd3baeb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 581.085163] env[63515]: ERROR nova.compute.manager [instance: 48908cb0-a989-4bad-84a4-25a4bdd3baeb] return self._exit_event.wait() [ 581.085163] env[63515]: ERROR nova.compute.manager [instance: 48908cb0-a989-4bad-84a4-25a4bdd3baeb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 581.085829] env[63515]: ERROR nova.compute.manager [instance: 48908cb0-a989-4bad-84a4-25a4bdd3baeb] result = hub.switch() [ 581.085829] env[63515]: ERROR nova.compute.manager [instance: 48908cb0-a989-4bad-84a4-25a4bdd3baeb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 581.085829] env[63515]: ERROR nova.compute.manager [instance: 48908cb0-a989-4bad-84a4-25a4bdd3baeb] return self.greenlet.switch() [ 581.085829] env[63515]: ERROR nova.compute.manager [instance: 48908cb0-a989-4bad-84a4-25a4bdd3baeb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 581.085829] env[63515]: ERROR nova.compute.manager [instance: 48908cb0-a989-4bad-84a4-25a4bdd3baeb] result = function(*args, **kwargs) [ 581.085829] env[63515]: ERROR nova.compute.manager [instance: 48908cb0-a989-4bad-84a4-25a4bdd3baeb] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 581.085829] env[63515]: ERROR nova.compute.manager [instance: 48908cb0-a989-4bad-84a4-25a4bdd3baeb] return func(*args, **kwargs) [ 581.085829] env[63515]: ERROR nova.compute.manager [instance: 48908cb0-a989-4bad-84a4-25a4bdd3baeb] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 581.085829] env[63515]: ERROR nova.compute.manager [instance: 48908cb0-a989-4bad-84a4-25a4bdd3baeb] raise e [ 581.085829] env[63515]: ERROR nova.compute.manager [instance: 48908cb0-a989-4bad-84a4-25a4bdd3baeb] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 581.085829] env[63515]: ERROR nova.compute.manager [instance: 48908cb0-a989-4bad-84a4-25a4bdd3baeb] nwinfo = self.network_api.allocate_for_instance( [ 581.085829] env[63515]: ERROR nova.compute.manager [instance: 48908cb0-a989-4bad-84a4-25a4bdd3baeb] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 581.085829] env[63515]: ERROR nova.compute.manager [instance: 48908cb0-a989-4bad-84a4-25a4bdd3baeb] created_port_ids = self._update_ports_for_instance( [ 581.086272] env[63515]: ERROR nova.compute.manager [instance: 48908cb0-a989-4bad-84a4-25a4bdd3baeb] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 581.086272] env[63515]: ERROR nova.compute.manager [instance: 48908cb0-a989-4bad-84a4-25a4bdd3baeb] with excutils.save_and_reraise_exception(): [ 581.086272] env[63515]: ERROR nova.compute.manager [instance: 48908cb0-a989-4bad-84a4-25a4bdd3baeb] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 581.086272] env[63515]: ERROR nova.compute.manager [instance: 48908cb0-a989-4bad-84a4-25a4bdd3baeb] self.force_reraise() [ 581.086272] env[63515]: ERROR nova.compute.manager [instance: 48908cb0-a989-4bad-84a4-25a4bdd3baeb] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 581.086272] env[63515]: ERROR nova.compute.manager [instance: 48908cb0-a989-4bad-84a4-25a4bdd3baeb] raise self.value [ 581.086272] env[63515]: ERROR nova.compute.manager [instance: 48908cb0-a989-4bad-84a4-25a4bdd3baeb] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 581.086272] env[63515]: ERROR nova.compute.manager [instance: 48908cb0-a989-4bad-84a4-25a4bdd3baeb] updated_port = self._update_port( [ 581.086272] env[63515]: ERROR nova.compute.manager [instance: 48908cb0-a989-4bad-84a4-25a4bdd3baeb] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 581.086272] env[63515]: ERROR nova.compute.manager [instance: 48908cb0-a989-4bad-84a4-25a4bdd3baeb] _ensure_no_port_binding_failure(port) [ 581.086272] env[63515]: ERROR nova.compute.manager [instance: 48908cb0-a989-4bad-84a4-25a4bdd3baeb] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 581.086272] env[63515]: ERROR nova.compute.manager [instance: 48908cb0-a989-4bad-84a4-25a4bdd3baeb] raise exception.PortBindingFailed(port_id=port['id']) [ 581.086642] env[63515]: ERROR nova.compute.manager [instance: 48908cb0-a989-4bad-84a4-25a4bdd3baeb] nova.exception.PortBindingFailed: Binding failed for port e7173eb8-86a4-4d43-8391-1d565f1dfaa9, please check neutron logs for more information. [ 581.086642] env[63515]: ERROR nova.compute.manager [instance: 48908cb0-a989-4bad-84a4-25a4bdd3baeb] [ 581.086642] env[63515]: INFO nova.compute.manager [None req-5e279136-7e72-43a2-85c4-9cd11158a72b tempest-ListImageFiltersTestJSON-1100101045 tempest-ListImageFiltersTestJSON-1100101045-project-member] [instance: 48908cb0-a989-4bad-84a4-25a4bdd3baeb] Terminating instance [ 581.087282] env[63515]: DEBUG oslo_concurrency.lockutils [None req-5e279136-7e72-43a2-85c4-9cd11158a72b tempest-ListImageFiltersTestJSON-1100101045 tempest-ListImageFiltersTestJSON-1100101045-project-member] Acquiring lock "refresh_cache-48908cb0-a989-4bad-84a4-25a4bdd3baeb" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 581.108994] env[63515]: DEBUG nova.network.neutron [req-6809d388-0d7d-46fe-adf3-c1d10bbb9a98 req-1f7249e4-7a80-4d0b-9cce-e7e10f725265 service nova] [instance: 48908cb0-a989-4bad-84a4-25a4bdd3baeb] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 581.309608] env[63515]: INFO nova.scheduler.client.report [None req-757a72d0-3dae-4296-830e-18ff5e948610 tempest-ServerDiagnosticsNegativeTest-1799249066 tempest-ServerDiagnosticsNegativeTest-1799249066-project-member] Deleted allocations for instance 1c5d87c7-6747-45a7-8488-6f107eeaeb78 [ 581.393253] env[63515]: DEBUG oslo_concurrency.lockutils [None req-a5c600bc-b1c1-4a6d-9386-2b0dbd0395e4 tempest-InstanceActionsNegativeTestJSON-1061405461 tempest-InstanceActionsNegativeTestJSON-1061405461-project-member] Releasing lock "refresh_cache-b641d2d1-de70-48b6-9137-fe18c3bc7511" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 581.393779] env[63515]: DEBUG nova.compute.manager [None req-a5c600bc-b1c1-4a6d-9386-2b0dbd0395e4 tempest-InstanceActionsNegativeTestJSON-1061405461 tempest-InstanceActionsNegativeTestJSON-1061405461-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63515) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 581.393909] env[63515]: DEBUG nova.compute.manager [None req-a5c600bc-b1c1-4a6d-9386-2b0dbd0395e4 tempest-InstanceActionsNegativeTestJSON-1061405461 tempest-InstanceActionsNegativeTestJSON-1061405461-project-member] [instance: b641d2d1-de70-48b6-9137-fe18c3bc7511] Deallocating network for instance {{(pid=63515) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 581.394178] env[63515]: DEBUG nova.network.neutron [None req-a5c600bc-b1c1-4a6d-9386-2b0dbd0395e4 tempest-InstanceActionsNegativeTestJSON-1061405461 tempest-InstanceActionsNegativeTestJSON-1061405461-project-member] [instance: b641d2d1-de70-48b6-9137-fe18c3bc7511] deallocate_for_instance() {{(pid=63515) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 581.495215] env[63515]: DEBUG nova.network.neutron [None req-a5c600bc-b1c1-4a6d-9386-2b0dbd0395e4 tempest-InstanceActionsNegativeTestJSON-1061405461 tempest-InstanceActionsNegativeTestJSON-1061405461-project-member] [instance: b641d2d1-de70-48b6-9137-fe18c3bc7511] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 581.530496] env[63515]: DEBUG oslo_concurrency.lockutils [None req-b1f9566f-fcad-4d5d-9d53-b04b591d6c49 tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 581.582608] env[63515]: DEBUG nova.scheduler.client.report [None req-50ea3e37-671c-4a4b-9c1a-d782ff1b578c tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 581.611103] env[63515]: DEBUG oslo_concurrency.lockutils [req-6809d388-0d7d-46fe-adf3-c1d10bbb9a98 req-1f7249e4-7a80-4d0b-9cce-e7e10f725265 service nova] Releasing lock "refresh_cache-48908cb0-a989-4bad-84a4-25a4bdd3baeb" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 581.611583] env[63515]: DEBUG oslo_concurrency.lockutils [None req-5e279136-7e72-43a2-85c4-9cd11158a72b tempest-ListImageFiltersTestJSON-1100101045 tempest-ListImageFiltersTestJSON-1100101045-project-member] Acquired lock "refresh_cache-48908cb0-a989-4bad-84a4-25a4bdd3baeb" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 581.611738] env[63515]: DEBUG nova.network.neutron [None req-5e279136-7e72-43a2-85c4-9cd11158a72b tempest-ListImageFiltersTestJSON-1100101045 tempest-ListImageFiltersTestJSON-1100101045-project-member] [instance: 48908cb0-a989-4bad-84a4-25a4bdd3baeb] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 581.824337] env[63515]: DEBUG oslo_concurrency.lockutils [None req-757a72d0-3dae-4296-830e-18ff5e948610 tempest-ServerDiagnosticsNegativeTest-1799249066 tempest-ServerDiagnosticsNegativeTest-1799249066-project-member] Lock "1c5d87c7-6747-45a7-8488-6f107eeaeb78" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 44.577s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 581.998600] env[63515]: DEBUG nova.network.neutron [None req-a5c600bc-b1c1-4a6d-9386-2b0dbd0395e4 tempest-InstanceActionsNegativeTestJSON-1061405461 tempest-InstanceActionsNegativeTestJSON-1061405461-project-member] [instance: b641d2d1-de70-48b6-9137-fe18c3bc7511] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 582.081241] env[63515]: INFO nova.compute.manager [None req-ae27ad25-8527-4c20-9aca-0f14006c890e tempest-ServersAdmin275Test-220579119 tempest-ServersAdmin275Test-220579119-project-admin] [instance: 93b28142-8454-43c0-b0a7-d61aa95c8fc8] Rebuilding instance [ 582.088618] env[63515]: DEBUG oslo_concurrency.lockutils [None req-50ea3e37-671c-4a4b-9c1a-d782ff1b578c tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.108s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 582.089421] env[63515]: ERROR nova.compute.manager [None req-50ea3e37-671c-4a4b-9c1a-d782ff1b578c tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] [instance: cfc6c8d5-c9e8-46fb-99e7-d245664f9652] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 589525f2-8185-44ca-9cdc-7c1b03dd8ee3, please check neutron logs for more information. [ 582.089421] env[63515]: ERROR nova.compute.manager [instance: cfc6c8d5-c9e8-46fb-99e7-d245664f9652] Traceback (most recent call last): [ 582.089421] env[63515]: ERROR nova.compute.manager [instance: cfc6c8d5-c9e8-46fb-99e7-d245664f9652] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 582.089421] env[63515]: ERROR nova.compute.manager [instance: cfc6c8d5-c9e8-46fb-99e7-d245664f9652] self.driver.spawn(context, instance, image_meta, [ 582.089421] env[63515]: ERROR nova.compute.manager [instance: cfc6c8d5-c9e8-46fb-99e7-d245664f9652] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 582.089421] env[63515]: ERROR nova.compute.manager [instance: cfc6c8d5-c9e8-46fb-99e7-d245664f9652] self._vmops.spawn(context, instance, image_meta, injected_files, [ 582.089421] env[63515]: ERROR nova.compute.manager [instance: cfc6c8d5-c9e8-46fb-99e7-d245664f9652] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 582.089421] env[63515]: ERROR nova.compute.manager [instance: cfc6c8d5-c9e8-46fb-99e7-d245664f9652] vm_ref = self.build_virtual_machine(instance, [ 582.089421] env[63515]: ERROR nova.compute.manager [instance: cfc6c8d5-c9e8-46fb-99e7-d245664f9652] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 582.089421] env[63515]: ERROR nova.compute.manager [instance: cfc6c8d5-c9e8-46fb-99e7-d245664f9652] vif_infos = vmwarevif.get_vif_info(self._session, [ 582.089421] env[63515]: ERROR nova.compute.manager [instance: cfc6c8d5-c9e8-46fb-99e7-d245664f9652] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 582.089817] env[63515]: ERROR nova.compute.manager [instance: cfc6c8d5-c9e8-46fb-99e7-d245664f9652] for vif in network_info: [ 582.089817] env[63515]: ERROR nova.compute.manager [instance: cfc6c8d5-c9e8-46fb-99e7-d245664f9652] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 582.089817] env[63515]: ERROR nova.compute.manager [instance: cfc6c8d5-c9e8-46fb-99e7-d245664f9652] return self._sync_wrapper(fn, *args, **kwargs) [ 582.089817] env[63515]: ERROR nova.compute.manager [instance: cfc6c8d5-c9e8-46fb-99e7-d245664f9652] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 582.089817] env[63515]: ERROR nova.compute.manager [instance: cfc6c8d5-c9e8-46fb-99e7-d245664f9652] self.wait() [ 582.089817] env[63515]: ERROR nova.compute.manager [instance: cfc6c8d5-c9e8-46fb-99e7-d245664f9652] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 582.089817] env[63515]: ERROR nova.compute.manager [instance: cfc6c8d5-c9e8-46fb-99e7-d245664f9652] self[:] = self._gt.wait() [ 582.089817] env[63515]: ERROR nova.compute.manager [instance: cfc6c8d5-c9e8-46fb-99e7-d245664f9652] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 582.089817] env[63515]: ERROR nova.compute.manager [instance: cfc6c8d5-c9e8-46fb-99e7-d245664f9652] return self._exit_event.wait() [ 582.089817] env[63515]: ERROR nova.compute.manager [instance: cfc6c8d5-c9e8-46fb-99e7-d245664f9652] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 582.089817] env[63515]: ERROR nova.compute.manager [instance: cfc6c8d5-c9e8-46fb-99e7-d245664f9652] result = hub.switch() [ 582.089817] env[63515]: ERROR nova.compute.manager [instance: cfc6c8d5-c9e8-46fb-99e7-d245664f9652] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 582.089817] env[63515]: ERROR nova.compute.manager [instance: cfc6c8d5-c9e8-46fb-99e7-d245664f9652] return self.greenlet.switch() [ 582.090225] env[63515]: ERROR nova.compute.manager [instance: cfc6c8d5-c9e8-46fb-99e7-d245664f9652] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 582.090225] env[63515]: ERROR nova.compute.manager [instance: cfc6c8d5-c9e8-46fb-99e7-d245664f9652] result = function(*args, **kwargs) [ 582.090225] env[63515]: ERROR nova.compute.manager [instance: cfc6c8d5-c9e8-46fb-99e7-d245664f9652] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 582.090225] env[63515]: ERROR nova.compute.manager [instance: cfc6c8d5-c9e8-46fb-99e7-d245664f9652] return func(*args, **kwargs) [ 582.090225] env[63515]: ERROR nova.compute.manager [instance: cfc6c8d5-c9e8-46fb-99e7-d245664f9652] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 582.090225] env[63515]: ERROR nova.compute.manager [instance: cfc6c8d5-c9e8-46fb-99e7-d245664f9652] raise e [ 582.090225] env[63515]: ERROR nova.compute.manager [instance: cfc6c8d5-c9e8-46fb-99e7-d245664f9652] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 582.090225] env[63515]: ERROR nova.compute.manager [instance: cfc6c8d5-c9e8-46fb-99e7-d245664f9652] nwinfo = self.network_api.allocate_for_instance( [ 582.090225] env[63515]: ERROR nova.compute.manager [instance: cfc6c8d5-c9e8-46fb-99e7-d245664f9652] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 582.090225] env[63515]: ERROR nova.compute.manager [instance: cfc6c8d5-c9e8-46fb-99e7-d245664f9652] created_port_ids = self._update_ports_for_instance( [ 582.090225] env[63515]: ERROR nova.compute.manager [instance: cfc6c8d5-c9e8-46fb-99e7-d245664f9652] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 582.090225] env[63515]: ERROR nova.compute.manager [instance: cfc6c8d5-c9e8-46fb-99e7-d245664f9652] with excutils.save_and_reraise_exception(): [ 582.090225] env[63515]: ERROR nova.compute.manager [instance: cfc6c8d5-c9e8-46fb-99e7-d245664f9652] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 582.090631] env[63515]: ERROR nova.compute.manager [instance: cfc6c8d5-c9e8-46fb-99e7-d245664f9652] self.force_reraise() [ 582.090631] env[63515]: ERROR nova.compute.manager [instance: cfc6c8d5-c9e8-46fb-99e7-d245664f9652] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 582.090631] env[63515]: ERROR nova.compute.manager [instance: cfc6c8d5-c9e8-46fb-99e7-d245664f9652] raise self.value [ 582.090631] env[63515]: ERROR nova.compute.manager [instance: cfc6c8d5-c9e8-46fb-99e7-d245664f9652] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 582.090631] env[63515]: ERROR nova.compute.manager [instance: cfc6c8d5-c9e8-46fb-99e7-d245664f9652] updated_port = self._update_port( [ 582.090631] env[63515]: ERROR nova.compute.manager [instance: cfc6c8d5-c9e8-46fb-99e7-d245664f9652] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 582.090631] env[63515]: ERROR nova.compute.manager [instance: cfc6c8d5-c9e8-46fb-99e7-d245664f9652] _ensure_no_port_binding_failure(port) [ 582.090631] env[63515]: ERROR nova.compute.manager [instance: cfc6c8d5-c9e8-46fb-99e7-d245664f9652] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 582.090631] env[63515]: ERROR nova.compute.manager [instance: cfc6c8d5-c9e8-46fb-99e7-d245664f9652] raise exception.PortBindingFailed(port_id=port['id']) [ 582.090631] env[63515]: ERROR nova.compute.manager [instance: cfc6c8d5-c9e8-46fb-99e7-d245664f9652] nova.exception.PortBindingFailed: Binding failed for port 589525f2-8185-44ca-9cdc-7c1b03dd8ee3, please check neutron logs for more information. [ 582.090631] env[63515]: ERROR nova.compute.manager [instance: cfc6c8d5-c9e8-46fb-99e7-d245664f9652] [ 582.093037] env[63515]: DEBUG nova.compute.utils [None req-50ea3e37-671c-4a4b-9c1a-d782ff1b578c tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] [instance: cfc6c8d5-c9e8-46fb-99e7-d245664f9652] Binding failed for port 589525f2-8185-44ca-9cdc-7c1b03dd8ee3, please check neutron logs for more information. {{(pid=63515) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 582.095076] env[63515]: DEBUG nova.compute.manager [None req-50ea3e37-671c-4a4b-9c1a-d782ff1b578c tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] [instance: cfc6c8d5-c9e8-46fb-99e7-d245664f9652] Build of instance cfc6c8d5-c9e8-46fb-99e7-d245664f9652 was re-scheduled: Binding failed for port 589525f2-8185-44ca-9cdc-7c1b03dd8ee3, please check neutron logs for more information. {{(pid=63515) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 582.098017] env[63515]: DEBUG nova.compute.manager [None req-50ea3e37-671c-4a4b-9c1a-d782ff1b578c tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] [instance: cfc6c8d5-c9e8-46fb-99e7-d245664f9652] Unplugging VIFs for instance {{(pid=63515) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 582.098017] env[63515]: DEBUG oslo_concurrency.lockutils [None req-50ea3e37-671c-4a4b-9c1a-d782ff1b578c tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Acquiring lock "refresh_cache-cfc6c8d5-c9e8-46fb-99e7-d245664f9652" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 582.098017] env[63515]: DEBUG oslo_concurrency.lockutils [None req-50ea3e37-671c-4a4b-9c1a-d782ff1b578c tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Acquired lock "refresh_cache-cfc6c8d5-c9e8-46fb-99e7-d245664f9652" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 582.098017] env[63515]: DEBUG nova.network.neutron [None req-50ea3e37-671c-4a4b-9c1a-d782ff1b578c tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] [instance: cfc6c8d5-c9e8-46fb-99e7-d245664f9652] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 582.099442] env[63515]: DEBUG oslo_concurrency.lockutils [None req-12d73cc4-118d-457e-b393-cc4362d15b0d tempest-ListImageFiltersTestJSON-1100101045 tempest-ListImageFiltersTestJSON-1100101045-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.959s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 582.101289] env[63515]: INFO nova.compute.claims [None req-12d73cc4-118d-457e-b393-cc4362d15b0d tempest-ListImageFiltersTestJSON-1100101045 tempest-ListImageFiltersTestJSON-1100101045-project-member] [instance: ab1539ba-33a4-4139-a2f3-1f8c507beede] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 582.162087] env[63515]: DEBUG nova.compute.manager [None req-ae27ad25-8527-4c20-9aca-0f14006c890e tempest-ServersAdmin275Test-220579119 tempest-ServersAdmin275Test-220579119-project-admin] [instance: 93b28142-8454-43c0-b0a7-d61aa95c8fc8] Checking state {{(pid=63515) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 582.162087] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5867ec0f-3603-42fe-9ff1-252469d5de21 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 582.168882] env[63515]: DEBUG nova.network.neutron [None req-5e279136-7e72-43a2-85c4-9cd11158a72b tempest-ListImageFiltersTestJSON-1100101045 tempest-ListImageFiltersTestJSON-1100101045-project-member] [instance: 48908cb0-a989-4bad-84a4-25a4bdd3baeb] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 582.287101] env[63515]: DEBUG nova.network.neutron [None req-5e279136-7e72-43a2-85c4-9cd11158a72b tempest-ListImageFiltersTestJSON-1100101045 tempest-ListImageFiltersTestJSON-1100101045-project-member] [instance: 48908cb0-a989-4bad-84a4-25a4bdd3baeb] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 582.331072] env[63515]: DEBUG nova.compute.manager [None req-24636c84-3869-4a1b-acfb-9c10d33a9826 tempest-ServersWithSpecificFlavorTestJSON-691988195 tempest-ServersWithSpecificFlavorTestJSON-691988195-project-member] [instance: ebc737a3-e6cd-4f4c-a4b6-fce5924de4c6] Starting instance... {{(pid=63515) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 582.501576] env[63515]: INFO nova.compute.manager [None req-a5c600bc-b1c1-4a6d-9386-2b0dbd0395e4 tempest-InstanceActionsNegativeTestJSON-1061405461 tempest-InstanceActionsNegativeTestJSON-1061405461-project-member] [instance: b641d2d1-de70-48b6-9137-fe18c3bc7511] Took 1.11 seconds to deallocate network for instance. [ 582.550155] env[63515]: DEBUG nova.compute.manager [req-48c124c4-0123-44f7-aa7c-2d29e7a54b89 req-031e7fab-323a-4668-bcfa-fcff18c991e4 service nova] [instance: 48908cb0-a989-4bad-84a4-25a4bdd3baeb] Received event network-vif-deleted-e7173eb8-86a4-4d43-8391-1d565f1dfaa9 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 582.638828] env[63515]: DEBUG nova.network.neutron [None req-50ea3e37-671c-4a4b-9c1a-d782ff1b578c tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] [instance: cfc6c8d5-c9e8-46fb-99e7-d245664f9652] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 582.681168] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-ae27ad25-8527-4c20-9aca-0f14006c890e tempest-ServersAdmin275Test-220579119 tempest-ServersAdmin275Test-220579119-project-admin] [instance: 93b28142-8454-43c0-b0a7-d61aa95c8fc8] Powering off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 582.681523] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d9089050-d761-4f19-a403-0ab46db28cf3 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 582.690301] env[63515]: DEBUG oslo_vmware.api [None req-ae27ad25-8527-4c20-9aca-0f14006c890e tempest-ServersAdmin275Test-220579119 tempest-ServersAdmin275Test-220579119-project-admin] Waiting for the task: (returnval){ [ 582.690301] env[63515]: value = "task-1110874" [ 582.690301] env[63515]: _type = "Task" [ 582.690301] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 582.701384] env[63515]: DEBUG oslo_vmware.api [None req-ae27ad25-8527-4c20-9aca-0f14006c890e tempest-ServersAdmin275Test-220579119 tempest-ServersAdmin275Test-220579119-project-admin] Task: {'id': task-1110874, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 582.718661] env[63515]: DEBUG nova.network.neutron [None req-50ea3e37-671c-4a4b-9c1a-d782ff1b578c tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] [instance: cfc6c8d5-c9e8-46fb-99e7-d245664f9652] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 582.789284] env[63515]: DEBUG oslo_concurrency.lockutils [None req-5e279136-7e72-43a2-85c4-9cd11158a72b tempest-ListImageFiltersTestJSON-1100101045 tempest-ListImageFiltersTestJSON-1100101045-project-member] Releasing lock "refresh_cache-48908cb0-a989-4bad-84a4-25a4bdd3baeb" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 582.789749] env[63515]: DEBUG nova.compute.manager [None req-5e279136-7e72-43a2-85c4-9cd11158a72b tempest-ListImageFiltersTestJSON-1100101045 tempest-ListImageFiltersTestJSON-1100101045-project-member] [instance: 48908cb0-a989-4bad-84a4-25a4bdd3baeb] Start destroying the instance on the hypervisor. {{(pid=63515) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 582.789946] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-5e279136-7e72-43a2-85c4-9cd11158a72b tempest-ListImageFiltersTestJSON-1100101045 tempest-ListImageFiltersTestJSON-1100101045-project-member] [instance: 48908cb0-a989-4bad-84a4-25a4bdd3baeb] Destroying instance {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 582.790280] env[63515]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-85c5917b-1451-468e-aebf-b32518b1fb6a {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 582.799854] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41fa8d7f-9810-499c-837c-f37a0c947a42 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 582.823944] env[63515]: WARNING nova.virt.vmwareapi.vmops [None req-5e279136-7e72-43a2-85c4-9cd11158a72b tempest-ListImageFiltersTestJSON-1100101045 tempest-ListImageFiltersTestJSON-1100101045-project-member] [instance: 48908cb0-a989-4bad-84a4-25a4bdd3baeb] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 48908cb0-a989-4bad-84a4-25a4bdd3baeb could not be found. [ 582.824207] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-5e279136-7e72-43a2-85c4-9cd11158a72b tempest-ListImageFiltersTestJSON-1100101045 tempest-ListImageFiltersTestJSON-1100101045-project-member] [instance: 48908cb0-a989-4bad-84a4-25a4bdd3baeb] Instance destroyed {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 582.824441] env[63515]: INFO nova.compute.manager [None req-5e279136-7e72-43a2-85c4-9cd11158a72b tempest-ListImageFiltersTestJSON-1100101045 tempest-ListImageFiltersTestJSON-1100101045-project-member] [instance: 48908cb0-a989-4bad-84a4-25a4bdd3baeb] Took 0.03 seconds to destroy the instance on the hypervisor. [ 582.824695] env[63515]: DEBUG oslo.service.loopingcall [None req-5e279136-7e72-43a2-85c4-9cd11158a72b tempest-ListImageFiltersTestJSON-1100101045 tempest-ListImageFiltersTestJSON-1100101045-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 582.824898] env[63515]: DEBUG nova.compute.manager [-] [instance: 48908cb0-a989-4bad-84a4-25a4bdd3baeb] Deallocating network for instance {{(pid=63515) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 582.824993] env[63515]: DEBUG nova.network.neutron [-] [instance: 48908cb0-a989-4bad-84a4-25a4bdd3baeb] deallocate_for_instance() {{(pid=63515) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 582.844848] env[63515]: DEBUG nova.network.neutron [-] [instance: 48908cb0-a989-4bad-84a4-25a4bdd3baeb] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 582.851928] env[63515]: DEBUG oslo_concurrency.lockutils [None req-24636c84-3869-4a1b-acfb-9c10d33a9826 tempest-ServersWithSpecificFlavorTestJSON-691988195 tempest-ServersWithSpecificFlavorTestJSON-691988195-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 583.206600] env[63515]: DEBUG oslo_vmware.api [None req-ae27ad25-8527-4c20-9aca-0f14006c890e tempest-ServersAdmin275Test-220579119 tempest-ServersAdmin275Test-220579119-project-admin] Task: {'id': task-1110874, 'name': PowerOffVM_Task, 'duration_secs': 0.232376} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 583.207724] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-ae27ad25-8527-4c20-9aca-0f14006c890e tempest-ServersAdmin275Test-220579119 tempest-ServersAdmin275Test-220579119-project-admin] [instance: 93b28142-8454-43c0-b0a7-d61aa95c8fc8] Powered off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 583.207724] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-ae27ad25-8527-4c20-9aca-0f14006c890e tempest-ServersAdmin275Test-220579119 tempest-ServersAdmin275Test-220579119-project-admin] [instance: 93b28142-8454-43c0-b0a7-d61aa95c8fc8] Destroying instance {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 583.208075] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e363920-a327-41bc-9c2b-31fa317a8166 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 583.217943] env[63515]: DEBUG oslo_concurrency.lockutils [None req-9ecc0501-dc42-4f19-a331-8b337de58eaf tempest-AttachInterfacesUnderV243Test-1943594058 tempest-AttachInterfacesUnderV243Test-1943594058-project-member] Acquiring lock "b911a5b5-9617-4fb3-9b5e-fb8c492e4931" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 583.218192] env[63515]: DEBUG oslo_concurrency.lockutils [None req-9ecc0501-dc42-4f19-a331-8b337de58eaf tempest-AttachInterfacesUnderV243Test-1943594058 tempest-AttachInterfacesUnderV243Test-1943594058-project-member] Lock "b911a5b5-9617-4fb3-9b5e-fb8c492e4931" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 583.218423] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-ae27ad25-8527-4c20-9aca-0f14006c890e tempest-ServersAdmin275Test-220579119 tempest-ServersAdmin275Test-220579119-project-admin] [instance: 93b28142-8454-43c0-b0a7-d61aa95c8fc8] Unregistering the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 583.220950] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-396ad546-86e5-4334-a19c-18111fd12424 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 583.225541] env[63515]: DEBUG oslo_concurrency.lockutils [None req-50ea3e37-671c-4a4b-9c1a-d782ff1b578c tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Releasing lock "refresh_cache-cfc6c8d5-c9e8-46fb-99e7-d245664f9652" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 583.225723] env[63515]: DEBUG nova.compute.manager [None req-50ea3e37-671c-4a4b-9c1a-d782ff1b578c tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63515) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 583.225944] env[63515]: DEBUG nova.compute.manager [None req-50ea3e37-671c-4a4b-9c1a-d782ff1b578c tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] [instance: cfc6c8d5-c9e8-46fb-99e7-d245664f9652] Deallocating network for instance {{(pid=63515) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 583.226070] env[63515]: DEBUG nova.network.neutron [None req-50ea3e37-671c-4a4b-9c1a-d782ff1b578c tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] [instance: cfc6c8d5-c9e8-46fb-99e7-d245664f9652] deallocate_for_instance() {{(pid=63515) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 583.247078] env[63515]: DEBUG nova.network.neutron [None req-50ea3e37-671c-4a4b-9c1a-d782ff1b578c tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] [instance: cfc6c8d5-c9e8-46fb-99e7-d245664f9652] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 583.250420] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-ae27ad25-8527-4c20-9aca-0f14006c890e tempest-ServersAdmin275Test-220579119 tempest-ServersAdmin275Test-220579119-project-admin] [instance: 93b28142-8454-43c0-b0a7-d61aa95c8fc8] Unregistered the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 583.250420] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-ae27ad25-8527-4c20-9aca-0f14006c890e tempest-ServersAdmin275Test-220579119 tempest-ServersAdmin275Test-220579119-project-admin] [instance: 93b28142-8454-43c0-b0a7-d61aa95c8fc8] Deleting contents of the VM from datastore datastore1 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 583.250627] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-ae27ad25-8527-4c20-9aca-0f14006c890e tempest-ServersAdmin275Test-220579119 tempest-ServersAdmin275Test-220579119-project-admin] Deleting the datastore file [datastore1] 93b28142-8454-43c0-b0a7-d61aa95c8fc8 {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 583.251381] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ab20636e-59d9-4eb7-92e7-154ebdb030a7 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 583.259137] env[63515]: DEBUG oslo_vmware.api [None req-ae27ad25-8527-4c20-9aca-0f14006c890e tempest-ServersAdmin275Test-220579119 tempest-ServersAdmin275Test-220579119-project-admin] Waiting for the task: (returnval){ [ 583.259137] env[63515]: value = "task-1110876" [ 583.259137] env[63515]: _type = "Task" [ 583.259137] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 583.271607] env[63515]: DEBUG oslo_vmware.api [None req-ae27ad25-8527-4c20-9aca-0f14006c890e tempest-ServersAdmin275Test-220579119 tempest-ServersAdmin275Test-220579119-project-admin] Task: {'id': task-1110876, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 583.347463] env[63515]: DEBUG nova.network.neutron [-] [instance: 48908cb0-a989-4bad-84a4-25a4bdd3baeb] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 583.533087] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f7f7939-31a9-4483-9dee-661d0968838a {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 583.541067] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9991667f-4c31-42b0-8c10-08ea27a6ab5f {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 583.545431] env[63515]: INFO nova.scheduler.client.report [None req-a5c600bc-b1c1-4a6d-9386-2b0dbd0395e4 tempest-InstanceActionsNegativeTestJSON-1061405461 tempest-InstanceActionsNegativeTestJSON-1061405461-project-member] Deleted allocations for instance b641d2d1-de70-48b6-9137-fe18c3bc7511 [ 583.581684] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d32a811d-febf-401c-bfb1-573618fc0d05 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 583.589791] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1377931c-827d-4437-a7b2-abc78f57b7da {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 583.607310] env[63515]: DEBUG nova.compute.provider_tree [None req-12d73cc4-118d-457e-b393-cc4362d15b0d tempest-ListImageFiltersTestJSON-1100101045 tempest-ListImageFiltersTestJSON-1100101045-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 583.704948] env[63515]: DEBUG oslo_concurrency.lockutils [None req-95251dc9-cf22-4a89-a3b0-fce88f9e12db tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Acquiring lock "3ad04a9f-62cf-4b1c-aab2-7dc4a219edc6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 583.705204] env[63515]: DEBUG oslo_concurrency.lockutils [None req-95251dc9-cf22-4a89-a3b0-fce88f9e12db tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Lock "3ad04a9f-62cf-4b1c-aab2-7dc4a219edc6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 583.753563] env[63515]: DEBUG nova.network.neutron [None req-50ea3e37-671c-4a4b-9c1a-d782ff1b578c tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] [instance: cfc6c8d5-c9e8-46fb-99e7-d245664f9652] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 583.771315] env[63515]: DEBUG oslo_vmware.api [None req-ae27ad25-8527-4c20-9aca-0f14006c890e tempest-ServersAdmin275Test-220579119 tempest-ServersAdmin275Test-220579119-project-admin] Task: {'id': task-1110876, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.11545} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 583.771570] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-ae27ad25-8527-4c20-9aca-0f14006c890e tempest-ServersAdmin275Test-220579119 tempest-ServersAdmin275Test-220579119-project-admin] Deleted the datastore file {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 583.771749] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-ae27ad25-8527-4c20-9aca-0f14006c890e tempest-ServersAdmin275Test-220579119 tempest-ServersAdmin275Test-220579119-project-admin] [instance: 93b28142-8454-43c0-b0a7-d61aa95c8fc8] Deleted contents of the VM from datastore datastore1 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 583.771921] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-ae27ad25-8527-4c20-9aca-0f14006c890e tempest-ServersAdmin275Test-220579119 tempest-ServersAdmin275Test-220579119-project-admin] [instance: 93b28142-8454-43c0-b0a7-d61aa95c8fc8] Instance destroyed {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 583.849580] env[63515]: INFO nova.compute.manager [-] [instance: 48908cb0-a989-4bad-84a4-25a4bdd3baeb] Took 1.02 seconds to deallocate network for instance. [ 583.851900] env[63515]: DEBUG nova.compute.claims [None req-5e279136-7e72-43a2-85c4-9cd11158a72b tempest-ListImageFiltersTestJSON-1100101045 tempest-ListImageFiltersTestJSON-1100101045-project-member] [instance: 48908cb0-a989-4bad-84a4-25a4bdd3baeb] Aborting claim: {{(pid=63515) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 583.852097] env[63515]: DEBUG oslo_concurrency.lockutils [None req-5e279136-7e72-43a2-85c4-9cd11158a72b tempest-ListImageFiltersTestJSON-1100101045 tempest-ListImageFiltersTestJSON-1100101045-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 584.053892] env[63515]: DEBUG oslo_concurrency.lockutils [None req-a5c600bc-b1c1-4a6d-9386-2b0dbd0395e4 tempest-InstanceActionsNegativeTestJSON-1061405461 tempest-InstanceActionsNegativeTestJSON-1061405461-project-member] Lock "b641d2d1-de70-48b6-9137-fe18c3bc7511" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 45.923s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 584.110531] env[63515]: DEBUG nova.scheduler.client.report [None req-12d73cc4-118d-457e-b393-cc4362d15b0d tempest-ListImageFiltersTestJSON-1100101045 tempest-ListImageFiltersTestJSON-1100101045-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 584.257700] env[63515]: INFO nova.compute.manager [None req-50ea3e37-671c-4a4b-9c1a-d782ff1b578c tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] [instance: cfc6c8d5-c9e8-46fb-99e7-d245664f9652] Took 1.03 seconds to deallocate network for instance. [ 584.562959] env[63515]: DEBUG nova.compute.manager [None req-d9ffa9d5-6856-47e5-8926-d090bbea2c09 tempest-DeleteServersAdminTestJSON-944542165 tempest-DeleteServersAdminTestJSON-944542165-project-member] [instance: ae5c6e69-a885-48e3-9c9e-017f33430edd] Starting instance... {{(pid=63515) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 584.619301] env[63515]: DEBUG oslo_concurrency.lockutils [None req-12d73cc4-118d-457e-b393-cc4362d15b0d tempest-ListImageFiltersTestJSON-1100101045 tempest-ListImageFiltersTestJSON-1100101045-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.520s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 584.619911] env[63515]: DEBUG nova.compute.manager [None req-12d73cc4-118d-457e-b393-cc4362d15b0d tempest-ListImageFiltersTestJSON-1100101045 tempest-ListImageFiltersTestJSON-1100101045-project-member] [instance: ab1539ba-33a4-4139-a2f3-1f8c507beede] Start building networks asynchronously for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 584.624647] env[63515]: DEBUG oslo_concurrency.lockutils [None req-eaa7f189-a861-45f8-872f-29d870dc784b tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 13.234s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 584.840695] env[63515]: DEBUG nova.virt.hardware [None req-ae27ad25-8527-4c20-9aca-0f14006c890e tempest-ServersAdmin275Test-220579119 tempest-ServersAdmin275Test-220579119-project-admin] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T02:52:14Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T02:51:56Z,direct_url=,disk_format='vmdk',id=8a120570-cb06-4099-b262-554ca0ad15c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b569255cc43e42e7ae3f2b4ad37c6ef4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T02:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 584.840940] env[63515]: DEBUG nova.virt.hardware [None req-ae27ad25-8527-4c20-9aca-0f14006c890e tempest-ServersAdmin275Test-220579119 tempest-ServersAdmin275Test-220579119-project-admin] Flavor limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 584.841125] env[63515]: DEBUG nova.virt.hardware [None req-ae27ad25-8527-4c20-9aca-0f14006c890e tempest-ServersAdmin275Test-220579119 tempest-ServersAdmin275Test-220579119-project-admin] Image limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 584.841369] env[63515]: DEBUG nova.virt.hardware [None req-ae27ad25-8527-4c20-9aca-0f14006c890e tempest-ServersAdmin275Test-220579119 tempest-ServersAdmin275Test-220579119-project-admin] Flavor pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 584.841528] env[63515]: DEBUG nova.virt.hardware [None req-ae27ad25-8527-4c20-9aca-0f14006c890e tempest-ServersAdmin275Test-220579119 tempest-ServersAdmin275Test-220579119-project-admin] Image pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 584.841659] env[63515]: DEBUG nova.virt.hardware [None req-ae27ad25-8527-4c20-9aca-0f14006c890e tempest-ServersAdmin275Test-220579119 tempest-ServersAdmin275Test-220579119-project-admin] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 584.841864] env[63515]: DEBUG nova.virt.hardware [None req-ae27ad25-8527-4c20-9aca-0f14006c890e tempest-ServersAdmin275Test-220579119 tempest-ServersAdmin275Test-220579119-project-admin] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 584.842189] env[63515]: DEBUG nova.virt.hardware [None req-ae27ad25-8527-4c20-9aca-0f14006c890e tempest-ServersAdmin275Test-220579119 tempest-ServersAdmin275Test-220579119-project-admin] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 584.842445] env[63515]: DEBUG nova.virt.hardware [None req-ae27ad25-8527-4c20-9aca-0f14006c890e tempest-ServersAdmin275Test-220579119 tempest-ServersAdmin275Test-220579119-project-admin] Got 1 possible topologies {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 584.842645] env[63515]: DEBUG nova.virt.hardware [None req-ae27ad25-8527-4c20-9aca-0f14006c890e tempest-ServersAdmin275Test-220579119 tempest-ServersAdmin275Test-220579119-project-admin] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 584.842857] env[63515]: DEBUG nova.virt.hardware [None req-ae27ad25-8527-4c20-9aca-0f14006c890e tempest-ServersAdmin275Test-220579119 tempest-ServersAdmin275Test-220579119-project-admin] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 584.843901] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42bf9b5b-870d-4a8c-8e9c-99ada8f28cf3 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 584.854279] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a0e9ac3-5cf0-49b9-870f-5e102b02bef7 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 584.870777] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-ae27ad25-8527-4c20-9aca-0f14006c890e tempest-ServersAdmin275Test-220579119 tempest-ServersAdmin275Test-220579119-project-admin] [instance: 93b28142-8454-43c0-b0a7-d61aa95c8fc8] Instance VIF info [] {{(pid=63515) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 584.876898] env[63515]: DEBUG oslo.service.loopingcall [None req-ae27ad25-8527-4c20-9aca-0f14006c890e tempest-ServersAdmin275Test-220579119 tempest-ServersAdmin275Test-220579119-project-admin] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 584.878641] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 93b28142-8454-43c0-b0a7-d61aa95c8fc8] Creating VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 584.881597] env[63515]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-8845a546-6ff1-4338-8bbf-6b0097b665be {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 584.899825] env[63515]: DEBUG oslo_concurrency.lockutils [None req-c7572d27-5b85-4acb-89ec-c56462911d65 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] Acquiring lock "5b42f744-fdd6-45b1-8563-896869648c23" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 584.899825] env[63515]: DEBUG oslo_concurrency.lockutils [None req-c7572d27-5b85-4acb-89ec-c56462911d65 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] Lock "5b42f744-fdd6-45b1-8563-896869648c23" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 584.909873] env[63515]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 584.909873] env[63515]: value = "task-1110878" [ 584.909873] env[63515]: _type = "Task" [ 584.909873] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 584.919042] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1110878, 'name': CreateVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 585.087250] env[63515]: DEBUG oslo_concurrency.lockutils [None req-d9ffa9d5-6856-47e5-8926-d090bbea2c09 tempest-DeleteServersAdminTestJSON-944542165 tempest-DeleteServersAdminTestJSON-944542165-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 585.127420] env[63515]: DEBUG nova.compute.utils [None req-12d73cc4-118d-457e-b393-cc4362d15b0d tempest-ListImageFiltersTestJSON-1100101045 tempest-ListImageFiltersTestJSON-1100101045-project-member] Using /dev/sd instead of None {{(pid=63515) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 585.128987] env[63515]: DEBUG nova.compute.manager [None req-12d73cc4-118d-457e-b393-cc4362d15b0d tempest-ListImageFiltersTestJSON-1100101045 tempest-ListImageFiltersTestJSON-1100101045-project-member] [instance: ab1539ba-33a4-4139-a2f3-1f8c507beede] Allocating IP information in the background. {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 585.129377] env[63515]: DEBUG nova.network.neutron [None req-12d73cc4-118d-457e-b393-cc4362d15b0d tempest-ListImageFiltersTestJSON-1100101045 tempest-ListImageFiltersTestJSON-1100101045-project-member] [instance: ab1539ba-33a4-4139-a2f3-1f8c507beede] allocate_for_instance() {{(pid=63515) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 585.253721] env[63515]: DEBUG nova.policy [None req-12d73cc4-118d-457e-b393-cc4362d15b0d tempest-ListImageFiltersTestJSON-1100101045 tempest-ListImageFiltersTestJSON-1100101045-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'fb6aad1c2ef9400fb9e7cfe4572e5cac', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '09ce0d978c1b4ecfb2b91fa1b1c20622', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63515) authorize /opt/stack/nova/nova/policy.py:201}} [ 585.305284] env[63515]: INFO nova.scheduler.client.report [None req-50ea3e37-671c-4a4b-9c1a-d782ff1b578c tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Deleted allocations for instance cfc6c8d5-c9e8-46fb-99e7-d245664f9652 [ 585.425478] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1110878, 'name': CreateVM_Task, 'duration_secs': 0.272301} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 585.425478] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 93b28142-8454-43c0-b0a7-d61aa95c8fc8] Created VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 585.425478] env[63515]: DEBUG oslo_concurrency.lockutils [None req-ae27ad25-8527-4c20-9aca-0f14006c890e tempest-ServersAdmin275Test-220579119 tempest-ServersAdmin275Test-220579119-project-admin] Acquiring lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 585.425478] env[63515]: DEBUG oslo_concurrency.lockutils [None req-ae27ad25-8527-4c20-9aca-0f14006c890e tempest-ServersAdmin275Test-220579119 tempest-ServersAdmin275Test-220579119-project-admin] Acquired lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 585.425775] env[63515]: DEBUG oslo_concurrency.lockutils [None req-ae27ad25-8527-4c20-9aca-0f14006c890e tempest-ServersAdmin275Test-220579119 tempest-ServersAdmin275Test-220579119-project-admin] Acquired external semaphore "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 585.426110] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6ff80f6f-d6c7-43c1-9fa3-eb4e675ecf26 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 585.435692] env[63515]: DEBUG oslo_vmware.api [None req-ae27ad25-8527-4c20-9aca-0f14006c890e tempest-ServersAdmin275Test-220579119 tempest-ServersAdmin275Test-220579119-project-admin] Waiting for the task: (returnval){ [ 585.435692] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]525f3cba-6a82-4364-3cc2-d4ee83782223" [ 585.435692] env[63515]: _type = "Task" [ 585.435692] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 585.443280] env[63515]: DEBUG oslo_vmware.api [None req-ae27ad25-8527-4c20-9aca-0f14006c890e tempest-ServersAdmin275Test-220579119 tempest-ServersAdmin275Test-220579119-project-admin] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]525f3cba-6a82-4364-3cc2-d4ee83782223, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 585.503346] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d22a084-21fc-4a1f-95b7-942ca7def7b7 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 585.513218] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ccce8f9-1246-4772-9ded-087f2002941b {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 585.546515] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb26b593-da86-42a0-8827-497d651be97c {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 585.555948] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d858136f-93ff-46b8-a8a5-4a98e8cf7e0c {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 585.572593] env[63515]: DEBUG nova.compute.provider_tree [None req-eaa7f189-a861-45f8-872f-29d870dc784b tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 585.635115] env[63515]: DEBUG nova.compute.manager [None req-12d73cc4-118d-457e-b393-cc4362d15b0d tempest-ListImageFiltersTestJSON-1100101045 tempest-ListImageFiltersTestJSON-1100101045-project-member] [instance: ab1539ba-33a4-4139-a2f3-1f8c507beede] Start building block device mappings for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 585.783036] env[63515]: DEBUG nova.network.neutron [None req-12d73cc4-118d-457e-b393-cc4362d15b0d tempest-ListImageFiltersTestJSON-1100101045 tempest-ListImageFiltersTestJSON-1100101045-project-member] [instance: ab1539ba-33a4-4139-a2f3-1f8c507beede] Successfully created port: bd569ca1-70ab-4c56-b6af-9ae3816c1d2c {{(pid=63515) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 585.817331] env[63515]: DEBUG oslo_concurrency.lockutils [None req-50ea3e37-671c-4a4b-9c1a-d782ff1b578c tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Lock "cfc6c8d5-c9e8-46fb-99e7-d245664f9652" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 46.328s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 585.948845] env[63515]: DEBUG oslo_vmware.api [None req-ae27ad25-8527-4c20-9aca-0f14006c890e tempest-ServersAdmin275Test-220579119 tempest-ServersAdmin275Test-220579119-project-admin] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]525f3cba-6a82-4364-3cc2-d4ee83782223, 'name': SearchDatastore_Task, 'duration_secs': 0.008958} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 585.948845] env[63515]: DEBUG oslo_concurrency.lockutils [None req-ae27ad25-8527-4c20-9aca-0f14006c890e tempest-ServersAdmin275Test-220579119 tempest-ServersAdmin275Test-220579119-project-admin] Releasing lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 585.948845] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-ae27ad25-8527-4c20-9aca-0f14006c890e tempest-ServersAdmin275Test-220579119 tempest-ServersAdmin275Test-220579119-project-admin] [instance: 93b28142-8454-43c0-b0a7-d61aa95c8fc8] Processing image 8a120570-cb06-4099-b262-554ca0ad15c5 {{(pid=63515) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 585.948845] env[63515]: DEBUG oslo_concurrency.lockutils [None req-ae27ad25-8527-4c20-9aca-0f14006c890e tempest-ServersAdmin275Test-220579119 tempest-ServersAdmin275Test-220579119-project-admin] Acquiring lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 585.949090] env[63515]: DEBUG oslo_concurrency.lockutils [None req-ae27ad25-8527-4c20-9aca-0f14006c890e tempest-ServersAdmin275Test-220579119 tempest-ServersAdmin275Test-220579119-project-admin] Acquired lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 585.949090] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-ae27ad25-8527-4c20-9aca-0f14006c890e tempest-ServersAdmin275Test-220579119 tempest-ServersAdmin275Test-220579119-project-admin] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 585.949090] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d0e01818-cfad-49ef-9ec6-0b92cc41caee {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 585.957570] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-ae27ad25-8527-4c20-9aca-0f14006c890e tempest-ServersAdmin275Test-220579119 tempest-ServersAdmin275Test-220579119-project-admin] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 585.957777] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-ae27ad25-8527-4c20-9aca-0f14006c890e tempest-ServersAdmin275Test-220579119 tempest-ServersAdmin275Test-220579119-project-admin] Folder [datastore1] devstack-image-cache_base created. {{(pid=63515) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 585.958965] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a572ec98-0074-4ab7-9534-4bf839245f3f {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 585.964350] env[63515]: DEBUG oslo_vmware.api [None req-ae27ad25-8527-4c20-9aca-0f14006c890e tempest-ServersAdmin275Test-220579119 tempest-ServersAdmin275Test-220579119-project-admin] Waiting for the task: (returnval){ [ 585.964350] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]52effcdb-67c1-23a7-4279-7cd9c7f7ba21" [ 585.964350] env[63515]: _type = "Task" [ 585.964350] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 585.976547] env[63515]: DEBUG oslo_vmware.api [None req-ae27ad25-8527-4c20-9aca-0f14006c890e tempest-ServersAdmin275Test-220579119 tempest-ServersAdmin275Test-220579119-project-admin] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52effcdb-67c1-23a7-4279-7cd9c7f7ba21, 'name': SearchDatastore_Task, 'duration_secs': 0.008932} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 585.978345] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5cd98d92-c13a-4511-991f-cd0bb474ac07 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 585.984133] env[63515]: DEBUG oslo_vmware.api [None req-ae27ad25-8527-4c20-9aca-0f14006c890e tempest-ServersAdmin275Test-220579119 tempest-ServersAdmin275Test-220579119-project-admin] Waiting for the task: (returnval){ [ 585.984133] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]52359d39-86a1-b089-02c4-f0bef628d0e2" [ 585.984133] env[63515]: _type = "Task" [ 585.984133] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 585.991730] env[63515]: DEBUG oslo_vmware.api [None req-ae27ad25-8527-4c20-9aca-0f14006c890e tempest-ServersAdmin275Test-220579119 tempest-ServersAdmin275Test-220579119-project-admin] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52359d39-86a1-b089-02c4-f0bef628d0e2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 586.073977] env[63515]: DEBUG nova.scheduler.client.report [None req-eaa7f189-a861-45f8-872f-29d870dc784b tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 586.319828] env[63515]: DEBUG nova.compute.manager [None req-9f5c50b0-6bbe-4982-8818-c7255a8c23ab tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: b53754b1-ec31-4dc3-a839-771f71affe1e] Starting instance... {{(pid=63515) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 586.496179] env[63515]: DEBUG oslo_vmware.api [None req-ae27ad25-8527-4c20-9aca-0f14006c890e tempest-ServersAdmin275Test-220579119 tempest-ServersAdmin275Test-220579119-project-admin] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52359d39-86a1-b089-02c4-f0bef628d0e2, 'name': SearchDatastore_Task, 'duration_secs': 0.008873} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 586.496443] env[63515]: DEBUG oslo_concurrency.lockutils [None req-ae27ad25-8527-4c20-9aca-0f14006c890e tempest-ServersAdmin275Test-220579119 tempest-ServersAdmin275Test-220579119-project-admin] Releasing lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 586.496709] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-ae27ad25-8527-4c20-9aca-0f14006c890e tempest-ServersAdmin275Test-220579119 tempest-ServersAdmin275Test-220579119-project-admin] Copying Virtual Disk [datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk to [datastore1] 93b28142-8454-43c0-b0a7-d61aa95c8fc8/93b28142-8454-43c0-b0a7-d61aa95c8fc8.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 586.496940] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3b908bf9-9362-4bdc-9aa6-3e92cb81dcd3 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 586.505977] env[63515]: DEBUG oslo_vmware.api [None req-ae27ad25-8527-4c20-9aca-0f14006c890e tempest-ServersAdmin275Test-220579119 tempest-ServersAdmin275Test-220579119-project-admin] Waiting for the task: (returnval){ [ 586.505977] env[63515]: value = "task-1110879" [ 586.505977] env[63515]: _type = "Task" [ 586.505977] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 586.515883] env[63515]: DEBUG oslo_vmware.api [None req-ae27ad25-8527-4c20-9aca-0f14006c890e tempest-ServersAdmin275Test-220579119 tempest-ServersAdmin275Test-220579119-project-admin] Task: {'id': task-1110879, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 586.579398] env[63515]: DEBUG oslo_concurrency.lockutils [None req-eaa7f189-a861-45f8-872f-29d870dc784b tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.954s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 586.579995] env[63515]: ERROR nova.compute.manager [None req-eaa7f189-a861-45f8-872f-29d870dc784b tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] [instance: e39bc55e-b677-4e2d-9aa4-7bc1d5ffadfa] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port ffab05fb-3a66-444a-9cca-115828ea7f22, please check neutron logs for more information. [ 586.579995] env[63515]: ERROR nova.compute.manager [instance: e39bc55e-b677-4e2d-9aa4-7bc1d5ffadfa] Traceback (most recent call last): [ 586.579995] env[63515]: ERROR nova.compute.manager [instance: e39bc55e-b677-4e2d-9aa4-7bc1d5ffadfa] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 586.579995] env[63515]: ERROR nova.compute.manager [instance: e39bc55e-b677-4e2d-9aa4-7bc1d5ffadfa] self.driver.spawn(context, instance, image_meta, [ 586.579995] env[63515]: ERROR nova.compute.manager [instance: e39bc55e-b677-4e2d-9aa4-7bc1d5ffadfa] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 586.579995] env[63515]: ERROR nova.compute.manager [instance: e39bc55e-b677-4e2d-9aa4-7bc1d5ffadfa] self._vmops.spawn(context, instance, image_meta, injected_files, [ 586.579995] env[63515]: ERROR nova.compute.manager [instance: e39bc55e-b677-4e2d-9aa4-7bc1d5ffadfa] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 586.579995] env[63515]: ERROR nova.compute.manager [instance: e39bc55e-b677-4e2d-9aa4-7bc1d5ffadfa] vm_ref = self.build_virtual_machine(instance, [ 586.579995] env[63515]: ERROR nova.compute.manager [instance: e39bc55e-b677-4e2d-9aa4-7bc1d5ffadfa] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 586.579995] env[63515]: ERROR nova.compute.manager [instance: e39bc55e-b677-4e2d-9aa4-7bc1d5ffadfa] vif_infos = vmwarevif.get_vif_info(self._session, [ 586.579995] env[63515]: ERROR nova.compute.manager [instance: e39bc55e-b677-4e2d-9aa4-7bc1d5ffadfa] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 586.580486] env[63515]: ERROR nova.compute.manager [instance: e39bc55e-b677-4e2d-9aa4-7bc1d5ffadfa] for vif in network_info: [ 586.580486] env[63515]: ERROR nova.compute.manager [instance: e39bc55e-b677-4e2d-9aa4-7bc1d5ffadfa] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 586.580486] env[63515]: ERROR nova.compute.manager [instance: e39bc55e-b677-4e2d-9aa4-7bc1d5ffadfa] return self._sync_wrapper(fn, *args, **kwargs) [ 586.580486] env[63515]: ERROR nova.compute.manager [instance: e39bc55e-b677-4e2d-9aa4-7bc1d5ffadfa] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 586.580486] env[63515]: ERROR nova.compute.manager [instance: e39bc55e-b677-4e2d-9aa4-7bc1d5ffadfa] self.wait() [ 586.580486] env[63515]: ERROR nova.compute.manager [instance: e39bc55e-b677-4e2d-9aa4-7bc1d5ffadfa] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 586.580486] env[63515]: ERROR nova.compute.manager [instance: e39bc55e-b677-4e2d-9aa4-7bc1d5ffadfa] self[:] = self._gt.wait() [ 586.580486] env[63515]: ERROR nova.compute.manager [instance: e39bc55e-b677-4e2d-9aa4-7bc1d5ffadfa] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 586.580486] env[63515]: ERROR nova.compute.manager [instance: e39bc55e-b677-4e2d-9aa4-7bc1d5ffadfa] return self._exit_event.wait() [ 586.580486] env[63515]: ERROR nova.compute.manager [instance: e39bc55e-b677-4e2d-9aa4-7bc1d5ffadfa] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 586.580486] env[63515]: ERROR nova.compute.manager [instance: e39bc55e-b677-4e2d-9aa4-7bc1d5ffadfa] result = hub.switch() [ 586.580486] env[63515]: ERROR nova.compute.manager [instance: e39bc55e-b677-4e2d-9aa4-7bc1d5ffadfa] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 586.580486] env[63515]: ERROR nova.compute.manager [instance: e39bc55e-b677-4e2d-9aa4-7bc1d5ffadfa] return self.greenlet.switch() [ 586.580988] env[63515]: ERROR nova.compute.manager [instance: e39bc55e-b677-4e2d-9aa4-7bc1d5ffadfa] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 586.580988] env[63515]: ERROR nova.compute.manager [instance: e39bc55e-b677-4e2d-9aa4-7bc1d5ffadfa] result = function(*args, **kwargs) [ 586.580988] env[63515]: ERROR nova.compute.manager [instance: e39bc55e-b677-4e2d-9aa4-7bc1d5ffadfa] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 586.580988] env[63515]: ERROR nova.compute.manager [instance: e39bc55e-b677-4e2d-9aa4-7bc1d5ffadfa] return func(*args, **kwargs) [ 586.580988] env[63515]: ERROR nova.compute.manager [instance: e39bc55e-b677-4e2d-9aa4-7bc1d5ffadfa] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 586.580988] env[63515]: ERROR nova.compute.manager [instance: e39bc55e-b677-4e2d-9aa4-7bc1d5ffadfa] raise e [ 586.580988] env[63515]: ERROR nova.compute.manager [instance: e39bc55e-b677-4e2d-9aa4-7bc1d5ffadfa] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 586.580988] env[63515]: ERROR nova.compute.manager [instance: e39bc55e-b677-4e2d-9aa4-7bc1d5ffadfa] nwinfo = self.network_api.allocate_for_instance( [ 586.580988] env[63515]: ERROR nova.compute.manager [instance: e39bc55e-b677-4e2d-9aa4-7bc1d5ffadfa] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 586.580988] env[63515]: ERROR nova.compute.manager [instance: e39bc55e-b677-4e2d-9aa4-7bc1d5ffadfa] created_port_ids = self._update_ports_for_instance( [ 586.580988] env[63515]: ERROR nova.compute.manager [instance: e39bc55e-b677-4e2d-9aa4-7bc1d5ffadfa] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 586.580988] env[63515]: ERROR nova.compute.manager [instance: e39bc55e-b677-4e2d-9aa4-7bc1d5ffadfa] with excutils.save_and_reraise_exception(): [ 586.580988] env[63515]: ERROR nova.compute.manager [instance: e39bc55e-b677-4e2d-9aa4-7bc1d5ffadfa] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 586.582874] env[63515]: ERROR nova.compute.manager [instance: e39bc55e-b677-4e2d-9aa4-7bc1d5ffadfa] self.force_reraise() [ 586.582874] env[63515]: ERROR nova.compute.manager [instance: e39bc55e-b677-4e2d-9aa4-7bc1d5ffadfa] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 586.582874] env[63515]: ERROR nova.compute.manager [instance: e39bc55e-b677-4e2d-9aa4-7bc1d5ffadfa] raise self.value [ 586.582874] env[63515]: ERROR nova.compute.manager [instance: e39bc55e-b677-4e2d-9aa4-7bc1d5ffadfa] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 586.582874] env[63515]: ERROR nova.compute.manager [instance: e39bc55e-b677-4e2d-9aa4-7bc1d5ffadfa] updated_port = self._update_port( [ 586.582874] env[63515]: ERROR nova.compute.manager [instance: e39bc55e-b677-4e2d-9aa4-7bc1d5ffadfa] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 586.582874] env[63515]: ERROR nova.compute.manager [instance: e39bc55e-b677-4e2d-9aa4-7bc1d5ffadfa] _ensure_no_port_binding_failure(port) [ 586.582874] env[63515]: ERROR nova.compute.manager [instance: e39bc55e-b677-4e2d-9aa4-7bc1d5ffadfa] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 586.582874] env[63515]: ERROR nova.compute.manager [instance: e39bc55e-b677-4e2d-9aa4-7bc1d5ffadfa] raise exception.PortBindingFailed(port_id=port['id']) [ 586.582874] env[63515]: ERROR nova.compute.manager [instance: e39bc55e-b677-4e2d-9aa4-7bc1d5ffadfa] nova.exception.PortBindingFailed: Binding failed for port ffab05fb-3a66-444a-9cca-115828ea7f22, please check neutron logs for more information. [ 586.582874] env[63515]: ERROR nova.compute.manager [instance: e39bc55e-b677-4e2d-9aa4-7bc1d5ffadfa] [ 586.583357] env[63515]: DEBUG nova.compute.utils [None req-eaa7f189-a861-45f8-872f-29d870dc784b tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] [instance: e39bc55e-b677-4e2d-9aa4-7bc1d5ffadfa] Binding failed for port ffab05fb-3a66-444a-9cca-115828ea7f22, please check neutron logs for more information. {{(pid=63515) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 586.583357] env[63515]: DEBUG oslo_concurrency.lockutils [None req-c80cd50f-4127-4451-9298-83a67d8f0ab6 tempest-ServersTestFqdnHostnames-1452522528 tempest-ServersTestFqdnHostnames-1452522528-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.929s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 586.583861] env[63515]: INFO nova.compute.claims [None req-c80cd50f-4127-4451-9298-83a67d8f0ab6 tempest-ServersTestFqdnHostnames-1452522528 tempest-ServersTestFqdnHostnames-1452522528-project-member] [instance: b84c6c8d-e358-4a14-8425-c4f1d8adf4ee] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 586.590417] env[63515]: DEBUG nova.compute.manager [None req-eaa7f189-a861-45f8-872f-29d870dc784b tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] [instance: e39bc55e-b677-4e2d-9aa4-7bc1d5ffadfa] Build of instance e39bc55e-b677-4e2d-9aa4-7bc1d5ffadfa was re-scheduled: Binding failed for port ffab05fb-3a66-444a-9cca-115828ea7f22, please check neutron logs for more information. {{(pid=63515) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 586.590417] env[63515]: DEBUG nova.compute.manager [None req-eaa7f189-a861-45f8-872f-29d870dc784b tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] [instance: e39bc55e-b677-4e2d-9aa4-7bc1d5ffadfa] Unplugging VIFs for instance {{(pid=63515) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 586.590417] env[63515]: DEBUG oslo_concurrency.lockutils [None req-eaa7f189-a861-45f8-872f-29d870dc784b tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] Acquiring lock "refresh_cache-e39bc55e-b677-4e2d-9aa4-7bc1d5ffadfa" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 586.590417] env[63515]: DEBUG oslo_concurrency.lockutils [None req-eaa7f189-a861-45f8-872f-29d870dc784b tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] Acquired lock "refresh_cache-e39bc55e-b677-4e2d-9aa4-7bc1d5ffadfa" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 586.591679] env[63515]: DEBUG nova.network.neutron [None req-eaa7f189-a861-45f8-872f-29d870dc784b tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] [instance: e39bc55e-b677-4e2d-9aa4-7bc1d5ffadfa] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 586.647864] env[63515]: DEBUG nova.compute.manager [None req-12d73cc4-118d-457e-b393-cc4362d15b0d tempest-ListImageFiltersTestJSON-1100101045 tempest-ListImageFiltersTestJSON-1100101045-project-member] [instance: ab1539ba-33a4-4139-a2f3-1f8c507beede] Start spawning the instance on the hypervisor. {{(pid=63515) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 586.682927] env[63515]: DEBUG nova.virt.hardware [None req-12d73cc4-118d-457e-b393-cc4362d15b0d tempest-ListImageFiltersTestJSON-1100101045 tempest-ListImageFiltersTestJSON-1100101045-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T02:52:14Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T02:51:56Z,direct_url=,disk_format='vmdk',id=8a120570-cb06-4099-b262-554ca0ad15c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b569255cc43e42e7ae3f2b4ad37c6ef4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T02:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 586.683358] env[63515]: DEBUG nova.virt.hardware [None req-12d73cc4-118d-457e-b393-cc4362d15b0d tempest-ListImageFiltersTestJSON-1100101045 tempest-ListImageFiltersTestJSON-1100101045-project-member] Flavor limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 586.683537] env[63515]: DEBUG nova.virt.hardware [None req-12d73cc4-118d-457e-b393-cc4362d15b0d tempest-ListImageFiltersTestJSON-1100101045 tempest-ListImageFiltersTestJSON-1100101045-project-member] Image limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 586.683728] env[63515]: DEBUG nova.virt.hardware [None req-12d73cc4-118d-457e-b393-cc4362d15b0d tempest-ListImageFiltersTestJSON-1100101045 tempest-ListImageFiltersTestJSON-1100101045-project-member] Flavor pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 586.684104] env[63515]: DEBUG nova.virt.hardware [None req-12d73cc4-118d-457e-b393-cc4362d15b0d tempest-ListImageFiltersTestJSON-1100101045 tempest-ListImageFiltersTestJSON-1100101045-project-member] Image pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 586.684104] env[63515]: DEBUG nova.virt.hardware [None req-12d73cc4-118d-457e-b393-cc4362d15b0d tempest-ListImageFiltersTestJSON-1100101045 tempest-ListImageFiltersTestJSON-1100101045-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 586.684311] env[63515]: DEBUG nova.virt.hardware [None req-12d73cc4-118d-457e-b393-cc4362d15b0d tempest-ListImageFiltersTestJSON-1100101045 tempest-ListImageFiltersTestJSON-1100101045-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 586.684895] env[63515]: DEBUG nova.virt.hardware [None req-12d73cc4-118d-457e-b393-cc4362d15b0d tempest-ListImageFiltersTestJSON-1100101045 tempest-ListImageFiltersTestJSON-1100101045-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 586.684895] env[63515]: DEBUG nova.virt.hardware [None req-12d73cc4-118d-457e-b393-cc4362d15b0d tempest-ListImageFiltersTestJSON-1100101045 tempest-ListImageFiltersTestJSON-1100101045-project-member] Got 1 possible topologies {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 586.685040] env[63515]: DEBUG nova.virt.hardware [None req-12d73cc4-118d-457e-b393-cc4362d15b0d tempest-ListImageFiltersTestJSON-1100101045 tempest-ListImageFiltersTestJSON-1100101045-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 586.685230] env[63515]: DEBUG nova.virt.hardware [None req-12d73cc4-118d-457e-b393-cc4362d15b0d tempest-ListImageFiltersTestJSON-1100101045 tempest-ListImageFiltersTestJSON-1100101045-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 586.686181] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ae2c5f9-7e42-4429-aa63-b79ff4a835aa {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 586.697136] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-649212b5-46f9-4fd2-818b-88a0a0b1534c {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 586.849121] env[63515]: DEBUG oslo_concurrency.lockutils [None req-9f5c50b0-6bbe-4982-8818-c7255a8c23ab tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 587.017319] env[63515]: DEBUG oslo_vmware.api [None req-ae27ad25-8527-4c20-9aca-0f14006c890e tempest-ServersAdmin275Test-220579119 tempest-ServersAdmin275Test-220579119-project-admin] Task: {'id': task-1110879, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.499391} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 587.017655] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-ae27ad25-8527-4c20-9aca-0f14006c890e tempest-ServersAdmin275Test-220579119 tempest-ServersAdmin275Test-220579119-project-admin] Copied Virtual Disk [datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk to [datastore1] 93b28142-8454-43c0-b0a7-d61aa95c8fc8/93b28142-8454-43c0-b0a7-d61aa95c8fc8.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 587.017790] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-ae27ad25-8527-4c20-9aca-0f14006c890e tempest-ServersAdmin275Test-220579119 tempest-ServersAdmin275Test-220579119-project-admin] [instance: 93b28142-8454-43c0-b0a7-d61aa95c8fc8] Extending root virtual disk to 1048576 {{(pid=63515) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 587.018071] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-2ba8027e-be85-42bc-924b-6135c826f6c6 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 587.024736] env[63515]: DEBUG oslo_vmware.api [None req-ae27ad25-8527-4c20-9aca-0f14006c890e tempest-ServersAdmin275Test-220579119 tempest-ServersAdmin275Test-220579119-project-admin] Waiting for the task: (returnval){ [ 587.024736] env[63515]: value = "task-1110881" [ 587.024736] env[63515]: _type = "Task" [ 587.024736] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 587.033113] env[63515]: DEBUG oslo_vmware.api [None req-ae27ad25-8527-4c20-9aca-0f14006c890e tempest-ServersAdmin275Test-220579119 tempest-ServersAdmin275Test-220579119-project-admin] Task: {'id': task-1110881, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 587.128283] env[63515]: DEBUG nova.network.neutron [None req-eaa7f189-a861-45f8-872f-29d870dc784b tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] [instance: e39bc55e-b677-4e2d-9aa4-7bc1d5ffadfa] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 587.324867] env[63515]: DEBUG nova.network.neutron [None req-eaa7f189-a861-45f8-872f-29d870dc784b tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] [instance: e39bc55e-b677-4e2d-9aa4-7bc1d5ffadfa] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 587.538330] env[63515]: DEBUG oslo_vmware.api [None req-ae27ad25-8527-4c20-9aca-0f14006c890e tempest-ServersAdmin275Test-220579119 tempest-ServersAdmin275Test-220579119-project-admin] Task: {'id': task-1110881, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.067269} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 587.539296] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-ae27ad25-8527-4c20-9aca-0f14006c890e tempest-ServersAdmin275Test-220579119 tempest-ServersAdmin275Test-220579119-project-admin] [instance: 93b28142-8454-43c0-b0a7-d61aa95c8fc8] Extended root virtual disk {{(pid=63515) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 587.539596] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-597a7691-27c0-4d44-a7db-67132270a66d {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 587.569535] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-ae27ad25-8527-4c20-9aca-0f14006c890e tempest-ServersAdmin275Test-220579119 tempest-ServersAdmin275Test-220579119-project-admin] [instance: 93b28142-8454-43c0-b0a7-d61aa95c8fc8] Reconfiguring VM instance instance-0000000a to attach disk [datastore1] 93b28142-8454-43c0-b0a7-d61aa95c8fc8/93b28142-8454-43c0-b0a7-d61aa95c8fc8.vmdk or device None with type sparse {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 587.570070] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cda88362-8734-4940-8906-8574bd5b425e {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 587.592924] env[63515]: DEBUG oslo_vmware.api [None req-ae27ad25-8527-4c20-9aca-0f14006c890e tempest-ServersAdmin275Test-220579119 tempest-ServersAdmin275Test-220579119-project-admin] Waiting for the task: (returnval){ [ 587.592924] env[63515]: value = "task-1110882" [ 587.592924] env[63515]: _type = "Task" [ 587.592924] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 587.605695] env[63515]: DEBUG oslo_vmware.api [None req-ae27ad25-8527-4c20-9aca-0f14006c890e tempest-ServersAdmin275Test-220579119 tempest-ServersAdmin275Test-220579119-project-admin] Task: {'id': task-1110882, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 587.832519] env[63515]: DEBUG oslo_concurrency.lockutils [None req-eaa7f189-a861-45f8-872f-29d870dc784b tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] Releasing lock "refresh_cache-e39bc55e-b677-4e2d-9aa4-7bc1d5ffadfa" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 587.832519] env[63515]: DEBUG nova.compute.manager [None req-eaa7f189-a861-45f8-872f-29d870dc784b tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63515) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 587.832519] env[63515]: DEBUG nova.compute.manager [None req-eaa7f189-a861-45f8-872f-29d870dc784b tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] [instance: e39bc55e-b677-4e2d-9aa4-7bc1d5ffadfa] Deallocating network for instance {{(pid=63515) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 587.834583] env[63515]: DEBUG nova.network.neutron [None req-eaa7f189-a861-45f8-872f-29d870dc784b tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] [instance: e39bc55e-b677-4e2d-9aa4-7bc1d5ffadfa] deallocate_for_instance() {{(pid=63515) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 587.879276] env[63515]: DEBUG nova.network.neutron [None req-eaa7f189-a861-45f8-872f-29d870dc784b tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] [instance: e39bc55e-b677-4e2d-9aa4-7bc1d5ffadfa] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 587.979181] env[63515]: DEBUG nova.compute.manager [req-b93a78c0-10f9-4419-a75f-44bc4e6e417e req-5f1ecf75-8db5-45af-a409-6a59c0607be7 service nova] [instance: ab1539ba-33a4-4139-a2f3-1f8c507beede] Received event network-changed-bd569ca1-70ab-4c56-b6af-9ae3816c1d2c {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 587.979430] env[63515]: DEBUG nova.compute.manager [req-b93a78c0-10f9-4419-a75f-44bc4e6e417e req-5f1ecf75-8db5-45af-a409-6a59c0607be7 service nova] [instance: ab1539ba-33a4-4139-a2f3-1f8c507beede] Refreshing instance network info cache due to event network-changed-bd569ca1-70ab-4c56-b6af-9ae3816c1d2c. {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 587.979628] env[63515]: DEBUG oslo_concurrency.lockutils [req-b93a78c0-10f9-4419-a75f-44bc4e6e417e req-5f1ecf75-8db5-45af-a409-6a59c0607be7 service nova] Acquiring lock "refresh_cache-ab1539ba-33a4-4139-a2f3-1f8c507beede" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 587.982726] env[63515]: DEBUG oslo_concurrency.lockutils [req-b93a78c0-10f9-4419-a75f-44bc4e6e417e req-5f1ecf75-8db5-45af-a409-6a59c0607be7 service nova] Acquired lock "refresh_cache-ab1539ba-33a4-4139-a2f3-1f8c507beede" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 587.982726] env[63515]: DEBUG nova.network.neutron [req-b93a78c0-10f9-4419-a75f-44bc4e6e417e req-5f1ecf75-8db5-45af-a409-6a59c0607be7 service nova] [instance: ab1539ba-33a4-4139-a2f3-1f8c507beede] Refreshing network info cache for port bd569ca1-70ab-4c56-b6af-9ae3816c1d2c {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 588.053081] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02c29d55-edca-4bca-bd15-1204ceb47b0e {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.063017] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2e8040a-4732-4efb-a72c-d0c1f14535ed {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.093943] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-364af790-d950-4069-9ac8-aeea83ab6d74 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.107442] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af558151-b769-434f-b972-94346c83e350 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.111908] env[63515]: DEBUG oslo_vmware.api [None req-ae27ad25-8527-4c20-9aca-0f14006c890e tempest-ServersAdmin275Test-220579119 tempest-ServersAdmin275Test-220579119-project-admin] Task: {'id': task-1110882, 'name': ReconfigVM_Task, 'duration_secs': 0.282409} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 588.112241] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-ae27ad25-8527-4c20-9aca-0f14006c890e tempest-ServersAdmin275Test-220579119 tempest-ServersAdmin275Test-220579119-project-admin] [instance: 93b28142-8454-43c0-b0a7-d61aa95c8fc8] Reconfigured VM instance instance-0000000a to attach disk [datastore1] 93b28142-8454-43c0-b0a7-d61aa95c8fc8/93b28142-8454-43c0-b0a7-d61aa95c8fc8.vmdk or device None with type sparse {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 588.113240] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-58f531c6-8692-48f5-b94b-677be7d7ab3a {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.124614] env[63515]: DEBUG nova.compute.provider_tree [None req-c80cd50f-4127-4451-9298-83a67d8f0ab6 tempest-ServersTestFqdnHostnames-1452522528 tempest-ServersTestFqdnHostnames-1452522528-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 588.131116] env[63515]: DEBUG oslo_vmware.api [None req-ae27ad25-8527-4c20-9aca-0f14006c890e tempest-ServersAdmin275Test-220579119 tempest-ServersAdmin275Test-220579119-project-admin] Waiting for the task: (returnval){ [ 588.131116] env[63515]: value = "task-1110883" [ 588.131116] env[63515]: _type = "Task" [ 588.131116] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 588.141158] env[63515]: DEBUG oslo_vmware.api [None req-ae27ad25-8527-4c20-9aca-0f14006c890e tempest-ServersAdmin275Test-220579119 tempest-ServersAdmin275Test-220579119-project-admin] Task: {'id': task-1110883, 'name': Rename_Task} progress is 5%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 588.238583] env[63515]: ERROR nova.compute.manager [None req-12d73cc4-118d-457e-b393-cc4362d15b0d tempest-ListImageFiltersTestJSON-1100101045 tempest-ListImageFiltersTestJSON-1100101045-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port bd569ca1-70ab-4c56-b6af-9ae3816c1d2c, please check neutron logs for more information. [ 588.238583] env[63515]: ERROR nova.compute.manager Traceback (most recent call last): [ 588.238583] env[63515]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 588.238583] env[63515]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 588.238583] env[63515]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 588.238583] env[63515]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 588.238583] env[63515]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 588.238583] env[63515]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 588.238583] env[63515]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 588.238583] env[63515]: ERROR nova.compute.manager self.force_reraise() [ 588.238583] env[63515]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 588.238583] env[63515]: ERROR nova.compute.manager raise self.value [ 588.238583] env[63515]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 588.238583] env[63515]: ERROR nova.compute.manager updated_port = self._update_port( [ 588.238583] env[63515]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 588.238583] env[63515]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 588.239247] env[63515]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 588.239247] env[63515]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 588.239247] env[63515]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port bd569ca1-70ab-4c56-b6af-9ae3816c1d2c, please check neutron logs for more information. [ 588.239247] env[63515]: ERROR nova.compute.manager [ 588.239247] env[63515]: Traceback (most recent call last): [ 588.239247] env[63515]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 588.239247] env[63515]: listener.cb(fileno) [ 588.239247] env[63515]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 588.239247] env[63515]: result = function(*args, **kwargs) [ 588.239247] env[63515]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 588.239247] env[63515]: return func(*args, **kwargs) [ 588.239247] env[63515]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 588.239247] env[63515]: raise e [ 588.239247] env[63515]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 588.239247] env[63515]: nwinfo = self.network_api.allocate_for_instance( [ 588.239247] env[63515]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 588.239247] env[63515]: created_port_ids = self._update_ports_for_instance( [ 588.239247] env[63515]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 588.239247] env[63515]: with excutils.save_and_reraise_exception(): [ 588.239247] env[63515]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 588.239247] env[63515]: self.force_reraise() [ 588.239247] env[63515]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 588.239247] env[63515]: raise self.value [ 588.239247] env[63515]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 588.239247] env[63515]: updated_port = self._update_port( [ 588.239247] env[63515]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 588.239247] env[63515]: _ensure_no_port_binding_failure(port) [ 588.239247] env[63515]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 588.239247] env[63515]: raise exception.PortBindingFailed(port_id=port['id']) [ 588.240157] env[63515]: nova.exception.PortBindingFailed: Binding failed for port bd569ca1-70ab-4c56-b6af-9ae3816c1d2c, please check neutron logs for more information. [ 588.240157] env[63515]: Removing descriptor: 14 [ 588.240157] env[63515]: ERROR nova.compute.manager [None req-12d73cc4-118d-457e-b393-cc4362d15b0d tempest-ListImageFiltersTestJSON-1100101045 tempest-ListImageFiltersTestJSON-1100101045-project-member] [instance: ab1539ba-33a4-4139-a2f3-1f8c507beede] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port bd569ca1-70ab-4c56-b6af-9ae3816c1d2c, please check neutron logs for more information. [ 588.240157] env[63515]: ERROR nova.compute.manager [instance: ab1539ba-33a4-4139-a2f3-1f8c507beede] Traceback (most recent call last): [ 588.240157] env[63515]: ERROR nova.compute.manager [instance: ab1539ba-33a4-4139-a2f3-1f8c507beede] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 588.240157] env[63515]: ERROR nova.compute.manager [instance: ab1539ba-33a4-4139-a2f3-1f8c507beede] yield resources [ 588.240157] env[63515]: ERROR nova.compute.manager [instance: ab1539ba-33a4-4139-a2f3-1f8c507beede] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 588.240157] env[63515]: ERROR nova.compute.manager [instance: ab1539ba-33a4-4139-a2f3-1f8c507beede] self.driver.spawn(context, instance, image_meta, [ 588.240157] env[63515]: ERROR nova.compute.manager [instance: ab1539ba-33a4-4139-a2f3-1f8c507beede] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 588.240157] env[63515]: ERROR nova.compute.manager [instance: ab1539ba-33a4-4139-a2f3-1f8c507beede] self._vmops.spawn(context, instance, image_meta, injected_files, [ 588.240157] env[63515]: ERROR nova.compute.manager [instance: ab1539ba-33a4-4139-a2f3-1f8c507beede] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 588.240157] env[63515]: ERROR nova.compute.manager [instance: ab1539ba-33a4-4139-a2f3-1f8c507beede] vm_ref = self.build_virtual_machine(instance, [ 588.240532] env[63515]: ERROR nova.compute.manager [instance: ab1539ba-33a4-4139-a2f3-1f8c507beede] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 588.240532] env[63515]: ERROR nova.compute.manager [instance: ab1539ba-33a4-4139-a2f3-1f8c507beede] vif_infos = vmwarevif.get_vif_info(self._session, [ 588.240532] env[63515]: ERROR nova.compute.manager [instance: ab1539ba-33a4-4139-a2f3-1f8c507beede] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 588.240532] env[63515]: ERROR nova.compute.manager [instance: ab1539ba-33a4-4139-a2f3-1f8c507beede] for vif in network_info: [ 588.240532] env[63515]: ERROR nova.compute.manager [instance: ab1539ba-33a4-4139-a2f3-1f8c507beede] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 588.240532] env[63515]: ERROR nova.compute.manager [instance: ab1539ba-33a4-4139-a2f3-1f8c507beede] return self._sync_wrapper(fn, *args, **kwargs) [ 588.240532] env[63515]: ERROR nova.compute.manager [instance: ab1539ba-33a4-4139-a2f3-1f8c507beede] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 588.240532] env[63515]: ERROR nova.compute.manager [instance: ab1539ba-33a4-4139-a2f3-1f8c507beede] self.wait() [ 588.240532] env[63515]: ERROR nova.compute.manager [instance: ab1539ba-33a4-4139-a2f3-1f8c507beede] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 588.240532] env[63515]: ERROR nova.compute.manager [instance: ab1539ba-33a4-4139-a2f3-1f8c507beede] self[:] = self._gt.wait() [ 588.240532] env[63515]: ERROR nova.compute.manager [instance: ab1539ba-33a4-4139-a2f3-1f8c507beede] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 588.240532] env[63515]: ERROR nova.compute.manager [instance: ab1539ba-33a4-4139-a2f3-1f8c507beede] return self._exit_event.wait() [ 588.240532] env[63515]: ERROR nova.compute.manager [instance: ab1539ba-33a4-4139-a2f3-1f8c507beede] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 588.241043] env[63515]: ERROR nova.compute.manager [instance: ab1539ba-33a4-4139-a2f3-1f8c507beede] result = hub.switch() [ 588.241043] env[63515]: ERROR nova.compute.manager [instance: ab1539ba-33a4-4139-a2f3-1f8c507beede] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 588.241043] env[63515]: ERROR nova.compute.manager [instance: ab1539ba-33a4-4139-a2f3-1f8c507beede] return self.greenlet.switch() [ 588.241043] env[63515]: ERROR nova.compute.manager [instance: ab1539ba-33a4-4139-a2f3-1f8c507beede] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 588.241043] env[63515]: ERROR nova.compute.manager [instance: ab1539ba-33a4-4139-a2f3-1f8c507beede] result = function(*args, **kwargs) [ 588.241043] env[63515]: ERROR nova.compute.manager [instance: ab1539ba-33a4-4139-a2f3-1f8c507beede] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 588.241043] env[63515]: ERROR nova.compute.manager [instance: ab1539ba-33a4-4139-a2f3-1f8c507beede] return func(*args, **kwargs) [ 588.241043] env[63515]: ERROR nova.compute.manager [instance: ab1539ba-33a4-4139-a2f3-1f8c507beede] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 588.241043] env[63515]: ERROR nova.compute.manager [instance: ab1539ba-33a4-4139-a2f3-1f8c507beede] raise e [ 588.241043] env[63515]: ERROR nova.compute.manager [instance: ab1539ba-33a4-4139-a2f3-1f8c507beede] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 588.241043] env[63515]: ERROR nova.compute.manager [instance: ab1539ba-33a4-4139-a2f3-1f8c507beede] nwinfo = self.network_api.allocate_for_instance( [ 588.241043] env[63515]: ERROR nova.compute.manager [instance: ab1539ba-33a4-4139-a2f3-1f8c507beede] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 588.241043] env[63515]: ERROR nova.compute.manager [instance: ab1539ba-33a4-4139-a2f3-1f8c507beede] created_port_ids = self._update_ports_for_instance( [ 588.241485] env[63515]: ERROR nova.compute.manager [instance: ab1539ba-33a4-4139-a2f3-1f8c507beede] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 588.241485] env[63515]: ERROR nova.compute.manager [instance: ab1539ba-33a4-4139-a2f3-1f8c507beede] with excutils.save_and_reraise_exception(): [ 588.241485] env[63515]: ERROR nova.compute.manager [instance: ab1539ba-33a4-4139-a2f3-1f8c507beede] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 588.241485] env[63515]: ERROR nova.compute.manager [instance: ab1539ba-33a4-4139-a2f3-1f8c507beede] self.force_reraise() [ 588.241485] env[63515]: ERROR nova.compute.manager [instance: ab1539ba-33a4-4139-a2f3-1f8c507beede] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 588.241485] env[63515]: ERROR nova.compute.manager [instance: ab1539ba-33a4-4139-a2f3-1f8c507beede] raise self.value [ 588.241485] env[63515]: ERROR nova.compute.manager [instance: ab1539ba-33a4-4139-a2f3-1f8c507beede] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 588.241485] env[63515]: ERROR nova.compute.manager [instance: ab1539ba-33a4-4139-a2f3-1f8c507beede] updated_port = self._update_port( [ 588.241485] env[63515]: ERROR nova.compute.manager [instance: ab1539ba-33a4-4139-a2f3-1f8c507beede] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 588.241485] env[63515]: ERROR nova.compute.manager [instance: ab1539ba-33a4-4139-a2f3-1f8c507beede] _ensure_no_port_binding_failure(port) [ 588.241485] env[63515]: ERROR nova.compute.manager [instance: ab1539ba-33a4-4139-a2f3-1f8c507beede] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 588.241485] env[63515]: ERROR nova.compute.manager [instance: ab1539ba-33a4-4139-a2f3-1f8c507beede] raise exception.PortBindingFailed(port_id=port['id']) [ 588.242054] env[63515]: ERROR nova.compute.manager [instance: ab1539ba-33a4-4139-a2f3-1f8c507beede] nova.exception.PortBindingFailed: Binding failed for port bd569ca1-70ab-4c56-b6af-9ae3816c1d2c, please check neutron logs for more information. [ 588.242054] env[63515]: ERROR nova.compute.manager [instance: ab1539ba-33a4-4139-a2f3-1f8c507beede] [ 588.242054] env[63515]: INFO nova.compute.manager [None req-12d73cc4-118d-457e-b393-cc4362d15b0d tempest-ListImageFiltersTestJSON-1100101045 tempest-ListImageFiltersTestJSON-1100101045-project-member] [instance: ab1539ba-33a4-4139-a2f3-1f8c507beede] Terminating instance [ 588.242265] env[63515]: DEBUG oslo_concurrency.lockutils [None req-12d73cc4-118d-457e-b393-cc4362d15b0d tempest-ListImageFiltersTestJSON-1100101045 tempest-ListImageFiltersTestJSON-1100101045-project-member] Acquiring lock "refresh_cache-ab1539ba-33a4-4139-a2f3-1f8c507beede" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 588.271528] env[63515]: DEBUG oslo_concurrency.lockutils [None req-89884269-25d4-4981-bd3b-a346ece7b681 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Acquiring lock "f1d01b75-ac9d-458d-8cc2-ae64cffca4e8" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 588.271801] env[63515]: DEBUG oslo_concurrency.lockutils [None req-89884269-25d4-4981-bd3b-a346ece7b681 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Lock "f1d01b75-ac9d-458d-8cc2-ae64cffca4e8" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 588.384595] env[63515]: DEBUG nova.network.neutron [None req-eaa7f189-a861-45f8-872f-29d870dc784b tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] [instance: e39bc55e-b677-4e2d-9aa4-7bc1d5ffadfa] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 588.510108] env[63515]: DEBUG nova.network.neutron [req-b93a78c0-10f9-4419-a75f-44bc4e6e417e req-5f1ecf75-8db5-45af-a409-6a59c0607be7 service nova] [instance: ab1539ba-33a4-4139-a2f3-1f8c507beede] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 588.592514] env[63515]: DEBUG nova.network.neutron [req-b93a78c0-10f9-4419-a75f-44bc4e6e417e req-5f1ecf75-8db5-45af-a409-6a59c0607be7 service nova] [instance: ab1539ba-33a4-4139-a2f3-1f8c507beede] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 588.627843] env[63515]: DEBUG nova.scheduler.client.report [None req-c80cd50f-4127-4451-9298-83a67d8f0ab6 tempest-ServersTestFqdnHostnames-1452522528 tempest-ServersTestFqdnHostnames-1452522528-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 588.642172] env[63515]: DEBUG oslo_vmware.api [None req-ae27ad25-8527-4c20-9aca-0f14006c890e tempest-ServersAdmin275Test-220579119 tempest-ServersAdmin275Test-220579119-project-admin] Task: {'id': task-1110883, 'name': Rename_Task, 'duration_secs': 0.372315} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 588.642452] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-ae27ad25-8527-4c20-9aca-0f14006c890e tempest-ServersAdmin275Test-220579119 tempest-ServersAdmin275Test-220579119-project-admin] [instance: 93b28142-8454-43c0-b0a7-d61aa95c8fc8] Powering on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 588.642691] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5cbbcd36-92c4-4dbe-80c7-4ee31e25ec21 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.651588] env[63515]: DEBUG oslo_vmware.api [None req-ae27ad25-8527-4c20-9aca-0f14006c890e tempest-ServersAdmin275Test-220579119 tempest-ServersAdmin275Test-220579119-project-admin] Waiting for the task: (returnval){ [ 588.651588] env[63515]: value = "task-1110884" [ 588.651588] env[63515]: _type = "Task" [ 588.651588] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 588.662684] env[63515]: DEBUG oslo_vmware.api [None req-ae27ad25-8527-4c20-9aca-0f14006c890e tempest-ServersAdmin275Test-220579119 tempest-ServersAdmin275Test-220579119-project-admin] Task: {'id': task-1110884, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 588.887664] env[63515]: INFO nova.compute.manager [None req-eaa7f189-a861-45f8-872f-29d870dc784b tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] [instance: e39bc55e-b677-4e2d-9aa4-7bc1d5ffadfa] Took 1.05 seconds to deallocate network for instance. [ 589.104580] env[63515]: DEBUG oslo_concurrency.lockutils [req-b93a78c0-10f9-4419-a75f-44bc4e6e417e req-5f1ecf75-8db5-45af-a409-6a59c0607be7 service nova] Releasing lock "refresh_cache-ab1539ba-33a4-4139-a2f3-1f8c507beede" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 589.104889] env[63515]: DEBUG oslo_concurrency.lockutils [None req-12d73cc4-118d-457e-b393-cc4362d15b0d tempest-ListImageFiltersTestJSON-1100101045 tempest-ListImageFiltersTestJSON-1100101045-project-member] Acquired lock "refresh_cache-ab1539ba-33a4-4139-a2f3-1f8c507beede" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 589.105251] env[63515]: DEBUG nova.network.neutron [None req-12d73cc4-118d-457e-b393-cc4362d15b0d tempest-ListImageFiltersTestJSON-1100101045 tempest-ListImageFiltersTestJSON-1100101045-project-member] [instance: ab1539ba-33a4-4139-a2f3-1f8c507beede] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 589.138223] env[63515]: DEBUG oslo_concurrency.lockutils [None req-c80cd50f-4127-4451-9298-83a67d8f0ab6 tempest-ServersTestFqdnHostnames-1452522528 tempest-ServersTestFqdnHostnames-1452522528-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.556s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 589.138792] env[63515]: DEBUG nova.compute.manager [None req-c80cd50f-4127-4451-9298-83a67d8f0ab6 tempest-ServersTestFqdnHostnames-1452522528 tempest-ServersTestFqdnHostnames-1452522528-project-member] [instance: b84c6c8d-e358-4a14-8425-c4f1d8adf4ee] Start building networks asynchronously for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 589.142522] env[63515]: DEBUG oslo_concurrency.lockutils [None req-22f02712-96d1-4031-8b8f-5bc880f34c33 tempest-ImagesNegativeTestJSON-2041328384 tempest-ImagesNegativeTestJSON-2041328384-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.802s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 589.144166] env[63515]: INFO nova.compute.claims [None req-22f02712-96d1-4031-8b8f-5bc880f34c33 tempest-ImagesNegativeTestJSON-2041328384 tempest-ImagesNegativeTestJSON-2041328384-project-member] [instance: d5f988f2-765b-4ecf-bad5-4a70fc95d74c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 589.166036] env[63515]: DEBUG oslo_vmware.api [None req-ae27ad25-8527-4c20-9aca-0f14006c890e tempest-ServersAdmin275Test-220579119 tempest-ServersAdmin275Test-220579119-project-admin] Task: {'id': task-1110884, 'name': PowerOnVM_Task, 'duration_secs': 0.454273} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 589.167036] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-ae27ad25-8527-4c20-9aca-0f14006c890e tempest-ServersAdmin275Test-220579119 tempest-ServersAdmin275Test-220579119-project-admin] [instance: 93b28142-8454-43c0-b0a7-d61aa95c8fc8] Powered on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 589.167036] env[63515]: DEBUG nova.compute.manager [None req-ae27ad25-8527-4c20-9aca-0f14006c890e tempest-ServersAdmin275Test-220579119 tempest-ServersAdmin275Test-220579119-project-admin] [instance: 93b28142-8454-43c0-b0a7-d61aa95c8fc8] Checking state {{(pid=63515) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 589.167441] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22d97cb4-d20e-4f54-9db5-c9c6f1e22d04 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 589.646691] env[63515]: DEBUG nova.network.neutron [None req-12d73cc4-118d-457e-b393-cc4362d15b0d tempest-ListImageFiltersTestJSON-1100101045 tempest-ListImageFiltersTestJSON-1100101045-project-member] [instance: ab1539ba-33a4-4139-a2f3-1f8c507beede] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 589.651582] env[63515]: DEBUG nova.compute.utils [None req-c80cd50f-4127-4451-9298-83a67d8f0ab6 tempest-ServersTestFqdnHostnames-1452522528 tempest-ServersTestFqdnHostnames-1452522528-project-member] Using /dev/sd instead of None {{(pid=63515) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 589.658085] env[63515]: DEBUG nova.compute.manager [None req-c80cd50f-4127-4451-9298-83a67d8f0ab6 tempest-ServersTestFqdnHostnames-1452522528 tempest-ServersTestFqdnHostnames-1452522528-project-member] [instance: b84c6c8d-e358-4a14-8425-c4f1d8adf4ee] Allocating IP information in the background. {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 589.658085] env[63515]: DEBUG nova.network.neutron [None req-c80cd50f-4127-4451-9298-83a67d8f0ab6 tempest-ServersTestFqdnHostnames-1452522528 tempest-ServersTestFqdnHostnames-1452522528-project-member] [instance: b84c6c8d-e358-4a14-8425-c4f1d8adf4ee] allocate_for_instance() {{(pid=63515) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 589.688672] env[63515]: DEBUG oslo_concurrency.lockutils [None req-ae27ad25-8527-4c20-9aca-0f14006c890e tempest-ServersAdmin275Test-220579119 tempest-ServersAdmin275Test-220579119-project-admin] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 589.779239] env[63515]: DEBUG nova.policy [None req-c80cd50f-4127-4451-9298-83a67d8f0ab6 tempest-ServersTestFqdnHostnames-1452522528 tempest-ServersTestFqdnHostnames-1452522528-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e5239c37501c49989fdda2696f98ea96', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a5a4ab0a8d38434da361d4977ff19ecd', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63515) authorize /opt/stack/nova/nova/policy.py:201}} [ 589.796148] env[63515]: DEBUG nova.network.neutron [None req-12d73cc4-118d-457e-b393-cc4362d15b0d tempest-ListImageFiltersTestJSON-1100101045 tempest-ListImageFiltersTestJSON-1100101045-project-member] [instance: ab1539ba-33a4-4139-a2f3-1f8c507beede] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 589.939531] env[63515]: INFO nova.scheduler.client.report [None req-eaa7f189-a861-45f8-872f-29d870dc784b tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] Deleted allocations for instance e39bc55e-b677-4e2d-9aa4-7bc1d5ffadfa [ 590.157219] env[63515]: DEBUG nova.compute.manager [None req-c80cd50f-4127-4451-9298-83a67d8f0ab6 tempest-ServersTestFqdnHostnames-1452522528 tempest-ServersTestFqdnHostnames-1452522528-project-member] [instance: b84c6c8d-e358-4a14-8425-c4f1d8adf4ee] Start building block device mappings for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 590.300211] env[63515]: DEBUG oslo_concurrency.lockutils [None req-12d73cc4-118d-457e-b393-cc4362d15b0d tempest-ListImageFiltersTestJSON-1100101045 tempest-ListImageFiltersTestJSON-1100101045-project-member] Releasing lock "refresh_cache-ab1539ba-33a4-4139-a2f3-1f8c507beede" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 590.301343] env[63515]: DEBUG nova.compute.manager [None req-12d73cc4-118d-457e-b393-cc4362d15b0d tempest-ListImageFiltersTestJSON-1100101045 tempest-ListImageFiltersTestJSON-1100101045-project-member] [instance: ab1539ba-33a4-4139-a2f3-1f8c507beede] Start destroying the instance on the hypervisor. {{(pid=63515) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 590.301343] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-12d73cc4-118d-457e-b393-cc4362d15b0d tempest-ListImageFiltersTestJSON-1100101045 tempest-ListImageFiltersTestJSON-1100101045-project-member] [instance: ab1539ba-33a4-4139-a2f3-1f8c507beede] Destroying instance {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 590.301957] env[63515]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-3b48f91b-7dfb-4932-ac22-99dca3286828 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 590.313025] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3e77752-dd12-4c5d-95f9-75e5e763dd5e {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 590.345241] env[63515]: WARNING nova.virt.vmwareapi.vmops [None req-12d73cc4-118d-457e-b393-cc4362d15b0d tempest-ListImageFiltersTestJSON-1100101045 tempest-ListImageFiltersTestJSON-1100101045-project-member] [instance: ab1539ba-33a4-4139-a2f3-1f8c507beede] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance ab1539ba-33a4-4139-a2f3-1f8c507beede could not be found. [ 590.346150] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-12d73cc4-118d-457e-b393-cc4362d15b0d tempest-ListImageFiltersTestJSON-1100101045 tempest-ListImageFiltersTestJSON-1100101045-project-member] [instance: ab1539ba-33a4-4139-a2f3-1f8c507beede] Instance destroyed {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 590.346150] env[63515]: INFO nova.compute.manager [None req-12d73cc4-118d-457e-b393-cc4362d15b0d tempest-ListImageFiltersTestJSON-1100101045 tempest-ListImageFiltersTestJSON-1100101045-project-member] [instance: ab1539ba-33a4-4139-a2f3-1f8c507beede] Took 0.04 seconds to destroy the instance on the hypervisor. [ 590.346150] env[63515]: DEBUG oslo.service.loopingcall [None req-12d73cc4-118d-457e-b393-cc4362d15b0d tempest-ListImageFiltersTestJSON-1100101045 tempest-ListImageFiltersTestJSON-1100101045-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 590.347161] env[63515]: DEBUG nova.network.neutron [None req-c80cd50f-4127-4451-9298-83a67d8f0ab6 tempest-ServersTestFqdnHostnames-1452522528 tempest-ServersTestFqdnHostnames-1452522528-project-member] [instance: b84c6c8d-e358-4a14-8425-c4f1d8adf4ee] Successfully created port: 3d61ba1c-4556-41b6-a699-f5b65b0de93f {{(pid=63515) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 590.356046] env[63515]: DEBUG nova.compute.manager [-] [instance: ab1539ba-33a4-4139-a2f3-1f8c507beede] Deallocating network for instance {{(pid=63515) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 590.356046] env[63515]: DEBUG nova.network.neutron [-] [instance: ab1539ba-33a4-4139-a2f3-1f8c507beede] deallocate_for_instance() {{(pid=63515) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 590.388800] env[63515]: DEBUG nova.network.neutron [-] [instance: ab1539ba-33a4-4139-a2f3-1f8c507beede] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 590.452511] env[63515]: DEBUG oslo_concurrency.lockutils [None req-eaa7f189-a861-45f8-872f-29d870dc784b tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] Lock "e39bc55e-b677-4e2d-9aa4-7bc1d5ffadfa" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 48.141s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 590.586764] env[63515]: DEBUG nova.compute.manager [req-2a5371d6-dfec-4862-8a01-d258dcf64eb3 req-aebffe14-db98-4ee8-a34d-f0fcabf4e708 service nova] [instance: ab1539ba-33a4-4139-a2f3-1f8c507beede] Received event network-vif-deleted-bd569ca1-70ab-4c56-b6af-9ae3816c1d2c {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 590.633439] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bada8adf-7f75-46d6-8d0e-49cbf8bac9a4 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 590.642292] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc74c962-3786-447d-82ed-b90592c3be1f {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 590.682085] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7585146-6d7e-4db8-b179-de9c67b52b62 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 590.690119] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f215ac39-fd66-4c26-a038-b23ec5cf6019 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 590.703779] env[63515]: DEBUG nova.compute.provider_tree [None req-22f02712-96d1-4031-8b8f-5bc880f34c33 tempest-ImagesNegativeTestJSON-2041328384 tempest-ImagesNegativeTestJSON-2041328384-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 590.895063] env[63515]: DEBUG nova.network.neutron [-] [instance: ab1539ba-33a4-4139-a2f3-1f8c507beede] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 590.956109] env[63515]: DEBUG nova.compute.manager [None req-886adfc0-007d-4f83-bc30-b755d00d04c7 tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] [instance: 8b65c223-03a8-41e0-917e-45dd76b6e57f] Starting instance... {{(pid=63515) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 591.187417] env[63515]: DEBUG nova.compute.manager [None req-c80cd50f-4127-4451-9298-83a67d8f0ab6 tempest-ServersTestFqdnHostnames-1452522528 tempest-ServersTestFqdnHostnames-1452522528-project-member] [instance: b84c6c8d-e358-4a14-8425-c4f1d8adf4ee] Start spawning the instance on the hypervisor. {{(pid=63515) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 591.207092] env[63515]: DEBUG nova.scheduler.client.report [None req-22f02712-96d1-4031-8b8f-5bc880f34c33 tempest-ImagesNegativeTestJSON-2041328384 tempest-ImagesNegativeTestJSON-2041328384-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 591.221011] env[63515]: DEBUG nova.virt.hardware [None req-c80cd50f-4127-4451-9298-83a67d8f0ab6 tempest-ServersTestFqdnHostnames-1452522528 tempest-ServersTestFqdnHostnames-1452522528-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T02:52:14Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T02:51:56Z,direct_url=,disk_format='vmdk',id=8a120570-cb06-4099-b262-554ca0ad15c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b569255cc43e42e7ae3f2b4ad37c6ef4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T02:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 591.221815] env[63515]: DEBUG nova.virt.hardware [None req-c80cd50f-4127-4451-9298-83a67d8f0ab6 tempest-ServersTestFqdnHostnames-1452522528 tempest-ServersTestFqdnHostnames-1452522528-project-member] Flavor limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 591.221815] env[63515]: DEBUG nova.virt.hardware [None req-c80cd50f-4127-4451-9298-83a67d8f0ab6 tempest-ServersTestFqdnHostnames-1452522528 tempest-ServersTestFqdnHostnames-1452522528-project-member] Image limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 591.221815] env[63515]: DEBUG nova.virt.hardware [None req-c80cd50f-4127-4451-9298-83a67d8f0ab6 tempest-ServersTestFqdnHostnames-1452522528 tempest-ServersTestFqdnHostnames-1452522528-project-member] Flavor pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 591.221995] env[63515]: DEBUG nova.virt.hardware [None req-c80cd50f-4127-4451-9298-83a67d8f0ab6 tempest-ServersTestFqdnHostnames-1452522528 tempest-ServersTestFqdnHostnames-1452522528-project-member] Image pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 591.221995] env[63515]: DEBUG nova.virt.hardware [None req-c80cd50f-4127-4451-9298-83a67d8f0ab6 tempest-ServersTestFqdnHostnames-1452522528 tempest-ServersTestFqdnHostnames-1452522528-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 591.222366] env[63515]: DEBUG nova.virt.hardware [None req-c80cd50f-4127-4451-9298-83a67d8f0ab6 tempest-ServersTestFqdnHostnames-1452522528 tempest-ServersTestFqdnHostnames-1452522528-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 591.222802] env[63515]: DEBUG nova.virt.hardware [None req-c80cd50f-4127-4451-9298-83a67d8f0ab6 tempest-ServersTestFqdnHostnames-1452522528 tempest-ServersTestFqdnHostnames-1452522528-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 591.223779] env[63515]: DEBUG nova.virt.hardware [None req-c80cd50f-4127-4451-9298-83a67d8f0ab6 tempest-ServersTestFqdnHostnames-1452522528 tempest-ServersTestFqdnHostnames-1452522528-project-member] Got 1 possible topologies {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 591.223779] env[63515]: DEBUG nova.virt.hardware [None req-c80cd50f-4127-4451-9298-83a67d8f0ab6 tempest-ServersTestFqdnHostnames-1452522528 tempest-ServersTestFqdnHostnames-1452522528-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 591.223779] env[63515]: DEBUG nova.virt.hardware [None req-c80cd50f-4127-4451-9298-83a67d8f0ab6 tempest-ServersTestFqdnHostnames-1452522528 tempest-ServersTestFqdnHostnames-1452522528-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 591.224327] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-230a4ea6-202e-472c-8ef5-3749aa89c3ed {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 591.237564] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec428c2a-09df-4bea-a266-1fe0870f26d4 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 591.397391] env[63515]: INFO nova.compute.manager [-] [instance: ab1539ba-33a4-4139-a2f3-1f8c507beede] Took 1.04 seconds to deallocate network for instance. [ 591.404827] env[63515]: DEBUG nova.compute.claims [None req-12d73cc4-118d-457e-b393-cc4362d15b0d tempest-ListImageFiltersTestJSON-1100101045 tempest-ListImageFiltersTestJSON-1100101045-project-member] [instance: ab1539ba-33a4-4139-a2f3-1f8c507beede] Aborting claim: {{(pid=63515) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 591.405066] env[63515]: DEBUG oslo_concurrency.lockutils [None req-12d73cc4-118d-457e-b393-cc4362d15b0d tempest-ListImageFiltersTestJSON-1100101045 tempest-ListImageFiltersTestJSON-1100101045-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 591.483604] env[63515]: DEBUG oslo_concurrency.lockutils [None req-886adfc0-007d-4f83-bc30-b755d00d04c7 tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 591.714767] env[63515]: DEBUG oslo_concurrency.lockutils [None req-22f02712-96d1-4031-8b8f-5bc880f34c33 tempest-ImagesNegativeTestJSON-2041328384 tempest-ImagesNegativeTestJSON-2041328384-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.572s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 591.715320] env[63515]: DEBUG nova.compute.manager [None req-22f02712-96d1-4031-8b8f-5bc880f34c33 tempest-ImagesNegativeTestJSON-2041328384 tempest-ImagesNegativeTestJSON-2041328384-project-member] [instance: d5f988f2-765b-4ecf-bad5-4a70fc95d74c] Start building networks asynchronously for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 591.721029] env[63515]: DEBUG oslo_concurrency.lockutils [None req-9e973647-d524-4c71-b05a-f1a253758bab tempest-ServerAddressesTestJSON-9863250 tempest-ServerAddressesTestJSON-9863250-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.262s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 591.721343] env[63515]: INFO nova.compute.claims [None req-9e973647-d524-4c71-b05a-f1a253758bab tempest-ServerAddressesTestJSON-9863250 tempest-ServerAddressesTestJSON-9863250-project-member] [instance: 7fee0afb-88c9-4e93-a4d9-94312394f927] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 591.899357] env[63515]: DEBUG oslo_concurrency.lockutils [None req-c26675e2-4152-417b-99dd-0c18de4dc137 tempest-ServersTestBootFromVolume-2042560423 tempest-ServersTestBootFromVolume-2042560423-project-member] Acquiring lock "90f4930b-aaa0-4c4b-9ab8-92aed45e200b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 591.899976] env[63515]: DEBUG oslo_concurrency.lockutils [None req-c26675e2-4152-417b-99dd-0c18de4dc137 tempest-ServersTestBootFromVolume-2042560423 tempest-ServersTestBootFromVolume-2042560423-project-member] Lock "90f4930b-aaa0-4c4b-9ab8-92aed45e200b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 591.988910] env[63515]: ERROR nova.compute.manager [None req-c80cd50f-4127-4451-9298-83a67d8f0ab6 tempest-ServersTestFqdnHostnames-1452522528 tempest-ServersTestFqdnHostnames-1452522528-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 3d61ba1c-4556-41b6-a699-f5b65b0de93f, please check neutron logs for more information. [ 591.988910] env[63515]: ERROR nova.compute.manager Traceback (most recent call last): [ 591.988910] env[63515]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 591.988910] env[63515]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 591.988910] env[63515]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 591.988910] env[63515]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 591.988910] env[63515]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 591.988910] env[63515]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 591.988910] env[63515]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 591.988910] env[63515]: ERROR nova.compute.manager self.force_reraise() [ 591.988910] env[63515]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 591.988910] env[63515]: ERROR nova.compute.manager raise self.value [ 591.988910] env[63515]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 591.988910] env[63515]: ERROR nova.compute.manager updated_port = self._update_port( [ 591.988910] env[63515]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 591.988910] env[63515]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 591.989611] env[63515]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 591.989611] env[63515]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 591.989611] env[63515]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 3d61ba1c-4556-41b6-a699-f5b65b0de93f, please check neutron logs for more information. [ 591.989611] env[63515]: ERROR nova.compute.manager [ 591.989611] env[63515]: Traceback (most recent call last): [ 591.989910] env[63515]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 591.989910] env[63515]: listener.cb(fileno) [ 591.989910] env[63515]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 591.989910] env[63515]: result = function(*args, **kwargs) [ 591.989910] env[63515]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 591.989910] env[63515]: return func(*args, **kwargs) [ 591.989910] env[63515]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 591.989910] env[63515]: raise e [ 591.989910] env[63515]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 591.989910] env[63515]: nwinfo = self.network_api.allocate_for_instance( [ 591.989910] env[63515]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 591.989910] env[63515]: created_port_ids = self._update_ports_for_instance( [ 591.989910] env[63515]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 591.989910] env[63515]: with excutils.save_and_reraise_exception(): [ 591.989910] env[63515]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 591.989910] env[63515]: self.force_reraise() [ 591.989910] env[63515]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 591.989910] env[63515]: raise self.value [ 591.989910] env[63515]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 591.989910] env[63515]: updated_port = self._update_port( [ 591.989910] env[63515]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 591.989910] env[63515]: _ensure_no_port_binding_failure(port) [ 591.989910] env[63515]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 591.989910] env[63515]: raise exception.PortBindingFailed(port_id=port['id']) [ 591.989910] env[63515]: nova.exception.PortBindingFailed: Binding failed for port 3d61ba1c-4556-41b6-a699-f5b65b0de93f, please check neutron logs for more information. [ 591.989910] env[63515]: Removing descriptor: 16 [ 591.990965] env[63515]: ERROR nova.compute.manager [None req-c80cd50f-4127-4451-9298-83a67d8f0ab6 tempest-ServersTestFqdnHostnames-1452522528 tempest-ServersTestFqdnHostnames-1452522528-project-member] [instance: b84c6c8d-e358-4a14-8425-c4f1d8adf4ee] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 3d61ba1c-4556-41b6-a699-f5b65b0de93f, please check neutron logs for more information. [ 591.990965] env[63515]: ERROR nova.compute.manager [instance: b84c6c8d-e358-4a14-8425-c4f1d8adf4ee] Traceback (most recent call last): [ 591.990965] env[63515]: ERROR nova.compute.manager [instance: b84c6c8d-e358-4a14-8425-c4f1d8adf4ee] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 591.990965] env[63515]: ERROR nova.compute.manager [instance: b84c6c8d-e358-4a14-8425-c4f1d8adf4ee] yield resources [ 591.990965] env[63515]: ERROR nova.compute.manager [instance: b84c6c8d-e358-4a14-8425-c4f1d8adf4ee] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 591.990965] env[63515]: ERROR nova.compute.manager [instance: b84c6c8d-e358-4a14-8425-c4f1d8adf4ee] self.driver.spawn(context, instance, image_meta, [ 591.990965] env[63515]: ERROR nova.compute.manager [instance: b84c6c8d-e358-4a14-8425-c4f1d8adf4ee] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 591.990965] env[63515]: ERROR nova.compute.manager [instance: b84c6c8d-e358-4a14-8425-c4f1d8adf4ee] self._vmops.spawn(context, instance, image_meta, injected_files, [ 591.990965] env[63515]: ERROR nova.compute.manager [instance: b84c6c8d-e358-4a14-8425-c4f1d8adf4ee] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 591.990965] env[63515]: ERROR nova.compute.manager [instance: b84c6c8d-e358-4a14-8425-c4f1d8adf4ee] vm_ref = self.build_virtual_machine(instance, [ 591.990965] env[63515]: ERROR nova.compute.manager [instance: b84c6c8d-e358-4a14-8425-c4f1d8adf4ee] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 591.992527] env[63515]: ERROR nova.compute.manager [instance: b84c6c8d-e358-4a14-8425-c4f1d8adf4ee] vif_infos = vmwarevif.get_vif_info(self._session, [ 591.992527] env[63515]: ERROR nova.compute.manager [instance: b84c6c8d-e358-4a14-8425-c4f1d8adf4ee] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 591.992527] env[63515]: ERROR nova.compute.manager [instance: b84c6c8d-e358-4a14-8425-c4f1d8adf4ee] for vif in network_info: [ 591.992527] env[63515]: ERROR nova.compute.manager [instance: b84c6c8d-e358-4a14-8425-c4f1d8adf4ee] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 591.992527] env[63515]: ERROR nova.compute.manager [instance: b84c6c8d-e358-4a14-8425-c4f1d8adf4ee] return self._sync_wrapper(fn, *args, **kwargs) [ 591.992527] env[63515]: ERROR nova.compute.manager [instance: b84c6c8d-e358-4a14-8425-c4f1d8adf4ee] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 591.992527] env[63515]: ERROR nova.compute.manager [instance: b84c6c8d-e358-4a14-8425-c4f1d8adf4ee] self.wait() [ 591.992527] env[63515]: ERROR nova.compute.manager [instance: b84c6c8d-e358-4a14-8425-c4f1d8adf4ee] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 591.992527] env[63515]: ERROR nova.compute.manager [instance: b84c6c8d-e358-4a14-8425-c4f1d8adf4ee] self[:] = self._gt.wait() [ 591.992527] env[63515]: ERROR nova.compute.manager [instance: b84c6c8d-e358-4a14-8425-c4f1d8adf4ee] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 591.992527] env[63515]: ERROR nova.compute.manager [instance: b84c6c8d-e358-4a14-8425-c4f1d8adf4ee] return self._exit_event.wait() [ 591.992527] env[63515]: ERROR nova.compute.manager [instance: b84c6c8d-e358-4a14-8425-c4f1d8adf4ee] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 591.992527] env[63515]: ERROR nova.compute.manager [instance: b84c6c8d-e358-4a14-8425-c4f1d8adf4ee] result = hub.switch() [ 591.993094] env[63515]: ERROR nova.compute.manager [instance: b84c6c8d-e358-4a14-8425-c4f1d8adf4ee] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 591.993094] env[63515]: ERROR nova.compute.manager [instance: b84c6c8d-e358-4a14-8425-c4f1d8adf4ee] return self.greenlet.switch() [ 591.993094] env[63515]: ERROR nova.compute.manager [instance: b84c6c8d-e358-4a14-8425-c4f1d8adf4ee] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 591.993094] env[63515]: ERROR nova.compute.manager [instance: b84c6c8d-e358-4a14-8425-c4f1d8adf4ee] result = function(*args, **kwargs) [ 591.993094] env[63515]: ERROR nova.compute.manager [instance: b84c6c8d-e358-4a14-8425-c4f1d8adf4ee] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 591.993094] env[63515]: ERROR nova.compute.manager [instance: b84c6c8d-e358-4a14-8425-c4f1d8adf4ee] return func(*args, **kwargs) [ 591.993094] env[63515]: ERROR nova.compute.manager [instance: b84c6c8d-e358-4a14-8425-c4f1d8adf4ee] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 591.993094] env[63515]: ERROR nova.compute.manager [instance: b84c6c8d-e358-4a14-8425-c4f1d8adf4ee] raise e [ 591.993094] env[63515]: ERROR nova.compute.manager [instance: b84c6c8d-e358-4a14-8425-c4f1d8adf4ee] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 591.993094] env[63515]: ERROR nova.compute.manager [instance: b84c6c8d-e358-4a14-8425-c4f1d8adf4ee] nwinfo = self.network_api.allocate_for_instance( [ 591.993094] env[63515]: ERROR nova.compute.manager [instance: b84c6c8d-e358-4a14-8425-c4f1d8adf4ee] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 591.993094] env[63515]: ERROR nova.compute.manager [instance: b84c6c8d-e358-4a14-8425-c4f1d8adf4ee] created_port_ids = self._update_ports_for_instance( [ 591.993094] env[63515]: ERROR nova.compute.manager [instance: b84c6c8d-e358-4a14-8425-c4f1d8adf4ee] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 591.993604] env[63515]: ERROR nova.compute.manager [instance: b84c6c8d-e358-4a14-8425-c4f1d8adf4ee] with excutils.save_and_reraise_exception(): [ 591.993604] env[63515]: ERROR nova.compute.manager [instance: b84c6c8d-e358-4a14-8425-c4f1d8adf4ee] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 591.993604] env[63515]: ERROR nova.compute.manager [instance: b84c6c8d-e358-4a14-8425-c4f1d8adf4ee] self.force_reraise() [ 591.993604] env[63515]: ERROR nova.compute.manager [instance: b84c6c8d-e358-4a14-8425-c4f1d8adf4ee] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 591.993604] env[63515]: ERROR nova.compute.manager [instance: b84c6c8d-e358-4a14-8425-c4f1d8adf4ee] raise self.value [ 591.993604] env[63515]: ERROR nova.compute.manager [instance: b84c6c8d-e358-4a14-8425-c4f1d8adf4ee] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 591.993604] env[63515]: ERROR nova.compute.manager [instance: b84c6c8d-e358-4a14-8425-c4f1d8adf4ee] updated_port = self._update_port( [ 591.993604] env[63515]: ERROR nova.compute.manager [instance: b84c6c8d-e358-4a14-8425-c4f1d8adf4ee] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 591.993604] env[63515]: ERROR nova.compute.manager [instance: b84c6c8d-e358-4a14-8425-c4f1d8adf4ee] _ensure_no_port_binding_failure(port) [ 591.993604] env[63515]: ERROR nova.compute.manager [instance: b84c6c8d-e358-4a14-8425-c4f1d8adf4ee] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 591.993604] env[63515]: ERROR nova.compute.manager [instance: b84c6c8d-e358-4a14-8425-c4f1d8adf4ee] raise exception.PortBindingFailed(port_id=port['id']) [ 591.993604] env[63515]: ERROR nova.compute.manager [instance: b84c6c8d-e358-4a14-8425-c4f1d8adf4ee] nova.exception.PortBindingFailed: Binding failed for port 3d61ba1c-4556-41b6-a699-f5b65b0de93f, please check neutron logs for more information. [ 591.993604] env[63515]: ERROR nova.compute.manager [instance: b84c6c8d-e358-4a14-8425-c4f1d8adf4ee] [ 591.994173] env[63515]: INFO nova.compute.manager [None req-c80cd50f-4127-4451-9298-83a67d8f0ab6 tempest-ServersTestFqdnHostnames-1452522528 tempest-ServersTestFqdnHostnames-1452522528-project-member] [instance: b84c6c8d-e358-4a14-8425-c4f1d8adf4ee] Terminating instance [ 591.994719] env[63515]: DEBUG oslo_concurrency.lockutils [None req-c80cd50f-4127-4451-9298-83a67d8f0ab6 tempest-ServersTestFqdnHostnames-1452522528 tempest-ServersTestFqdnHostnames-1452522528-project-member] Acquiring lock "refresh_cache-b84c6c8d-e358-4a14-8425-c4f1d8adf4ee" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 591.994888] env[63515]: DEBUG oslo_concurrency.lockutils [None req-c80cd50f-4127-4451-9298-83a67d8f0ab6 tempest-ServersTestFqdnHostnames-1452522528 tempest-ServersTestFqdnHostnames-1452522528-project-member] Acquired lock "refresh_cache-b84c6c8d-e358-4a14-8425-c4f1d8adf4ee" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 591.995067] env[63515]: DEBUG nova.network.neutron [None req-c80cd50f-4127-4451-9298-83a67d8f0ab6 tempest-ServersTestFqdnHostnames-1452522528 tempest-ServersTestFqdnHostnames-1452522528-project-member] [instance: b84c6c8d-e358-4a14-8425-c4f1d8adf4ee] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 592.221679] env[63515]: DEBUG nova.compute.utils [None req-22f02712-96d1-4031-8b8f-5bc880f34c33 tempest-ImagesNegativeTestJSON-2041328384 tempest-ImagesNegativeTestJSON-2041328384-project-member] Using /dev/sd instead of None {{(pid=63515) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 592.223165] env[63515]: DEBUG nova.compute.manager [None req-22f02712-96d1-4031-8b8f-5bc880f34c33 tempest-ImagesNegativeTestJSON-2041328384 tempest-ImagesNegativeTestJSON-2041328384-project-member] [instance: d5f988f2-765b-4ecf-bad5-4a70fc95d74c] Allocating IP information in the background. {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 592.223361] env[63515]: DEBUG nova.network.neutron [None req-22f02712-96d1-4031-8b8f-5bc880f34c33 tempest-ImagesNegativeTestJSON-2041328384 tempest-ImagesNegativeTestJSON-2041328384-project-member] [instance: d5f988f2-765b-4ecf-bad5-4a70fc95d74c] allocate_for_instance() {{(pid=63515) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 592.299186] env[63515]: DEBUG nova.policy [None req-22f02712-96d1-4031-8b8f-5bc880f34c33 tempest-ImagesNegativeTestJSON-2041328384 tempest-ImagesNegativeTestJSON-2041328384-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '3c4d97c1e61c4c2eabb131ae6f2c84b0', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'eda248e44fa0417089969f86a7573fa1', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63515) authorize /opt/stack/nova/nova/policy.py:201}} [ 592.529869] env[63515]: DEBUG nova.network.neutron [None req-c80cd50f-4127-4451-9298-83a67d8f0ab6 tempest-ServersTestFqdnHostnames-1452522528 tempest-ServersTestFqdnHostnames-1452522528-project-member] [instance: b84c6c8d-e358-4a14-8425-c4f1d8adf4ee] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 592.577676] env[63515]: DEBUG oslo_concurrency.lockutils [None req-eb536733-3a8a-4dae-88ae-0ad9d450b524 tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] Acquiring lock "93b28142-8454-43c0-b0a7-d61aa95c8fc8" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 592.577971] env[63515]: DEBUG oslo_concurrency.lockutils [None req-eb536733-3a8a-4dae-88ae-0ad9d450b524 tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] Lock "93b28142-8454-43c0-b0a7-d61aa95c8fc8" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 592.578208] env[63515]: DEBUG oslo_concurrency.lockutils [None req-eb536733-3a8a-4dae-88ae-0ad9d450b524 tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] Acquiring lock "93b28142-8454-43c0-b0a7-d61aa95c8fc8-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 592.578392] env[63515]: DEBUG oslo_concurrency.lockutils [None req-eb536733-3a8a-4dae-88ae-0ad9d450b524 tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] Lock "93b28142-8454-43c0-b0a7-d61aa95c8fc8-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 592.578560] env[63515]: DEBUG oslo_concurrency.lockutils [None req-eb536733-3a8a-4dae-88ae-0ad9d450b524 tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] Lock "93b28142-8454-43c0-b0a7-d61aa95c8fc8-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 592.581608] env[63515]: INFO nova.compute.manager [None req-eb536733-3a8a-4dae-88ae-0ad9d450b524 tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] [instance: 93b28142-8454-43c0-b0a7-d61aa95c8fc8] Terminating instance [ 592.586759] env[63515]: DEBUG oslo_concurrency.lockutils [None req-eb536733-3a8a-4dae-88ae-0ad9d450b524 tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] Acquiring lock "refresh_cache-93b28142-8454-43c0-b0a7-d61aa95c8fc8" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 592.586759] env[63515]: DEBUG oslo_concurrency.lockutils [None req-eb536733-3a8a-4dae-88ae-0ad9d450b524 tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] Acquired lock "refresh_cache-93b28142-8454-43c0-b0a7-d61aa95c8fc8" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 592.586759] env[63515]: DEBUG nova.network.neutron [None req-eb536733-3a8a-4dae-88ae-0ad9d450b524 tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] [instance: 93b28142-8454-43c0-b0a7-d61aa95c8fc8] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 592.715440] env[63515]: DEBUG nova.network.neutron [None req-c80cd50f-4127-4451-9298-83a67d8f0ab6 tempest-ServersTestFqdnHostnames-1452522528 tempest-ServersTestFqdnHostnames-1452522528-project-member] [instance: b84c6c8d-e358-4a14-8425-c4f1d8adf4ee] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 592.730443] env[63515]: DEBUG nova.compute.manager [None req-22f02712-96d1-4031-8b8f-5bc880f34c33 tempest-ImagesNegativeTestJSON-2041328384 tempest-ImagesNegativeTestJSON-2041328384-project-member] [instance: d5f988f2-765b-4ecf-bad5-4a70fc95d74c] Start building block device mappings for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 592.740178] env[63515]: DEBUG nova.network.neutron [None req-22f02712-96d1-4031-8b8f-5bc880f34c33 tempest-ImagesNegativeTestJSON-2041328384 tempest-ImagesNegativeTestJSON-2041328384-project-member] [instance: d5f988f2-765b-4ecf-bad5-4a70fc95d74c] Successfully created port: fca51b1b-a915-4365-a1b6-ccc0a5a43604 {{(pid=63515) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 593.121010] env[63515]: DEBUG nova.network.neutron [None req-eb536733-3a8a-4dae-88ae-0ad9d450b524 tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] [instance: 93b28142-8454-43c0-b0a7-d61aa95c8fc8] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 593.149648] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-106ffdcd-b411-485e-b2c1-237856a8fd98 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 593.160741] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38f335fe-c1ac-4e3c-b140-dc761f2d0ff0 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 593.199429] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8282b62b-ecad-4397-bbac-5cd03efabb53 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 593.207218] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4f658b1-1bc7-4490-afd7-db09490d969a {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 593.212794] env[63515]: DEBUG nova.compute.manager [req-dad27614-a688-416e-9a95-f66417b24a6a req-df39a862-095f-4c0a-a316-1b76b773110e service nova] [instance: b84c6c8d-e358-4a14-8425-c4f1d8adf4ee] Received event network-changed-3d61ba1c-4556-41b6-a699-f5b65b0de93f {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 593.212907] env[63515]: DEBUG nova.compute.manager [req-dad27614-a688-416e-9a95-f66417b24a6a req-df39a862-095f-4c0a-a316-1b76b773110e service nova] [instance: b84c6c8d-e358-4a14-8425-c4f1d8adf4ee] Refreshing instance network info cache due to event network-changed-3d61ba1c-4556-41b6-a699-f5b65b0de93f. {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 593.213113] env[63515]: DEBUG oslo_concurrency.lockutils [req-dad27614-a688-416e-9a95-f66417b24a6a req-df39a862-095f-4c0a-a316-1b76b773110e service nova] Acquiring lock "refresh_cache-b84c6c8d-e358-4a14-8425-c4f1d8adf4ee" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 593.223121] env[63515]: DEBUG oslo_concurrency.lockutils [None req-c80cd50f-4127-4451-9298-83a67d8f0ab6 tempest-ServersTestFqdnHostnames-1452522528 tempest-ServersTestFqdnHostnames-1452522528-project-member] Releasing lock "refresh_cache-b84c6c8d-e358-4a14-8425-c4f1d8adf4ee" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 593.223549] env[63515]: DEBUG nova.compute.manager [None req-c80cd50f-4127-4451-9298-83a67d8f0ab6 tempest-ServersTestFqdnHostnames-1452522528 tempest-ServersTestFqdnHostnames-1452522528-project-member] [instance: b84c6c8d-e358-4a14-8425-c4f1d8adf4ee] Start destroying the instance on the hypervisor. {{(pid=63515) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 593.223844] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-c80cd50f-4127-4451-9298-83a67d8f0ab6 tempest-ServersTestFqdnHostnames-1452522528 tempest-ServersTestFqdnHostnames-1452522528-project-member] [instance: b84c6c8d-e358-4a14-8425-c4f1d8adf4ee] Destroying instance {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 593.224199] env[63515]: DEBUG nova.compute.provider_tree [None req-9e973647-d524-4c71-b05a-f1a253758bab tempest-ServerAddressesTestJSON-9863250 tempest-ServerAddressesTestJSON-9863250-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 593.225302] env[63515]: DEBUG oslo_concurrency.lockutils [req-dad27614-a688-416e-9a95-f66417b24a6a req-df39a862-095f-4c0a-a316-1b76b773110e service nova] Acquired lock "refresh_cache-b84c6c8d-e358-4a14-8425-c4f1d8adf4ee" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 593.225481] env[63515]: DEBUG nova.network.neutron [req-dad27614-a688-416e-9a95-f66417b24a6a req-df39a862-095f-4c0a-a316-1b76b773110e service nova] [instance: b84c6c8d-e358-4a14-8425-c4f1d8adf4ee] Refreshing network info cache for port 3d61ba1c-4556-41b6-a699-f5b65b0de93f {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 593.226386] env[63515]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d9e3b41a-fc6e-4ba9-8f40-3802de9cc914 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 593.236465] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bdbfdc8f-e926-4a87-9648-49679795e047 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 593.262527] env[63515]: WARNING nova.virt.vmwareapi.vmops [None req-c80cd50f-4127-4451-9298-83a67d8f0ab6 tempest-ServersTestFqdnHostnames-1452522528 tempest-ServersTestFqdnHostnames-1452522528-project-member] [instance: b84c6c8d-e358-4a14-8425-c4f1d8adf4ee] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance b84c6c8d-e358-4a14-8425-c4f1d8adf4ee could not be found. [ 593.262764] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-c80cd50f-4127-4451-9298-83a67d8f0ab6 tempest-ServersTestFqdnHostnames-1452522528 tempest-ServersTestFqdnHostnames-1452522528-project-member] [instance: b84c6c8d-e358-4a14-8425-c4f1d8adf4ee] Instance destroyed {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 593.262926] env[63515]: INFO nova.compute.manager [None req-c80cd50f-4127-4451-9298-83a67d8f0ab6 tempest-ServersTestFqdnHostnames-1452522528 tempest-ServersTestFqdnHostnames-1452522528-project-member] [instance: b84c6c8d-e358-4a14-8425-c4f1d8adf4ee] Took 0.04 seconds to destroy the instance on the hypervisor. [ 593.263181] env[63515]: DEBUG oslo.service.loopingcall [None req-c80cd50f-4127-4451-9298-83a67d8f0ab6 tempest-ServersTestFqdnHostnames-1452522528 tempest-ServersTestFqdnHostnames-1452522528-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 593.263437] env[63515]: DEBUG nova.compute.manager [-] [instance: b84c6c8d-e358-4a14-8425-c4f1d8adf4ee] Deallocating network for instance {{(pid=63515) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 593.263533] env[63515]: DEBUG nova.network.neutron [-] [instance: b84c6c8d-e358-4a14-8425-c4f1d8adf4ee] deallocate_for_instance() {{(pid=63515) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 593.273263] env[63515]: DEBUG nova.network.neutron [None req-eb536733-3a8a-4dae-88ae-0ad9d450b524 tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] [instance: 93b28142-8454-43c0-b0a7-d61aa95c8fc8] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 593.306977] env[63515]: DEBUG nova.network.neutron [-] [instance: b84c6c8d-e358-4a14-8425-c4f1d8adf4ee] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 593.736009] env[63515]: DEBUG nova.scheduler.client.report [None req-9e973647-d524-4c71-b05a-f1a253758bab tempest-ServerAddressesTestJSON-9863250 tempest-ServerAddressesTestJSON-9863250-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 593.752785] env[63515]: DEBUG nova.compute.manager [None req-22f02712-96d1-4031-8b8f-5bc880f34c33 tempest-ImagesNegativeTestJSON-2041328384 tempest-ImagesNegativeTestJSON-2041328384-project-member] [instance: d5f988f2-765b-4ecf-bad5-4a70fc95d74c] Start spawning the instance on the hypervisor. {{(pid=63515) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 593.767702] env[63515]: DEBUG nova.network.neutron [req-dad27614-a688-416e-9a95-f66417b24a6a req-df39a862-095f-4c0a-a316-1b76b773110e service nova] [instance: b84c6c8d-e358-4a14-8425-c4f1d8adf4ee] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 593.776850] env[63515]: DEBUG oslo_concurrency.lockutils [None req-eb536733-3a8a-4dae-88ae-0ad9d450b524 tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] Releasing lock "refresh_cache-93b28142-8454-43c0-b0a7-d61aa95c8fc8" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 593.777336] env[63515]: DEBUG nova.compute.manager [None req-eb536733-3a8a-4dae-88ae-0ad9d450b524 tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] [instance: 93b28142-8454-43c0-b0a7-d61aa95c8fc8] Start destroying the instance on the hypervisor. {{(pid=63515) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 593.777632] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-eb536733-3a8a-4dae-88ae-0ad9d450b524 tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] [instance: 93b28142-8454-43c0-b0a7-d61aa95c8fc8] Destroying instance {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 593.779298] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc2a798c-0f07-4cca-8c4e-9806efc6c958 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 593.784183] env[63515]: DEBUG nova.virt.hardware [None req-22f02712-96d1-4031-8b8f-5bc880f34c33 tempest-ImagesNegativeTestJSON-2041328384 tempest-ImagesNegativeTestJSON-2041328384-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T02:52:14Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T02:51:56Z,direct_url=,disk_format='vmdk',id=8a120570-cb06-4099-b262-554ca0ad15c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b569255cc43e42e7ae3f2b4ad37c6ef4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T02:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 593.784481] env[63515]: DEBUG nova.virt.hardware [None req-22f02712-96d1-4031-8b8f-5bc880f34c33 tempest-ImagesNegativeTestJSON-2041328384 tempest-ImagesNegativeTestJSON-2041328384-project-member] Flavor limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 593.784692] env[63515]: DEBUG nova.virt.hardware [None req-22f02712-96d1-4031-8b8f-5bc880f34c33 tempest-ImagesNegativeTestJSON-2041328384 tempest-ImagesNegativeTestJSON-2041328384-project-member] Image limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 593.784935] env[63515]: DEBUG nova.virt.hardware [None req-22f02712-96d1-4031-8b8f-5bc880f34c33 tempest-ImagesNegativeTestJSON-2041328384 tempest-ImagesNegativeTestJSON-2041328384-project-member] Flavor pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 593.785113] env[63515]: DEBUG nova.virt.hardware [None req-22f02712-96d1-4031-8b8f-5bc880f34c33 tempest-ImagesNegativeTestJSON-2041328384 tempest-ImagesNegativeTestJSON-2041328384-project-member] Image pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 593.785311] env[63515]: DEBUG nova.virt.hardware [None req-22f02712-96d1-4031-8b8f-5bc880f34c33 tempest-ImagesNegativeTestJSON-2041328384 tempest-ImagesNegativeTestJSON-2041328384-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 593.785646] env[63515]: DEBUG nova.virt.hardware [None req-22f02712-96d1-4031-8b8f-5bc880f34c33 tempest-ImagesNegativeTestJSON-2041328384 tempest-ImagesNegativeTestJSON-2041328384-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 593.785766] env[63515]: DEBUG nova.virt.hardware [None req-22f02712-96d1-4031-8b8f-5bc880f34c33 tempest-ImagesNegativeTestJSON-2041328384 tempest-ImagesNegativeTestJSON-2041328384-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 593.786014] env[63515]: DEBUG nova.virt.hardware [None req-22f02712-96d1-4031-8b8f-5bc880f34c33 tempest-ImagesNegativeTestJSON-2041328384 tempest-ImagesNegativeTestJSON-2041328384-project-member] Got 1 possible topologies {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 593.786175] env[63515]: DEBUG nova.virt.hardware [None req-22f02712-96d1-4031-8b8f-5bc880f34c33 tempest-ImagesNegativeTestJSON-2041328384 tempest-ImagesNegativeTestJSON-2041328384-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 593.786612] env[63515]: DEBUG nova.virt.hardware [None req-22f02712-96d1-4031-8b8f-5bc880f34c33 tempest-ImagesNegativeTestJSON-2041328384 tempest-ImagesNegativeTestJSON-2041328384-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 593.787459] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94e78cfd-8532-4205-b072-2fa85815a4da {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 593.798622] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e903f06b-d2e1-43f4-811f-cefeefeee0ad {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 593.802944] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-eb536733-3a8a-4dae-88ae-0ad9d450b524 tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] [instance: 93b28142-8454-43c0-b0a7-d61aa95c8fc8] Powering off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 593.803238] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0299a687-0296-4780-a472-33d131e535c2 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 593.814999] env[63515]: DEBUG nova.network.neutron [-] [instance: b84c6c8d-e358-4a14-8425-c4f1d8adf4ee] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 593.820874] env[63515]: DEBUG oslo_vmware.api [None req-eb536733-3a8a-4dae-88ae-0ad9d450b524 tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] Waiting for the task: (returnval){ [ 593.820874] env[63515]: value = "task-1110885" [ 593.820874] env[63515]: _type = "Task" [ 593.820874] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 593.829979] env[63515]: DEBUG oslo_vmware.api [None req-eb536733-3a8a-4dae-88ae-0ad9d450b524 tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] Task: {'id': task-1110885, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 593.929644] env[63515]: DEBUG nova.network.neutron [req-dad27614-a688-416e-9a95-f66417b24a6a req-df39a862-095f-4c0a-a316-1b76b773110e service nova] [instance: b84c6c8d-e358-4a14-8425-c4f1d8adf4ee] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 594.242307] env[63515]: DEBUG oslo_concurrency.lockutils [None req-9e973647-d524-4c71-b05a-f1a253758bab tempest-ServerAddressesTestJSON-9863250 tempest-ServerAddressesTestJSON-9863250-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.523s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 594.242949] env[63515]: DEBUG nova.compute.manager [None req-9e973647-d524-4c71-b05a-f1a253758bab tempest-ServerAddressesTestJSON-9863250 tempest-ServerAddressesTestJSON-9863250-project-member] [instance: 7fee0afb-88c9-4e93-a4d9-94312394f927] Start building networks asynchronously for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 594.246285] env[63515]: DEBUG oslo_concurrency.lockutils [None req-7f5043f7-e6d7-46dd-88fd-99ee049c2b4b tempest-VolumesAssistedSnapshotsTest-453349191 tempest-VolumesAssistedSnapshotsTest-453349191-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.530s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 594.318424] env[63515]: INFO nova.compute.manager [-] [instance: b84c6c8d-e358-4a14-8425-c4f1d8adf4ee] Took 1.05 seconds to deallocate network for instance. [ 594.324233] env[63515]: DEBUG nova.compute.claims [None req-c80cd50f-4127-4451-9298-83a67d8f0ab6 tempest-ServersTestFqdnHostnames-1452522528 tempest-ServersTestFqdnHostnames-1452522528-project-member] [instance: b84c6c8d-e358-4a14-8425-c4f1d8adf4ee] Aborting claim: {{(pid=63515) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 594.324415] env[63515]: DEBUG oslo_concurrency.lockutils [None req-c80cd50f-4127-4451-9298-83a67d8f0ab6 tempest-ServersTestFqdnHostnames-1452522528 tempest-ServersTestFqdnHostnames-1452522528-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 594.330697] env[63515]: DEBUG oslo_vmware.api [None req-eb536733-3a8a-4dae-88ae-0ad9d450b524 tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] Task: {'id': task-1110885, 'name': PowerOffVM_Task, 'duration_secs': 0.121367} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 594.332711] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-eb536733-3a8a-4dae-88ae-0ad9d450b524 tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] [instance: 93b28142-8454-43c0-b0a7-d61aa95c8fc8] Powered off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 594.332891] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-eb536733-3a8a-4dae-88ae-0ad9d450b524 tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] [instance: 93b28142-8454-43c0-b0a7-d61aa95c8fc8] Unregistering the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 594.333154] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-22dc390b-d216-4b5c-966b-a849b5a58f53 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.361305] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-eb536733-3a8a-4dae-88ae-0ad9d450b524 tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] [instance: 93b28142-8454-43c0-b0a7-d61aa95c8fc8] Unregistered the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 594.361305] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-eb536733-3a8a-4dae-88ae-0ad9d450b524 tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] [instance: 93b28142-8454-43c0-b0a7-d61aa95c8fc8] Deleting contents of the VM from datastore datastore1 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 594.361305] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-eb536733-3a8a-4dae-88ae-0ad9d450b524 tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] Deleting the datastore file [datastore1] 93b28142-8454-43c0-b0a7-d61aa95c8fc8 {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 594.361305] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-64f14cf6-56bb-448c-8d11-e7fc009de905 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.367562] env[63515]: DEBUG oslo_vmware.api [None req-eb536733-3a8a-4dae-88ae-0ad9d450b524 tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] Waiting for the task: (returnval){ [ 594.367562] env[63515]: value = "task-1110887" [ 594.367562] env[63515]: _type = "Task" [ 594.367562] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 594.379776] env[63515]: DEBUG oslo_vmware.api [None req-eb536733-3a8a-4dae-88ae-0ad9d450b524 tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] Task: {'id': task-1110887, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 594.432097] env[63515]: DEBUG oslo_concurrency.lockutils [req-dad27614-a688-416e-9a95-f66417b24a6a req-df39a862-095f-4c0a-a316-1b76b773110e service nova] Releasing lock "refresh_cache-b84c6c8d-e358-4a14-8425-c4f1d8adf4ee" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 594.432381] env[63515]: DEBUG nova.compute.manager [req-dad27614-a688-416e-9a95-f66417b24a6a req-df39a862-095f-4c0a-a316-1b76b773110e service nova] [instance: b84c6c8d-e358-4a14-8425-c4f1d8adf4ee] Received event network-vif-deleted-3d61ba1c-4556-41b6-a699-f5b65b0de93f {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 594.538015] env[63515]: DEBUG oslo_concurrency.lockutils [None req-33f5b097-4c20-4377-b5be-7e558874e5d3 tempest-ServersV294TestFqdnHostnames-2112993702 tempest-ServersV294TestFqdnHostnames-2112993702-project-member] Acquiring lock "396e49dd-48c0-496b-a1ec-190c33a22c5e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 594.538276] env[63515]: DEBUG oslo_concurrency.lockutils [None req-33f5b097-4c20-4377-b5be-7e558874e5d3 tempest-ServersV294TestFqdnHostnames-2112993702 tempest-ServersV294TestFqdnHostnames-2112993702-project-member] Lock "396e49dd-48c0-496b-a1ec-190c33a22c5e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 594.678705] env[63515]: ERROR nova.compute.manager [None req-22f02712-96d1-4031-8b8f-5bc880f34c33 tempest-ImagesNegativeTestJSON-2041328384 tempest-ImagesNegativeTestJSON-2041328384-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port fca51b1b-a915-4365-a1b6-ccc0a5a43604, please check neutron logs for more information. [ 594.678705] env[63515]: ERROR nova.compute.manager Traceback (most recent call last): [ 594.678705] env[63515]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 594.678705] env[63515]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 594.678705] env[63515]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 594.678705] env[63515]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 594.678705] env[63515]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 594.678705] env[63515]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 594.678705] env[63515]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 594.678705] env[63515]: ERROR nova.compute.manager self.force_reraise() [ 594.678705] env[63515]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 594.678705] env[63515]: ERROR nova.compute.manager raise self.value [ 594.678705] env[63515]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 594.678705] env[63515]: ERROR nova.compute.manager updated_port = self._update_port( [ 594.678705] env[63515]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 594.678705] env[63515]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 594.679263] env[63515]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 594.679263] env[63515]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 594.679263] env[63515]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port fca51b1b-a915-4365-a1b6-ccc0a5a43604, please check neutron logs for more information. [ 594.679263] env[63515]: ERROR nova.compute.manager [ 594.679263] env[63515]: Traceback (most recent call last): [ 594.679263] env[63515]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 594.679263] env[63515]: listener.cb(fileno) [ 594.679263] env[63515]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 594.679263] env[63515]: result = function(*args, **kwargs) [ 594.679263] env[63515]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 594.679263] env[63515]: return func(*args, **kwargs) [ 594.679263] env[63515]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 594.679263] env[63515]: raise e [ 594.679263] env[63515]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 594.679263] env[63515]: nwinfo = self.network_api.allocate_for_instance( [ 594.679263] env[63515]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 594.679263] env[63515]: created_port_ids = self._update_ports_for_instance( [ 594.679263] env[63515]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 594.679263] env[63515]: with excutils.save_and_reraise_exception(): [ 594.679263] env[63515]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 594.679263] env[63515]: self.force_reraise() [ 594.679263] env[63515]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 594.679263] env[63515]: raise self.value [ 594.679263] env[63515]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 594.679263] env[63515]: updated_port = self._update_port( [ 594.679263] env[63515]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 594.679263] env[63515]: _ensure_no_port_binding_failure(port) [ 594.680059] env[63515]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 594.680059] env[63515]: raise exception.PortBindingFailed(port_id=port['id']) [ 594.680059] env[63515]: nova.exception.PortBindingFailed: Binding failed for port fca51b1b-a915-4365-a1b6-ccc0a5a43604, please check neutron logs for more information. [ 594.680059] env[63515]: Removing descriptor: 16 [ 594.680305] env[63515]: ERROR nova.compute.manager [None req-22f02712-96d1-4031-8b8f-5bc880f34c33 tempest-ImagesNegativeTestJSON-2041328384 tempest-ImagesNegativeTestJSON-2041328384-project-member] [instance: d5f988f2-765b-4ecf-bad5-4a70fc95d74c] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port fca51b1b-a915-4365-a1b6-ccc0a5a43604, please check neutron logs for more information. [ 594.680305] env[63515]: ERROR nova.compute.manager [instance: d5f988f2-765b-4ecf-bad5-4a70fc95d74c] Traceback (most recent call last): [ 594.680305] env[63515]: ERROR nova.compute.manager [instance: d5f988f2-765b-4ecf-bad5-4a70fc95d74c] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 594.680305] env[63515]: ERROR nova.compute.manager [instance: d5f988f2-765b-4ecf-bad5-4a70fc95d74c] yield resources [ 594.680305] env[63515]: ERROR nova.compute.manager [instance: d5f988f2-765b-4ecf-bad5-4a70fc95d74c] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 594.680305] env[63515]: ERROR nova.compute.manager [instance: d5f988f2-765b-4ecf-bad5-4a70fc95d74c] self.driver.spawn(context, instance, image_meta, [ 594.680305] env[63515]: ERROR nova.compute.manager [instance: d5f988f2-765b-4ecf-bad5-4a70fc95d74c] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 594.680305] env[63515]: ERROR nova.compute.manager [instance: d5f988f2-765b-4ecf-bad5-4a70fc95d74c] self._vmops.spawn(context, instance, image_meta, injected_files, [ 594.680305] env[63515]: ERROR nova.compute.manager [instance: d5f988f2-765b-4ecf-bad5-4a70fc95d74c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 594.680305] env[63515]: ERROR nova.compute.manager [instance: d5f988f2-765b-4ecf-bad5-4a70fc95d74c] vm_ref = self.build_virtual_machine(instance, [ 594.680305] env[63515]: ERROR nova.compute.manager [instance: d5f988f2-765b-4ecf-bad5-4a70fc95d74c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 594.680697] env[63515]: ERROR nova.compute.manager [instance: d5f988f2-765b-4ecf-bad5-4a70fc95d74c] vif_infos = vmwarevif.get_vif_info(self._session, [ 594.680697] env[63515]: ERROR nova.compute.manager [instance: d5f988f2-765b-4ecf-bad5-4a70fc95d74c] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 594.680697] env[63515]: ERROR nova.compute.manager [instance: d5f988f2-765b-4ecf-bad5-4a70fc95d74c] for vif in network_info: [ 594.680697] env[63515]: ERROR nova.compute.manager [instance: d5f988f2-765b-4ecf-bad5-4a70fc95d74c] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 594.680697] env[63515]: ERROR nova.compute.manager [instance: d5f988f2-765b-4ecf-bad5-4a70fc95d74c] return self._sync_wrapper(fn, *args, **kwargs) [ 594.680697] env[63515]: ERROR nova.compute.manager [instance: d5f988f2-765b-4ecf-bad5-4a70fc95d74c] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 594.680697] env[63515]: ERROR nova.compute.manager [instance: d5f988f2-765b-4ecf-bad5-4a70fc95d74c] self.wait() [ 594.680697] env[63515]: ERROR nova.compute.manager [instance: d5f988f2-765b-4ecf-bad5-4a70fc95d74c] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 594.680697] env[63515]: ERROR nova.compute.manager [instance: d5f988f2-765b-4ecf-bad5-4a70fc95d74c] self[:] = self._gt.wait() [ 594.680697] env[63515]: ERROR nova.compute.manager [instance: d5f988f2-765b-4ecf-bad5-4a70fc95d74c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 594.680697] env[63515]: ERROR nova.compute.manager [instance: d5f988f2-765b-4ecf-bad5-4a70fc95d74c] return self._exit_event.wait() [ 594.680697] env[63515]: ERROR nova.compute.manager [instance: d5f988f2-765b-4ecf-bad5-4a70fc95d74c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 594.680697] env[63515]: ERROR nova.compute.manager [instance: d5f988f2-765b-4ecf-bad5-4a70fc95d74c] result = hub.switch() [ 594.681128] env[63515]: ERROR nova.compute.manager [instance: d5f988f2-765b-4ecf-bad5-4a70fc95d74c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 594.681128] env[63515]: ERROR nova.compute.manager [instance: d5f988f2-765b-4ecf-bad5-4a70fc95d74c] return self.greenlet.switch() [ 594.681128] env[63515]: ERROR nova.compute.manager [instance: d5f988f2-765b-4ecf-bad5-4a70fc95d74c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 594.681128] env[63515]: ERROR nova.compute.manager [instance: d5f988f2-765b-4ecf-bad5-4a70fc95d74c] result = function(*args, **kwargs) [ 594.681128] env[63515]: ERROR nova.compute.manager [instance: d5f988f2-765b-4ecf-bad5-4a70fc95d74c] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 594.681128] env[63515]: ERROR nova.compute.manager [instance: d5f988f2-765b-4ecf-bad5-4a70fc95d74c] return func(*args, **kwargs) [ 594.681128] env[63515]: ERROR nova.compute.manager [instance: d5f988f2-765b-4ecf-bad5-4a70fc95d74c] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 594.681128] env[63515]: ERROR nova.compute.manager [instance: d5f988f2-765b-4ecf-bad5-4a70fc95d74c] raise e [ 594.681128] env[63515]: ERROR nova.compute.manager [instance: d5f988f2-765b-4ecf-bad5-4a70fc95d74c] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 594.681128] env[63515]: ERROR nova.compute.manager [instance: d5f988f2-765b-4ecf-bad5-4a70fc95d74c] nwinfo = self.network_api.allocate_for_instance( [ 594.681128] env[63515]: ERROR nova.compute.manager [instance: d5f988f2-765b-4ecf-bad5-4a70fc95d74c] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 594.681128] env[63515]: ERROR nova.compute.manager [instance: d5f988f2-765b-4ecf-bad5-4a70fc95d74c] created_port_ids = self._update_ports_for_instance( [ 594.681128] env[63515]: ERROR nova.compute.manager [instance: d5f988f2-765b-4ecf-bad5-4a70fc95d74c] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 594.681581] env[63515]: ERROR nova.compute.manager [instance: d5f988f2-765b-4ecf-bad5-4a70fc95d74c] with excutils.save_and_reraise_exception(): [ 594.681581] env[63515]: ERROR nova.compute.manager [instance: d5f988f2-765b-4ecf-bad5-4a70fc95d74c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 594.681581] env[63515]: ERROR nova.compute.manager [instance: d5f988f2-765b-4ecf-bad5-4a70fc95d74c] self.force_reraise() [ 594.681581] env[63515]: ERROR nova.compute.manager [instance: d5f988f2-765b-4ecf-bad5-4a70fc95d74c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 594.681581] env[63515]: ERROR nova.compute.manager [instance: d5f988f2-765b-4ecf-bad5-4a70fc95d74c] raise self.value [ 594.681581] env[63515]: ERROR nova.compute.manager [instance: d5f988f2-765b-4ecf-bad5-4a70fc95d74c] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 594.681581] env[63515]: ERROR nova.compute.manager [instance: d5f988f2-765b-4ecf-bad5-4a70fc95d74c] updated_port = self._update_port( [ 594.681581] env[63515]: ERROR nova.compute.manager [instance: d5f988f2-765b-4ecf-bad5-4a70fc95d74c] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 594.681581] env[63515]: ERROR nova.compute.manager [instance: d5f988f2-765b-4ecf-bad5-4a70fc95d74c] _ensure_no_port_binding_failure(port) [ 594.681581] env[63515]: ERROR nova.compute.manager [instance: d5f988f2-765b-4ecf-bad5-4a70fc95d74c] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 594.681581] env[63515]: ERROR nova.compute.manager [instance: d5f988f2-765b-4ecf-bad5-4a70fc95d74c] raise exception.PortBindingFailed(port_id=port['id']) [ 594.681581] env[63515]: ERROR nova.compute.manager [instance: d5f988f2-765b-4ecf-bad5-4a70fc95d74c] nova.exception.PortBindingFailed: Binding failed for port fca51b1b-a915-4365-a1b6-ccc0a5a43604, please check neutron logs for more information. [ 594.681581] env[63515]: ERROR nova.compute.manager [instance: d5f988f2-765b-4ecf-bad5-4a70fc95d74c] [ 594.681987] env[63515]: INFO nova.compute.manager [None req-22f02712-96d1-4031-8b8f-5bc880f34c33 tempest-ImagesNegativeTestJSON-2041328384 tempest-ImagesNegativeTestJSON-2041328384-project-member] [instance: d5f988f2-765b-4ecf-bad5-4a70fc95d74c] Terminating instance [ 594.683648] env[63515]: DEBUG oslo_concurrency.lockutils [None req-22f02712-96d1-4031-8b8f-5bc880f34c33 tempest-ImagesNegativeTestJSON-2041328384 tempest-ImagesNegativeTestJSON-2041328384-project-member] Acquiring lock "refresh_cache-d5f988f2-765b-4ecf-bad5-4a70fc95d74c" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 594.683915] env[63515]: DEBUG oslo_concurrency.lockutils [None req-22f02712-96d1-4031-8b8f-5bc880f34c33 tempest-ImagesNegativeTestJSON-2041328384 tempest-ImagesNegativeTestJSON-2041328384-project-member] Acquired lock "refresh_cache-d5f988f2-765b-4ecf-bad5-4a70fc95d74c" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 594.684128] env[63515]: DEBUG nova.network.neutron [None req-22f02712-96d1-4031-8b8f-5bc880f34c33 tempest-ImagesNegativeTestJSON-2041328384 tempest-ImagesNegativeTestJSON-2041328384-project-member] [instance: d5f988f2-765b-4ecf-bad5-4a70fc95d74c] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 594.752367] env[63515]: DEBUG nova.compute.utils [None req-9e973647-d524-4c71-b05a-f1a253758bab tempest-ServerAddressesTestJSON-9863250 tempest-ServerAddressesTestJSON-9863250-project-member] Using /dev/sd instead of None {{(pid=63515) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 594.757537] env[63515]: DEBUG nova.compute.manager [None req-9e973647-d524-4c71-b05a-f1a253758bab tempest-ServerAddressesTestJSON-9863250 tempest-ServerAddressesTestJSON-9863250-project-member] [instance: 7fee0afb-88c9-4e93-a4d9-94312394f927] Allocating IP information in the background. {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 594.757759] env[63515]: DEBUG nova.network.neutron [None req-9e973647-d524-4c71-b05a-f1a253758bab tempest-ServerAddressesTestJSON-9863250 tempest-ServerAddressesTestJSON-9863250-project-member] [instance: 7fee0afb-88c9-4e93-a4d9-94312394f927] allocate_for_instance() {{(pid=63515) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 594.830445] env[63515]: DEBUG nova.policy [None req-9e973647-d524-4c71-b05a-f1a253758bab tempest-ServerAddressesTestJSON-9863250 tempest-ServerAddressesTestJSON-9863250-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e97e45c73018441d91b08d6e1fdd8999', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c817423ea0f4459d84b64e50ef4b3b50', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63515) authorize /opt/stack/nova/nova/policy.py:201}} [ 594.868814] env[63515]: DEBUG oslo_concurrency.lockutils [None req-4fa686a6-dc45-4603-a4ee-7107c810126f tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] Acquiring lock "4087b2c4-6ed1-4b68-8b78-a36e34d935b1" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 594.868814] env[63515]: DEBUG oslo_concurrency.lockutils [None req-4fa686a6-dc45-4603-a4ee-7107c810126f tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] Lock "4087b2c4-6ed1-4b68-8b78-a36e34d935b1" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 594.880154] env[63515]: DEBUG oslo_vmware.api [None req-eb536733-3a8a-4dae-88ae-0ad9d450b524 tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] Task: {'id': task-1110887, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.089354} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 594.883449] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-eb536733-3a8a-4dae-88ae-0ad9d450b524 tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] Deleted the datastore file {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 594.883669] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-eb536733-3a8a-4dae-88ae-0ad9d450b524 tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] [instance: 93b28142-8454-43c0-b0a7-d61aa95c8fc8] Deleted contents of the VM from datastore datastore1 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 594.883869] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-eb536733-3a8a-4dae-88ae-0ad9d450b524 tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] [instance: 93b28142-8454-43c0-b0a7-d61aa95c8fc8] Instance destroyed {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 594.884737] env[63515]: INFO nova.compute.manager [None req-eb536733-3a8a-4dae-88ae-0ad9d450b524 tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] [instance: 93b28142-8454-43c0-b0a7-d61aa95c8fc8] Took 1.11 seconds to destroy the instance on the hypervisor. [ 594.885060] env[63515]: DEBUG oslo.service.loopingcall [None req-eb536733-3a8a-4dae-88ae-0ad9d450b524 tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 594.886033] env[63515]: DEBUG nova.compute.manager [-] [instance: 93b28142-8454-43c0-b0a7-d61aa95c8fc8] Deallocating network for instance {{(pid=63515) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 594.886033] env[63515]: DEBUG nova.network.neutron [-] [instance: 93b28142-8454-43c0-b0a7-d61aa95c8fc8] deallocate_for_instance() {{(pid=63515) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 594.909812] env[63515]: DEBUG nova.network.neutron [-] [instance: 93b28142-8454-43c0-b0a7-d61aa95c8fc8] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 595.165349] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5c52e36-e9e5-43d3-be34-81bd9d8a50fc {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.174177] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a14a6f0-4087-4089-8ffd-1fe5ae4ba831 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.210802] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32d27e4e-1050-46f5-9652-9a665c66aef6 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.219524] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec8d7434-d887-422a-923b-48c488aeb3e6 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.237194] env[63515]: DEBUG nova.compute.provider_tree [None req-7f5043f7-e6d7-46dd-88fd-99ee049c2b4b tempest-VolumesAssistedSnapshotsTest-453349191 tempest-VolumesAssistedSnapshotsTest-453349191-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 595.247256] env[63515]: DEBUG nova.network.neutron [None req-22f02712-96d1-4031-8b8f-5bc880f34c33 tempest-ImagesNegativeTestJSON-2041328384 tempest-ImagesNegativeTestJSON-2041328384-project-member] [instance: d5f988f2-765b-4ecf-bad5-4a70fc95d74c] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 595.258157] env[63515]: DEBUG nova.compute.manager [None req-9e973647-d524-4c71-b05a-f1a253758bab tempest-ServerAddressesTestJSON-9863250 tempest-ServerAddressesTestJSON-9863250-project-member] [instance: 7fee0afb-88c9-4e93-a4d9-94312394f927] Start building block device mappings for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 595.400726] env[63515]: DEBUG oslo_concurrency.lockutils [None req-44a3080f-f2e2-4f0c-b244-122beee2aea9 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Acquiring lock "87c468d9-9594-4804-b461-527f01f6118f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 595.400967] env[63515]: DEBUG oslo_concurrency.lockutils [None req-44a3080f-f2e2-4f0c-b244-122beee2aea9 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Lock "87c468d9-9594-4804-b461-527f01f6118f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 595.415572] env[63515]: DEBUG nova.network.neutron [-] [instance: 93b28142-8454-43c0-b0a7-d61aa95c8fc8] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 595.430218] env[63515]: DEBUG nova.network.neutron [None req-9e973647-d524-4c71-b05a-f1a253758bab tempest-ServerAddressesTestJSON-9863250 tempest-ServerAddressesTestJSON-9863250-project-member] [instance: 7fee0afb-88c9-4e93-a4d9-94312394f927] Successfully created port: 79794b46-d801-404a-94f2-a223e825fcfc {{(pid=63515) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 595.543040] env[63515]: DEBUG nova.network.neutron [None req-22f02712-96d1-4031-8b8f-5bc880f34c33 tempest-ImagesNegativeTestJSON-2041328384 tempest-ImagesNegativeTestJSON-2041328384-project-member] [instance: d5f988f2-765b-4ecf-bad5-4a70fc95d74c] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 595.620614] env[63515]: DEBUG nova.compute.manager [req-401b3b15-b183-4efa-884e-530522f60457 req-409f492e-85a2-4ffb-9ad8-e86a4ba8c9d3 service nova] [instance: d5f988f2-765b-4ecf-bad5-4a70fc95d74c] Received event network-changed-fca51b1b-a915-4365-a1b6-ccc0a5a43604 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 595.620809] env[63515]: DEBUG nova.compute.manager [req-401b3b15-b183-4efa-884e-530522f60457 req-409f492e-85a2-4ffb-9ad8-e86a4ba8c9d3 service nova] [instance: d5f988f2-765b-4ecf-bad5-4a70fc95d74c] Refreshing instance network info cache due to event network-changed-fca51b1b-a915-4365-a1b6-ccc0a5a43604. {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 595.621008] env[63515]: DEBUG oslo_concurrency.lockutils [req-401b3b15-b183-4efa-884e-530522f60457 req-409f492e-85a2-4ffb-9ad8-e86a4ba8c9d3 service nova] Acquiring lock "refresh_cache-d5f988f2-765b-4ecf-bad5-4a70fc95d74c" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 595.742203] env[63515]: DEBUG nova.scheduler.client.report [None req-7f5043f7-e6d7-46dd-88fd-99ee049c2b4b tempest-VolumesAssistedSnapshotsTest-453349191 tempest-VolumesAssistedSnapshotsTest-453349191-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 595.917645] env[63515]: INFO nova.compute.manager [-] [instance: 93b28142-8454-43c0-b0a7-d61aa95c8fc8] Took 1.03 seconds to deallocate network for instance. [ 596.045585] env[63515]: DEBUG oslo_concurrency.lockutils [None req-22f02712-96d1-4031-8b8f-5bc880f34c33 tempest-ImagesNegativeTestJSON-2041328384 tempest-ImagesNegativeTestJSON-2041328384-project-member] Releasing lock "refresh_cache-d5f988f2-765b-4ecf-bad5-4a70fc95d74c" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 596.046031] env[63515]: DEBUG nova.compute.manager [None req-22f02712-96d1-4031-8b8f-5bc880f34c33 tempest-ImagesNegativeTestJSON-2041328384 tempest-ImagesNegativeTestJSON-2041328384-project-member] [instance: d5f988f2-765b-4ecf-bad5-4a70fc95d74c] Start destroying the instance on the hypervisor. {{(pid=63515) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 596.046229] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-22f02712-96d1-4031-8b8f-5bc880f34c33 tempest-ImagesNegativeTestJSON-2041328384 tempest-ImagesNegativeTestJSON-2041328384-project-member] [instance: d5f988f2-765b-4ecf-bad5-4a70fc95d74c] Destroying instance {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 596.046564] env[63515]: DEBUG oslo_concurrency.lockutils [req-401b3b15-b183-4efa-884e-530522f60457 req-409f492e-85a2-4ffb-9ad8-e86a4ba8c9d3 service nova] Acquired lock "refresh_cache-d5f988f2-765b-4ecf-bad5-4a70fc95d74c" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 596.046747] env[63515]: DEBUG nova.network.neutron [req-401b3b15-b183-4efa-884e-530522f60457 req-409f492e-85a2-4ffb-9ad8-e86a4ba8c9d3 service nova] [instance: d5f988f2-765b-4ecf-bad5-4a70fc95d74c] Refreshing network info cache for port fca51b1b-a915-4365-a1b6-ccc0a5a43604 {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 596.052935] env[63515]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-53a9b55e-3021-4f92-835e-933de7ec6e93 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.060974] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-195773aa-4586-43ad-b025-65df59f997a0 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.083427] env[63515]: WARNING nova.virt.vmwareapi.vmops [None req-22f02712-96d1-4031-8b8f-5bc880f34c33 tempest-ImagesNegativeTestJSON-2041328384 tempest-ImagesNegativeTestJSON-2041328384-project-member] [instance: d5f988f2-765b-4ecf-bad5-4a70fc95d74c] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance d5f988f2-765b-4ecf-bad5-4a70fc95d74c could not be found. [ 596.083739] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-22f02712-96d1-4031-8b8f-5bc880f34c33 tempest-ImagesNegativeTestJSON-2041328384 tempest-ImagesNegativeTestJSON-2041328384-project-member] [instance: d5f988f2-765b-4ecf-bad5-4a70fc95d74c] Instance destroyed {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 596.084176] env[63515]: INFO nova.compute.manager [None req-22f02712-96d1-4031-8b8f-5bc880f34c33 tempest-ImagesNegativeTestJSON-2041328384 tempest-ImagesNegativeTestJSON-2041328384-project-member] [instance: d5f988f2-765b-4ecf-bad5-4a70fc95d74c] Took 0.04 seconds to destroy the instance on the hypervisor. [ 596.084417] env[63515]: DEBUG oslo.service.loopingcall [None req-22f02712-96d1-4031-8b8f-5bc880f34c33 tempest-ImagesNegativeTestJSON-2041328384 tempest-ImagesNegativeTestJSON-2041328384-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 596.084693] env[63515]: DEBUG nova.compute.manager [-] [instance: d5f988f2-765b-4ecf-bad5-4a70fc95d74c] Deallocating network for instance {{(pid=63515) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 596.084837] env[63515]: DEBUG nova.network.neutron [-] [instance: d5f988f2-765b-4ecf-bad5-4a70fc95d74c] deallocate_for_instance() {{(pid=63515) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 596.108100] env[63515]: DEBUG nova.network.neutron [-] [instance: d5f988f2-765b-4ecf-bad5-4a70fc95d74c] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 596.247341] env[63515]: DEBUG oslo_concurrency.lockutils [None req-7f5043f7-e6d7-46dd-88fd-99ee049c2b4b tempest-VolumesAssistedSnapshotsTest-453349191 tempest-VolumesAssistedSnapshotsTest-453349191-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.002s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 596.247992] env[63515]: ERROR nova.compute.manager [None req-7f5043f7-e6d7-46dd-88fd-99ee049c2b4b tempest-VolumesAssistedSnapshotsTest-453349191 tempest-VolumesAssistedSnapshotsTest-453349191-project-member] [instance: db682708-36fa-4126-a848-bfb609df96d7] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port ee30f5c5-4733-4b41-ad1e-b45f0ae502cb, please check neutron logs for more information. [ 596.247992] env[63515]: ERROR nova.compute.manager [instance: db682708-36fa-4126-a848-bfb609df96d7] Traceback (most recent call last): [ 596.247992] env[63515]: ERROR nova.compute.manager [instance: db682708-36fa-4126-a848-bfb609df96d7] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 596.247992] env[63515]: ERROR nova.compute.manager [instance: db682708-36fa-4126-a848-bfb609df96d7] self.driver.spawn(context, instance, image_meta, [ 596.247992] env[63515]: ERROR nova.compute.manager [instance: db682708-36fa-4126-a848-bfb609df96d7] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 596.247992] env[63515]: ERROR nova.compute.manager [instance: db682708-36fa-4126-a848-bfb609df96d7] self._vmops.spawn(context, instance, image_meta, injected_files, [ 596.247992] env[63515]: ERROR nova.compute.manager [instance: db682708-36fa-4126-a848-bfb609df96d7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 596.247992] env[63515]: ERROR nova.compute.manager [instance: db682708-36fa-4126-a848-bfb609df96d7] vm_ref = self.build_virtual_machine(instance, [ 596.247992] env[63515]: ERROR nova.compute.manager [instance: db682708-36fa-4126-a848-bfb609df96d7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 596.247992] env[63515]: ERROR nova.compute.manager [instance: db682708-36fa-4126-a848-bfb609df96d7] vif_infos = vmwarevif.get_vif_info(self._session, [ 596.247992] env[63515]: ERROR nova.compute.manager [instance: db682708-36fa-4126-a848-bfb609df96d7] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 596.248445] env[63515]: ERROR nova.compute.manager [instance: db682708-36fa-4126-a848-bfb609df96d7] for vif in network_info: [ 596.248445] env[63515]: ERROR nova.compute.manager [instance: db682708-36fa-4126-a848-bfb609df96d7] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 596.248445] env[63515]: ERROR nova.compute.manager [instance: db682708-36fa-4126-a848-bfb609df96d7] return self._sync_wrapper(fn, *args, **kwargs) [ 596.248445] env[63515]: ERROR nova.compute.manager [instance: db682708-36fa-4126-a848-bfb609df96d7] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 596.248445] env[63515]: ERROR nova.compute.manager [instance: db682708-36fa-4126-a848-bfb609df96d7] self.wait() [ 596.248445] env[63515]: ERROR nova.compute.manager [instance: db682708-36fa-4126-a848-bfb609df96d7] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 596.248445] env[63515]: ERROR nova.compute.manager [instance: db682708-36fa-4126-a848-bfb609df96d7] self[:] = self._gt.wait() [ 596.248445] env[63515]: ERROR nova.compute.manager [instance: db682708-36fa-4126-a848-bfb609df96d7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 596.248445] env[63515]: ERROR nova.compute.manager [instance: db682708-36fa-4126-a848-bfb609df96d7] return self._exit_event.wait() [ 596.248445] env[63515]: ERROR nova.compute.manager [instance: db682708-36fa-4126-a848-bfb609df96d7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 596.248445] env[63515]: ERROR nova.compute.manager [instance: db682708-36fa-4126-a848-bfb609df96d7] result = hub.switch() [ 596.248445] env[63515]: ERROR nova.compute.manager [instance: db682708-36fa-4126-a848-bfb609df96d7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 596.248445] env[63515]: ERROR nova.compute.manager [instance: db682708-36fa-4126-a848-bfb609df96d7] return self.greenlet.switch() [ 596.248860] env[63515]: ERROR nova.compute.manager [instance: db682708-36fa-4126-a848-bfb609df96d7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 596.248860] env[63515]: ERROR nova.compute.manager [instance: db682708-36fa-4126-a848-bfb609df96d7] result = function(*args, **kwargs) [ 596.248860] env[63515]: ERROR nova.compute.manager [instance: db682708-36fa-4126-a848-bfb609df96d7] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 596.248860] env[63515]: ERROR nova.compute.manager [instance: db682708-36fa-4126-a848-bfb609df96d7] return func(*args, **kwargs) [ 596.248860] env[63515]: ERROR nova.compute.manager [instance: db682708-36fa-4126-a848-bfb609df96d7] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 596.248860] env[63515]: ERROR nova.compute.manager [instance: db682708-36fa-4126-a848-bfb609df96d7] raise e [ 596.248860] env[63515]: ERROR nova.compute.manager [instance: db682708-36fa-4126-a848-bfb609df96d7] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 596.248860] env[63515]: ERROR nova.compute.manager [instance: db682708-36fa-4126-a848-bfb609df96d7] nwinfo = self.network_api.allocate_for_instance( [ 596.248860] env[63515]: ERROR nova.compute.manager [instance: db682708-36fa-4126-a848-bfb609df96d7] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 596.248860] env[63515]: ERROR nova.compute.manager [instance: db682708-36fa-4126-a848-bfb609df96d7] created_port_ids = self._update_ports_for_instance( [ 596.248860] env[63515]: ERROR nova.compute.manager [instance: db682708-36fa-4126-a848-bfb609df96d7] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 596.248860] env[63515]: ERROR nova.compute.manager [instance: db682708-36fa-4126-a848-bfb609df96d7] with excutils.save_and_reraise_exception(): [ 596.248860] env[63515]: ERROR nova.compute.manager [instance: db682708-36fa-4126-a848-bfb609df96d7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 596.249291] env[63515]: ERROR nova.compute.manager [instance: db682708-36fa-4126-a848-bfb609df96d7] self.force_reraise() [ 596.249291] env[63515]: ERROR nova.compute.manager [instance: db682708-36fa-4126-a848-bfb609df96d7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 596.249291] env[63515]: ERROR nova.compute.manager [instance: db682708-36fa-4126-a848-bfb609df96d7] raise self.value [ 596.249291] env[63515]: ERROR nova.compute.manager [instance: db682708-36fa-4126-a848-bfb609df96d7] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 596.249291] env[63515]: ERROR nova.compute.manager [instance: db682708-36fa-4126-a848-bfb609df96d7] updated_port = self._update_port( [ 596.249291] env[63515]: ERROR nova.compute.manager [instance: db682708-36fa-4126-a848-bfb609df96d7] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 596.249291] env[63515]: ERROR nova.compute.manager [instance: db682708-36fa-4126-a848-bfb609df96d7] _ensure_no_port_binding_failure(port) [ 596.249291] env[63515]: ERROR nova.compute.manager [instance: db682708-36fa-4126-a848-bfb609df96d7] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 596.249291] env[63515]: ERROR nova.compute.manager [instance: db682708-36fa-4126-a848-bfb609df96d7] raise exception.PortBindingFailed(port_id=port['id']) [ 596.249291] env[63515]: ERROR nova.compute.manager [instance: db682708-36fa-4126-a848-bfb609df96d7] nova.exception.PortBindingFailed: Binding failed for port ee30f5c5-4733-4b41-ad1e-b45f0ae502cb, please check neutron logs for more information. [ 596.249291] env[63515]: ERROR nova.compute.manager [instance: db682708-36fa-4126-a848-bfb609df96d7] [ 596.249634] env[63515]: DEBUG nova.compute.utils [None req-7f5043f7-e6d7-46dd-88fd-99ee049c2b4b tempest-VolumesAssistedSnapshotsTest-453349191 tempest-VolumesAssistedSnapshotsTest-453349191-project-member] [instance: db682708-36fa-4126-a848-bfb609df96d7] Binding failed for port ee30f5c5-4733-4b41-ad1e-b45f0ae502cb, please check neutron logs for more information. {{(pid=63515) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 596.250032] env[63515]: DEBUG oslo_concurrency.lockutils [None req-b1f9566f-fcad-4d5d-9d53-b04b591d6c49 tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 14.720s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 596.250230] env[63515]: DEBUG nova.objects.instance [None req-b1f9566f-fcad-4d5d-9d53-b04b591d6c49 tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] [instance: 93b28142-8454-43c0-b0a7-d61aa95c8fc8] Trying to apply a migration context that does not seem to be set for this instance {{(pid=63515) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 596.252750] env[63515]: DEBUG nova.compute.manager [None req-7f5043f7-e6d7-46dd-88fd-99ee049c2b4b tempest-VolumesAssistedSnapshotsTest-453349191 tempest-VolumesAssistedSnapshotsTest-453349191-project-member] [instance: db682708-36fa-4126-a848-bfb609df96d7] Build of instance db682708-36fa-4126-a848-bfb609df96d7 was re-scheduled: Binding failed for port ee30f5c5-4733-4b41-ad1e-b45f0ae502cb, please check neutron logs for more information. {{(pid=63515) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 596.253452] env[63515]: DEBUG nova.compute.manager [None req-7f5043f7-e6d7-46dd-88fd-99ee049c2b4b tempest-VolumesAssistedSnapshotsTest-453349191 tempest-VolumesAssistedSnapshotsTest-453349191-project-member] [instance: db682708-36fa-4126-a848-bfb609df96d7] Unplugging VIFs for instance {{(pid=63515) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 596.254094] env[63515]: DEBUG oslo_concurrency.lockutils [None req-7f5043f7-e6d7-46dd-88fd-99ee049c2b4b tempest-VolumesAssistedSnapshotsTest-453349191 tempest-VolumesAssistedSnapshotsTest-453349191-project-member] Acquiring lock "refresh_cache-db682708-36fa-4126-a848-bfb609df96d7" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 596.254094] env[63515]: DEBUG oslo_concurrency.lockutils [None req-7f5043f7-e6d7-46dd-88fd-99ee049c2b4b tempest-VolumesAssistedSnapshotsTest-453349191 tempest-VolumesAssistedSnapshotsTest-453349191-project-member] Acquired lock "refresh_cache-db682708-36fa-4126-a848-bfb609df96d7" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 596.254094] env[63515]: DEBUG nova.network.neutron [None req-7f5043f7-e6d7-46dd-88fd-99ee049c2b4b tempest-VolumesAssistedSnapshotsTest-453349191 tempest-VolumesAssistedSnapshotsTest-453349191-project-member] [instance: db682708-36fa-4126-a848-bfb609df96d7] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 596.270194] env[63515]: DEBUG nova.compute.manager [None req-9e973647-d524-4c71-b05a-f1a253758bab tempest-ServerAddressesTestJSON-9863250 tempest-ServerAddressesTestJSON-9863250-project-member] [instance: 7fee0afb-88c9-4e93-a4d9-94312394f927] Start spawning the instance on the hypervisor. {{(pid=63515) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 596.305441] env[63515]: DEBUG nova.virt.hardware [None req-9e973647-d524-4c71-b05a-f1a253758bab tempest-ServerAddressesTestJSON-9863250 tempest-ServerAddressesTestJSON-9863250-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T02:52:14Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T02:51:56Z,direct_url=,disk_format='vmdk',id=8a120570-cb06-4099-b262-554ca0ad15c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b569255cc43e42e7ae3f2b4ad37c6ef4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T02:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 596.305719] env[63515]: DEBUG nova.virt.hardware [None req-9e973647-d524-4c71-b05a-f1a253758bab tempest-ServerAddressesTestJSON-9863250 tempest-ServerAddressesTestJSON-9863250-project-member] Flavor limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 596.305878] env[63515]: DEBUG nova.virt.hardware [None req-9e973647-d524-4c71-b05a-f1a253758bab tempest-ServerAddressesTestJSON-9863250 tempest-ServerAddressesTestJSON-9863250-project-member] Image limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 596.306068] env[63515]: DEBUG nova.virt.hardware [None req-9e973647-d524-4c71-b05a-f1a253758bab tempest-ServerAddressesTestJSON-9863250 tempest-ServerAddressesTestJSON-9863250-project-member] Flavor pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 596.306219] env[63515]: DEBUG nova.virt.hardware [None req-9e973647-d524-4c71-b05a-f1a253758bab tempest-ServerAddressesTestJSON-9863250 tempest-ServerAddressesTestJSON-9863250-project-member] Image pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 596.306364] env[63515]: DEBUG nova.virt.hardware [None req-9e973647-d524-4c71-b05a-f1a253758bab tempest-ServerAddressesTestJSON-9863250 tempest-ServerAddressesTestJSON-9863250-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 596.306570] env[63515]: DEBUG nova.virt.hardware [None req-9e973647-d524-4c71-b05a-f1a253758bab tempest-ServerAddressesTestJSON-9863250 tempest-ServerAddressesTestJSON-9863250-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 596.306726] env[63515]: DEBUG nova.virt.hardware [None req-9e973647-d524-4c71-b05a-f1a253758bab tempest-ServerAddressesTestJSON-9863250 tempest-ServerAddressesTestJSON-9863250-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 596.306933] env[63515]: DEBUG nova.virt.hardware [None req-9e973647-d524-4c71-b05a-f1a253758bab tempest-ServerAddressesTestJSON-9863250 tempest-ServerAddressesTestJSON-9863250-project-member] Got 1 possible topologies {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 596.307117] env[63515]: DEBUG nova.virt.hardware [None req-9e973647-d524-4c71-b05a-f1a253758bab tempest-ServerAddressesTestJSON-9863250 tempest-ServerAddressesTestJSON-9863250-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 596.307290] env[63515]: DEBUG nova.virt.hardware [None req-9e973647-d524-4c71-b05a-f1a253758bab tempest-ServerAddressesTestJSON-9863250 tempest-ServerAddressesTestJSON-9863250-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 596.308144] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-490ec748-f823-48ec-8cfb-04c90116383f {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.318462] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4940bd8b-ee7b-4256-961c-1f59dc89df49 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.425091] env[63515]: DEBUG oslo_concurrency.lockutils [None req-eb536733-3a8a-4dae-88ae-0ad9d450b524 tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 596.581494] env[63515]: DEBUG nova.network.neutron [req-401b3b15-b183-4efa-884e-530522f60457 req-409f492e-85a2-4ffb-9ad8-e86a4ba8c9d3 service nova] [instance: d5f988f2-765b-4ecf-bad5-4a70fc95d74c] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 596.610866] env[63515]: DEBUG nova.network.neutron [-] [instance: d5f988f2-765b-4ecf-bad5-4a70fc95d74c] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 596.751486] env[63515]: DEBUG nova.network.neutron [req-401b3b15-b183-4efa-884e-530522f60457 req-409f492e-85a2-4ffb-9ad8-e86a4ba8c9d3 service nova] [instance: d5f988f2-765b-4ecf-bad5-4a70fc95d74c] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 596.774707] env[63515]: DEBUG nova.network.neutron [None req-7f5043f7-e6d7-46dd-88fd-99ee049c2b4b tempest-VolumesAssistedSnapshotsTest-453349191 tempest-VolumesAssistedSnapshotsTest-453349191-project-member] [instance: db682708-36fa-4126-a848-bfb609df96d7] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 596.884633] env[63515]: ERROR nova.compute.manager [None req-9e973647-d524-4c71-b05a-f1a253758bab tempest-ServerAddressesTestJSON-9863250 tempest-ServerAddressesTestJSON-9863250-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 79794b46-d801-404a-94f2-a223e825fcfc, please check neutron logs for more information. [ 596.884633] env[63515]: ERROR nova.compute.manager Traceback (most recent call last): [ 596.884633] env[63515]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 596.884633] env[63515]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 596.884633] env[63515]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 596.884633] env[63515]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 596.884633] env[63515]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 596.884633] env[63515]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 596.884633] env[63515]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 596.884633] env[63515]: ERROR nova.compute.manager self.force_reraise() [ 596.884633] env[63515]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 596.884633] env[63515]: ERROR nova.compute.manager raise self.value [ 596.884633] env[63515]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 596.884633] env[63515]: ERROR nova.compute.manager updated_port = self._update_port( [ 596.884633] env[63515]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 596.884633] env[63515]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 596.885288] env[63515]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 596.885288] env[63515]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 596.885288] env[63515]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 79794b46-d801-404a-94f2-a223e825fcfc, please check neutron logs for more information. [ 596.885288] env[63515]: ERROR nova.compute.manager [ 596.885288] env[63515]: Traceback (most recent call last): [ 596.885288] env[63515]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 596.885288] env[63515]: listener.cb(fileno) [ 596.885288] env[63515]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 596.885288] env[63515]: result = function(*args, **kwargs) [ 596.885288] env[63515]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 596.885288] env[63515]: return func(*args, **kwargs) [ 596.885288] env[63515]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 596.885288] env[63515]: raise e [ 596.885288] env[63515]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 596.885288] env[63515]: nwinfo = self.network_api.allocate_for_instance( [ 596.885288] env[63515]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 596.885288] env[63515]: created_port_ids = self._update_ports_for_instance( [ 596.885288] env[63515]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 596.885288] env[63515]: with excutils.save_and_reraise_exception(): [ 596.885288] env[63515]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 596.885288] env[63515]: self.force_reraise() [ 596.885288] env[63515]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 596.885288] env[63515]: raise self.value [ 596.885288] env[63515]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 596.885288] env[63515]: updated_port = self._update_port( [ 596.885288] env[63515]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 596.885288] env[63515]: _ensure_no_port_binding_failure(port) [ 596.885288] env[63515]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 596.885288] env[63515]: raise exception.PortBindingFailed(port_id=port['id']) [ 596.886361] env[63515]: nova.exception.PortBindingFailed: Binding failed for port 79794b46-d801-404a-94f2-a223e825fcfc, please check neutron logs for more information. [ 596.886361] env[63515]: Removing descriptor: 16 [ 596.886651] env[63515]: ERROR nova.compute.manager [None req-9e973647-d524-4c71-b05a-f1a253758bab tempest-ServerAddressesTestJSON-9863250 tempest-ServerAddressesTestJSON-9863250-project-member] [instance: 7fee0afb-88c9-4e93-a4d9-94312394f927] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 79794b46-d801-404a-94f2-a223e825fcfc, please check neutron logs for more information. [ 596.886651] env[63515]: ERROR nova.compute.manager [instance: 7fee0afb-88c9-4e93-a4d9-94312394f927] Traceback (most recent call last): [ 596.886651] env[63515]: ERROR nova.compute.manager [instance: 7fee0afb-88c9-4e93-a4d9-94312394f927] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 596.886651] env[63515]: ERROR nova.compute.manager [instance: 7fee0afb-88c9-4e93-a4d9-94312394f927] yield resources [ 596.886651] env[63515]: ERROR nova.compute.manager [instance: 7fee0afb-88c9-4e93-a4d9-94312394f927] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 596.886651] env[63515]: ERROR nova.compute.manager [instance: 7fee0afb-88c9-4e93-a4d9-94312394f927] self.driver.spawn(context, instance, image_meta, [ 596.886651] env[63515]: ERROR nova.compute.manager [instance: 7fee0afb-88c9-4e93-a4d9-94312394f927] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 596.886651] env[63515]: ERROR nova.compute.manager [instance: 7fee0afb-88c9-4e93-a4d9-94312394f927] self._vmops.spawn(context, instance, image_meta, injected_files, [ 596.886651] env[63515]: ERROR nova.compute.manager [instance: 7fee0afb-88c9-4e93-a4d9-94312394f927] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 596.886651] env[63515]: ERROR nova.compute.manager [instance: 7fee0afb-88c9-4e93-a4d9-94312394f927] vm_ref = self.build_virtual_machine(instance, [ 596.886651] env[63515]: ERROR nova.compute.manager [instance: 7fee0afb-88c9-4e93-a4d9-94312394f927] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 596.887085] env[63515]: ERROR nova.compute.manager [instance: 7fee0afb-88c9-4e93-a4d9-94312394f927] vif_infos = vmwarevif.get_vif_info(self._session, [ 596.887085] env[63515]: ERROR nova.compute.manager [instance: 7fee0afb-88c9-4e93-a4d9-94312394f927] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 596.887085] env[63515]: ERROR nova.compute.manager [instance: 7fee0afb-88c9-4e93-a4d9-94312394f927] for vif in network_info: [ 596.887085] env[63515]: ERROR nova.compute.manager [instance: 7fee0afb-88c9-4e93-a4d9-94312394f927] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 596.887085] env[63515]: ERROR nova.compute.manager [instance: 7fee0afb-88c9-4e93-a4d9-94312394f927] return self._sync_wrapper(fn, *args, **kwargs) [ 596.887085] env[63515]: ERROR nova.compute.manager [instance: 7fee0afb-88c9-4e93-a4d9-94312394f927] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 596.887085] env[63515]: ERROR nova.compute.manager [instance: 7fee0afb-88c9-4e93-a4d9-94312394f927] self.wait() [ 596.887085] env[63515]: ERROR nova.compute.manager [instance: 7fee0afb-88c9-4e93-a4d9-94312394f927] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 596.887085] env[63515]: ERROR nova.compute.manager [instance: 7fee0afb-88c9-4e93-a4d9-94312394f927] self[:] = self._gt.wait() [ 596.887085] env[63515]: ERROR nova.compute.manager [instance: 7fee0afb-88c9-4e93-a4d9-94312394f927] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 596.887085] env[63515]: ERROR nova.compute.manager [instance: 7fee0afb-88c9-4e93-a4d9-94312394f927] return self._exit_event.wait() [ 596.887085] env[63515]: ERROR nova.compute.manager [instance: 7fee0afb-88c9-4e93-a4d9-94312394f927] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 596.887085] env[63515]: ERROR nova.compute.manager [instance: 7fee0afb-88c9-4e93-a4d9-94312394f927] result = hub.switch() [ 596.887573] env[63515]: ERROR nova.compute.manager [instance: 7fee0afb-88c9-4e93-a4d9-94312394f927] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 596.887573] env[63515]: ERROR nova.compute.manager [instance: 7fee0afb-88c9-4e93-a4d9-94312394f927] return self.greenlet.switch() [ 596.887573] env[63515]: ERROR nova.compute.manager [instance: 7fee0afb-88c9-4e93-a4d9-94312394f927] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 596.887573] env[63515]: ERROR nova.compute.manager [instance: 7fee0afb-88c9-4e93-a4d9-94312394f927] result = function(*args, **kwargs) [ 596.887573] env[63515]: ERROR nova.compute.manager [instance: 7fee0afb-88c9-4e93-a4d9-94312394f927] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 596.887573] env[63515]: ERROR nova.compute.manager [instance: 7fee0afb-88c9-4e93-a4d9-94312394f927] return func(*args, **kwargs) [ 596.887573] env[63515]: ERROR nova.compute.manager [instance: 7fee0afb-88c9-4e93-a4d9-94312394f927] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 596.887573] env[63515]: ERROR nova.compute.manager [instance: 7fee0afb-88c9-4e93-a4d9-94312394f927] raise e [ 596.887573] env[63515]: ERROR nova.compute.manager [instance: 7fee0afb-88c9-4e93-a4d9-94312394f927] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 596.887573] env[63515]: ERROR nova.compute.manager [instance: 7fee0afb-88c9-4e93-a4d9-94312394f927] nwinfo = self.network_api.allocate_for_instance( [ 596.887573] env[63515]: ERROR nova.compute.manager [instance: 7fee0afb-88c9-4e93-a4d9-94312394f927] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 596.887573] env[63515]: ERROR nova.compute.manager [instance: 7fee0afb-88c9-4e93-a4d9-94312394f927] created_port_ids = self._update_ports_for_instance( [ 596.887573] env[63515]: ERROR nova.compute.manager [instance: 7fee0afb-88c9-4e93-a4d9-94312394f927] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 596.888045] env[63515]: ERROR nova.compute.manager [instance: 7fee0afb-88c9-4e93-a4d9-94312394f927] with excutils.save_and_reraise_exception(): [ 596.888045] env[63515]: ERROR nova.compute.manager [instance: 7fee0afb-88c9-4e93-a4d9-94312394f927] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 596.888045] env[63515]: ERROR nova.compute.manager [instance: 7fee0afb-88c9-4e93-a4d9-94312394f927] self.force_reraise() [ 596.888045] env[63515]: ERROR nova.compute.manager [instance: 7fee0afb-88c9-4e93-a4d9-94312394f927] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 596.888045] env[63515]: ERROR nova.compute.manager [instance: 7fee0afb-88c9-4e93-a4d9-94312394f927] raise self.value [ 596.888045] env[63515]: ERROR nova.compute.manager [instance: 7fee0afb-88c9-4e93-a4d9-94312394f927] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 596.888045] env[63515]: ERROR nova.compute.manager [instance: 7fee0afb-88c9-4e93-a4d9-94312394f927] updated_port = self._update_port( [ 596.888045] env[63515]: ERROR nova.compute.manager [instance: 7fee0afb-88c9-4e93-a4d9-94312394f927] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 596.888045] env[63515]: ERROR nova.compute.manager [instance: 7fee0afb-88c9-4e93-a4d9-94312394f927] _ensure_no_port_binding_failure(port) [ 596.888045] env[63515]: ERROR nova.compute.manager [instance: 7fee0afb-88c9-4e93-a4d9-94312394f927] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 596.888045] env[63515]: ERROR nova.compute.manager [instance: 7fee0afb-88c9-4e93-a4d9-94312394f927] raise exception.PortBindingFailed(port_id=port['id']) [ 596.888045] env[63515]: ERROR nova.compute.manager [instance: 7fee0afb-88c9-4e93-a4d9-94312394f927] nova.exception.PortBindingFailed: Binding failed for port 79794b46-d801-404a-94f2-a223e825fcfc, please check neutron logs for more information. [ 596.888045] env[63515]: ERROR nova.compute.manager [instance: 7fee0afb-88c9-4e93-a4d9-94312394f927] [ 596.888584] env[63515]: INFO nova.compute.manager [None req-9e973647-d524-4c71-b05a-f1a253758bab tempest-ServerAddressesTestJSON-9863250 tempest-ServerAddressesTestJSON-9863250-project-member] [instance: 7fee0afb-88c9-4e93-a4d9-94312394f927] Terminating instance [ 596.893152] env[63515]: DEBUG oslo_concurrency.lockutils [None req-9e973647-d524-4c71-b05a-f1a253758bab tempest-ServerAddressesTestJSON-9863250 tempest-ServerAddressesTestJSON-9863250-project-member] Acquiring lock "refresh_cache-7fee0afb-88c9-4e93-a4d9-94312394f927" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 596.893152] env[63515]: DEBUG oslo_concurrency.lockutils [None req-9e973647-d524-4c71-b05a-f1a253758bab tempest-ServerAddressesTestJSON-9863250 tempest-ServerAddressesTestJSON-9863250-project-member] Acquired lock "refresh_cache-7fee0afb-88c9-4e93-a4d9-94312394f927" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 596.893152] env[63515]: DEBUG nova.network.neutron [None req-9e973647-d524-4c71-b05a-f1a253758bab tempest-ServerAddressesTestJSON-9863250 tempest-ServerAddressesTestJSON-9863250-project-member] [instance: 7fee0afb-88c9-4e93-a4d9-94312394f927] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 596.913426] env[63515]: DEBUG nova.network.neutron [None req-7f5043f7-e6d7-46dd-88fd-99ee049c2b4b tempest-VolumesAssistedSnapshotsTest-453349191 tempest-VolumesAssistedSnapshotsTest-453349191-project-member] [instance: db682708-36fa-4126-a848-bfb609df96d7] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 597.116288] env[63515]: INFO nova.compute.manager [-] [instance: d5f988f2-765b-4ecf-bad5-4a70fc95d74c] Took 1.03 seconds to deallocate network for instance. [ 597.117143] env[63515]: DEBUG nova.compute.claims [None req-22f02712-96d1-4031-8b8f-5bc880f34c33 tempest-ImagesNegativeTestJSON-2041328384 tempest-ImagesNegativeTestJSON-2041328384-project-member] [instance: d5f988f2-765b-4ecf-bad5-4a70fc95d74c] Aborting claim: {{(pid=63515) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 597.117578] env[63515]: DEBUG oslo_concurrency.lockutils [None req-22f02712-96d1-4031-8b8f-5bc880f34c33 tempest-ImagesNegativeTestJSON-2041328384 tempest-ImagesNegativeTestJSON-2041328384-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 597.255111] env[63515]: DEBUG oslo_concurrency.lockutils [req-401b3b15-b183-4efa-884e-530522f60457 req-409f492e-85a2-4ffb-9ad8-e86a4ba8c9d3 service nova] Releasing lock "refresh_cache-d5f988f2-765b-4ecf-bad5-4a70fc95d74c" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 597.255111] env[63515]: DEBUG nova.compute.manager [req-401b3b15-b183-4efa-884e-530522f60457 req-409f492e-85a2-4ffb-9ad8-e86a4ba8c9d3 service nova] [instance: d5f988f2-765b-4ecf-bad5-4a70fc95d74c] Received event network-vif-deleted-fca51b1b-a915-4365-a1b6-ccc0a5a43604 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 597.261208] env[63515]: DEBUG oslo_concurrency.lockutils [None req-b1f9566f-fcad-4d5d-9d53-b04b591d6c49 tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.011s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 597.262323] env[63515]: DEBUG oslo_concurrency.lockutils [None req-24636c84-3869-4a1b-acfb-9c10d33a9826 tempest-ServersWithSpecificFlavorTestJSON-691988195 tempest-ServersWithSpecificFlavorTestJSON-691988195-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.411s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 597.263840] env[63515]: INFO nova.compute.claims [None req-24636c84-3869-4a1b-acfb-9c10d33a9826 tempest-ServersWithSpecificFlavorTestJSON-691988195 tempest-ServersWithSpecificFlavorTestJSON-691988195-project-member] [instance: ebc737a3-e6cd-4f4c-a4b6-fce5924de4c6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 597.417050] env[63515]: DEBUG oslo_concurrency.lockutils [None req-7f5043f7-e6d7-46dd-88fd-99ee049c2b4b tempest-VolumesAssistedSnapshotsTest-453349191 tempest-VolumesAssistedSnapshotsTest-453349191-project-member] Releasing lock "refresh_cache-db682708-36fa-4126-a848-bfb609df96d7" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 597.417050] env[63515]: DEBUG nova.compute.manager [None req-7f5043f7-e6d7-46dd-88fd-99ee049c2b4b tempest-VolumesAssistedSnapshotsTest-453349191 tempest-VolumesAssistedSnapshotsTest-453349191-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63515) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 597.417050] env[63515]: DEBUG nova.compute.manager [None req-7f5043f7-e6d7-46dd-88fd-99ee049c2b4b tempest-VolumesAssistedSnapshotsTest-453349191 tempest-VolumesAssistedSnapshotsTest-453349191-project-member] [instance: db682708-36fa-4126-a848-bfb609df96d7] Deallocating network for instance {{(pid=63515) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 597.417050] env[63515]: DEBUG nova.network.neutron [None req-7f5043f7-e6d7-46dd-88fd-99ee049c2b4b tempest-VolumesAssistedSnapshotsTest-453349191 tempest-VolumesAssistedSnapshotsTest-453349191-project-member] [instance: db682708-36fa-4126-a848-bfb609df96d7] deallocate_for_instance() {{(pid=63515) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 597.420350] env[63515]: DEBUG nova.network.neutron [None req-9e973647-d524-4c71-b05a-f1a253758bab tempest-ServerAddressesTestJSON-9863250 tempest-ServerAddressesTestJSON-9863250-project-member] [instance: 7fee0afb-88c9-4e93-a4d9-94312394f927] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 597.436500] env[63515]: DEBUG nova.network.neutron [None req-7f5043f7-e6d7-46dd-88fd-99ee049c2b4b tempest-VolumesAssistedSnapshotsTest-453349191 tempest-VolumesAssistedSnapshotsTest-453349191-project-member] [instance: db682708-36fa-4126-a848-bfb609df96d7] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 597.526322] env[63515]: DEBUG nova.network.neutron [None req-9e973647-d524-4c71-b05a-f1a253758bab tempest-ServerAddressesTestJSON-9863250 tempest-ServerAddressesTestJSON-9863250-project-member] [instance: 7fee0afb-88c9-4e93-a4d9-94312394f927] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 597.689546] env[63515]: DEBUG nova.compute.manager [req-7e7dc84a-03a1-4ba7-9f1a-9fd821f4b78b req-08b94d07-fab6-4386-adfc-a8308360a4e5 service nova] [instance: 7fee0afb-88c9-4e93-a4d9-94312394f927] Received event network-changed-79794b46-d801-404a-94f2-a223e825fcfc {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 597.690903] env[63515]: DEBUG nova.compute.manager [req-7e7dc84a-03a1-4ba7-9f1a-9fd821f4b78b req-08b94d07-fab6-4386-adfc-a8308360a4e5 service nova] [instance: 7fee0afb-88c9-4e93-a4d9-94312394f927] Refreshing instance network info cache due to event network-changed-79794b46-d801-404a-94f2-a223e825fcfc. {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 597.690903] env[63515]: DEBUG oslo_concurrency.lockutils [req-7e7dc84a-03a1-4ba7-9f1a-9fd821f4b78b req-08b94d07-fab6-4386-adfc-a8308360a4e5 service nova] Acquiring lock "refresh_cache-7fee0afb-88c9-4e93-a4d9-94312394f927" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 597.941925] env[63515]: DEBUG nova.network.neutron [None req-7f5043f7-e6d7-46dd-88fd-99ee049c2b4b tempest-VolumesAssistedSnapshotsTest-453349191 tempest-VolumesAssistedSnapshotsTest-453349191-project-member] [instance: db682708-36fa-4126-a848-bfb609df96d7] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 598.029783] env[63515]: DEBUG oslo_concurrency.lockutils [None req-9e973647-d524-4c71-b05a-f1a253758bab tempest-ServerAddressesTestJSON-9863250 tempest-ServerAddressesTestJSON-9863250-project-member] Releasing lock "refresh_cache-7fee0afb-88c9-4e93-a4d9-94312394f927" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 598.030252] env[63515]: DEBUG nova.compute.manager [None req-9e973647-d524-4c71-b05a-f1a253758bab tempest-ServerAddressesTestJSON-9863250 tempest-ServerAddressesTestJSON-9863250-project-member] [instance: 7fee0afb-88c9-4e93-a4d9-94312394f927] Start destroying the instance on the hypervisor. {{(pid=63515) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 598.030449] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-9e973647-d524-4c71-b05a-f1a253758bab tempest-ServerAddressesTestJSON-9863250 tempest-ServerAddressesTestJSON-9863250-project-member] [instance: 7fee0afb-88c9-4e93-a4d9-94312394f927] Destroying instance {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 598.031492] env[63515]: DEBUG oslo_concurrency.lockutils [req-7e7dc84a-03a1-4ba7-9f1a-9fd821f4b78b req-08b94d07-fab6-4386-adfc-a8308360a4e5 service nova] Acquired lock "refresh_cache-7fee0afb-88c9-4e93-a4d9-94312394f927" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 598.031492] env[63515]: DEBUG nova.network.neutron [req-7e7dc84a-03a1-4ba7-9f1a-9fd821f4b78b req-08b94d07-fab6-4386-adfc-a8308360a4e5 service nova] [instance: 7fee0afb-88c9-4e93-a4d9-94312394f927] Refreshing network info cache for port 79794b46-d801-404a-94f2-a223e825fcfc {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 598.032542] env[63515]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-0b22755d-dbce-450e-a0c3-40cfbded213a {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.042167] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9742dfaa-653d-44e7-9fa7-deeb5398004f {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.065947] env[63515]: WARNING nova.virt.vmwareapi.vmops [None req-9e973647-d524-4c71-b05a-f1a253758bab tempest-ServerAddressesTestJSON-9863250 tempest-ServerAddressesTestJSON-9863250-project-member] [instance: 7fee0afb-88c9-4e93-a4d9-94312394f927] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 7fee0afb-88c9-4e93-a4d9-94312394f927 could not be found. [ 598.066188] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-9e973647-d524-4c71-b05a-f1a253758bab tempest-ServerAddressesTestJSON-9863250 tempest-ServerAddressesTestJSON-9863250-project-member] [instance: 7fee0afb-88c9-4e93-a4d9-94312394f927] Instance destroyed {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 598.066367] env[63515]: INFO nova.compute.manager [None req-9e973647-d524-4c71-b05a-f1a253758bab tempest-ServerAddressesTestJSON-9863250 tempest-ServerAddressesTestJSON-9863250-project-member] [instance: 7fee0afb-88c9-4e93-a4d9-94312394f927] Took 0.04 seconds to destroy the instance on the hypervisor. [ 598.066607] env[63515]: DEBUG oslo.service.loopingcall [None req-9e973647-d524-4c71-b05a-f1a253758bab tempest-ServerAddressesTestJSON-9863250 tempest-ServerAddressesTestJSON-9863250-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 598.066818] env[63515]: DEBUG nova.compute.manager [-] [instance: 7fee0afb-88c9-4e93-a4d9-94312394f927] Deallocating network for instance {{(pid=63515) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 598.066913] env[63515]: DEBUG nova.network.neutron [-] [instance: 7fee0afb-88c9-4e93-a4d9-94312394f927] deallocate_for_instance() {{(pid=63515) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 598.088632] env[63515]: DEBUG nova.network.neutron [-] [instance: 7fee0afb-88c9-4e93-a4d9-94312394f927] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 598.446240] env[63515]: INFO nova.compute.manager [None req-7f5043f7-e6d7-46dd-88fd-99ee049c2b4b tempest-VolumesAssistedSnapshotsTest-453349191 tempest-VolumesAssistedSnapshotsTest-453349191-project-member] [instance: db682708-36fa-4126-a848-bfb609df96d7] Took 1.03 seconds to deallocate network for instance. [ 598.559032] env[63515]: DEBUG nova.network.neutron [req-7e7dc84a-03a1-4ba7-9f1a-9fd821f4b78b req-08b94d07-fab6-4386-adfc-a8308360a4e5 service nova] [instance: 7fee0afb-88c9-4e93-a4d9-94312394f927] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 598.595652] env[63515]: DEBUG nova.network.neutron [-] [instance: 7fee0afb-88c9-4e93-a4d9-94312394f927] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 598.656704] env[63515]: DEBUG nova.network.neutron [req-7e7dc84a-03a1-4ba7-9f1a-9fd821f4b78b req-08b94d07-fab6-4386-adfc-a8308360a4e5 service nova] [instance: 7fee0afb-88c9-4e93-a4d9-94312394f927] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 598.700738] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41895c6c-cbf5-44a5-8561-e3d2eaa189d3 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.710116] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee237823-a94e-48ae-a88f-e6a9d7814a7f {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.744623] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c203e6c-8ad6-4af2-b8da-66208517bc6f {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.754168] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5ce1000-4f35-424b-939e-77bc8a3b50da {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.768924] env[63515]: DEBUG nova.compute.provider_tree [None req-24636c84-3869-4a1b-acfb-9c10d33a9826 tempest-ServersWithSpecificFlavorTestJSON-691988195 tempest-ServersWithSpecificFlavorTestJSON-691988195-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 599.097138] env[63515]: INFO nova.compute.manager [-] [instance: 7fee0afb-88c9-4e93-a4d9-94312394f927] Took 1.03 seconds to deallocate network for instance. [ 599.100095] env[63515]: DEBUG nova.compute.claims [None req-9e973647-d524-4c71-b05a-f1a253758bab tempest-ServerAddressesTestJSON-9863250 tempest-ServerAddressesTestJSON-9863250-project-member] [instance: 7fee0afb-88c9-4e93-a4d9-94312394f927] Aborting claim: {{(pid=63515) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 599.100261] env[63515]: DEBUG oslo_concurrency.lockutils [None req-9e973647-d524-4c71-b05a-f1a253758bab tempest-ServerAddressesTestJSON-9863250 tempest-ServerAddressesTestJSON-9863250-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 599.160774] env[63515]: DEBUG oslo_concurrency.lockutils [req-7e7dc84a-03a1-4ba7-9f1a-9fd821f4b78b req-08b94d07-fab6-4386-adfc-a8308360a4e5 service nova] Releasing lock "refresh_cache-7fee0afb-88c9-4e93-a4d9-94312394f927" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 599.160774] env[63515]: DEBUG nova.compute.manager [req-7e7dc84a-03a1-4ba7-9f1a-9fd821f4b78b req-08b94d07-fab6-4386-adfc-a8308360a4e5 service nova] [instance: 7fee0afb-88c9-4e93-a4d9-94312394f927] Received event network-vif-deleted-79794b46-d801-404a-94f2-a223e825fcfc {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 599.272273] env[63515]: DEBUG nova.scheduler.client.report [None req-24636c84-3869-4a1b-acfb-9c10d33a9826 tempest-ServersWithSpecificFlavorTestJSON-691988195 tempest-ServersWithSpecificFlavorTestJSON-691988195-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 599.481738] env[63515]: INFO nova.scheduler.client.report [None req-7f5043f7-e6d7-46dd-88fd-99ee049c2b4b tempest-VolumesAssistedSnapshotsTest-453349191 tempest-VolumesAssistedSnapshotsTest-453349191-project-member] Deleted allocations for instance db682708-36fa-4126-a848-bfb609df96d7 [ 599.778636] env[63515]: DEBUG oslo_concurrency.lockutils [None req-24636c84-3869-4a1b-acfb-9c10d33a9826 tempest-ServersWithSpecificFlavorTestJSON-691988195 tempest-ServersWithSpecificFlavorTestJSON-691988195-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.516s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 599.779173] env[63515]: DEBUG nova.compute.manager [None req-24636c84-3869-4a1b-acfb-9c10d33a9826 tempest-ServersWithSpecificFlavorTestJSON-691988195 tempest-ServersWithSpecificFlavorTestJSON-691988195-project-member] [instance: ebc737a3-e6cd-4f4c-a4b6-fce5924de4c6] Start building networks asynchronously for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 599.781794] env[63515]: DEBUG oslo_concurrency.lockutils [None req-5e279136-7e72-43a2-85c4-9cd11158a72b tempest-ListImageFiltersTestJSON-1100101045 tempest-ListImageFiltersTestJSON-1100101045-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.930s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 599.790681] env[63515]: DEBUG oslo_service.periodic_task [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=63515) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 599.790681] env[63515]: DEBUG oslo_service.periodic_task [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=63515) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 599.990436] env[63515]: DEBUG oslo_concurrency.lockutils [None req-7f5043f7-e6d7-46dd-88fd-99ee049c2b4b tempest-VolumesAssistedSnapshotsTest-453349191 tempest-VolumesAssistedSnapshotsTest-453349191-project-member] Lock "db682708-36fa-4126-a848-bfb609df96d7" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 56.573s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 600.286476] env[63515]: DEBUG nova.compute.utils [None req-24636c84-3869-4a1b-acfb-9c10d33a9826 tempest-ServersWithSpecificFlavorTestJSON-691988195 tempest-ServersWithSpecificFlavorTestJSON-691988195-project-member] Using /dev/sd instead of None {{(pid=63515) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 600.291311] env[63515]: DEBUG nova.compute.manager [None req-24636c84-3869-4a1b-acfb-9c10d33a9826 tempest-ServersWithSpecificFlavorTestJSON-691988195 tempest-ServersWithSpecificFlavorTestJSON-691988195-project-member] [instance: ebc737a3-e6cd-4f4c-a4b6-fce5924de4c6] Allocating IP information in the background. {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 600.291461] env[63515]: DEBUG nova.network.neutron [None req-24636c84-3869-4a1b-acfb-9c10d33a9826 tempest-ServersWithSpecificFlavorTestJSON-691988195 tempest-ServersWithSpecificFlavorTestJSON-691988195-project-member] [instance: ebc737a3-e6cd-4f4c-a4b6-fce5924de4c6] allocate_for_instance() {{(pid=63515) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 600.296958] env[63515]: DEBUG oslo_service.periodic_task [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=63515) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 600.297119] env[63515]: DEBUG nova.compute.manager [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Starting heal instance info cache {{(pid=63515) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9926}} [ 600.297246] env[63515]: DEBUG nova.compute.manager [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Rebuilding the list of instances to heal {{(pid=63515) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9930}} [ 600.346016] env[63515]: DEBUG nova.policy [None req-24636c84-3869-4a1b-acfb-9c10d33a9826 tempest-ServersWithSpecificFlavorTestJSON-691988195 tempest-ServersWithSpecificFlavorTestJSON-691988195-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '62e2d4ecc9a64d068e9856efc7f17282', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7eb8587e6ef34a6aa6931f5c0fc39271', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63515) authorize /opt/stack/nova/nova/policy.py:201}} [ 600.493987] env[63515]: DEBUG nova.compute.manager [None req-9131a685-703c-44a1-81f8-7c157d95e037 tempest-ServerAddressesNegativeTestJSON-1262066083 tempest-ServerAddressesNegativeTestJSON-1262066083-project-member] [instance: 5b201ab6-c7ec-4a5c-a310-982649f34be4] Starting instance... {{(pid=63515) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 600.694903] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-506222db-a81e-4704-9c16-82a59fb53529 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 600.703146] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2aa626e-15f6-492b-8a8d-56627d8845e4 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 600.753156] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6862bb45-db1c-4485-94c0-c0cfb5e373fb {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 600.765022] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa3d47e1-8249-496c-b884-0882ce24d5c2 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 600.766314] env[63515]: DEBUG nova.network.neutron [None req-24636c84-3869-4a1b-acfb-9c10d33a9826 tempest-ServersWithSpecificFlavorTestJSON-691988195 tempest-ServersWithSpecificFlavorTestJSON-691988195-project-member] [instance: ebc737a3-e6cd-4f4c-a4b6-fce5924de4c6] Successfully created port: 168387a2-8a0f-41c4-b727-bad7d761613f {{(pid=63515) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 600.777367] env[63515]: DEBUG nova.compute.provider_tree [None req-5e279136-7e72-43a2-85c4-9cd11158a72b tempest-ListImageFiltersTestJSON-1100101045 tempest-ListImageFiltersTestJSON-1100101045-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 600.791546] env[63515]: DEBUG nova.compute.manager [None req-24636c84-3869-4a1b-acfb-9c10d33a9826 tempest-ServersWithSpecificFlavorTestJSON-691988195 tempest-ServersWithSpecificFlavorTestJSON-691988195-project-member] [instance: ebc737a3-e6cd-4f4c-a4b6-fce5924de4c6] Start building block device mappings for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 600.803416] env[63515]: DEBUG nova.compute.manager [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] [instance: ab1539ba-33a4-4139-a2f3-1f8c507beede] Skipping network cache update for instance because it is Building. {{(pid=63515) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 600.803695] env[63515]: DEBUG nova.compute.manager [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] [instance: b84c6c8d-e358-4a14-8425-c4f1d8adf4ee] Skipping network cache update for instance because it is Building. {{(pid=63515) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 600.803852] env[63515]: DEBUG nova.compute.manager [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] [instance: d5f988f2-765b-4ecf-bad5-4a70fc95d74c] Skipping network cache update for instance because it is Building. {{(pid=63515) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 600.803960] env[63515]: DEBUG nova.compute.manager [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] [instance: 7fee0afb-88c9-4e93-a4d9-94312394f927] Skipping network cache update for instance because it is Building. {{(pid=63515) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 600.804095] env[63515]: DEBUG nova.compute.manager [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] [instance: ebc737a3-e6cd-4f4c-a4b6-fce5924de4c6] Skipping network cache update for instance because it is Building. {{(pid=63515) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 600.830233] env[63515]: DEBUG oslo_concurrency.lockutils [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Acquiring lock "refresh_cache-93b28142-8454-43c0-b0a7-d61aa95c8fc8" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 600.830233] env[63515]: DEBUG oslo_concurrency.lockutils [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Acquired lock "refresh_cache-93b28142-8454-43c0-b0a7-d61aa95c8fc8" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 600.830233] env[63515]: DEBUG nova.network.neutron [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] [instance: 93b28142-8454-43c0-b0a7-d61aa95c8fc8] Forcefully refreshing network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2004}} [ 600.830233] env[63515]: DEBUG nova.objects.instance [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Lazy-loading 'info_cache' on Instance uuid 93b28142-8454-43c0-b0a7-d61aa95c8fc8 {{(pid=63515) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 601.023456] env[63515]: DEBUG oslo_concurrency.lockutils [None req-9131a685-703c-44a1-81f8-7c157d95e037 tempest-ServerAddressesNegativeTestJSON-1262066083 tempest-ServerAddressesNegativeTestJSON-1262066083-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 601.280869] env[63515]: DEBUG nova.scheduler.client.report [None req-5e279136-7e72-43a2-85c4-9cd11158a72b tempest-ListImageFiltersTestJSON-1100101045 tempest-ListImageFiltersTestJSON-1100101045-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 601.681795] env[63515]: DEBUG nova.compute.manager [req-6f92f102-f77d-45b3-84dc-a7190b04f22d req-d87e6c88-acfc-49a3-9d0b-b0540ccd9308 service nova] [instance: ebc737a3-e6cd-4f4c-a4b6-fce5924de4c6] Received event network-changed-168387a2-8a0f-41c4-b727-bad7d761613f {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 601.682085] env[63515]: DEBUG nova.compute.manager [req-6f92f102-f77d-45b3-84dc-a7190b04f22d req-d87e6c88-acfc-49a3-9d0b-b0540ccd9308 service nova] [instance: ebc737a3-e6cd-4f4c-a4b6-fce5924de4c6] Refreshing instance network info cache due to event network-changed-168387a2-8a0f-41c4-b727-bad7d761613f. {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 601.682242] env[63515]: DEBUG oslo_concurrency.lockutils [req-6f92f102-f77d-45b3-84dc-a7190b04f22d req-d87e6c88-acfc-49a3-9d0b-b0540ccd9308 service nova] Acquiring lock "refresh_cache-ebc737a3-e6cd-4f4c-a4b6-fce5924de4c6" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 601.682352] env[63515]: DEBUG oslo_concurrency.lockutils [req-6f92f102-f77d-45b3-84dc-a7190b04f22d req-d87e6c88-acfc-49a3-9d0b-b0540ccd9308 service nova] Acquired lock "refresh_cache-ebc737a3-e6cd-4f4c-a4b6-fce5924de4c6" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 601.682523] env[63515]: DEBUG nova.network.neutron [req-6f92f102-f77d-45b3-84dc-a7190b04f22d req-d87e6c88-acfc-49a3-9d0b-b0540ccd9308 service nova] [instance: ebc737a3-e6cd-4f4c-a4b6-fce5924de4c6] Refreshing network info cache for port 168387a2-8a0f-41c4-b727-bad7d761613f {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 601.786299] env[63515]: DEBUG oslo_concurrency.lockutils [None req-5e279136-7e72-43a2-85c4-9cd11158a72b tempest-ListImageFiltersTestJSON-1100101045 tempest-ListImageFiltersTestJSON-1100101045-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.004s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 601.786945] env[63515]: ERROR nova.compute.manager [None req-5e279136-7e72-43a2-85c4-9cd11158a72b tempest-ListImageFiltersTestJSON-1100101045 tempest-ListImageFiltersTestJSON-1100101045-project-member] [instance: 48908cb0-a989-4bad-84a4-25a4bdd3baeb] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port e7173eb8-86a4-4d43-8391-1d565f1dfaa9, please check neutron logs for more information. [ 601.786945] env[63515]: ERROR nova.compute.manager [instance: 48908cb0-a989-4bad-84a4-25a4bdd3baeb] Traceback (most recent call last): [ 601.786945] env[63515]: ERROR nova.compute.manager [instance: 48908cb0-a989-4bad-84a4-25a4bdd3baeb] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 601.786945] env[63515]: ERROR nova.compute.manager [instance: 48908cb0-a989-4bad-84a4-25a4bdd3baeb] self.driver.spawn(context, instance, image_meta, [ 601.786945] env[63515]: ERROR nova.compute.manager [instance: 48908cb0-a989-4bad-84a4-25a4bdd3baeb] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 601.786945] env[63515]: ERROR nova.compute.manager [instance: 48908cb0-a989-4bad-84a4-25a4bdd3baeb] self._vmops.spawn(context, instance, image_meta, injected_files, [ 601.786945] env[63515]: ERROR nova.compute.manager [instance: 48908cb0-a989-4bad-84a4-25a4bdd3baeb] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 601.786945] env[63515]: ERROR nova.compute.manager [instance: 48908cb0-a989-4bad-84a4-25a4bdd3baeb] vm_ref = self.build_virtual_machine(instance, [ 601.786945] env[63515]: ERROR nova.compute.manager [instance: 48908cb0-a989-4bad-84a4-25a4bdd3baeb] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 601.786945] env[63515]: ERROR nova.compute.manager [instance: 48908cb0-a989-4bad-84a4-25a4bdd3baeb] vif_infos = vmwarevif.get_vif_info(self._session, [ 601.786945] env[63515]: ERROR nova.compute.manager [instance: 48908cb0-a989-4bad-84a4-25a4bdd3baeb] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 601.787526] env[63515]: ERROR nova.compute.manager [instance: 48908cb0-a989-4bad-84a4-25a4bdd3baeb] for vif in network_info: [ 601.787526] env[63515]: ERROR nova.compute.manager [instance: 48908cb0-a989-4bad-84a4-25a4bdd3baeb] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 601.787526] env[63515]: ERROR nova.compute.manager [instance: 48908cb0-a989-4bad-84a4-25a4bdd3baeb] return self._sync_wrapper(fn, *args, **kwargs) [ 601.787526] env[63515]: ERROR nova.compute.manager [instance: 48908cb0-a989-4bad-84a4-25a4bdd3baeb] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 601.787526] env[63515]: ERROR nova.compute.manager [instance: 48908cb0-a989-4bad-84a4-25a4bdd3baeb] self.wait() [ 601.787526] env[63515]: ERROR nova.compute.manager [instance: 48908cb0-a989-4bad-84a4-25a4bdd3baeb] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 601.787526] env[63515]: ERROR nova.compute.manager [instance: 48908cb0-a989-4bad-84a4-25a4bdd3baeb] self[:] = self._gt.wait() [ 601.787526] env[63515]: ERROR nova.compute.manager [instance: 48908cb0-a989-4bad-84a4-25a4bdd3baeb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 601.787526] env[63515]: ERROR nova.compute.manager [instance: 48908cb0-a989-4bad-84a4-25a4bdd3baeb] return self._exit_event.wait() [ 601.787526] env[63515]: ERROR nova.compute.manager [instance: 48908cb0-a989-4bad-84a4-25a4bdd3baeb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 601.787526] env[63515]: ERROR nova.compute.manager [instance: 48908cb0-a989-4bad-84a4-25a4bdd3baeb] result = hub.switch() [ 601.787526] env[63515]: ERROR nova.compute.manager [instance: 48908cb0-a989-4bad-84a4-25a4bdd3baeb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 601.787526] env[63515]: ERROR nova.compute.manager [instance: 48908cb0-a989-4bad-84a4-25a4bdd3baeb] return self.greenlet.switch() [ 601.788130] env[63515]: ERROR nova.compute.manager [instance: 48908cb0-a989-4bad-84a4-25a4bdd3baeb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 601.788130] env[63515]: ERROR nova.compute.manager [instance: 48908cb0-a989-4bad-84a4-25a4bdd3baeb] result = function(*args, **kwargs) [ 601.788130] env[63515]: ERROR nova.compute.manager [instance: 48908cb0-a989-4bad-84a4-25a4bdd3baeb] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 601.788130] env[63515]: ERROR nova.compute.manager [instance: 48908cb0-a989-4bad-84a4-25a4bdd3baeb] return func(*args, **kwargs) [ 601.788130] env[63515]: ERROR nova.compute.manager [instance: 48908cb0-a989-4bad-84a4-25a4bdd3baeb] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 601.788130] env[63515]: ERROR nova.compute.manager [instance: 48908cb0-a989-4bad-84a4-25a4bdd3baeb] raise e [ 601.788130] env[63515]: ERROR nova.compute.manager [instance: 48908cb0-a989-4bad-84a4-25a4bdd3baeb] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 601.788130] env[63515]: ERROR nova.compute.manager [instance: 48908cb0-a989-4bad-84a4-25a4bdd3baeb] nwinfo = self.network_api.allocate_for_instance( [ 601.788130] env[63515]: ERROR nova.compute.manager [instance: 48908cb0-a989-4bad-84a4-25a4bdd3baeb] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 601.788130] env[63515]: ERROR nova.compute.manager [instance: 48908cb0-a989-4bad-84a4-25a4bdd3baeb] created_port_ids = self._update_ports_for_instance( [ 601.788130] env[63515]: ERROR nova.compute.manager [instance: 48908cb0-a989-4bad-84a4-25a4bdd3baeb] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 601.788130] env[63515]: ERROR nova.compute.manager [instance: 48908cb0-a989-4bad-84a4-25a4bdd3baeb] with excutils.save_and_reraise_exception(): [ 601.788130] env[63515]: ERROR nova.compute.manager [instance: 48908cb0-a989-4bad-84a4-25a4bdd3baeb] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 601.788728] env[63515]: ERROR nova.compute.manager [instance: 48908cb0-a989-4bad-84a4-25a4bdd3baeb] self.force_reraise() [ 601.788728] env[63515]: ERROR nova.compute.manager [instance: 48908cb0-a989-4bad-84a4-25a4bdd3baeb] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 601.788728] env[63515]: ERROR nova.compute.manager [instance: 48908cb0-a989-4bad-84a4-25a4bdd3baeb] raise self.value [ 601.788728] env[63515]: ERROR nova.compute.manager [instance: 48908cb0-a989-4bad-84a4-25a4bdd3baeb] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 601.788728] env[63515]: ERROR nova.compute.manager [instance: 48908cb0-a989-4bad-84a4-25a4bdd3baeb] updated_port = self._update_port( [ 601.788728] env[63515]: ERROR nova.compute.manager [instance: 48908cb0-a989-4bad-84a4-25a4bdd3baeb] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 601.788728] env[63515]: ERROR nova.compute.manager [instance: 48908cb0-a989-4bad-84a4-25a4bdd3baeb] _ensure_no_port_binding_failure(port) [ 601.788728] env[63515]: ERROR nova.compute.manager [instance: 48908cb0-a989-4bad-84a4-25a4bdd3baeb] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 601.788728] env[63515]: ERROR nova.compute.manager [instance: 48908cb0-a989-4bad-84a4-25a4bdd3baeb] raise exception.PortBindingFailed(port_id=port['id']) [ 601.788728] env[63515]: ERROR nova.compute.manager [instance: 48908cb0-a989-4bad-84a4-25a4bdd3baeb] nova.exception.PortBindingFailed: Binding failed for port e7173eb8-86a4-4d43-8391-1d565f1dfaa9, please check neutron logs for more information. [ 601.788728] env[63515]: ERROR nova.compute.manager [instance: 48908cb0-a989-4bad-84a4-25a4bdd3baeb] [ 601.789255] env[63515]: DEBUG nova.compute.utils [None req-5e279136-7e72-43a2-85c4-9cd11158a72b tempest-ListImageFiltersTestJSON-1100101045 tempest-ListImageFiltersTestJSON-1100101045-project-member] [instance: 48908cb0-a989-4bad-84a4-25a4bdd3baeb] Binding failed for port e7173eb8-86a4-4d43-8391-1d565f1dfaa9, please check neutron logs for more information. {{(pid=63515) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 601.789255] env[63515]: DEBUG oslo_concurrency.lockutils [None req-d9ffa9d5-6856-47e5-8926-d090bbea2c09 tempest-DeleteServersAdminTestJSON-944542165 tempest-DeleteServersAdminTestJSON-944542165-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.702s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 601.790652] env[63515]: INFO nova.compute.claims [None req-d9ffa9d5-6856-47e5-8926-d090bbea2c09 tempest-DeleteServersAdminTestJSON-944542165 tempest-DeleteServersAdminTestJSON-944542165-project-member] [instance: ae5c6e69-a885-48e3-9c9e-017f33430edd] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 601.793335] env[63515]: DEBUG nova.compute.manager [None req-5e279136-7e72-43a2-85c4-9cd11158a72b tempest-ListImageFiltersTestJSON-1100101045 tempest-ListImageFiltersTestJSON-1100101045-project-member] [instance: 48908cb0-a989-4bad-84a4-25a4bdd3baeb] Build of instance 48908cb0-a989-4bad-84a4-25a4bdd3baeb was re-scheduled: Binding failed for port e7173eb8-86a4-4d43-8391-1d565f1dfaa9, please check neutron logs for more information. {{(pid=63515) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 601.793879] env[63515]: DEBUG nova.compute.manager [None req-5e279136-7e72-43a2-85c4-9cd11158a72b tempest-ListImageFiltersTestJSON-1100101045 tempest-ListImageFiltersTestJSON-1100101045-project-member] [instance: 48908cb0-a989-4bad-84a4-25a4bdd3baeb] Unplugging VIFs for instance {{(pid=63515) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 601.794170] env[63515]: DEBUG oslo_concurrency.lockutils [None req-5e279136-7e72-43a2-85c4-9cd11158a72b tempest-ListImageFiltersTestJSON-1100101045 tempest-ListImageFiltersTestJSON-1100101045-project-member] Acquiring lock "refresh_cache-48908cb0-a989-4bad-84a4-25a4bdd3baeb" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 601.794324] env[63515]: DEBUG oslo_concurrency.lockutils [None req-5e279136-7e72-43a2-85c4-9cd11158a72b tempest-ListImageFiltersTestJSON-1100101045 tempest-ListImageFiltersTestJSON-1100101045-project-member] Acquired lock "refresh_cache-48908cb0-a989-4bad-84a4-25a4bdd3baeb" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 601.794502] env[63515]: DEBUG nova.network.neutron [None req-5e279136-7e72-43a2-85c4-9cd11158a72b tempest-ListImageFiltersTestJSON-1100101045 tempest-ListImageFiltersTestJSON-1100101045-project-member] [instance: 48908cb0-a989-4bad-84a4-25a4bdd3baeb] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 601.803332] env[63515]: DEBUG nova.compute.manager [None req-24636c84-3869-4a1b-acfb-9c10d33a9826 tempest-ServersWithSpecificFlavorTestJSON-691988195 tempest-ServersWithSpecificFlavorTestJSON-691988195-project-member] [instance: ebc737a3-e6cd-4f4c-a4b6-fce5924de4c6] Start spawning the instance on the hypervisor. {{(pid=63515) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 601.841850] env[63515]: DEBUG nova.virt.hardware [None req-24636c84-3869-4a1b-acfb-9c10d33a9826 tempest-ServersWithSpecificFlavorTestJSON-691988195 tempest-ServersWithSpecificFlavorTestJSON-691988195-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T02:54:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='2011243300',id=25,is_public=True,memory_mb=192,name='tempest-flavor_with_ephemeral_0-390491150',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T02:51:56Z,direct_url=,disk_format='vmdk',id=8a120570-cb06-4099-b262-554ca0ad15c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b569255cc43e42e7ae3f2b4ad37c6ef4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T02:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 601.842443] env[63515]: DEBUG nova.virt.hardware [None req-24636c84-3869-4a1b-acfb-9c10d33a9826 tempest-ServersWithSpecificFlavorTestJSON-691988195 tempest-ServersWithSpecificFlavorTestJSON-691988195-project-member] Flavor limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 601.842741] env[63515]: DEBUG nova.virt.hardware [None req-24636c84-3869-4a1b-acfb-9c10d33a9826 tempest-ServersWithSpecificFlavorTestJSON-691988195 tempest-ServersWithSpecificFlavorTestJSON-691988195-project-member] Image limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 601.843058] env[63515]: DEBUG nova.virt.hardware [None req-24636c84-3869-4a1b-acfb-9c10d33a9826 tempest-ServersWithSpecificFlavorTestJSON-691988195 tempest-ServersWithSpecificFlavorTestJSON-691988195-project-member] Flavor pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 601.843316] env[63515]: DEBUG nova.virt.hardware [None req-24636c84-3869-4a1b-acfb-9c10d33a9826 tempest-ServersWithSpecificFlavorTestJSON-691988195 tempest-ServersWithSpecificFlavorTestJSON-691988195-project-member] Image pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 601.843647] env[63515]: DEBUG nova.virt.hardware [None req-24636c84-3869-4a1b-acfb-9c10d33a9826 tempest-ServersWithSpecificFlavorTestJSON-691988195 tempest-ServersWithSpecificFlavorTestJSON-691988195-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 601.844024] env[63515]: DEBUG nova.virt.hardware [None req-24636c84-3869-4a1b-acfb-9c10d33a9826 tempest-ServersWithSpecificFlavorTestJSON-691988195 tempest-ServersWithSpecificFlavorTestJSON-691988195-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 601.844607] env[63515]: DEBUG nova.virt.hardware [None req-24636c84-3869-4a1b-acfb-9c10d33a9826 tempest-ServersWithSpecificFlavorTestJSON-691988195 tempest-ServersWithSpecificFlavorTestJSON-691988195-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 601.845023] env[63515]: DEBUG nova.virt.hardware [None req-24636c84-3869-4a1b-acfb-9c10d33a9826 tempest-ServersWithSpecificFlavorTestJSON-691988195 tempest-ServersWithSpecificFlavorTestJSON-691988195-project-member] Got 1 possible topologies {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 601.845426] env[63515]: DEBUG nova.virt.hardware [None req-24636c84-3869-4a1b-acfb-9c10d33a9826 tempest-ServersWithSpecificFlavorTestJSON-691988195 tempest-ServersWithSpecificFlavorTestJSON-691988195-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 601.848045] env[63515]: DEBUG nova.virt.hardware [None req-24636c84-3869-4a1b-acfb-9c10d33a9826 tempest-ServersWithSpecificFlavorTestJSON-691988195 tempest-ServersWithSpecificFlavorTestJSON-691988195-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 601.850016] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f099b42-8c2f-4903-8996-ea2c409aa81f {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.858580] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b458a26-d7d9-4b89-bc87-5dc300a7f1c9 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.882545] env[63515]: DEBUG nova.network.neutron [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] [instance: 93b28142-8454-43c0-b0a7-d61aa95c8fc8] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 601.889873] env[63515]: ERROR nova.compute.manager [None req-24636c84-3869-4a1b-acfb-9c10d33a9826 tempest-ServersWithSpecificFlavorTestJSON-691988195 tempest-ServersWithSpecificFlavorTestJSON-691988195-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 168387a2-8a0f-41c4-b727-bad7d761613f, please check neutron logs for more information. [ 601.889873] env[63515]: ERROR nova.compute.manager Traceback (most recent call last): [ 601.889873] env[63515]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 601.889873] env[63515]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 601.889873] env[63515]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 601.889873] env[63515]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 601.889873] env[63515]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 601.889873] env[63515]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 601.889873] env[63515]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 601.889873] env[63515]: ERROR nova.compute.manager self.force_reraise() [ 601.889873] env[63515]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 601.889873] env[63515]: ERROR nova.compute.manager raise self.value [ 601.889873] env[63515]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 601.889873] env[63515]: ERROR nova.compute.manager updated_port = self._update_port( [ 601.889873] env[63515]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 601.889873] env[63515]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 601.890488] env[63515]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 601.890488] env[63515]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 601.890488] env[63515]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 168387a2-8a0f-41c4-b727-bad7d761613f, please check neutron logs for more information. [ 601.890488] env[63515]: ERROR nova.compute.manager [ 601.890488] env[63515]: Traceback (most recent call last): [ 601.890488] env[63515]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 601.890488] env[63515]: listener.cb(fileno) [ 601.890488] env[63515]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 601.890488] env[63515]: result = function(*args, **kwargs) [ 601.890488] env[63515]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 601.890488] env[63515]: return func(*args, **kwargs) [ 601.890488] env[63515]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 601.890488] env[63515]: raise e [ 601.890488] env[63515]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 601.890488] env[63515]: nwinfo = self.network_api.allocate_for_instance( [ 601.890488] env[63515]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 601.890488] env[63515]: created_port_ids = self._update_ports_for_instance( [ 601.890488] env[63515]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 601.890488] env[63515]: with excutils.save_and_reraise_exception(): [ 601.890488] env[63515]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 601.890488] env[63515]: self.force_reraise() [ 601.890488] env[63515]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 601.890488] env[63515]: raise self.value [ 601.890488] env[63515]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 601.890488] env[63515]: updated_port = self._update_port( [ 601.890488] env[63515]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 601.890488] env[63515]: _ensure_no_port_binding_failure(port) [ 601.890488] env[63515]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 601.890488] env[63515]: raise exception.PortBindingFailed(port_id=port['id']) [ 601.891489] env[63515]: nova.exception.PortBindingFailed: Binding failed for port 168387a2-8a0f-41c4-b727-bad7d761613f, please check neutron logs for more information. [ 601.891489] env[63515]: Removing descriptor: 17 [ 601.891489] env[63515]: ERROR nova.compute.manager [None req-24636c84-3869-4a1b-acfb-9c10d33a9826 tempest-ServersWithSpecificFlavorTestJSON-691988195 tempest-ServersWithSpecificFlavorTestJSON-691988195-project-member] [instance: ebc737a3-e6cd-4f4c-a4b6-fce5924de4c6] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 168387a2-8a0f-41c4-b727-bad7d761613f, please check neutron logs for more information. [ 601.891489] env[63515]: ERROR nova.compute.manager [instance: ebc737a3-e6cd-4f4c-a4b6-fce5924de4c6] Traceback (most recent call last): [ 601.891489] env[63515]: ERROR nova.compute.manager [instance: ebc737a3-e6cd-4f4c-a4b6-fce5924de4c6] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 601.891489] env[63515]: ERROR nova.compute.manager [instance: ebc737a3-e6cd-4f4c-a4b6-fce5924de4c6] yield resources [ 601.891489] env[63515]: ERROR nova.compute.manager [instance: ebc737a3-e6cd-4f4c-a4b6-fce5924de4c6] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 601.891489] env[63515]: ERROR nova.compute.manager [instance: ebc737a3-e6cd-4f4c-a4b6-fce5924de4c6] self.driver.spawn(context, instance, image_meta, [ 601.891489] env[63515]: ERROR nova.compute.manager [instance: ebc737a3-e6cd-4f4c-a4b6-fce5924de4c6] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 601.891489] env[63515]: ERROR nova.compute.manager [instance: ebc737a3-e6cd-4f4c-a4b6-fce5924de4c6] self._vmops.spawn(context, instance, image_meta, injected_files, [ 601.891489] env[63515]: ERROR nova.compute.manager [instance: ebc737a3-e6cd-4f4c-a4b6-fce5924de4c6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 601.891489] env[63515]: ERROR nova.compute.manager [instance: ebc737a3-e6cd-4f4c-a4b6-fce5924de4c6] vm_ref = self.build_virtual_machine(instance, [ 601.891994] env[63515]: ERROR nova.compute.manager [instance: ebc737a3-e6cd-4f4c-a4b6-fce5924de4c6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 601.891994] env[63515]: ERROR nova.compute.manager [instance: ebc737a3-e6cd-4f4c-a4b6-fce5924de4c6] vif_infos = vmwarevif.get_vif_info(self._session, [ 601.891994] env[63515]: ERROR nova.compute.manager [instance: ebc737a3-e6cd-4f4c-a4b6-fce5924de4c6] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 601.891994] env[63515]: ERROR nova.compute.manager [instance: ebc737a3-e6cd-4f4c-a4b6-fce5924de4c6] for vif in network_info: [ 601.891994] env[63515]: ERROR nova.compute.manager [instance: ebc737a3-e6cd-4f4c-a4b6-fce5924de4c6] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 601.891994] env[63515]: ERROR nova.compute.manager [instance: ebc737a3-e6cd-4f4c-a4b6-fce5924de4c6] return self._sync_wrapper(fn, *args, **kwargs) [ 601.891994] env[63515]: ERROR nova.compute.manager [instance: ebc737a3-e6cd-4f4c-a4b6-fce5924de4c6] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 601.891994] env[63515]: ERROR nova.compute.manager [instance: ebc737a3-e6cd-4f4c-a4b6-fce5924de4c6] self.wait() [ 601.891994] env[63515]: ERROR nova.compute.manager [instance: ebc737a3-e6cd-4f4c-a4b6-fce5924de4c6] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 601.891994] env[63515]: ERROR nova.compute.manager [instance: ebc737a3-e6cd-4f4c-a4b6-fce5924de4c6] self[:] = self._gt.wait() [ 601.891994] env[63515]: ERROR nova.compute.manager [instance: ebc737a3-e6cd-4f4c-a4b6-fce5924de4c6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 601.891994] env[63515]: ERROR nova.compute.manager [instance: ebc737a3-e6cd-4f4c-a4b6-fce5924de4c6] return self._exit_event.wait() [ 601.891994] env[63515]: ERROR nova.compute.manager [instance: ebc737a3-e6cd-4f4c-a4b6-fce5924de4c6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 601.892464] env[63515]: ERROR nova.compute.manager [instance: ebc737a3-e6cd-4f4c-a4b6-fce5924de4c6] result = hub.switch() [ 601.892464] env[63515]: ERROR nova.compute.manager [instance: ebc737a3-e6cd-4f4c-a4b6-fce5924de4c6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 601.892464] env[63515]: ERROR nova.compute.manager [instance: ebc737a3-e6cd-4f4c-a4b6-fce5924de4c6] return self.greenlet.switch() [ 601.892464] env[63515]: ERROR nova.compute.manager [instance: ebc737a3-e6cd-4f4c-a4b6-fce5924de4c6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 601.892464] env[63515]: ERROR nova.compute.manager [instance: ebc737a3-e6cd-4f4c-a4b6-fce5924de4c6] result = function(*args, **kwargs) [ 601.892464] env[63515]: ERROR nova.compute.manager [instance: ebc737a3-e6cd-4f4c-a4b6-fce5924de4c6] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 601.892464] env[63515]: ERROR nova.compute.manager [instance: ebc737a3-e6cd-4f4c-a4b6-fce5924de4c6] return func(*args, **kwargs) [ 601.892464] env[63515]: ERROR nova.compute.manager [instance: ebc737a3-e6cd-4f4c-a4b6-fce5924de4c6] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 601.892464] env[63515]: ERROR nova.compute.manager [instance: ebc737a3-e6cd-4f4c-a4b6-fce5924de4c6] raise e [ 601.892464] env[63515]: ERROR nova.compute.manager [instance: ebc737a3-e6cd-4f4c-a4b6-fce5924de4c6] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 601.892464] env[63515]: ERROR nova.compute.manager [instance: ebc737a3-e6cd-4f4c-a4b6-fce5924de4c6] nwinfo = self.network_api.allocate_for_instance( [ 601.892464] env[63515]: ERROR nova.compute.manager [instance: ebc737a3-e6cd-4f4c-a4b6-fce5924de4c6] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 601.892464] env[63515]: ERROR nova.compute.manager [instance: ebc737a3-e6cd-4f4c-a4b6-fce5924de4c6] created_port_ids = self._update_ports_for_instance( [ 601.892928] env[63515]: ERROR nova.compute.manager [instance: ebc737a3-e6cd-4f4c-a4b6-fce5924de4c6] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 601.892928] env[63515]: ERROR nova.compute.manager [instance: ebc737a3-e6cd-4f4c-a4b6-fce5924de4c6] with excutils.save_and_reraise_exception(): [ 601.892928] env[63515]: ERROR nova.compute.manager [instance: ebc737a3-e6cd-4f4c-a4b6-fce5924de4c6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 601.892928] env[63515]: ERROR nova.compute.manager [instance: ebc737a3-e6cd-4f4c-a4b6-fce5924de4c6] self.force_reraise() [ 601.892928] env[63515]: ERROR nova.compute.manager [instance: ebc737a3-e6cd-4f4c-a4b6-fce5924de4c6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 601.892928] env[63515]: ERROR nova.compute.manager [instance: ebc737a3-e6cd-4f4c-a4b6-fce5924de4c6] raise self.value [ 601.892928] env[63515]: ERROR nova.compute.manager [instance: ebc737a3-e6cd-4f4c-a4b6-fce5924de4c6] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 601.892928] env[63515]: ERROR nova.compute.manager [instance: ebc737a3-e6cd-4f4c-a4b6-fce5924de4c6] updated_port = self._update_port( [ 601.892928] env[63515]: ERROR nova.compute.manager [instance: ebc737a3-e6cd-4f4c-a4b6-fce5924de4c6] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 601.892928] env[63515]: ERROR nova.compute.manager [instance: ebc737a3-e6cd-4f4c-a4b6-fce5924de4c6] _ensure_no_port_binding_failure(port) [ 601.892928] env[63515]: ERROR nova.compute.manager [instance: ebc737a3-e6cd-4f4c-a4b6-fce5924de4c6] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 601.892928] env[63515]: ERROR nova.compute.manager [instance: ebc737a3-e6cd-4f4c-a4b6-fce5924de4c6] raise exception.PortBindingFailed(port_id=port['id']) [ 601.894327] env[63515]: ERROR nova.compute.manager [instance: ebc737a3-e6cd-4f4c-a4b6-fce5924de4c6] nova.exception.PortBindingFailed: Binding failed for port 168387a2-8a0f-41c4-b727-bad7d761613f, please check neutron logs for more information. [ 601.894327] env[63515]: ERROR nova.compute.manager [instance: ebc737a3-e6cd-4f4c-a4b6-fce5924de4c6] [ 601.894327] env[63515]: INFO nova.compute.manager [None req-24636c84-3869-4a1b-acfb-9c10d33a9826 tempest-ServersWithSpecificFlavorTestJSON-691988195 tempest-ServersWithSpecificFlavorTestJSON-691988195-project-member] [instance: ebc737a3-e6cd-4f4c-a4b6-fce5924de4c6] Terminating instance [ 601.894327] env[63515]: DEBUG oslo_concurrency.lockutils [None req-24636c84-3869-4a1b-acfb-9c10d33a9826 tempest-ServersWithSpecificFlavorTestJSON-691988195 tempest-ServersWithSpecificFlavorTestJSON-691988195-project-member] Acquiring lock "refresh_cache-ebc737a3-e6cd-4f4c-a4b6-fce5924de4c6" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 602.207104] env[63515]: DEBUG nova.network.neutron [req-6f92f102-f77d-45b3-84dc-a7190b04f22d req-d87e6c88-acfc-49a3-9d0b-b0540ccd9308 service nova] [instance: ebc737a3-e6cd-4f4c-a4b6-fce5924de4c6] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 602.315118] env[63515]: DEBUG nova.network.neutron [req-6f92f102-f77d-45b3-84dc-a7190b04f22d req-d87e6c88-acfc-49a3-9d0b-b0540ccd9308 service nova] [instance: ebc737a3-e6cd-4f4c-a4b6-fce5924de4c6] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 602.323495] env[63515]: DEBUG nova.network.neutron [None req-5e279136-7e72-43a2-85c4-9cd11158a72b tempest-ListImageFiltersTestJSON-1100101045 tempest-ListImageFiltersTestJSON-1100101045-project-member] [instance: 48908cb0-a989-4bad-84a4-25a4bdd3baeb] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 602.579476] env[63515]: DEBUG nova.network.neutron [None req-5e279136-7e72-43a2-85c4-9cd11158a72b tempest-ListImageFiltersTestJSON-1100101045 tempest-ListImageFiltersTestJSON-1100101045-project-member] [instance: 48908cb0-a989-4bad-84a4-25a4bdd3baeb] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 602.585606] env[63515]: DEBUG nova.network.neutron [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] [instance: 93b28142-8454-43c0-b0a7-d61aa95c8fc8] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 602.819824] env[63515]: DEBUG oslo_concurrency.lockutils [req-6f92f102-f77d-45b3-84dc-a7190b04f22d req-d87e6c88-acfc-49a3-9d0b-b0540ccd9308 service nova] Releasing lock "refresh_cache-ebc737a3-e6cd-4f4c-a4b6-fce5924de4c6" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 602.819824] env[63515]: DEBUG oslo_concurrency.lockutils [None req-24636c84-3869-4a1b-acfb-9c10d33a9826 tempest-ServersWithSpecificFlavorTestJSON-691988195 tempest-ServersWithSpecificFlavorTestJSON-691988195-project-member] Acquired lock "refresh_cache-ebc737a3-e6cd-4f4c-a4b6-fce5924de4c6" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 602.819824] env[63515]: DEBUG nova.network.neutron [None req-24636c84-3869-4a1b-acfb-9c10d33a9826 tempest-ServersWithSpecificFlavorTestJSON-691988195 tempest-ServersWithSpecificFlavorTestJSON-691988195-project-member] [instance: ebc737a3-e6cd-4f4c-a4b6-fce5924de4c6] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 603.088652] env[63515]: DEBUG oslo_concurrency.lockutils [None req-5e279136-7e72-43a2-85c4-9cd11158a72b tempest-ListImageFiltersTestJSON-1100101045 tempest-ListImageFiltersTestJSON-1100101045-project-member] Releasing lock "refresh_cache-48908cb0-a989-4bad-84a4-25a4bdd3baeb" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 603.088652] env[63515]: DEBUG nova.compute.manager [None req-5e279136-7e72-43a2-85c4-9cd11158a72b tempest-ListImageFiltersTestJSON-1100101045 tempest-ListImageFiltersTestJSON-1100101045-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63515) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 603.088652] env[63515]: DEBUG nova.compute.manager [None req-5e279136-7e72-43a2-85c4-9cd11158a72b tempest-ListImageFiltersTestJSON-1100101045 tempest-ListImageFiltersTestJSON-1100101045-project-member] [instance: 48908cb0-a989-4bad-84a4-25a4bdd3baeb] Deallocating network for instance {{(pid=63515) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 603.088652] env[63515]: DEBUG nova.network.neutron [None req-5e279136-7e72-43a2-85c4-9cd11158a72b tempest-ListImageFiltersTestJSON-1100101045 tempest-ListImageFiltersTestJSON-1100101045-project-member] [instance: 48908cb0-a989-4bad-84a4-25a4bdd3baeb] deallocate_for_instance() {{(pid=63515) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 603.091801] env[63515]: DEBUG oslo_concurrency.lockutils [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Releasing lock "refresh_cache-93b28142-8454-43c0-b0a7-d61aa95c8fc8" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 603.091801] env[63515]: DEBUG nova.compute.manager [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] [instance: 93b28142-8454-43c0-b0a7-d61aa95c8fc8] Updated the network info_cache for instance {{(pid=63515) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9997}} [ 603.091801] env[63515]: DEBUG oslo_service.periodic_task [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=63515) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 603.091801] env[63515]: DEBUG oslo_service.periodic_task [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=63515) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 603.091801] env[63515]: DEBUG oslo_service.periodic_task [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=63515) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 603.091801] env[63515]: DEBUG oslo_service.periodic_task [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=63515) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 603.091957] env[63515]: DEBUG oslo_service.periodic_task [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=63515) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 603.091957] env[63515]: DEBUG oslo_service.periodic_task [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=63515) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 603.091957] env[63515]: DEBUG nova.compute.manager [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=63515) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10545}} [ 603.092042] env[63515]: DEBUG oslo_service.periodic_task [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Running periodic task ComputeManager.update_available_resource {{(pid=63515) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 603.111522] env[63515]: DEBUG nova.network.neutron [None req-5e279136-7e72-43a2-85c4-9cd11158a72b tempest-ListImageFiltersTestJSON-1100101045 tempest-ListImageFiltersTestJSON-1100101045-project-member] [instance: 48908cb0-a989-4bad-84a4-25a4bdd3baeb] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 603.261970] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58a3a5b8-4085-4cbe-b296-e9bc4466cd99 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.271167] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1c87d2f-4c09-40e2-b913-9b67400cb140 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.301662] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57893423-5d51-418a-9e75-5ade08e26794 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.310173] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab5cb60c-3e17-4d2a-bed6-b0858bf1aeb4 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.323945] env[63515]: DEBUG nova.compute.provider_tree [None req-d9ffa9d5-6856-47e5-8926-d090bbea2c09 tempest-DeleteServersAdminTestJSON-944542165 tempest-DeleteServersAdminTestJSON-944542165-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 603.350674] env[63515]: DEBUG nova.network.neutron [None req-24636c84-3869-4a1b-acfb-9c10d33a9826 tempest-ServersWithSpecificFlavorTestJSON-691988195 tempest-ServersWithSpecificFlavorTestJSON-691988195-project-member] [instance: ebc737a3-e6cd-4f4c-a4b6-fce5924de4c6] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 603.455550] env[63515]: DEBUG nova.network.neutron [None req-24636c84-3869-4a1b-acfb-9c10d33a9826 tempest-ServersWithSpecificFlavorTestJSON-691988195 tempest-ServersWithSpecificFlavorTestJSON-691988195-project-member] [instance: ebc737a3-e6cd-4f4c-a4b6-fce5924de4c6] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 603.544640] env[63515]: DEBUG oslo_concurrency.lockutils [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Acquiring lock "48668736-df27-4f2a-94d9-132f5b49701b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 603.546430] env[63515]: DEBUG oslo_concurrency.lockutils [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Lock "48668736-df27-4f2a-94d9-132f5b49701b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 603.574760] env[63515]: DEBUG oslo_concurrency.lockutils [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Acquiring lock "452f3ca2-6141-43b2-a77a-c9ab5754192d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 603.575083] env[63515]: DEBUG oslo_concurrency.lockutils [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Lock "452f3ca2-6141-43b2-a77a-c9ab5754192d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 603.599828] env[63515]: DEBUG oslo_concurrency.lockutils [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 603.602699] env[63515]: DEBUG oslo_concurrency.lockutils [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Acquiring lock "37e6c27e-317b-45d2-bd55-2fd78ccf009f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 603.602938] env[63515]: DEBUG oslo_concurrency.lockutils [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Lock "37e6c27e-317b-45d2-bd55-2fd78ccf009f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 603.613385] env[63515]: DEBUG nova.network.neutron [None req-5e279136-7e72-43a2-85c4-9cd11158a72b tempest-ListImageFiltersTestJSON-1100101045 tempest-ListImageFiltersTestJSON-1100101045-project-member] [instance: 48908cb0-a989-4bad-84a4-25a4bdd3baeb] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 603.714765] env[63515]: DEBUG nova.compute.manager [req-912a925f-f754-4dfd-88cf-69aa82f556c8 req-3df9c8b0-3a07-4ba2-9e2a-f962df00902d service nova] [instance: ebc737a3-e6cd-4f4c-a4b6-fce5924de4c6] Received event network-vif-deleted-168387a2-8a0f-41c4-b727-bad7d761613f {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 603.829457] env[63515]: DEBUG nova.scheduler.client.report [None req-d9ffa9d5-6856-47e5-8926-d090bbea2c09 tempest-DeleteServersAdminTestJSON-944542165 tempest-DeleteServersAdminTestJSON-944542165-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 603.971362] env[63515]: DEBUG oslo_concurrency.lockutils [None req-24636c84-3869-4a1b-acfb-9c10d33a9826 tempest-ServersWithSpecificFlavorTestJSON-691988195 tempest-ServersWithSpecificFlavorTestJSON-691988195-project-member] Releasing lock "refresh_cache-ebc737a3-e6cd-4f4c-a4b6-fce5924de4c6" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 603.971362] env[63515]: DEBUG nova.compute.manager [None req-24636c84-3869-4a1b-acfb-9c10d33a9826 tempest-ServersWithSpecificFlavorTestJSON-691988195 tempest-ServersWithSpecificFlavorTestJSON-691988195-project-member] [instance: ebc737a3-e6cd-4f4c-a4b6-fce5924de4c6] Start destroying the instance on the hypervisor. {{(pid=63515) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 603.971362] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-24636c84-3869-4a1b-acfb-9c10d33a9826 tempest-ServersWithSpecificFlavorTestJSON-691988195 tempest-ServersWithSpecificFlavorTestJSON-691988195-project-member] [instance: ebc737a3-e6cd-4f4c-a4b6-fce5924de4c6] Destroying instance {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 603.971362] env[63515]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d8556837-d4ec-4ac0-9928-8b34851214a9 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.985692] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-205207b3-94ff-46a1-855f-9196bf877d23 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.013223] env[63515]: WARNING nova.virt.vmwareapi.vmops [None req-24636c84-3869-4a1b-acfb-9c10d33a9826 tempest-ServersWithSpecificFlavorTestJSON-691988195 tempest-ServersWithSpecificFlavorTestJSON-691988195-project-member] [instance: ebc737a3-e6cd-4f4c-a4b6-fce5924de4c6] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance ebc737a3-e6cd-4f4c-a4b6-fce5924de4c6 could not be found. [ 604.013538] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-24636c84-3869-4a1b-acfb-9c10d33a9826 tempest-ServersWithSpecificFlavorTestJSON-691988195 tempest-ServersWithSpecificFlavorTestJSON-691988195-project-member] [instance: ebc737a3-e6cd-4f4c-a4b6-fce5924de4c6] Instance destroyed {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 604.013757] env[63515]: INFO nova.compute.manager [None req-24636c84-3869-4a1b-acfb-9c10d33a9826 tempest-ServersWithSpecificFlavorTestJSON-691988195 tempest-ServersWithSpecificFlavorTestJSON-691988195-project-member] [instance: ebc737a3-e6cd-4f4c-a4b6-fce5924de4c6] Took 0.04 seconds to destroy the instance on the hypervisor. [ 604.014055] env[63515]: DEBUG oslo.service.loopingcall [None req-24636c84-3869-4a1b-acfb-9c10d33a9826 tempest-ServersWithSpecificFlavorTestJSON-691988195 tempest-ServersWithSpecificFlavorTestJSON-691988195-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 604.014503] env[63515]: DEBUG nova.compute.manager [-] [instance: ebc737a3-e6cd-4f4c-a4b6-fce5924de4c6] Deallocating network for instance {{(pid=63515) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 604.014503] env[63515]: DEBUG nova.network.neutron [-] [instance: ebc737a3-e6cd-4f4c-a4b6-fce5924de4c6] deallocate_for_instance() {{(pid=63515) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 604.038989] env[63515]: DEBUG nova.network.neutron [-] [instance: ebc737a3-e6cd-4f4c-a4b6-fce5924de4c6] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 604.115899] env[63515]: INFO nova.compute.manager [None req-5e279136-7e72-43a2-85c4-9cd11158a72b tempest-ListImageFiltersTestJSON-1100101045 tempest-ListImageFiltersTestJSON-1100101045-project-member] [instance: 48908cb0-a989-4bad-84a4-25a4bdd3baeb] Took 1.03 seconds to deallocate network for instance. [ 604.334593] env[63515]: DEBUG oslo_concurrency.lockutils [None req-d9ffa9d5-6856-47e5-8926-d090bbea2c09 tempest-DeleteServersAdminTestJSON-944542165 tempest-DeleteServersAdminTestJSON-944542165-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.545s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 604.334718] env[63515]: DEBUG nova.compute.manager [None req-d9ffa9d5-6856-47e5-8926-d090bbea2c09 tempest-DeleteServersAdminTestJSON-944542165 tempest-DeleteServersAdminTestJSON-944542165-project-member] [instance: ae5c6e69-a885-48e3-9c9e-017f33430edd] Start building networks asynchronously for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 604.337766] env[63515]: DEBUG oslo_concurrency.lockutils [None req-9f5c50b0-6bbe-4982-8818-c7255a8c23ab tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.489s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 604.339130] env[63515]: INFO nova.compute.claims [None req-9f5c50b0-6bbe-4982-8818-c7255a8c23ab tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: b53754b1-ec31-4dc3-a839-771f71affe1e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 604.542776] env[63515]: DEBUG nova.network.neutron [-] [instance: ebc737a3-e6cd-4f4c-a4b6-fce5924de4c6] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 604.839191] env[63515]: DEBUG nova.compute.utils [None req-d9ffa9d5-6856-47e5-8926-d090bbea2c09 tempest-DeleteServersAdminTestJSON-944542165 tempest-DeleteServersAdminTestJSON-944542165-project-member] Using /dev/sd instead of None {{(pid=63515) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 604.840681] env[63515]: DEBUG nova.compute.manager [None req-d9ffa9d5-6856-47e5-8926-d090bbea2c09 tempest-DeleteServersAdminTestJSON-944542165 tempest-DeleteServersAdminTestJSON-944542165-project-member] [instance: ae5c6e69-a885-48e3-9c9e-017f33430edd] Allocating IP information in the background. {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 604.841370] env[63515]: DEBUG nova.network.neutron [None req-d9ffa9d5-6856-47e5-8926-d090bbea2c09 tempest-DeleteServersAdminTestJSON-944542165 tempest-DeleteServersAdminTestJSON-944542165-project-member] [instance: ae5c6e69-a885-48e3-9c9e-017f33430edd] allocate_for_instance() {{(pid=63515) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 604.907566] env[63515]: DEBUG nova.policy [None req-d9ffa9d5-6856-47e5-8926-d090bbea2c09 tempest-DeleteServersAdminTestJSON-944542165 tempest-DeleteServersAdminTestJSON-944542165-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '664c0b774cd64a8d9d798e724c206fe4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '48d812be60fd41179b1d5d43c39c335b', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63515) authorize /opt/stack/nova/nova/policy.py:201}} [ 605.047854] env[63515]: INFO nova.compute.manager [-] [instance: ebc737a3-e6cd-4f4c-a4b6-fce5924de4c6] Took 1.03 seconds to deallocate network for instance. [ 605.051983] env[63515]: DEBUG nova.compute.claims [None req-24636c84-3869-4a1b-acfb-9c10d33a9826 tempest-ServersWithSpecificFlavorTestJSON-691988195 tempest-ServersWithSpecificFlavorTestJSON-691988195-project-member] [instance: ebc737a3-e6cd-4f4c-a4b6-fce5924de4c6] Aborting claim: {{(pid=63515) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 605.053136] env[63515]: DEBUG oslo_concurrency.lockutils [None req-24636c84-3869-4a1b-acfb-9c10d33a9826 tempest-ServersWithSpecificFlavorTestJSON-691988195 tempest-ServersWithSpecificFlavorTestJSON-691988195-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 605.151522] env[63515]: INFO nova.scheduler.client.report [None req-5e279136-7e72-43a2-85c4-9cd11158a72b tempest-ListImageFiltersTestJSON-1100101045 tempest-ListImageFiltersTestJSON-1100101045-project-member] Deleted allocations for instance 48908cb0-a989-4bad-84a4-25a4bdd3baeb [ 605.272027] env[63515]: DEBUG nova.network.neutron [None req-d9ffa9d5-6856-47e5-8926-d090bbea2c09 tempest-DeleteServersAdminTestJSON-944542165 tempest-DeleteServersAdminTestJSON-944542165-project-member] [instance: ae5c6e69-a885-48e3-9c9e-017f33430edd] Successfully created port: c69d462c-5b89-4c1b-a518-3d862cdd355e {{(pid=63515) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 605.344387] env[63515]: DEBUG nova.compute.manager [None req-d9ffa9d5-6856-47e5-8926-d090bbea2c09 tempest-DeleteServersAdminTestJSON-944542165 tempest-DeleteServersAdminTestJSON-944542165-project-member] [instance: ae5c6e69-a885-48e3-9c9e-017f33430edd] Start building block device mappings for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 605.662858] env[63515]: DEBUG oslo_concurrency.lockutils [None req-5e279136-7e72-43a2-85c4-9cd11158a72b tempest-ListImageFiltersTestJSON-1100101045 tempest-ListImageFiltersTestJSON-1100101045-project-member] Lock "48908cb0-a989-4bad-84a4-25a4bdd3baeb" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 61.817s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 605.826663] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fadd54c9-3822-4072-9691-fec7712113a5 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.834279] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf8ebf41-7da0-4d41-b2b0-2c8591d18c1b {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.869807] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2255acb0-25e3-439c-8c02-bef953fb9c62 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.875130] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd15bc70-6308-4ea5-81dd-0565cdaa0134 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.889555] env[63515]: DEBUG nova.compute.provider_tree [None req-9f5c50b0-6bbe-4982-8818-c7255a8c23ab tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 606.166174] env[63515]: DEBUG nova.compute.manager [None req-292f2bc3-2911-4498-95fb-4e4a03681601 tempest-FloatingIPsAssociationTestJSON-629096942 tempest-FloatingIPsAssociationTestJSON-629096942-project-member] [instance: 2728c9f9-0d03-4313-ba2a-3dc22aff8538] Starting instance... {{(pid=63515) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 606.204281] env[63515]: DEBUG nova.compute.manager [req-074ee56f-49fd-4c9a-9c96-d347e32c76b7 req-0169a08a-5a22-436c-b0d1-9ed845f4ce65 service nova] [instance: ae5c6e69-a885-48e3-9c9e-017f33430edd] Received event network-changed-c69d462c-5b89-4c1b-a518-3d862cdd355e {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 606.204547] env[63515]: DEBUG nova.compute.manager [req-074ee56f-49fd-4c9a-9c96-d347e32c76b7 req-0169a08a-5a22-436c-b0d1-9ed845f4ce65 service nova] [instance: ae5c6e69-a885-48e3-9c9e-017f33430edd] Refreshing instance network info cache due to event network-changed-c69d462c-5b89-4c1b-a518-3d862cdd355e. {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 606.204723] env[63515]: DEBUG oslo_concurrency.lockutils [req-074ee56f-49fd-4c9a-9c96-d347e32c76b7 req-0169a08a-5a22-436c-b0d1-9ed845f4ce65 service nova] Acquiring lock "refresh_cache-ae5c6e69-a885-48e3-9c9e-017f33430edd" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 606.204832] env[63515]: DEBUG oslo_concurrency.lockutils [req-074ee56f-49fd-4c9a-9c96-d347e32c76b7 req-0169a08a-5a22-436c-b0d1-9ed845f4ce65 service nova] Acquired lock "refresh_cache-ae5c6e69-a885-48e3-9c9e-017f33430edd" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 606.204990] env[63515]: DEBUG nova.network.neutron [req-074ee56f-49fd-4c9a-9c96-d347e32c76b7 req-0169a08a-5a22-436c-b0d1-9ed845f4ce65 service nova] [instance: ae5c6e69-a885-48e3-9c9e-017f33430edd] Refreshing network info cache for port c69d462c-5b89-4c1b-a518-3d862cdd355e {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 606.378020] env[63515]: DEBUG nova.compute.manager [None req-d9ffa9d5-6856-47e5-8926-d090bbea2c09 tempest-DeleteServersAdminTestJSON-944542165 tempest-DeleteServersAdminTestJSON-944542165-project-member] [instance: ae5c6e69-a885-48e3-9c9e-017f33430edd] Start spawning the instance on the hypervisor. {{(pid=63515) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 606.392240] env[63515]: DEBUG nova.scheduler.client.report [None req-9f5c50b0-6bbe-4982-8818-c7255a8c23ab tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 606.406294] env[63515]: DEBUG nova.virt.hardware [None req-d9ffa9d5-6856-47e5-8926-d090bbea2c09 tempest-DeleteServersAdminTestJSON-944542165 tempest-DeleteServersAdminTestJSON-944542165-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T02:52:14Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T02:51:56Z,direct_url=,disk_format='vmdk',id=8a120570-cb06-4099-b262-554ca0ad15c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b569255cc43e42e7ae3f2b4ad37c6ef4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T02:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 606.406544] env[63515]: DEBUG nova.virt.hardware [None req-d9ffa9d5-6856-47e5-8926-d090bbea2c09 tempest-DeleteServersAdminTestJSON-944542165 tempest-DeleteServersAdminTestJSON-944542165-project-member] Flavor limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 606.406732] env[63515]: DEBUG nova.virt.hardware [None req-d9ffa9d5-6856-47e5-8926-d090bbea2c09 tempest-DeleteServersAdminTestJSON-944542165 tempest-DeleteServersAdminTestJSON-944542165-project-member] Image limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 606.407078] env[63515]: DEBUG nova.virt.hardware [None req-d9ffa9d5-6856-47e5-8926-d090bbea2c09 tempest-DeleteServersAdminTestJSON-944542165 tempest-DeleteServersAdminTestJSON-944542165-project-member] Flavor pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 606.407262] env[63515]: DEBUG nova.virt.hardware [None req-d9ffa9d5-6856-47e5-8926-d090bbea2c09 tempest-DeleteServersAdminTestJSON-944542165 tempest-DeleteServersAdminTestJSON-944542165-project-member] Image pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 606.407420] env[63515]: DEBUG nova.virt.hardware [None req-d9ffa9d5-6856-47e5-8926-d090bbea2c09 tempest-DeleteServersAdminTestJSON-944542165 tempest-DeleteServersAdminTestJSON-944542165-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 606.407639] env[63515]: DEBUG nova.virt.hardware [None req-d9ffa9d5-6856-47e5-8926-d090bbea2c09 tempest-DeleteServersAdminTestJSON-944542165 tempest-DeleteServersAdminTestJSON-944542165-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 606.407800] env[63515]: DEBUG nova.virt.hardware [None req-d9ffa9d5-6856-47e5-8926-d090bbea2c09 tempest-DeleteServersAdminTestJSON-944542165 tempest-DeleteServersAdminTestJSON-944542165-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 606.408753] env[63515]: DEBUG nova.virt.hardware [None req-d9ffa9d5-6856-47e5-8926-d090bbea2c09 tempest-DeleteServersAdminTestJSON-944542165 tempest-DeleteServersAdminTestJSON-944542165-project-member] Got 1 possible topologies {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 606.408753] env[63515]: DEBUG nova.virt.hardware [None req-d9ffa9d5-6856-47e5-8926-d090bbea2c09 tempest-DeleteServersAdminTestJSON-944542165 tempest-DeleteServersAdminTestJSON-944542165-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 606.408753] env[63515]: DEBUG nova.virt.hardware [None req-d9ffa9d5-6856-47e5-8926-d090bbea2c09 tempest-DeleteServersAdminTestJSON-944542165 tempest-DeleteServersAdminTestJSON-944542165-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 606.411397] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b1914f0-e8ee-4234-afc0-437c0e73220b {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.417440] env[63515]: ERROR nova.compute.manager [None req-d9ffa9d5-6856-47e5-8926-d090bbea2c09 tempest-DeleteServersAdminTestJSON-944542165 tempest-DeleteServersAdminTestJSON-944542165-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port c69d462c-5b89-4c1b-a518-3d862cdd355e, please check neutron logs for more information. [ 606.417440] env[63515]: ERROR nova.compute.manager Traceback (most recent call last): [ 606.417440] env[63515]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 606.417440] env[63515]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 606.417440] env[63515]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 606.417440] env[63515]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 606.417440] env[63515]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 606.417440] env[63515]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 606.417440] env[63515]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 606.417440] env[63515]: ERROR nova.compute.manager self.force_reraise() [ 606.417440] env[63515]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 606.417440] env[63515]: ERROR nova.compute.manager raise self.value [ 606.417440] env[63515]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 606.417440] env[63515]: ERROR nova.compute.manager updated_port = self._update_port( [ 606.417440] env[63515]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 606.417440] env[63515]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 606.417851] env[63515]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 606.417851] env[63515]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 606.417851] env[63515]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port c69d462c-5b89-4c1b-a518-3d862cdd355e, please check neutron logs for more information. [ 606.417851] env[63515]: ERROR nova.compute.manager [ 606.417851] env[63515]: Traceback (most recent call last): [ 606.417851] env[63515]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 606.417851] env[63515]: listener.cb(fileno) [ 606.417851] env[63515]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 606.417851] env[63515]: result = function(*args, **kwargs) [ 606.417851] env[63515]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 606.417851] env[63515]: return func(*args, **kwargs) [ 606.417851] env[63515]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 606.417851] env[63515]: raise e [ 606.417851] env[63515]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 606.417851] env[63515]: nwinfo = self.network_api.allocate_for_instance( [ 606.417851] env[63515]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 606.417851] env[63515]: created_port_ids = self._update_ports_for_instance( [ 606.417851] env[63515]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 606.417851] env[63515]: with excutils.save_and_reraise_exception(): [ 606.417851] env[63515]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 606.417851] env[63515]: self.force_reraise() [ 606.417851] env[63515]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 606.417851] env[63515]: raise self.value [ 606.417851] env[63515]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 606.417851] env[63515]: updated_port = self._update_port( [ 606.417851] env[63515]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 606.417851] env[63515]: _ensure_no_port_binding_failure(port) [ 606.417851] env[63515]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 606.417851] env[63515]: raise exception.PortBindingFailed(port_id=port['id']) [ 606.418521] env[63515]: nova.exception.PortBindingFailed: Binding failed for port c69d462c-5b89-4c1b-a518-3d862cdd355e, please check neutron logs for more information. [ 606.418521] env[63515]: Removing descriptor: 17 [ 606.418837] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-882838dd-11fe-42eb-b49e-e2399531d914 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.433628] env[63515]: ERROR nova.compute.manager [None req-d9ffa9d5-6856-47e5-8926-d090bbea2c09 tempest-DeleteServersAdminTestJSON-944542165 tempest-DeleteServersAdminTestJSON-944542165-project-member] [instance: ae5c6e69-a885-48e3-9c9e-017f33430edd] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port c69d462c-5b89-4c1b-a518-3d862cdd355e, please check neutron logs for more information. [ 606.433628] env[63515]: ERROR nova.compute.manager [instance: ae5c6e69-a885-48e3-9c9e-017f33430edd] Traceback (most recent call last): [ 606.433628] env[63515]: ERROR nova.compute.manager [instance: ae5c6e69-a885-48e3-9c9e-017f33430edd] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 606.433628] env[63515]: ERROR nova.compute.manager [instance: ae5c6e69-a885-48e3-9c9e-017f33430edd] yield resources [ 606.433628] env[63515]: ERROR nova.compute.manager [instance: ae5c6e69-a885-48e3-9c9e-017f33430edd] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 606.433628] env[63515]: ERROR nova.compute.manager [instance: ae5c6e69-a885-48e3-9c9e-017f33430edd] self.driver.spawn(context, instance, image_meta, [ 606.433628] env[63515]: ERROR nova.compute.manager [instance: ae5c6e69-a885-48e3-9c9e-017f33430edd] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 606.433628] env[63515]: ERROR nova.compute.manager [instance: ae5c6e69-a885-48e3-9c9e-017f33430edd] self._vmops.spawn(context, instance, image_meta, injected_files, [ 606.433628] env[63515]: ERROR nova.compute.manager [instance: ae5c6e69-a885-48e3-9c9e-017f33430edd] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 606.433628] env[63515]: ERROR nova.compute.manager [instance: ae5c6e69-a885-48e3-9c9e-017f33430edd] vm_ref = self.build_virtual_machine(instance, [ 606.433628] env[63515]: ERROR nova.compute.manager [instance: ae5c6e69-a885-48e3-9c9e-017f33430edd] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 606.434027] env[63515]: ERROR nova.compute.manager [instance: ae5c6e69-a885-48e3-9c9e-017f33430edd] vif_infos = vmwarevif.get_vif_info(self._session, [ 606.434027] env[63515]: ERROR nova.compute.manager [instance: ae5c6e69-a885-48e3-9c9e-017f33430edd] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 606.434027] env[63515]: ERROR nova.compute.manager [instance: ae5c6e69-a885-48e3-9c9e-017f33430edd] for vif in network_info: [ 606.434027] env[63515]: ERROR nova.compute.manager [instance: ae5c6e69-a885-48e3-9c9e-017f33430edd] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 606.434027] env[63515]: ERROR nova.compute.manager [instance: ae5c6e69-a885-48e3-9c9e-017f33430edd] return self._sync_wrapper(fn, *args, **kwargs) [ 606.434027] env[63515]: ERROR nova.compute.manager [instance: ae5c6e69-a885-48e3-9c9e-017f33430edd] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 606.434027] env[63515]: ERROR nova.compute.manager [instance: ae5c6e69-a885-48e3-9c9e-017f33430edd] self.wait() [ 606.434027] env[63515]: ERROR nova.compute.manager [instance: ae5c6e69-a885-48e3-9c9e-017f33430edd] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 606.434027] env[63515]: ERROR nova.compute.manager [instance: ae5c6e69-a885-48e3-9c9e-017f33430edd] self[:] = self._gt.wait() [ 606.434027] env[63515]: ERROR nova.compute.manager [instance: ae5c6e69-a885-48e3-9c9e-017f33430edd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 606.434027] env[63515]: ERROR nova.compute.manager [instance: ae5c6e69-a885-48e3-9c9e-017f33430edd] return self._exit_event.wait() [ 606.434027] env[63515]: ERROR nova.compute.manager [instance: ae5c6e69-a885-48e3-9c9e-017f33430edd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 606.434027] env[63515]: ERROR nova.compute.manager [instance: ae5c6e69-a885-48e3-9c9e-017f33430edd] current.throw(*self._exc) [ 606.434362] env[63515]: ERROR nova.compute.manager [instance: ae5c6e69-a885-48e3-9c9e-017f33430edd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 606.434362] env[63515]: ERROR nova.compute.manager [instance: ae5c6e69-a885-48e3-9c9e-017f33430edd] result = function(*args, **kwargs) [ 606.434362] env[63515]: ERROR nova.compute.manager [instance: ae5c6e69-a885-48e3-9c9e-017f33430edd] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 606.434362] env[63515]: ERROR nova.compute.manager [instance: ae5c6e69-a885-48e3-9c9e-017f33430edd] return func(*args, **kwargs) [ 606.434362] env[63515]: ERROR nova.compute.manager [instance: ae5c6e69-a885-48e3-9c9e-017f33430edd] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 606.434362] env[63515]: ERROR nova.compute.manager [instance: ae5c6e69-a885-48e3-9c9e-017f33430edd] raise e [ 606.434362] env[63515]: ERROR nova.compute.manager [instance: ae5c6e69-a885-48e3-9c9e-017f33430edd] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 606.434362] env[63515]: ERROR nova.compute.manager [instance: ae5c6e69-a885-48e3-9c9e-017f33430edd] nwinfo = self.network_api.allocate_for_instance( [ 606.434362] env[63515]: ERROR nova.compute.manager [instance: ae5c6e69-a885-48e3-9c9e-017f33430edd] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 606.434362] env[63515]: ERROR nova.compute.manager [instance: ae5c6e69-a885-48e3-9c9e-017f33430edd] created_port_ids = self._update_ports_for_instance( [ 606.434362] env[63515]: ERROR nova.compute.manager [instance: ae5c6e69-a885-48e3-9c9e-017f33430edd] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 606.434362] env[63515]: ERROR nova.compute.manager [instance: ae5c6e69-a885-48e3-9c9e-017f33430edd] with excutils.save_and_reraise_exception(): [ 606.434362] env[63515]: ERROR nova.compute.manager [instance: ae5c6e69-a885-48e3-9c9e-017f33430edd] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 606.434745] env[63515]: ERROR nova.compute.manager [instance: ae5c6e69-a885-48e3-9c9e-017f33430edd] self.force_reraise() [ 606.434745] env[63515]: ERROR nova.compute.manager [instance: ae5c6e69-a885-48e3-9c9e-017f33430edd] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 606.434745] env[63515]: ERROR nova.compute.manager [instance: ae5c6e69-a885-48e3-9c9e-017f33430edd] raise self.value [ 606.434745] env[63515]: ERROR nova.compute.manager [instance: ae5c6e69-a885-48e3-9c9e-017f33430edd] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 606.434745] env[63515]: ERROR nova.compute.manager [instance: ae5c6e69-a885-48e3-9c9e-017f33430edd] updated_port = self._update_port( [ 606.434745] env[63515]: ERROR nova.compute.manager [instance: ae5c6e69-a885-48e3-9c9e-017f33430edd] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 606.434745] env[63515]: ERROR nova.compute.manager [instance: ae5c6e69-a885-48e3-9c9e-017f33430edd] _ensure_no_port_binding_failure(port) [ 606.434745] env[63515]: ERROR nova.compute.manager [instance: ae5c6e69-a885-48e3-9c9e-017f33430edd] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 606.434745] env[63515]: ERROR nova.compute.manager [instance: ae5c6e69-a885-48e3-9c9e-017f33430edd] raise exception.PortBindingFailed(port_id=port['id']) [ 606.434745] env[63515]: ERROR nova.compute.manager [instance: ae5c6e69-a885-48e3-9c9e-017f33430edd] nova.exception.PortBindingFailed: Binding failed for port c69d462c-5b89-4c1b-a518-3d862cdd355e, please check neutron logs for more information. [ 606.434745] env[63515]: ERROR nova.compute.manager [instance: ae5c6e69-a885-48e3-9c9e-017f33430edd] [ 606.434745] env[63515]: INFO nova.compute.manager [None req-d9ffa9d5-6856-47e5-8926-d090bbea2c09 tempest-DeleteServersAdminTestJSON-944542165 tempest-DeleteServersAdminTestJSON-944542165-project-member] [instance: ae5c6e69-a885-48e3-9c9e-017f33430edd] Terminating instance [ 606.436932] env[63515]: DEBUG oslo_concurrency.lockutils [None req-d9ffa9d5-6856-47e5-8926-d090bbea2c09 tempest-DeleteServersAdminTestJSON-944542165 tempest-DeleteServersAdminTestJSON-944542165-project-member] Acquiring lock "refresh_cache-ae5c6e69-a885-48e3-9c9e-017f33430edd" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 606.692150] env[63515]: DEBUG oslo_concurrency.lockutils [None req-292f2bc3-2911-4498-95fb-4e4a03681601 tempest-FloatingIPsAssociationTestJSON-629096942 tempest-FloatingIPsAssociationTestJSON-629096942-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 606.729405] env[63515]: DEBUG nova.network.neutron [req-074ee56f-49fd-4c9a-9c96-d347e32c76b7 req-0169a08a-5a22-436c-b0d1-9ed845f4ce65 service nova] [instance: ae5c6e69-a885-48e3-9c9e-017f33430edd] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 606.785960] env[63515]: DEBUG nova.network.neutron [req-074ee56f-49fd-4c9a-9c96-d347e32c76b7 req-0169a08a-5a22-436c-b0d1-9ed845f4ce65 service nova] [instance: ae5c6e69-a885-48e3-9c9e-017f33430edd] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 606.900959] env[63515]: DEBUG oslo_concurrency.lockutils [None req-9f5c50b0-6bbe-4982-8818-c7255a8c23ab tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.562s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 606.900959] env[63515]: DEBUG nova.compute.manager [None req-9f5c50b0-6bbe-4982-8818-c7255a8c23ab tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: b53754b1-ec31-4dc3-a839-771f71affe1e] Start building networks asynchronously for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 606.904021] env[63515]: DEBUG oslo_concurrency.lockutils [None req-ae27ad25-8527-4c20-9aca-0f14006c890e tempest-ServersAdmin275Test-220579119 tempest-ServersAdmin275Test-220579119-project-admin] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 17.215s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 606.904021] env[63515]: DEBUG nova.objects.instance [None req-ae27ad25-8527-4c20-9aca-0f14006c890e tempest-ServersAdmin275Test-220579119 tempest-ServersAdmin275Test-220579119-project-admin] [instance: 93b28142-8454-43c0-b0a7-d61aa95c8fc8] Trying to apply a migration context that does not seem to be set for this instance {{(pid=63515) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 607.287789] env[63515]: DEBUG oslo_concurrency.lockutils [req-074ee56f-49fd-4c9a-9c96-d347e32c76b7 req-0169a08a-5a22-436c-b0d1-9ed845f4ce65 service nova] Releasing lock "refresh_cache-ae5c6e69-a885-48e3-9c9e-017f33430edd" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 607.288225] env[63515]: DEBUG oslo_concurrency.lockutils [None req-d9ffa9d5-6856-47e5-8926-d090bbea2c09 tempest-DeleteServersAdminTestJSON-944542165 tempest-DeleteServersAdminTestJSON-944542165-project-member] Acquired lock "refresh_cache-ae5c6e69-a885-48e3-9c9e-017f33430edd" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 607.288417] env[63515]: DEBUG nova.network.neutron [None req-d9ffa9d5-6856-47e5-8926-d090bbea2c09 tempest-DeleteServersAdminTestJSON-944542165 tempest-DeleteServersAdminTestJSON-944542165-project-member] [instance: ae5c6e69-a885-48e3-9c9e-017f33430edd] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 607.408919] env[63515]: DEBUG nova.compute.utils [None req-9f5c50b0-6bbe-4982-8818-c7255a8c23ab tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Using /dev/sd instead of None {{(pid=63515) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 607.413433] env[63515]: DEBUG nova.compute.manager [None req-9f5c50b0-6bbe-4982-8818-c7255a8c23ab tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: b53754b1-ec31-4dc3-a839-771f71affe1e] Allocating IP information in the background. {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 607.413433] env[63515]: DEBUG nova.network.neutron [None req-9f5c50b0-6bbe-4982-8818-c7255a8c23ab tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: b53754b1-ec31-4dc3-a839-771f71affe1e] allocate_for_instance() {{(pid=63515) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 607.468912] env[63515]: DEBUG nova.policy [None req-9f5c50b0-6bbe-4982-8818-c7255a8c23ab tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b5de3e36ae88482eb795894592e76c7c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f9c98e9f6020475490aaa7e76d907ab7', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63515) authorize /opt/stack/nova/nova/policy.py:201}} [ 607.770149] env[63515]: DEBUG nova.network.neutron [None req-9f5c50b0-6bbe-4982-8818-c7255a8c23ab tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: b53754b1-ec31-4dc3-a839-771f71affe1e] Successfully created port: 46784f84-e839-4fa3-88ea-2005d2779b07 {{(pid=63515) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 607.808863] env[63515]: DEBUG nova.network.neutron [None req-d9ffa9d5-6856-47e5-8926-d090bbea2c09 tempest-DeleteServersAdminTestJSON-944542165 tempest-DeleteServersAdminTestJSON-944542165-project-member] [instance: ae5c6e69-a885-48e3-9c9e-017f33430edd] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 607.892531] env[63515]: DEBUG nova.network.neutron [None req-d9ffa9d5-6856-47e5-8926-d090bbea2c09 tempest-DeleteServersAdminTestJSON-944542165 tempest-DeleteServersAdminTestJSON-944542165-project-member] [instance: ae5c6e69-a885-48e3-9c9e-017f33430edd] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 607.913974] env[63515]: DEBUG nova.compute.manager [None req-9f5c50b0-6bbe-4982-8818-c7255a8c23ab tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: b53754b1-ec31-4dc3-a839-771f71affe1e] Start building block device mappings for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 607.918297] env[63515]: DEBUG oslo_concurrency.lockutils [None req-ae27ad25-8527-4c20-9aca-0f14006c890e tempest-ServersAdmin275Test-220579119 tempest-ServersAdmin275Test-220579119-project-admin] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.015s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 607.921500] env[63515]: DEBUG oslo_concurrency.lockutils [None req-12d73cc4-118d-457e-b393-cc4362d15b0d tempest-ListImageFiltersTestJSON-1100101045 tempest-ListImageFiltersTestJSON-1100101045-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.515s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 608.233568] env[63515]: DEBUG nova.compute.manager [req-c77b235b-10af-48b6-a51b-d7b3a01246b4 req-ee4ccb0a-9b96-41cd-9e26-bfdeaa6747f5 service nova] [instance: ae5c6e69-a885-48e3-9c9e-017f33430edd] Received event network-vif-deleted-c69d462c-5b89-4c1b-a518-3d862cdd355e {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 608.395972] env[63515]: DEBUG oslo_concurrency.lockutils [None req-d9ffa9d5-6856-47e5-8926-d090bbea2c09 tempest-DeleteServersAdminTestJSON-944542165 tempest-DeleteServersAdminTestJSON-944542165-project-member] Releasing lock "refresh_cache-ae5c6e69-a885-48e3-9c9e-017f33430edd" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 608.396642] env[63515]: DEBUG nova.compute.manager [None req-d9ffa9d5-6856-47e5-8926-d090bbea2c09 tempest-DeleteServersAdminTestJSON-944542165 tempest-DeleteServersAdminTestJSON-944542165-project-member] [instance: ae5c6e69-a885-48e3-9c9e-017f33430edd] Start destroying the instance on the hypervisor. {{(pid=63515) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 608.396806] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-d9ffa9d5-6856-47e5-8926-d090bbea2c09 tempest-DeleteServersAdminTestJSON-944542165 tempest-DeleteServersAdminTestJSON-944542165-project-member] [instance: ae5c6e69-a885-48e3-9c9e-017f33430edd] Destroying instance {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 608.397081] env[63515]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-0833517b-326c-43d1-9dde-b07f8e040365 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.407119] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e371401-0542-4a57-8158-3391514cfb4e {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.447407] env[63515]: WARNING nova.virt.vmwareapi.vmops [None req-d9ffa9d5-6856-47e5-8926-d090bbea2c09 tempest-DeleteServersAdminTestJSON-944542165 tempest-DeleteServersAdminTestJSON-944542165-project-member] [instance: ae5c6e69-a885-48e3-9c9e-017f33430edd] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance ae5c6e69-a885-48e3-9c9e-017f33430edd could not be found. [ 608.447407] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-d9ffa9d5-6856-47e5-8926-d090bbea2c09 tempest-DeleteServersAdminTestJSON-944542165 tempest-DeleteServersAdminTestJSON-944542165-project-member] [instance: ae5c6e69-a885-48e3-9c9e-017f33430edd] Instance destroyed {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 608.447407] env[63515]: INFO nova.compute.manager [None req-d9ffa9d5-6856-47e5-8926-d090bbea2c09 tempest-DeleteServersAdminTestJSON-944542165 tempest-DeleteServersAdminTestJSON-944542165-project-member] [instance: ae5c6e69-a885-48e3-9c9e-017f33430edd] Took 0.05 seconds to destroy the instance on the hypervisor. [ 608.447407] env[63515]: DEBUG oslo.service.loopingcall [None req-d9ffa9d5-6856-47e5-8926-d090bbea2c09 tempest-DeleteServersAdminTestJSON-944542165 tempest-DeleteServersAdminTestJSON-944542165-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 608.447671] env[63515]: DEBUG nova.compute.manager [-] [instance: ae5c6e69-a885-48e3-9c9e-017f33430edd] Deallocating network for instance {{(pid=63515) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 608.447671] env[63515]: DEBUG nova.network.neutron [-] [instance: ae5c6e69-a885-48e3-9c9e-017f33430edd] deallocate_for_instance() {{(pid=63515) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 608.469159] env[63515]: DEBUG nova.network.neutron [-] [instance: ae5c6e69-a885-48e3-9c9e-017f33430edd] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 608.858023] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-219b4420-fd89-4d5f-b11a-8b357bfe4860 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.864548] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1dd10a1f-125a-4975-9838-7b3d77aec871 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.900305] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-506140e7-63e8-4267-8e7b-387514879f47 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.903853] env[63515]: ERROR nova.compute.manager [None req-9f5c50b0-6bbe-4982-8818-c7255a8c23ab tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 46784f84-e839-4fa3-88ea-2005d2779b07, please check neutron logs for more information. [ 608.903853] env[63515]: ERROR nova.compute.manager Traceback (most recent call last): [ 608.903853] env[63515]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 608.903853] env[63515]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 608.903853] env[63515]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 608.903853] env[63515]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 608.903853] env[63515]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 608.903853] env[63515]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 608.903853] env[63515]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 608.903853] env[63515]: ERROR nova.compute.manager self.force_reraise() [ 608.903853] env[63515]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 608.903853] env[63515]: ERROR nova.compute.manager raise self.value [ 608.903853] env[63515]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 608.903853] env[63515]: ERROR nova.compute.manager updated_port = self._update_port( [ 608.903853] env[63515]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 608.903853] env[63515]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 608.904311] env[63515]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 608.904311] env[63515]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 608.904311] env[63515]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 46784f84-e839-4fa3-88ea-2005d2779b07, please check neutron logs for more information. [ 608.904311] env[63515]: ERROR nova.compute.manager [ 608.904311] env[63515]: Traceback (most recent call last): [ 608.904311] env[63515]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 608.904311] env[63515]: listener.cb(fileno) [ 608.904311] env[63515]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 608.904311] env[63515]: result = function(*args, **kwargs) [ 608.904311] env[63515]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 608.904311] env[63515]: return func(*args, **kwargs) [ 608.904311] env[63515]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 608.904311] env[63515]: raise e [ 608.904311] env[63515]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 608.904311] env[63515]: nwinfo = self.network_api.allocate_for_instance( [ 608.904311] env[63515]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 608.904311] env[63515]: created_port_ids = self._update_ports_for_instance( [ 608.904311] env[63515]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 608.904311] env[63515]: with excutils.save_and_reraise_exception(): [ 608.904311] env[63515]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 608.904311] env[63515]: self.force_reraise() [ 608.904311] env[63515]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 608.904311] env[63515]: raise self.value [ 608.904311] env[63515]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 608.904311] env[63515]: updated_port = self._update_port( [ 608.904311] env[63515]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 608.904311] env[63515]: _ensure_no_port_binding_failure(port) [ 608.904311] env[63515]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 608.904311] env[63515]: raise exception.PortBindingFailed(port_id=port['id']) [ 608.905063] env[63515]: nova.exception.PortBindingFailed: Binding failed for port 46784f84-e839-4fa3-88ea-2005d2779b07, please check neutron logs for more information. [ 608.905063] env[63515]: Removing descriptor: 17 [ 608.907556] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58b964ac-af67-4c4e-adfb-c5ab46019d4e {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.921834] env[63515]: DEBUG nova.compute.provider_tree [None req-12d73cc4-118d-457e-b393-cc4362d15b0d tempest-ListImageFiltersTestJSON-1100101045 tempest-ListImageFiltersTestJSON-1100101045-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 608.936298] env[63515]: DEBUG nova.compute.manager [None req-9f5c50b0-6bbe-4982-8818-c7255a8c23ab tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: b53754b1-ec31-4dc3-a839-771f71affe1e] Start spawning the instance on the hypervisor. {{(pid=63515) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 608.961971] env[63515]: DEBUG nova.virt.hardware [None req-9f5c50b0-6bbe-4982-8818-c7255a8c23ab tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T02:52:14Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T02:51:56Z,direct_url=,disk_format='vmdk',id=8a120570-cb06-4099-b262-554ca0ad15c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b569255cc43e42e7ae3f2b4ad37c6ef4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T02:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 608.962247] env[63515]: DEBUG nova.virt.hardware [None req-9f5c50b0-6bbe-4982-8818-c7255a8c23ab tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Flavor limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 608.962401] env[63515]: DEBUG nova.virt.hardware [None req-9f5c50b0-6bbe-4982-8818-c7255a8c23ab tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Image limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 608.962600] env[63515]: DEBUG nova.virt.hardware [None req-9f5c50b0-6bbe-4982-8818-c7255a8c23ab tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Flavor pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 608.962766] env[63515]: DEBUG nova.virt.hardware [None req-9f5c50b0-6bbe-4982-8818-c7255a8c23ab tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Image pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 608.962926] env[63515]: DEBUG nova.virt.hardware [None req-9f5c50b0-6bbe-4982-8818-c7255a8c23ab tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 608.963146] env[63515]: DEBUG nova.virt.hardware [None req-9f5c50b0-6bbe-4982-8818-c7255a8c23ab tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 608.963301] env[63515]: DEBUG nova.virt.hardware [None req-9f5c50b0-6bbe-4982-8818-c7255a8c23ab tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 608.963462] env[63515]: DEBUG nova.virt.hardware [None req-9f5c50b0-6bbe-4982-8818-c7255a8c23ab tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Got 1 possible topologies {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 608.963654] env[63515]: DEBUG nova.virt.hardware [None req-9f5c50b0-6bbe-4982-8818-c7255a8c23ab tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 608.963878] env[63515]: DEBUG nova.virt.hardware [None req-9f5c50b0-6bbe-4982-8818-c7255a8c23ab tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 608.964760] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65ed574d-a714-407c-b89f-519c99b00210 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.972192] env[63515]: DEBUG nova.network.neutron [-] [instance: ae5c6e69-a885-48e3-9c9e-017f33430edd] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 608.974908] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3828cb5-7c69-4f4b-a196-51f363541246 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.990809] env[63515]: ERROR nova.compute.manager [None req-9f5c50b0-6bbe-4982-8818-c7255a8c23ab tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: b53754b1-ec31-4dc3-a839-771f71affe1e] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 46784f84-e839-4fa3-88ea-2005d2779b07, please check neutron logs for more information. [ 608.990809] env[63515]: ERROR nova.compute.manager [instance: b53754b1-ec31-4dc3-a839-771f71affe1e] Traceback (most recent call last): [ 608.990809] env[63515]: ERROR nova.compute.manager [instance: b53754b1-ec31-4dc3-a839-771f71affe1e] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 608.990809] env[63515]: ERROR nova.compute.manager [instance: b53754b1-ec31-4dc3-a839-771f71affe1e] yield resources [ 608.990809] env[63515]: ERROR nova.compute.manager [instance: b53754b1-ec31-4dc3-a839-771f71affe1e] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 608.990809] env[63515]: ERROR nova.compute.manager [instance: b53754b1-ec31-4dc3-a839-771f71affe1e] self.driver.spawn(context, instance, image_meta, [ 608.990809] env[63515]: ERROR nova.compute.manager [instance: b53754b1-ec31-4dc3-a839-771f71affe1e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 608.990809] env[63515]: ERROR nova.compute.manager [instance: b53754b1-ec31-4dc3-a839-771f71affe1e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 608.990809] env[63515]: ERROR nova.compute.manager [instance: b53754b1-ec31-4dc3-a839-771f71affe1e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 608.990809] env[63515]: ERROR nova.compute.manager [instance: b53754b1-ec31-4dc3-a839-771f71affe1e] vm_ref = self.build_virtual_machine(instance, [ 608.990809] env[63515]: ERROR nova.compute.manager [instance: b53754b1-ec31-4dc3-a839-771f71affe1e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 608.991145] env[63515]: ERROR nova.compute.manager [instance: b53754b1-ec31-4dc3-a839-771f71affe1e] vif_infos = vmwarevif.get_vif_info(self._session, [ 608.991145] env[63515]: ERROR nova.compute.manager [instance: b53754b1-ec31-4dc3-a839-771f71affe1e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 608.991145] env[63515]: ERROR nova.compute.manager [instance: b53754b1-ec31-4dc3-a839-771f71affe1e] for vif in network_info: [ 608.991145] env[63515]: ERROR nova.compute.manager [instance: b53754b1-ec31-4dc3-a839-771f71affe1e] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 608.991145] env[63515]: ERROR nova.compute.manager [instance: b53754b1-ec31-4dc3-a839-771f71affe1e] return self._sync_wrapper(fn, *args, **kwargs) [ 608.991145] env[63515]: ERROR nova.compute.manager [instance: b53754b1-ec31-4dc3-a839-771f71affe1e] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 608.991145] env[63515]: ERROR nova.compute.manager [instance: b53754b1-ec31-4dc3-a839-771f71affe1e] self.wait() [ 608.991145] env[63515]: ERROR nova.compute.manager [instance: b53754b1-ec31-4dc3-a839-771f71affe1e] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 608.991145] env[63515]: ERROR nova.compute.manager [instance: b53754b1-ec31-4dc3-a839-771f71affe1e] self[:] = self._gt.wait() [ 608.991145] env[63515]: ERROR nova.compute.manager [instance: b53754b1-ec31-4dc3-a839-771f71affe1e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 608.991145] env[63515]: ERROR nova.compute.manager [instance: b53754b1-ec31-4dc3-a839-771f71affe1e] return self._exit_event.wait() [ 608.991145] env[63515]: ERROR nova.compute.manager [instance: b53754b1-ec31-4dc3-a839-771f71affe1e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 608.991145] env[63515]: ERROR nova.compute.manager [instance: b53754b1-ec31-4dc3-a839-771f71affe1e] current.throw(*self._exc) [ 608.991542] env[63515]: ERROR nova.compute.manager [instance: b53754b1-ec31-4dc3-a839-771f71affe1e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 608.991542] env[63515]: ERROR nova.compute.manager [instance: b53754b1-ec31-4dc3-a839-771f71affe1e] result = function(*args, **kwargs) [ 608.991542] env[63515]: ERROR nova.compute.manager [instance: b53754b1-ec31-4dc3-a839-771f71affe1e] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 608.991542] env[63515]: ERROR nova.compute.manager [instance: b53754b1-ec31-4dc3-a839-771f71affe1e] return func(*args, **kwargs) [ 608.991542] env[63515]: ERROR nova.compute.manager [instance: b53754b1-ec31-4dc3-a839-771f71affe1e] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 608.991542] env[63515]: ERROR nova.compute.manager [instance: b53754b1-ec31-4dc3-a839-771f71affe1e] raise e [ 608.991542] env[63515]: ERROR nova.compute.manager [instance: b53754b1-ec31-4dc3-a839-771f71affe1e] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 608.991542] env[63515]: ERROR nova.compute.manager [instance: b53754b1-ec31-4dc3-a839-771f71affe1e] nwinfo = self.network_api.allocate_for_instance( [ 608.991542] env[63515]: ERROR nova.compute.manager [instance: b53754b1-ec31-4dc3-a839-771f71affe1e] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 608.991542] env[63515]: ERROR nova.compute.manager [instance: b53754b1-ec31-4dc3-a839-771f71affe1e] created_port_ids = self._update_ports_for_instance( [ 608.991542] env[63515]: ERROR nova.compute.manager [instance: b53754b1-ec31-4dc3-a839-771f71affe1e] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 608.991542] env[63515]: ERROR nova.compute.manager [instance: b53754b1-ec31-4dc3-a839-771f71affe1e] with excutils.save_and_reraise_exception(): [ 608.991542] env[63515]: ERROR nova.compute.manager [instance: b53754b1-ec31-4dc3-a839-771f71affe1e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 608.991886] env[63515]: ERROR nova.compute.manager [instance: b53754b1-ec31-4dc3-a839-771f71affe1e] self.force_reraise() [ 608.991886] env[63515]: ERROR nova.compute.manager [instance: b53754b1-ec31-4dc3-a839-771f71affe1e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 608.991886] env[63515]: ERROR nova.compute.manager [instance: b53754b1-ec31-4dc3-a839-771f71affe1e] raise self.value [ 608.991886] env[63515]: ERROR nova.compute.manager [instance: b53754b1-ec31-4dc3-a839-771f71affe1e] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 608.991886] env[63515]: ERROR nova.compute.manager [instance: b53754b1-ec31-4dc3-a839-771f71affe1e] updated_port = self._update_port( [ 608.991886] env[63515]: ERROR nova.compute.manager [instance: b53754b1-ec31-4dc3-a839-771f71affe1e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 608.991886] env[63515]: ERROR nova.compute.manager [instance: b53754b1-ec31-4dc3-a839-771f71affe1e] _ensure_no_port_binding_failure(port) [ 608.991886] env[63515]: ERROR nova.compute.manager [instance: b53754b1-ec31-4dc3-a839-771f71affe1e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 608.991886] env[63515]: ERROR nova.compute.manager [instance: b53754b1-ec31-4dc3-a839-771f71affe1e] raise exception.PortBindingFailed(port_id=port['id']) [ 608.991886] env[63515]: ERROR nova.compute.manager [instance: b53754b1-ec31-4dc3-a839-771f71affe1e] nova.exception.PortBindingFailed: Binding failed for port 46784f84-e839-4fa3-88ea-2005d2779b07, please check neutron logs for more information. [ 608.991886] env[63515]: ERROR nova.compute.manager [instance: b53754b1-ec31-4dc3-a839-771f71affe1e] [ 608.991886] env[63515]: INFO nova.compute.manager [None req-9f5c50b0-6bbe-4982-8818-c7255a8c23ab tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: b53754b1-ec31-4dc3-a839-771f71affe1e] Terminating instance [ 608.993639] env[63515]: DEBUG oslo_concurrency.lockutils [None req-9f5c50b0-6bbe-4982-8818-c7255a8c23ab tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Acquiring lock "refresh_cache-b53754b1-ec31-4dc3-a839-771f71affe1e" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 608.993816] env[63515]: DEBUG oslo_concurrency.lockutils [None req-9f5c50b0-6bbe-4982-8818-c7255a8c23ab tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Acquired lock "refresh_cache-b53754b1-ec31-4dc3-a839-771f71affe1e" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 608.993988] env[63515]: DEBUG nova.network.neutron [None req-9f5c50b0-6bbe-4982-8818-c7255a8c23ab tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: b53754b1-ec31-4dc3-a839-771f71affe1e] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 609.425295] env[63515]: DEBUG nova.scheduler.client.report [None req-12d73cc4-118d-457e-b393-cc4362d15b0d tempest-ListImageFiltersTestJSON-1100101045 tempest-ListImageFiltersTestJSON-1100101045-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 609.480445] env[63515]: INFO nova.compute.manager [-] [instance: ae5c6e69-a885-48e3-9c9e-017f33430edd] Took 1.03 seconds to deallocate network for instance. [ 609.481634] env[63515]: DEBUG nova.compute.claims [None req-d9ffa9d5-6856-47e5-8926-d090bbea2c09 tempest-DeleteServersAdminTestJSON-944542165 tempest-DeleteServersAdminTestJSON-944542165-project-member] [instance: ae5c6e69-a885-48e3-9c9e-017f33430edd] Aborting claim: {{(pid=63515) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 609.481634] env[63515]: DEBUG oslo_concurrency.lockutils [None req-d9ffa9d5-6856-47e5-8926-d090bbea2c09 tempest-DeleteServersAdminTestJSON-944542165 tempest-DeleteServersAdminTestJSON-944542165-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 609.512928] env[63515]: DEBUG nova.network.neutron [None req-9f5c50b0-6bbe-4982-8818-c7255a8c23ab tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: b53754b1-ec31-4dc3-a839-771f71affe1e] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 609.591445] env[63515]: DEBUG nova.network.neutron [None req-9f5c50b0-6bbe-4982-8818-c7255a8c23ab tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: b53754b1-ec31-4dc3-a839-771f71affe1e] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 609.933253] env[63515]: DEBUG oslo_concurrency.lockutils [None req-12d73cc4-118d-457e-b393-cc4362d15b0d tempest-ListImageFiltersTestJSON-1100101045 tempest-ListImageFiltersTestJSON-1100101045-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.013s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 609.933916] env[63515]: ERROR nova.compute.manager [None req-12d73cc4-118d-457e-b393-cc4362d15b0d tempest-ListImageFiltersTestJSON-1100101045 tempest-ListImageFiltersTestJSON-1100101045-project-member] [instance: ab1539ba-33a4-4139-a2f3-1f8c507beede] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port bd569ca1-70ab-4c56-b6af-9ae3816c1d2c, please check neutron logs for more information. [ 609.933916] env[63515]: ERROR nova.compute.manager [instance: ab1539ba-33a4-4139-a2f3-1f8c507beede] Traceback (most recent call last): [ 609.933916] env[63515]: ERROR nova.compute.manager [instance: ab1539ba-33a4-4139-a2f3-1f8c507beede] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 609.933916] env[63515]: ERROR nova.compute.manager [instance: ab1539ba-33a4-4139-a2f3-1f8c507beede] self.driver.spawn(context, instance, image_meta, [ 609.933916] env[63515]: ERROR nova.compute.manager [instance: ab1539ba-33a4-4139-a2f3-1f8c507beede] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 609.933916] env[63515]: ERROR nova.compute.manager [instance: ab1539ba-33a4-4139-a2f3-1f8c507beede] self._vmops.spawn(context, instance, image_meta, injected_files, [ 609.933916] env[63515]: ERROR nova.compute.manager [instance: ab1539ba-33a4-4139-a2f3-1f8c507beede] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 609.933916] env[63515]: ERROR nova.compute.manager [instance: ab1539ba-33a4-4139-a2f3-1f8c507beede] vm_ref = self.build_virtual_machine(instance, [ 609.933916] env[63515]: ERROR nova.compute.manager [instance: ab1539ba-33a4-4139-a2f3-1f8c507beede] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 609.933916] env[63515]: ERROR nova.compute.manager [instance: ab1539ba-33a4-4139-a2f3-1f8c507beede] vif_infos = vmwarevif.get_vif_info(self._session, [ 609.933916] env[63515]: ERROR nova.compute.manager [instance: ab1539ba-33a4-4139-a2f3-1f8c507beede] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 609.934220] env[63515]: ERROR nova.compute.manager [instance: ab1539ba-33a4-4139-a2f3-1f8c507beede] for vif in network_info: [ 609.934220] env[63515]: ERROR nova.compute.manager [instance: ab1539ba-33a4-4139-a2f3-1f8c507beede] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 609.934220] env[63515]: ERROR nova.compute.manager [instance: ab1539ba-33a4-4139-a2f3-1f8c507beede] return self._sync_wrapper(fn, *args, **kwargs) [ 609.934220] env[63515]: ERROR nova.compute.manager [instance: ab1539ba-33a4-4139-a2f3-1f8c507beede] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 609.934220] env[63515]: ERROR nova.compute.manager [instance: ab1539ba-33a4-4139-a2f3-1f8c507beede] self.wait() [ 609.934220] env[63515]: ERROR nova.compute.manager [instance: ab1539ba-33a4-4139-a2f3-1f8c507beede] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 609.934220] env[63515]: ERROR nova.compute.manager [instance: ab1539ba-33a4-4139-a2f3-1f8c507beede] self[:] = self._gt.wait() [ 609.934220] env[63515]: ERROR nova.compute.manager [instance: ab1539ba-33a4-4139-a2f3-1f8c507beede] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 609.934220] env[63515]: ERROR nova.compute.manager [instance: ab1539ba-33a4-4139-a2f3-1f8c507beede] return self._exit_event.wait() [ 609.934220] env[63515]: ERROR nova.compute.manager [instance: ab1539ba-33a4-4139-a2f3-1f8c507beede] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 609.934220] env[63515]: ERROR nova.compute.manager [instance: ab1539ba-33a4-4139-a2f3-1f8c507beede] result = hub.switch() [ 609.934220] env[63515]: ERROR nova.compute.manager [instance: ab1539ba-33a4-4139-a2f3-1f8c507beede] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 609.934220] env[63515]: ERROR nova.compute.manager [instance: ab1539ba-33a4-4139-a2f3-1f8c507beede] return self.greenlet.switch() [ 609.934545] env[63515]: ERROR nova.compute.manager [instance: ab1539ba-33a4-4139-a2f3-1f8c507beede] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 609.934545] env[63515]: ERROR nova.compute.manager [instance: ab1539ba-33a4-4139-a2f3-1f8c507beede] result = function(*args, **kwargs) [ 609.934545] env[63515]: ERROR nova.compute.manager [instance: ab1539ba-33a4-4139-a2f3-1f8c507beede] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 609.934545] env[63515]: ERROR nova.compute.manager [instance: ab1539ba-33a4-4139-a2f3-1f8c507beede] return func(*args, **kwargs) [ 609.934545] env[63515]: ERROR nova.compute.manager [instance: ab1539ba-33a4-4139-a2f3-1f8c507beede] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 609.934545] env[63515]: ERROR nova.compute.manager [instance: ab1539ba-33a4-4139-a2f3-1f8c507beede] raise e [ 609.934545] env[63515]: ERROR nova.compute.manager [instance: ab1539ba-33a4-4139-a2f3-1f8c507beede] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 609.934545] env[63515]: ERROR nova.compute.manager [instance: ab1539ba-33a4-4139-a2f3-1f8c507beede] nwinfo = self.network_api.allocate_for_instance( [ 609.934545] env[63515]: ERROR nova.compute.manager [instance: ab1539ba-33a4-4139-a2f3-1f8c507beede] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 609.934545] env[63515]: ERROR nova.compute.manager [instance: ab1539ba-33a4-4139-a2f3-1f8c507beede] created_port_ids = self._update_ports_for_instance( [ 609.934545] env[63515]: ERROR nova.compute.manager [instance: ab1539ba-33a4-4139-a2f3-1f8c507beede] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 609.934545] env[63515]: ERROR nova.compute.manager [instance: ab1539ba-33a4-4139-a2f3-1f8c507beede] with excutils.save_and_reraise_exception(): [ 609.934545] env[63515]: ERROR nova.compute.manager [instance: ab1539ba-33a4-4139-a2f3-1f8c507beede] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 609.934814] env[63515]: ERROR nova.compute.manager [instance: ab1539ba-33a4-4139-a2f3-1f8c507beede] self.force_reraise() [ 609.934814] env[63515]: ERROR nova.compute.manager [instance: ab1539ba-33a4-4139-a2f3-1f8c507beede] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 609.934814] env[63515]: ERROR nova.compute.manager [instance: ab1539ba-33a4-4139-a2f3-1f8c507beede] raise self.value [ 609.934814] env[63515]: ERROR nova.compute.manager [instance: ab1539ba-33a4-4139-a2f3-1f8c507beede] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 609.934814] env[63515]: ERROR nova.compute.manager [instance: ab1539ba-33a4-4139-a2f3-1f8c507beede] updated_port = self._update_port( [ 609.934814] env[63515]: ERROR nova.compute.manager [instance: ab1539ba-33a4-4139-a2f3-1f8c507beede] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 609.934814] env[63515]: ERROR nova.compute.manager [instance: ab1539ba-33a4-4139-a2f3-1f8c507beede] _ensure_no_port_binding_failure(port) [ 609.934814] env[63515]: ERROR nova.compute.manager [instance: ab1539ba-33a4-4139-a2f3-1f8c507beede] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 609.934814] env[63515]: ERROR nova.compute.manager [instance: ab1539ba-33a4-4139-a2f3-1f8c507beede] raise exception.PortBindingFailed(port_id=port['id']) [ 609.934814] env[63515]: ERROR nova.compute.manager [instance: ab1539ba-33a4-4139-a2f3-1f8c507beede] nova.exception.PortBindingFailed: Binding failed for port bd569ca1-70ab-4c56-b6af-9ae3816c1d2c, please check neutron logs for more information. [ 609.934814] env[63515]: ERROR nova.compute.manager [instance: ab1539ba-33a4-4139-a2f3-1f8c507beede] [ 609.935067] env[63515]: DEBUG nova.compute.utils [None req-12d73cc4-118d-457e-b393-cc4362d15b0d tempest-ListImageFiltersTestJSON-1100101045 tempest-ListImageFiltersTestJSON-1100101045-project-member] [instance: ab1539ba-33a4-4139-a2f3-1f8c507beede] Binding failed for port bd569ca1-70ab-4c56-b6af-9ae3816c1d2c, please check neutron logs for more information. {{(pid=63515) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 609.935918] env[63515]: DEBUG oslo_concurrency.lockutils [None req-886adfc0-007d-4f83-bc30-b755d00d04c7 tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.452s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 609.937542] env[63515]: INFO nova.compute.claims [None req-886adfc0-007d-4f83-bc30-b755d00d04c7 tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] [instance: 8b65c223-03a8-41e0-917e-45dd76b6e57f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 609.941019] env[63515]: DEBUG nova.compute.manager [None req-12d73cc4-118d-457e-b393-cc4362d15b0d tempest-ListImageFiltersTestJSON-1100101045 tempest-ListImageFiltersTestJSON-1100101045-project-member] [instance: ab1539ba-33a4-4139-a2f3-1f8c507beede] Build of instance ab1539ba-33a4-4139-a2f3-1f8c507beede was re-scheduled: Binding failed for port bd569ca1-70ab-4c56-b6af-9ae3816c1d2c, please check neutron logs for more information. {{(pid=63515) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 609.941019] env[63515]: DEBUG nova.compute.manager [None req-12d73cc4-118d-457e-b393-cc4362d15b0d tempest-ListImageFiltersTestJSON-1100101045 tempest-ListImageFiltersTestJSON-1100101045-project-member] [instance: ab1539ba-33a4-4139-a2f3-1f8c507beede] Unplugging VIFs for instance {{(pid=63515) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 609.941019] env[63515]: DEBUG oslo_concurrency.lockutils [None req-12d73cc4-118d-457e-b393-cc4362d15b0d tempest-ListImageFiltersTestJSON-1100101045 tempest-ListImageFiltersTestJSON-1100101045-project-member] Acquiring lock "refresh_cache-ab1539ba-33a4-4139-a2f3-1f8c507beede" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 609.941019] env[63515]: DEBUG oslo_concurrency.lockutils [None req-12d73cc4-118d-457e-b393-cc4362d15b0d tempest-ListImageFiltersTestJSON-1100101045 tempest-ListImageFiltersTestJSON-1100101045-project-member] Acquired lock "refresh_cache-ab1539ba-33a4-4139-a2f3-1f8c507beede" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 609.941220] env[63515]: DEBUG nova.network.neutron [None req-12d73cc4-118d-457e-b393-cc4362d15b0d tempest-ListImageFiltersTestJSON-1100101045 tempest-ListImageFiltersTestJSON-1100101045-project-member] [instance: ab1539ba-33a4-4139-a2f3-1f8c507beede] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 610.095552] env[63515]: DEBUG oslo_concurrency.lockutils [None req-9f5c50b0-6bbe-4982-8818-c7255a8c23ab tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Releasing lock "refresh_cache-b53754b1-ec31-4dc3-a839-771f71affe1e" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 610.096219] env[63515]: DEBUG nova.compute.manager [None req-9f5c50b0-6bbe-4982-8818-c7255a8c23ab tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: b53754b1-ec31-4dc3-a839-771f71affe1e] Start destroying the instance on the hypervisor. {{(pid=63515) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 610.096541] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-9f5c50b0-6bbe-4982-8818-c7255a8c23ab tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: b53754b1-ec31-4dc3-a839-771f71affe1e] Destroying instance {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 610.099016] env[63515]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9afd9d4e-47e8-419d-9221-1ef0ca655217 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.105957] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12182c6c-93e2-4c78-871f-ae50d4a81e1d {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.128186] env[63515]: WARNING nova.virt.vmwareapi.vmops [None req-9f5c50b0-6bbe-4982-8818-c7255a8c23ab tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: b53754b1-ec31-4dc3-a839-771f71affe1e] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance b53754b1-ec31-4dc3-a839-771f71affe1e could not be found. [ 610.130817] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-9f5c50b0-6bbe-4982-8818-c7255a8c23ab tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: b53754b1-ec31-4dc3-a839-771f71affe1e] Instance destroyed {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 610.130817] env[63515]: INFO nova.compute.manager [None req-9f5c50b0-6bbe-4982-8818-c7255a8c23ab tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: b53754b1-ec31-4dc3-a839-771f71affe1e] Took 0.03 seconds to destroy the instance on the hypervisor. [ 610.130817] env[63515]: DEBUG oslo.service.loopingcall [None req-9f5c50b0-6bbe-4982-8818-c7255a8c23ab tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 610.130817] env[63515]: DEBUG nova.compute.manager [-] [instance: b53754b1-ec31-4dc3-a839-771f71affe1e] Deallocating network for instance {{(pid=63515) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 610.130817] env[63515]: DEBUG nova.network.neutron [-] [instance: b53754b1-ec31-4dc3-a839-771f71affe1e] deallocate_for_instance() {{(pid=63515) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 610.145694] env[63515]: DEBUG nova.network.neutron [-] [instance: b53754b1-ec31-4dc3-a839-771f71affe1e] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 610.259452] env[63515]: DEBUG nova.compute.manager [req-9d8fe6be-3f0e-4fb4-b663-cdf6ba17bc25 req-cccdb0cd-0532-4229-beb1-aa4c3fedad5b service nova] [instance: b53754b1-ec31-4dc3-a839-771f71affe1e] Received event network-changed-46784f84-e839-4fa3-88ea-2005d2779b07 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 610.259452] env[63515]: DEBUG nova.compute.manager [req-9d8fe6be-3f0e-4fb4-b663-cdf6ba17bc25 req-cccdb0cd-0532-4229-beb1-aa4c3fedad5b service nova] [instance: b53754b1-ec31-4dc3-a839-771f71affe1e] Refreshing instance network info cache due to event network-changed-46784f84-e839-4fa3-88ea-2005d2779b07. {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 610.259452] env[63515]: DEBUG oslo_concurrency.lockutils [req-9d8fe6be-3f0e-4fb4-b663-cdf6ba17bc25 req-cccdb0cd-0532-4229-beb1-aa4c3fedad5b service nova] Acquiring lock "refresh_cache-b53754b1-ec31-4dc3-a839-771f71affe1e" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 610.259452] env[63515]: DEBUG oslo_concurrency.lockutils [req-9d8fe6be-3f0e-4fb4-b663-cdf6ba17bc25 req-cccdb0cd-0532-4229-beb1-aa4c3fedad5b service nova] Acquired lock "refresh_cache-b53754b1-ec31-4dc3-a839-771f71affe1e" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 610.259452] env[63515]: DEBUG nova.network.neutron [req-9d8fe6be-3f0e-4fb4-b663-cdf6ba17bc25 req-cccdb0cd-0532-4229-beb1-aa4c3fedad5b service nova] [instance: b53754b1-ec31-4dc3-a839-771f71affe1e] Refreshing network info cache for port 46784f84-e839-4fa3-88ea-2005d2779b07 {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 610.466121] env[63515]: DEBUG nova.network.neutron [None req-12d73cc4-118d-457e-b393-cc4362d15b0d tempest-ListImageFiltersTestJSON-1100101045 tempest-ListImageFiltersTestJSON-1100101045-project-member] [instance: ab1539ba-33a4-4139-a2f3-1f8c507beede] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 610.544945] env[63515]: DEBUG nova.network.neutron [None req-12d73cc4-118d-457e-b393-cc4362d15b0d tempest-ListImageFiltersTestJSON-1100101045 tempest-ListImageFiltersTestJSON-1100101045-project-member] [instance: ab1539ba-33a4-4139-a2f3-1f8c507beede] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 610.647509] env[63515]: DEBUG nova.network.neutron [-] [instance: b53754b1-ec31-4dc3-a839-771f71affe1e] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 610.790743] env[63515]: DEBUG nova.network.neutron [req-9d8fe6be-3f0e-4fb4-b663-cdf6ba17bc25 req-cccdb0cd-0532-4229-beb1-aa4c3fedad5b service nova] [instance: b53754b1-ec31-4dc3-a839-771f71affe1e] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 610.915550] env[63515]: DEBUG nova.network.neutron [req-9d8fe6be-3f0e-4fb4-b663-cdf6ba17bc25 req-cccdb0cd-0532-4229-beb1-aa4c3fedad5b service nova] [instance: b53754b1-ec31-4dc3-a839-771f71affe1e] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 611.047662] env[63515]: DEBUG oslo_concurrency.lockutils [None req-12d73cc4-118d-457e-b393-cc4362d15b0d tempest-ListImageFiltersTestJSON-1100101045 tempest-ListImageFiltersTestJSON-1100101045-project-member] Releasing lock "refresh_cache-ab1539ba-33a4-4139-a2f3-1f8c507beede" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 611.047979] env[63515]: DEBUG nova.compute.manager [None req-12d73cc4-118d-457e-b393-cc4362d15b0d tempest-ListImageFiltersTestJSON-1100101045 tempest-ListImageFiltersTestJSON-1100101045-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63515) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 611.048132] env[63515]: DEBUG nova.compute.manager [None req-12d73cc4-118d-457e-b393-cc4362d15b0d tempest-ListImageFiltersTestJSON-1100101045 tempest-ListImageFiltersTestJSON-1100101045-project-member] [instance: ab1539ba-33a4-4139-a2f3-1f8c507beede] Deallocating network for instance {{(pid=63515) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 611.048306] env[63515]: DEBUG nova.network.neutron [None req-12d73cc4-118d-457e-b393-cc4362d15b0d tempest-ListImageFiltersTestJSON-1100101045 tempest-ListImageFiltersTestJSON-1100101045-project-member] [instance: ab1539ba-33a4-4139-a2f3-1f8c507beede] deallocate_for_instance() {{(pid=63515) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 611.068137] env[63515]: DEBUG nova.network.neutron [None req-12d73cc4-118d-457e-b393-cc4362d15b0d tempest-ListImageFiltersTestJSON-1100101045 tempest-ListImageFiltersTestJSON-1100101045-project-member] [instance: ab1539ba-33a4-4139-a2f3-1f8c507beede] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 611.150580] env[63515]: INFO nova.compute.manager [-] [instance: b53754b1-ec31-4dc3-a839-771f71affe1e] Took 1.02 seconds to deallocate network for instance. [ 611.160418] env[63515]: DEBUG nova.compute.claims [None req-9f5c50b0-6bbe-4982-8818-c7255a8c23ab tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: b53754b1-ec31-4dc3-a839-771f71affe1e] Aborting claim: {{(pid=63515) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 611.160604] env[63515]: DEBUG oslo_concurrency.lockutils [None req-9f5c50b0-6bbe-4982-8818-c7255a8c23ab tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 611.356365] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09d352c8-e707-4c87-b0a2-fc5e7a3a2e6c {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.365645] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18dcd2f9-62bf-483d-adbd-a56034921ae4 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.397637] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9dc3887e-6ff8-4d85-8590-57251219dea1 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.404480] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e8448c2-1526-4771-8a90-c5d9a3a0f999 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.419717] env[63515]: DEBUG oslo_concurrency.lockutils [req-9d8fe6be-3f0e-4fb4-b663-cdf6ba17bc25 req-cccdb0cd-0532-4229-beb1-aa4c3fedad5b service nova] Releasing lock "refresh_cache-b53754b1-ec31-4dc3-a839-771f71affe1e" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 611.419717] env[63515]: DEBUG nova.compute.manager [req-9d8fe6be-3f0e-4fb4-b663-cdf6ba17bc25 req-cccdb0cd-0532-4229-beb1-aa4c3fedad5b service nova] [instance: b53754b1-ec31-4dc3-a839-771f71affe1e] Received event network-vif-deleted-46784f84-e839-4fa3-88ea-2005d2779b07 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 611.420022] env[63515]: DEBUG nova.compute.provider_tree [None req-886adfc0-007d-4f83-bc30-b755d00d04c7 tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 611.572279] env[63515]: DEBUG nova.network.neutron [None req-12d73cc4-118d-457e-b393-cc4362d15b0d tempest-ListImageFiltersTestJSON-1100101045 tempest-ListImageFiltersTestJSON-1100101045-project-member] [instance: ab1539ba-33a4-4139-a2f3-1f8c507beede] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 611.923126] env[63515]: DEBUG nova.scheduler.client.report [None req-886adfc0-007d-4f83-bc30-b755d00d04c7 tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 612.074806] env[63515]: INFO nova.compute.manager [None req-12d73cc4-118d-457e-b393-cc4362d15b0d tempest-ListImageFiltersTestJSON-1100101045 tempest-ListImageFiltersTestJSON-1100101045-project-member] [instance: ab1539ba-33a4-4139-a2f3-1f8c507beede] Took 1.03 seconds to deallocate network for instance. [ 612.428080] env[63515]: DEBUG oslo_concurrency.lockutils [None req-886adfc0-007d-4f83-bc30-b755d00d04c7 tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.492s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 612.428614] env[63515]: DEBUG nova.compute.manager [None req-886adfc0-007d-4f83-bc30-b755d00d04c7 tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] [instance: 8b65c223-03a8-41e0-917e-45dd76b6e57f] Start building networks asynchronously for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 612.431141] env[63515]: DEBUG oslo_concurrency.lockutils [None req-c80cd50f-4127-4451-9298-83a67d8f0ab6 tempest-ServersTestFqdnHostnames-1452522528 tempest-ServersTestFqdnHostnames-1452522528-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 18.107s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 612.937370] env[63515]: DEBUG nova.compute.utils [None req-886adfc0-007d-4f83-bc30-b755d00d04c7 tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] Using /dev/sd instead of None {{(pid=63515) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 612.943183] env[63515]: DEBUG nova.compute.manager [None req-886adfc0-007d-4f83-bc30-b755d00d04c7 tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] [instance: 8b65c223-03a8-41e0-917e-45dd76b6e57f] Allocating IP information in the background. {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 612.943362] env[63515]: DEBUG nova.network.neutron [None req-886adfc0-007d-4f83-bc30-b755d00d04c7 tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] [instance: 8b65c223-03a8-41e0-917e-45dd76b6e57f] allocate_for_instance() {{(pid=63515) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 613.003138] env[63515]: DEBUG nova.policy [None req-886adfc0-007d-4f83-bc30-b755d00d04c7 tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '03a2c6b25cd6405fb5cba6858d83e309', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '72b65c09ed6d4492864022352180ff49', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63515) authorize /opt/stack/nova/nova/policy.py:201}} [ 613.107251] env[63515]: INFO nova.scheduler.client.report [None req-12d73cc4-118d-457e-b393-cc4362d15b0d tempest-ListImageFiltersTestJSON-1100101045 tempest-ListImageFiltersTestJSON-1100101045-project-member] Deleted allocations for instance ab1539ba-33a4-4139-a2f3-1f8c507beede [ 613.343103] env[63515]: DEBUG nova.network.neutron [None req-886adfc0-007d-4f83-bc30-b755d00d04c7 tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] [instance: 8b65c223-03a8-41e0-917e-45dd76b6e57f] Successfully created port: b7d27a8e-e6d6-42b4-a4bf-50e668e77788 {{(pid=63515) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 613.375401] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3638bd42-0f00-4d4f-b488-46a50a04145d {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.383354] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-416cf48d-4695-4e60-b629-d45f6b09cb42 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.416950] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0414e3ba-69e4-4694-bc7d-40a26ce8b992 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.425068] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cae6c753-70b6-4bca-89e4-c1226d60eae4 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.438405] env[63515]: DEBUG nova.compute.provider_tree [None req-c80cd50f-4127-4451-9298-83a67d8f0ab6 tempest-ServersTestFqdnHostnames-1452522528 tempest-ServersTestFqdnHostnames-1452522528-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 613.446934] env[63515]: DEBUG nova.compute.manager [None req-886adfc0-007d-4f83-bc30-b755d00d04c7 tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] [instance: 8b65c223-03a8-41e0-917e-45dd76b6e57f] Start building block device mappings for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 613.618128] env[63515]: DEBUG oslo_concurrency.lockutils [None req-12d73cc4-118d-457e-b393-cc4362d15b0d tempest-ListImageFiltersTestJSON-1100101045 tempest-ListImageFiltersTestJSON-1100101045-project-member] Lock "ab1539ba-33a4-4139-a2f3-1f8c507beede" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 68.232s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 613.941538] env[63515]: DEBUG nova.scheduler.client.report [None req-c80cd50f-4127-4451-9298-83a67d8f0ab6 tempest-ServersTestFqdnHostnames-1452522528 tempest-ServersTestFqdnHostnames-1452522528-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 614.124787] env[63515]: DEBUG nova.compute.manager [None req-4334f93e-6967-4c78-b6ec-145141dc41da tempest-AttachInterfacesV270Test-352099301 tempest-AttachInterfacesV270Test-352099301-project-member] [instance: 6876f61c-2315-4e49-9944-4a5caddede5b] Starting instance... {{(pid=63515) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 614.288252] env[63515]: DEBUG nova.compute.manager [req-58c64030-463b-45ac-b989-4b3f55b05c46 req-99bbeeaa-eb41-43d5-9869-c26a8aa6a5c1 service nova] [instance: 8b65c223-03a8-41e0-917e-45dd76b6e57f] Received event network-changed-b7d27a8e-e6d6-42b4-a4bf-50e668e77788 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 614.288497] env[63515]: DEBUG nova.compute.manager [req-58c64030-463b-45ac-b989-4b3f55b05c46 req-99bbeeaa-eb41-43d5-9869-c26a8aa6a5c1 service nova] [instance: 8b65c223-03a8-41e0-917e-45dd76b6e57f] Refreshing instance network info cache due to event network-changed-b7d27a8e-e6d6-42b4-a4bf-50e668e77788. {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 614.288764] env[63515]: DEBUG oslo_concurrency.lockutils [req-58c64030-463b-45ac-b989-4b3f55b05c46 req-99bbeeaa-eb41-43d5-9869-c26a8aa6a5c1 service nova] Acquiring lock "refresh_cache-8b65c223-03a8-41e0-917e-45dd76b6e57f" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 614.288949] env[63515]: DEBUG oslo_concurrency.lockutils [req-58c64030-463b-45ac-b989-4b3f55b05c46 req-99bbeeaa-eb41-43d5-9869-c26a8aa6a5c1 service nova] Acquired lock "refresh_cache-8b65c223-03a8-41e0-917e-45dd76b6e57f" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 614.289139] env[63515]: DEBUG nova.network.neutron [req-58c64030-463b-45ac-b989-4b3f55b05c46 req-99bbeeaa-eb41-43d5-9869-c26a8aa6a5c1 service nova] [instance: 8b65c223-03a8-41e0-917e-45dd76b6e57f] Refreshing network info cache for port b7d27a8e-e6d6-42b4-a4bf-50e668e77788 {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 614.328128] env[63515]: ERROR nova.compute.manager [None req-886adfc0-007d-4f83-bc30-b755d00d04c7 tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port b7d27a8e-e6d6-42b4-a4bf-50e668e77788, please check neutron logs for more information. [ 614.328128] env[63515]: ERROR nova.compute.manager Traceback (most recent call last): [ 614.328128] env[63515]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 614.328128] env[63515]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 614.328128] env[63515]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 614.328128] env[63515]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 614.328128] env[63515]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 614.328128] env[63515]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 614.328128] env[63515]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 614.328128] env[63515]: ERROR nova.compute.manager self.force_reraise() [ 614.328128] env[63515]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 614.328128] env[63515]: ERROR nova.compute.manager raise self.value [ 614.328128] env[63515]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 614.328128] env[63515]: ERROR nova.compute.manager updated_port = self._update_port( [ 614.328128] env[63515]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 614.328128] env[63515]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 614.328719] env[63515]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 614.328719] env[63515]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 614.328719] env[63515]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port b7d27a8e-e6d6-42b4-a4bf-50e668e77788, please check neutron logs for more information. [ 614.328719] env[63515]: ERROR nova.compute.manager [ 614.328719] env[63515]: Traceback (most recent call last): [ 614.328719] env[63515]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 614.328719] env[63515]: listener.cb(fileno) [ 614.328719] env[63515]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 614.328719] env[63515]: result = function(*args, **kwargs) [ 614.328719] env[63515]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 614.328719] env[63515]: return func(*args, **kwargs) [ 614.328719] env[63515]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 614.328719] env[63515]: raise e [ 614.328719] env[63515]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 614.328719] env[63515]: nwinfo = self.network_api.allocate_for_instance( [ 614.328719] env[63515]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 614.328719] env[63515]: created_port_ids = self._update_ports_for_instance( [ 614.328719] env[63515]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 614.328719] env[63515]: with excutils.save_and_reraise_exception(): [ 614.328719] env[63515]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 614.328719] env[63515]: self.force_reraise() [ 614.328719] env[63515]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 614.328719] env[63515]: raise self.value [ 614.328719] env[63515]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 614.328719] env[63515]: updated_port = self._update_port( [ 614.328719] env[63515]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 614.328719] env[63515]: _ensure_no_port_binding_failure(port) [ 614.328719] env[63515]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 614.328719] env[63515]: raise exception.PortBindingFailed(port_id=port['id']) [ 614.329430] env[63515]: nova.exception.PortBindingFailed: Binding failed for port b7d27a8e-e6d6-42b4-a4bf-50e668e77788, please check neutron logs for more information. [ 614.329430] env[63515]: Removing descriptor: 17 [ 614.448745] env[63515]: DEBUG oslo_concurrency.lockutils [None req-c80cd50f-4127-4451-9298-83a67d8f0ab6 tempest-ServersTestFqdnHostnames-1452522528 tempest-ServersTestFqdnHostnames-1452522528-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.016s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 614.448745] env[63515]: ERROR nova.compute.manager [None req-c80cd50f-4127-4451-9298-83a67d8f0ab6 tempest-ServersTestFqdnHostnames-1452522528 tempest-ServersTestFqdnHostnames-1452522528-project-member] [instance: b84c6c8d-e358-4a14-8425-c4f1d8adf4ee] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 3d61ba1c-4556-41b6-a699-f5b65b0de93f, please check neutron logs for more information. [ 614.448745] env[63515]: ERROR nova.compute.manager [instance: b84c6c8d-e358-4a14-8425-c4f1d8adf4ee] Traceback (most recent call last): [ 614.448745] env[63515]: ERROR nova.compute.manager [instance: b84c6c8d-e358-4a14-8425-c4f1d8adf4ee] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 614.448745] env[63515]: ERROR nova.compute.manager [instance: b84c6c8d-e358-4a14-8425-c4f1d8adf4ee] self.driver.spawn(context, instance, image_meta, [ 614.448745] env[63515]: ERROR nova.compute.manager [instance: b84c6c8d-e358-4a14-8425-c4f1d8adf4ee] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 614.448745] env[63515]: ERROR nova.compute.manager [instance: b84c6c8d-e358-4a14-8425-c4f1d8adf4ee] self._vmops.spawn(context, instance, image_meta, injected_files, [ 614.448745] env[63515]: ERROR nova.compute.manager [instance: b84c6c8d-e358-4a14-8425-c4f1d8adf4ee] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 614.448745] env[63515]: ERROR nova.compute.manager [instance: b84c6c8d-e358-4a14-8425-c4f1d8adf4ee] vm_ref = self.build_virtual_machine(instance, [ 614.449128] env[63515]: ERROR nova.compute.manager [instance: b84c6c8d-e358-4a14-8425-c4f1d8adf4ee] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 614.449128] env[63515]: ERROR nova.compute.manager [instance: b84c6c8d-e358-4a14-8425-c4f1d8adf4ee] vif_infos = vmwarevif.get_vif_info(self._session, [ 614.449128] env[63515]: ERROR nova.compute.manager [instance: b84c6c8d-e358-4a14-8425-c4f1d8adf4ee] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 614.449128] env[63515]: ERROR nova.compute.manager [instance: b84c6c8d-e358-4a14-8425-c4f1d8adf4ee] for vif in network_info: [ 614.449128] env[63515]: ERROR nova.compute.manager [instance: b84c6c8d-e358-4a14-8425-c4f1d8adf4ee] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 614.449128] env[63515]: ERROR nova.compute.manager [instance: b84c6c8d-e358-4a14-8425-c4f1d8adf4ee] return self._sync_wrapper(fn, *args, **kwargs) [ 614.449128] env[63515]: ERROR nova.compute.manager [instance: b84c6c8d-e358-4a14-8425-c4f1d8adf4ee] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 614.449128] env[63515]: ERROR nova.compute.manager [instance: b84c6c8d-e358-4a14-8425-c4f1d8adf4ee] self.wait() [ 614.449128] env[63515]: ERROR nova.compute.manager [instance: b84c6c8d-e358-4a14-8425-c4f1d8adf4ee] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 614.449128] env[63515]: ERROR nova.compute.manager [instance: b84c6c8d-e358-4a14-8425-c4f1d8adf4ee] self[:] = self._gt.wait() [ 614.449128] env[63515]: ERROR nova.compute.manager [instance: b84c6c8d-e358-4a14-8425-c4f1d8adf4ee] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 614.449128] env[63515]: ERROR nova.compute.manager [instance: b84c6c8d-e358-4a14-8425-c4f1d8adf4ee] return self._exit_event.wait() [ 614.449128] env[63515]: ERROR nova.compute.manager [instance: b84c6c8d-e358-4a14-8425-c4f1d8adf4ee] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 614.449480] env[63515]: ERROR nova.compute.manager [instance: b84c6c8d-e358-4a14-8425-c4f1d8adf4ee] result = hub.switch() [ 614.449480] env[63515]: ERROR nova.compute.manager [instance: b84c6c8d-e358-4a14-8425-c4f1d8adf4ee] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 614.449480] env[63515]: ERROR nova.compute.manager [instance: b84c6c8d-e358-4a14-8425-c4f1d8adf4ee] return self.greenlet.switch() [ 614.449480] env[63515]: ERROR nova.compute.manager [instance: b84c6c8d-e358-4a14-8425-c4f1d8adf4ee] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 614.449480] env[63515]: ERROR nova.compute.manager [instance: b84c6c8d-e358-4a14-8425-c4f1d8adf4ee] result = function(*args, **kwargs) [ 614.449480] env[63515]: ERROR nova.compute.manager [instance: b84c6c8d-e358-4a14-8425-c4f1d8adf4ee] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 614.449480] env[63515]: ERROR nova.compute.manager [instance: b84c6c8d-e358-4a14-8425-c4f1d8adf4ee] return func(*args, **kwargs) [ 614.449480] env[63515]: ERROR nova.compute.manager [instance: b84c6c8d-e358-4a14-8425-c4f1d8adf4ee] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 614.449480] env[63515]: ERROR nova.compute.manager [instance: b84c6c8d-e358-4a14-8425-c4f1d8adf4ee] raise e [ 614.449480] env[63515]: ERROR nova.compute.manager [instance: b84c6c8d-e358-4a14-8425-c4f1d8adf4ee] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 614.449480] env[63515]: ERROR nova.compute.manager [instance: b84c6c8d-e358-4a14-8425-c4f1d8adf4ee] nwinfo = self.network_api.allocate_for_instance( [ 614.449480] env[63515]: ERROR nova.compute.manager [instance: b84c6c8d-e358-4a14-8425-c4f1d8adf4ee] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 614.449480] env[63515]: ERROR nova.compute.manager [instance: b84c6c8d-e358-4a14-8425-c4f1d8adf4ee] created_port_ids = self._update_ports_for_instance( [ 614.449831] env[63515]: ERROR nova.compute.manager [instance: b84c6c8d-e358-4a14-8425-c4f1d8adf4ee] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 614.449831] env[63515]: ERROR nova.compute.manager [instance: b84c6c8d-e358-4a14-8425-c4f1d8adf4ee] with excutils.save_and_reraise_exception(): [ 614.449831] env[63515]: ERROR nova.compute.manager [instance: b84c6c8d-e358-4a14-8425-c4f1d8adf4ee] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 614.449831] env[63515]: ERROR nova.compute.manager [instance: b84c6c8d-e358-4a14-8425-c4f1d8adf4ee] self.force_reraise() [ 614.449831] env[63515]: ERROR nova.compute.manager [instance: b84c6c8d-e358-4a14-8425-c4f1d8adf4ee] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 614.449831] env[63515]: ERROR nova.compute.manager [instance: b84c6c8d-e358-4a14-8425-c4f1d8adf4ee] raise self.value [ 614.449831] env[63515]: ERROR nova.compute.manager [instance: b84c6c8d-e358-4a14-8425-c4f1d8adf4ee] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 614.449831] env[63515]: ERROR nova.compute.manager [instance: b84c6c8d-e358-4a14-8425-c4f1d8adf4ee] updated_port = self._update_port( [ 614.449831] env[63515]: ERROR nova.compute.manager [instance: b84c6c8d-e358-4a14-8425-c4f1d8adf4ee] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 614.449831] env[63515]: ERROR nova.compute.manager [instance: b84c6c8d-e358-4a14-8425-c4f1d8adf4ee] _ensure_no_port_binding_failure(port) [ 614.449831] env[63515]: ERROR nova.compute.manager [instance: b84c6c8d-e358-4a14-8425-c4f1d8adf4ee] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 614.449831] env[63515]: ERROR nova.compute.manager [instance: b84c6c8d-e358-4a14-8425-c4f1d8adf4ee] raise exception.PortBindingFailed(port_id=port['id']) [ 614.450175] env[63515]: ERROR nova.compute.manager [instance: b84c6c8d-e358-4a14-8425-c4f1d8adf4ee] nova.exception.PortBindingFailed: Binding failed for port 3d61ba1c-4556-41b6-a699-f5b65b0de93f, please check neutron logs for more information. [ 614.450175] env[63515]: ERROR nova.compute.manager [instance: b84c6c8d-e358-4a14-8425-c4f1d8adf4ee] [ 614.450175] env[63515]: DEBUG nova.compute.utils [None req-c80cd50f-4127-4451-9298-83a67d8f0ab6 tempest-ServersTestFqdnHostnames-1452522528 tempest-ServersTestFqdnHostnames-1452522528-project-member] [instance: b84c6c8d-e358-4a14-8425-c4f1d8adf4ee] Binding failed for port 3d61ba1c-4556-41b6-a699-f5b65b0de93f, please check neutron logs for more information. {{(pid=63515) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 614.450424] env[63515]: DEBUG oslo_concurrency.lockutils [None req-eb536733-3a8a-4dae-88ae-0ad9d450b524 tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 18.025s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 614.450668] env[63515]: DEBUG nova.objects.instance [None req-eb536733-3a8a-4dae-88ae-0ad9d450b524 tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] Lazy-loading 'resources' on Instance uuid 93b28142-8454-43c0-b0a7-d61aa95c8fc8 {{(pid=63515) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 614.451994] env[63515]: DEBUG nova.compute.manager [None req-c80cd50f-4127-4451-9298-83a67d8f0ab6 tempest-ServersTestFqdnHostnames-1452522528 tempest-ServersTestFqdnHostnames-1452522528-project-member] [instance: b84c6c8d-e358-4a14-8425-c4f1d8adf4ee] Build of instance b84c6c8d-e358-4a14-8425-c4f1d8adf4ee was re-scheduled: Binding failed for port 3d61ba1c-4556-41b6-a699-f5b65b0de93f, please check neutron logs for more information. {{(pid=63515) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 614.452595] env[63515]: DEBUG nova.compute.manager [None req-c80cd50f-4127-4451-9298-83a67d8f0ab6 tempest-ServersTestFqdnHostnames-1452522528 tempest-ServersTestFqdnHostnames-1452522528-project-member] [instance: b84c6c8d-e358-4a14-8425-c4f1d8adf4ee] Unplugging VIFs for instance {{(pid=63515) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 614.452823] env[63515]: DEBUG oslo_concurrency.lockutils [None req-c80cd50f-4127-4451-9298-83a67d8f0ab6 tempest-ServersTestFqdnHostnames-1452522528 tempest-ServersTestFqdnHostnames-1452522528-project-member] Acquiring lock "refresh_cache-b84c6c8d-e358-4a14-8425-c4f1d8adf4ee" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 614.452984] env[63515]: DEBUG oslo_concurrency.lockutils [None req-c80cd50f-4127-4451-9298-83a67d8f0ab6 tempest-ServersTestFqdnHostnames-1452522528 tempest-ServersTestFqdnHostnames-1452522528-project-member] Acquired lock "refresh_cache-b84c6c8d-e358-4a14-8425-c4f1d8adf4ee" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 614.453220] env[63515]: DEBUG nova.network.neutron [None req-c80cd50f-4127-4451-9298-83a67d8f0ab6 tempest-ServersTestFqdnHostnames-1452522528 tempest-ServersTestFqdnHostnames-1452522528-project-member] [instance: b84c6c8d-e358-4a14-8425-c4f1d8adf4ee] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 614.455610] env[63515]: DEBUG nova.compute.manager [None req-886adfc0-007d-4f83-bc30-b755d00d04c7 tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] [instance: 8b65c223-03a8-41e0-917e-45dd76b6e57f] Start spawning the instance on the hypervisor. {{(pid=63515) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 614.480832] env[63515]: DEBUG nova.virt.hardware [None req-886adfc0-007d-4f83-bc30-b755d00d04c7 tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T02:52:14Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T02:51:56Z,direct_url=,disk_format='vmdk',id=8a120570-cb06-4099-b262-554ca0ad15c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b569255cc43e42e7ae3f2b4ad37c6ef4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T02:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 614.481639] env[63515]: DEBUG nova.virt.hardware [None req-886adfc0-007d-4f83-bc30-b755d00d04c7 tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] Flavor limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 614.481639] env[63515]: DEBUG nova.virt.hardware [None req-886adfc0-007d-4f83-bc30-b755d00d04c7 tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] Image limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 614.481639] env[63515]: DEBUG nova.virt.hardware [None req-886adfc0-007d-4f83-bc30-b755d00d04c7 tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] Flavor pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 614.481639] env[63515]: DEBUG nova.virt.hardware [None req-886adfc0-007d-4f83-bc30-b755d00d04c7 tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] Image pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 614.481825] env[63515]: DEBUG nova.virt.hardware [None req-886adfc0-007d-4f83-bc30-b755d00d04c7 tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 614.481854] env[63515]: DEBUG nova.virt.hardware [None req-886adfc0-007d-4f83-bc30-b755d00d04c7 tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 614.482278] env[63515]: DEBUG nova.virt.hardware [None req-886adfc0-007d-4f83-bc30-b755d00d04c7 tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 614.482278] env[63515]: DEBUG nova.virt.hardware [None req-886adfc0-007d-4f83-bc30-b755d00d04c7 tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] Got 1 possible topologies {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 614.482379] env[63515]: DEBUG nova.virt.hardware [None req-886adfc0-007d-4f83-bc30-b755d00d04c7 tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 614.482494] env[63515]: DEBUG nova.virt.hardware [None req-886adfc0-007d-4f83-bc30-b755d00d04c7 tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 614.484530] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a00cc2e-c29d-4fae-aab6-decf6b91c245 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.491363] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c6dc9d5-3731-4a0f-b268-7d2fe9755ee6 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.505798] env[63515]: ERROR nova.compute.manager [None req-886adfc0-007d-4f83-bc30-b755d00d04c7 tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] [instance: 8b65c223-03a8-41e0-917e-45dd76b6e57f] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port b7d27a8e-e6d6-42b4-a4bf-50e668e77788, please check neutron logs for more information. [ 614.505798] env[63515]: ERROR nova.compute.manager [instance: 8b65c223-03a8-41e0-917e-45dd76b6e57f] Traceback (most recent call last): [ 614.505798] env[63515]: ERROR nova.compute.manager [instance: 8b65c223-03a8-41e0-917e-45dd76b6e57f] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 614.505798] env[63515]: ERROR nova.compute.manager [instance: 8b65c223-03a8-41e0-917e-45dd76b6e57f] yield resources [ 614.505798] env[63515]: ERROR nova.compute.manager [instance: 8b65c223-03a8-41e0-917e-45dd76b6e57f] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 614.505798] env[63515]: ERROR nova.compute.manager [instance: 8b65c223-03a8-41e0-917e-45dd76b6e57f] self.driver.spawn(context, instance, image_meta, [ 614.505798] env[63515]: ERROR nova.compute.manager [instance: 8b65c223-03a8-41e0-917e-45dd76b6e57f] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 614.505798] env[63515]: ERROR nova.compute.manager [instance: 8b65c223-03a8-41e0-917e-45dd76b6e57f] self._vmops.spawn(context, instance, image_meta, injected_files, [ 614.505798] env[63515]: ERROR nova.compute.manager [instance: 8b65c223-03a8-41e0-917e-45dd76b6e57f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 614.505798] env[63515]: ERROR nova.compute.manager [instance: 8b65c223-03a8-41e0-917e-45dd76b6e57f] vm_ref = self.build_virtual_machine(instance, [ 614.505798] env[63515]: ERROR nova.compute.manager [instance: 8b65c223-03a8-41e0-917e-45dd76b6e57f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 614.506111] env[63515]: ERROR nova.compute.manager [instance: 8b65c223-03a8-41e0-917e-45dd76b6e57f] vif_infos = vmwarevif.get_vif_info(self._session, [ 614.506111] env[63515]: ERROR nova.compute.manager [instance: 8b65c223-03a8-41e0-917e-45dd76b6e57f] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 614.506111] env[63515]: ERROR nova.compute.manager [instance: 8b65c223-03a8-41e0-917e-45dd76b6e57f] for vif in network_info: [ 614.506111] env[63515]: ERROR nova.compute.manager [instance: 8b65c223-03a8-41e0-917e-45dd76b6e57f] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 614.506111] env[63515]: ERROR nova.compute.manager [instance: 8b65c223-03a8-41e0-917e-45dd76b6e57f] return self._sync_wrapper(fn, *args, **kwargs) [ 614.506111] env[63515]: ERROR nova.compute.manager [instance: 8b65c223-03a8-41e0-917e-45dd76b6e57f] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 614.506111] env[63515]: ERROR nova.compute.manager [instance: 8b65c223-03a8-41e0-917e-45dd76b6e57f] self.wait() [ 614.506111] env[63515]: ERROR nova.compute.manager [instance: 8b65c223-03a8-41e0-917e-45dd76b6e57f] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 614.506111] env[63515]: ERROR nova.compute.manager [instance: 8b65c223-03a8-41e0-917e-45dd76b6e57f] self[:] = self._gt.wait() [ 614.506111] env[63515]: ERROR nova.compute.manager [instance: 8b65c223-03a8-41e0-917e-45dd76b6e57f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 614.506111] env[63515]: ERROR nova.compute.manager [instance: 8b65c223-03a8-41e0-917e-45dd76b6e57f] return self._exit_event.wait() [ 614.506111] env[63515]: ERROR nova.compute.manager [instance: 8b65c223-03a8-41e0-917e-45dd76b6e57f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 614.506111] env[63515]: ERROR nova.compute.manager [instance: 8b65c223-03a8-41e0-917e-45dd76b6e57f] current.throw(*self._exc) [ 614.506423] env[63515]: ERROR nova.compute.manager [instance: 8b65c223-03a8-41e0-917e-45dd76b6e57f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 614.506423] env[63515]: ERROR nova.compute.manager [instance: 8b65c223-03a8-41e0-917e-45dd76b6e57f] result = function(*args, **kwargs) [ 614.506423] env[63515]: ERROR nova.compute.manager [instance: 8b65c223-03a8-41e0-917e-45dd76b6e57f] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 614.506423] env[63515]: ERROR nova.compute.manager [instance: 8b65c223-03a8-41e0-917e-45dd76b6e57f] return func(*args, **kwargs) [ 614.506423] env[63515]: ERROR nova.compute.manager [instance: 8b65c223-03a8-41e0-917e-45dd76b6e57f] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 614.506423] env[63515]: ERROR nova.compute.manager [instance: 8b65c223-03a8-41e0-917e-45dd76b6e57f] raise e [ 614.506423] env[63515]: ERROR nova.compute.manager [instance: 8b65c223-03a8-41e0-917e-45dd76b6e57f] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 614.506423] env[63515]: ERROR nova.compute.manager [instance: 8b65c223-03a8-41e0-917e-45dd76b6e57f] nwinfo = self.network_api.allocate_for_instance( [ 614.506423] env[63515]: ERROR nova.compute.manager [instance: 8b65c223-03a8-41e0-917e-45dd76b6e57f] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 614.506423] env[63515]: ERROR nova.compute.manager [instance: 8b65c223-03a8-41e0-917e-45dd76b6e57f] created_port_ids = self._update_ports_for_instance( [ 614.506423] env[63515]: ERROR nova.compute.manager [instance: 8b65c223-03a8-41e0-917e-45dd76b6e57f] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 614.506423] env[63515]: ERROR nova.compute.manager [instance: 8b65c223-03a8-41e0-917e-45dd76b6e57f] with excutils.save_and_reraise_exception(): [ 614.506423] env[63515]: ERROR nova.compute.manager [instance: 8b65c223-03a8-41e0-917e-45dd76b6e57f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 614.506731] env[63515]: ERROR nova.compute.manager [instance: 8b65c223-03a8-41e0-917e-45dd76b6e57f] self.force_reraise() [ 614.506731] env[63515]: ERROR nova.compute.manager [instance: 8b65c223-03a8-41e0-917e-45dd76b6e57f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 614.506731] env[63515]: ERROR nova.compute.manager [instance: 8b65c223-03a8-41e0-917e-45dd76b6e57f] raise self.value [ 614.506731] env[63515]: ERROR nova.compute.manager [instance: 8b65c223-03a8-41e0-917e-45dd76b6e57f] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 614.506731] env[63515]: ERROR nova.compute.manager [instance: 8b65c223-03a8-41e0-917e-45dd76b6e57f] updated_port = self._update_port( [ 614.506731] env[63515]: ERROR nova.compute.manager [instance: 8b65c223-03a8-41e0-917e-45dd76b6e57f] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 614.506731] env[63515]: ERROR nova.compute.manager [instance: 8b65c223-03a8-41e0-917e-45dd76b6e57f] _ensure_no_port_binding_failure(port) [ 614.506731] env[63515]: ERROR nova.compute.manager [instance: 8b65c223-03a8-41e0-917e-45dd76b6e57f] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 614.506731] env[63515]: ERROR nova.compute.manager [instance: 8b65c223-03a8-41e0-917e-45dd76b6e57f] raise exception.PortBindingFailed(port_id=port['id']) [ 614.506731] env[63515]: ERROR nova.compute.manager [instance: 8b65c223-03a8-41e0-917e-45dd76b6e57f] nova.exception.PortBindingFailed: Binding failed for port b7d27a8e-e6d6-42b4-a4bf-50e668e77788, please check neutron logs for more information. [ 614.506731] env[63515]: ERROR nova.compute.manager [instance: 8b65c223-03a8-41e0-917e-45dd76b6e57f] [ 614.506731] env[63515]: INFO nova.compute.manager [None req-886adfc0-007d-4f83-bc30-b755d00d04c7 tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] [instance: 8b65c223-03a8-41e0-917e-45dd76b6e57f] Terminating instance [ 614.507982] env[63515]: DEBUG oslo_concurrency.lockutils [None req-886adfc0-007d-4f83-bc30-b755d00d04c7 tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] Acquiring lock "refresh_cache-8b65c223-03a8-41e0-917e-45dd76b6e57f" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 614.648893] env[63515]: DEBUG oslo_concurrency.lockutils [None req-4334f93e-6967-4c78-b6ec-145141dc41da tempest-AttachInterfacesV270Test-352099301 tempest-AttachInterfacesV270Test-352099301-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 614.807644] env[63515]: DEBUG nova.network.neutron [req-58c64030-463b-45ac-b989-4b3f55b05c46 req-99bbeeaa-eb41-43d5-9869-c26a8aa6a5c1 service nova] [instance: 8b65c223-03a8-41e0-917e-45dd76b6e57f] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 614.899526] env[63515]: DEBUG nova.network.neutron [req-58c64030-463b-45ac-b989-4b3f55b05c46 req-99bbeeaa-eb41-43d5-9869-c26a8aa6a5c1 service nova] [instance: 8b65c223-03a8-41e0-917e-45dd76b6e57f] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 614.981390] env[63515]: DEBUG nova.network.neutron [None req-c80cd50f-4127-4451-9298-83a67d8f0ab6 tempest-ServersTestFqdnHostnames-1452522528 tempest-ServersTestFqdnHostnames-1452522528-project-member] [instance: b84c6c8d-e358-4a14-8425-c4f1d8adf4ee] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 615.082860] env[63515]: DEBUG nova.network.neutron [None req-c80cd50f-4127-4451-9298-83a67d8f0ab6 tempest-ServersTestFqdnHostnames-1452522528 tempest-ServersTestFqdnHostnames-1452522528-project-member] [instance: b84c6c8d-e358-4a14-8425-c4f1d8adf4ee] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 615.316081] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6baf825c-58c6-4a1f-8adb-3abbffc32672 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.324202] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6a014e9-ece9-44f7-8a79-227a3f873f9e {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.355742] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d9b839e-edd0-4e85-ac10-8853845dbfd6 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.362903] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3dcff6b-2ec3-4359-b0b2-0b0d61f5ca2f {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.375686] env[63515]: DEBUG nova.compute.provider_tree [None req-eb536733-3a8a-4dae-88ae-0ad9d450b524 tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 615.402554] env[63515]: DEBUG oslo_concurrency.lockutils [req-58c64030-463b-45ac-b989-4b3f55b05c46 req-99bbeeaa-eb41-43d5-9869-c26a8aa6a5c1 service nova] Releasing lock "refresh_cache-8b65c223-03a8-41e0-917e-45dd76b6e57f" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 615.403250] env[63515]: DEBUG oslo_concurrency.lockutils [None req-886adfc0-007d-4f83-bc30-b755d00d04c7 tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] Acquired lock "refresh_cache-8b65c223-03a8-41e0-917e-45dd76b6e57f" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 615.403380] env[63515]: DEBUG nova.network.neutron [None req-886adfc0-007d-4f83-bc30-b755d00d04c7 tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] [instance: 8b65c223-03a8-41e0-917e-45dd76b6e57f] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 615.585629] env[63515]: DEBUG oslo_concurrency.lockutils [None req-c80cd50f-4127-4451-9298-83a67d8f0ab6 tempest-ServersTestFqdnHostnames-1452522528 tempest-ServersTestFqdnHostnames-1452522528-project-member] Releasing lock "refresh_cache-b84c6c8d-e358-4a14-8425-c4f1d8adf4ee" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 615.585908] env[63515]: DEBUG nova.compute.manager [None req-c80cd50f-4127-4451-9298-83a67d8f0ab6 tempest-ServersTestFqdnHostnames-1452522528 tempest-ServersTestFqdnHostnames-1452522528-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63515) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 615.586086] env[63515]: DEBUG nova.compute.manager [None req-c80cd50f-4127-4451-9298-83a67d8f0ab6 tempest-ServersTestFqdnHostnames-1452522528 tempest-ServersTestFqdnHostnames-1452522528-project-member] [instance: b84c6c8d-e358-4a14-8425-c4f1d8adf4ee] Deallocating network for instance {{(pid=63515) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 615.586263] env[63515]: DEBUG nova.network.neutron [None req-c80cd50f-4127-4451-9298-83a67d8f0ab6 tempest-ServersTestFqdnHostnames-1452522528 tempest-ServersTestFqdnHostnames-1452522528-project-member] [instance: b84c6c8d-e358-4a14-8425-c4f1d8adf4ee] deallocate_for_instance() {{(pid=63515) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 615.605684] env[63515]: DEBUG nova.network.neutron [None req-c80cd50f-4127-4451-9298-83a67d8f0ab6 tempest-ServersTestFqdnHostnames-1452522528 tempest-ServersTestFqdnHostnames-1452522528-project-member] [instance: b84c6c8d-e358-4a14-8425-c4f1d8adf4ee] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 615.880962] env[63515]: DEBUG nova.scheduler.client.report [None req-eb536733-3a8a-4dae-88ae-0ad9d450b524 tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 615.921737] env[63515]: DEBUG nova.network.neutron [None req-886adfc0-007d-4f83-bc30-b755d00d04c7 tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] [instance: 8b65c223-03a8-41e0-917e-45dd76b6e57f] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 616.032459] env[63515]: DEBUG nova.network.neutron [None req-886adfc0-007d-4f83-bc30-b755d00d04c7 tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] [instance: 8b65c223-03a8-41e0-917e-45dd76b6e57f] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 616.108935] env[63515]: DEBUG nova.network.neutron [None req-c80cd50f-4127-4451-9298-83a67d8f0ab6 tempest-ServersTestFqdnHostnames-1452522528 tempest-ServersTestFqdnHostnames-1452522528-project-member] [instance: b84c6c8d-e358-4a14-8425-c4f1d8adf4ee] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 616.312763] env[63515]: DEBUG nova.compute.manager [req-64aa267c-ad4a-452f-b81c-9ffc093c03a9 req-a3b38a1a-5e76-4fe0-85c0-091835da125d service nova] [instance: 8b65c223-03a8-41e0-917e-45dd76b6e57f] Received event network-vif-deleted-b7d27a8e-e6d6-42b4-a4bf-50e668e77788 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 616.387099] env[63515]: DEBUG oslo_concurrency.lockutils [None req-eb536733-3a8a-4dae-88ae-0ad9d450b524 tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.936s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 616.389545] env[63515]: DEBUG oslo_concurrency.lockutils [None req-22f02712-96d1-4031-8b8f-5bc880f34c33 tempest-ImagesNegativeTestJSON-2041328384 tempest-ImagesNegativeTestJSON-2041328384-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 19.272s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 616.410459] env[63515]: INFO nova.scheduler.client.report [None req-eb536733-3a8a-4dae-88ae-0ad9d450b524 tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] Deleted allocations for instance 93b28142-8454-43c0-b0a7-d61aa95c8fc8 [ 616.534761] env[63515]: DEBUG oslo_concurrency.lockutils [None req-886adfc0-007d-4f83-bc30-b755d00d04c7 tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] Releasing lock "refresh_cache-8b65c223-03a8-41e0-917e-45dd76b6e57f" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 616.535209] env[63515]: DEBUG nova.compute.manager [None req-886adfc0-007d-4f83-bc30-b755d00d04c7 tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] [instance: 8b65c223-03a8-41e0-917e-45dd76b6e57f] Start destroying the instance on the hypervisor. {{(pid=63515) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 616.535404] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-886adfc0-007d-4f83-bc30-b755d00d04c7 tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] [instance: 8b65c223-03a8-41e0-917e-45dd76b6e57f] Destroying instance {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 616.535694] env[63515]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e0c8b128-28ca-4c8b-90cd-245e966beab8 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.545499] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c312969-a379-45c4-b5d5-878d34eeed82 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.570888] env[63515]: DEBUG oslo_concurrency.lockutils [None req-746f5a16-d8b1-448e-a35c-34720548f33b tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Acquiring lock "b4477e66-ae12-4929-90ed-b7b652e0f207" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 616.571155] env[63515]: DEBUG oslo_concurrency.lockutils [None req-746f5a16-d8b1-448e-a35c-34720548f33b tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Lock "b4477e66-ae12-4929-90ed-b7b652e0f207" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 616.571669] env[63515]: WARNING nova.virt.vmwareapi.vmops [None req-886adfc0-007d-4f83-bc30-b755d00d04c7 tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] [instance: 8b65c223-03a8-41e0-917e-45dd76b6e57f] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 8b65c223-03a8-41e0-917e-45dd76b6e57f could not be found. [ 616.571890] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-886adfc0-007d-4f83-bc30-b755d00d04c7 tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] [instance: 8b65c223-03a8-41e0-917e-45dd76b6e57f] Instance destroyed {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 616.572078] env[63515]: INFO nova.compute.manager [None req-886adfc0-007d-4f83-bc30-b755d00d04c7 tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] [instance: 8b65c223-03a8-41e0-917e-45dd76b6e57f] Took 0.04 seconds to destroy the instance on the hypervisor. [ 616.572310] env[63515]: DEBUG oslo.service.loopingcall [None req-886adfc0-007d-4f83-bc30-b755d00d04c7 tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 616.572527] env[63515]: DEBUG nova.compute.manager [-] [instance: 8b65c223-03a8-41e0-917e-45dd76b6e57f] Deallocating network for instance {{(pid=63515) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 616.572619] env[63515]: DEBUG nova.network.neutron [-] [instance: 8b65c223-03a8-41e0-917e-45dd76b6e57f] deallocate_for_instance() {{(pid=63515) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 616.588466] env[63515]: DEBUG nova.network.neutron [-] [instance: 8b65c223-03a8-41e0-917e-45dd76b6e57f] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 616.612159] env[63515]: INFO nova.compute.manager [None req-c80cd50f-4127-4451-9298-83a67d8f0ab6 tempest-ServersTestFqdnHostnames-1452522528 tempest-ServersTestFqdnHostnames-1452522528-project-member] [instance: b84c6c8d-e358-4a14-8425-c4f1d8adf4ee] Took 1.03 seconds to deallocate network for instance. [ 616.858403] env[63515]: DEBUG oslo_concurrency.lockutils [None req-74f7de08-cdcf-4b98-bfdb-a2dd77c03e6d tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Acquiring lock "b85e9a70-7f5b-4d32-b616-f2a97e3186c8" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 616.858745] env[63515]: DEBUG oslo_concurrency.lockutils [None req-74f7de08-cdcf-4b98-bfdb-a2dd77c03e6d tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Lock "b85e9a70-7f5b-4d32-b616-f2a97e3186c8" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 616.922777] env[63515]: DEBUG oslo_concurrency.lockutils [None req-eb536733-3a8a-4dae-88ae-0ad9d450b524 tempest-ServersAdmin275Test-2128375172 tempest-ServersAdmin275Test-2128375172-project-member] Lock "93b28142-8454-43c0-b0a7-d61aa95c8fc8" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 24.345s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 617.094923] env[63515]: DEBUG nova.network.neutron [-] [instance: 8b65c223-03a8-41e0-917e-45dd76b6e57f] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 617.277981] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d0173e0-2514-4528-9860-a031dadf197c {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.288856] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a782cc78-ae62-440c-9b0f-2cfb2d861072 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.319799] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7101724-341e-414f-8637-b9f907413df7 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.327259] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0bfc8761-9473-492f-998d-d733db0c72d3 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.340545] env[63515]: DEBUG nova.compute.provider_tree [None req-22f02712-96d1-4031-8b8f-5bc880f34c33 tempest-ImagesNegativeTestJSON-2041328384 tempest-ImagesNegativeTestJSON-2041328384-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 617.597518] env[63515]: INFO nova.compute.manager [-] [instance: 8b65c223-03a8-41e0-917e-45dd76b6e57f] Took 1.02 seconds to deallocate network for instance. [ 617.600222] env[63515]: DEBUG nova.compute.claims [None req-886adfc0-007d-4f83-bc30-b755d00d04c7 tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] [instance: 8b65c223-03a8-41e0-917e-45dd76b6e57f] Aborting claim: {{(pid=63515) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 617.600222] env[63515]: DEBUG oslo_concurrency.lockutils [None req-886adfc0-007d-4f83-bc30-b755d00d04c7 tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 617.647928] env[63515]: INFO nova.scheduler.client.report [None req-c80cd50f-4127-4451-9298-83a67d8f0ab6 tempest-ServersTestFqdnHostnames-1452522528 tempest-ServersTestFqdnHostnames-1452522528-project-member] Deleted allocations for instance b84c6c8d-e358-4a14-8425-c4f1d8adf4ee [ 617.846031] env[63515]: DEBUG nova.scheduler.client.report [None req-22f02712-96d1-4031-8b8f-5bc880f34c33 tempest-ImagesNegativeTestJSON-2041328384 tempest-ImagesNegativeTestJSON-2041328384-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 618.158443] env[63515]: DEBUG oslo_concurrency.lockutils [None req-c80cd50f-4127-4451-9298-83a67d8f0ab6 tempest-ServersTestFqdnHostnames-1452522528 tempest-ServersTestFqdnHostnames-1452522528-project-member] Lock "b84c6c8d-e358-4a14-8425-c4f1d8adf4ee" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 71.701s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 618.350699] env[63515]: DEBUG oslo_concurrency.lockutils [None req-22f02712-96d1-4031-8b8f-5bc880f34c33 tempest-ImagesNegativeTestJSON-2041328384 tempest-ImagesNegativeTestJSON-2041328384-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.961s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 618.351432] env[63515]: ERROR nova.compute.manager [None req-22f02712-96d1-4031-8b8f-5bc880f34c33 tempest-ImagesNegativeTestJSON-2041328384 tempest-ImagesNegativeTestJSON-2041328384-project-member] [instance: d5f988f2-765b-4ecf-bad5-4a70fc95d74c] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port fca51b1b-a915-4365-a1b6-ccc0a5a43604, please check neutron logs for more information. [ 618.351432] env[63515]: ERROR nova.compute.manager [instance: d5f988f2-765b-4ecf-bad5-4a70fc95d74c] Traceback (most recent call last): [ 618.351432] env[63515]: ERROR nova.compute.manager [instance: d5f988f2-765b-4ecf-bad5-4a70fc95d74c] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 618.351432] env[63515]: ERROR nova.compute.manager [instance: d5f988f2-765b-4ecf-bad5-4a70fc95d74c] self.driver.spawn(context, instance, image_meta, [ 618.351432] env[63515]: ERROR nova.compute.manager [instance: d5f988f2-765b-4ecf-bad5-4a70fc95d74c] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 618.351432] env[63515]: ERROR nova.compute.manager [instance: d5f988f2-765b-4ecf-bad5-4a70fc95d74c] self._vmops.spawn(context, instance, image_meta, injected_files, [ 618.351432] env[63515]: ERROR nova.compute.manager [instance: d5f988f2-765b-4ecf-bad5-4a70fc95d74c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 618.351432] env[63515]: ERROR nova.compute.manager [instance: d5f988f2-765b-4ecf-bad5-4a70fc95d74c] vm_ref = self.build_virtual_machine(instance, [ 618.351432] env[63515]: ERROR nova.compute.manager [instance: d5f988f2-765b-4ecf-bad5-4a70fc95d74c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 618.351432] env[63515]: ERROR nova.compute.manager [instance: d5f988f2-765b-4ecf-bad5-4a70fc95d74c] vif_infos = vmwarevif.get_vif_info(self._session, [ 618.351432] env[63515]: ERROR nova.compute.manager [instance: d5f988f2-765b-4ecf-bad5-4a70fc95d74c] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 618.351751] env[63515]: ERROR nova.compute.manager [instance: d5f988f2-765b-4ecf-bad5-4a70fc95d74c] for vif in network_info: [ 618.351751] env[63515]: ERROR nova.compute.manager [instance: d5f988f2-765b-4ecf-bad5-4a70fc95d74c] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 618.351751] env[63515]: ERROR nova.compute.manager [instance: d5f988f2-765b-4ecf-bad5-4a70fc95d74c] return self._sync_wrapper(fn, *args, **kwargs) [ 618.351751] env[63515]: ERROR nova.compute.manager [instance: d5f988f2-765b-4ecf-bad5-4a70fc95d74c] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 618.351751] env[63515]: ERROR nova.compute.manager [instance: d5f988f2-765b-4ecf-bad5-4a70fc95d74c] self.wait() [ 618.351751] env[63515]: ERROR nova.compute.manager [instance: d5f988f2-765b-4ecf-bad5-4a70fc95d74c] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 618.351751] env[63515]: ERROR nova.compute.manager [instance: d5f988f2-765b-4ecf-bad5-4a70fc95d74c] self[:] = self._gt.wait() [ 618.351751] env[63515]: ERROR nova.compute.manager [instance: d5f988f2-765b-4ecf-bad5-4a70fc95d74c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 618.351751] env[63515]: ERROR nova.compute.manager [instance: d5f988f2-765b-4ecf-bad5-4a70fc95d74c] return self._exit_event.wait() [ 618.351751] env[63515]: ERROR nova.compute.manager [instance: d5f988f2-765b-4ecf-bad5-4a70fc95d74c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 618.351751] env[63515]: ERROR nova.compute.manager [instance: d5f988f2-765b-4ecf-bad5-4a70fc95d74c] result = hub.switch() [ 618.351751] env[63515]: ERROR nova.compute.manager [instance: d5f988f2-765b-4ecf-bad5-4a70fc95d74c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 618.351751] env[63515]: ERROR nova.compute.manager [instance: d5f988f2-765b-4ecf-bad5-4a70fc95d74c] return self.greenlet.switch() [ 618.352102] env[63515]: ERROR nova.compute.manager [instance: d5f988f2-765b-4ecf-bad5-4a70fc95d74c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 618.352102] env[63515]: ERROR nova.compute.manager [instance: d5f988f2-765b-4ecf-bad5-4a70fc95d74c] result = function(*args, **kwargs) [ 618.352102] env[63515]: ERROR nova.compute.manager [instance: d5f988f2-765b-4ecf-bad5-4a70fc95d74c] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 618.352102] env[63515]: ERROR nova.compute.manager [instance: d5f988f2-765b-4ecf-bad5-4a70fc95d74c] return func(*args, **kwargs) [ 618.352102] env[63515]: ERROR nova.compute.manager [instance: d5f988f2-765b-4ecf-bad5-4a70fc95d74c] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 618.352102] env[63515]: ERROR nova.compute.manager [instance: d5f988f2-765b-4ecf-bad5-4a70fc95d74c] raise e [ 618.352102] env[63515]: ERROR nova.compute.manager [instance: d5f988f2-765b-4ecf-bad5-4a70fc95d74c] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 618.352102] env[63515]: ERROR nova.compute.manager [instance: d5f988f2-765b-4ecf-bad5-4a70fc95d74c] nwinfo = self.network_api.allocate_for_instance( [ 618.352102] env[63515]: ERROR nova.compute.manager [instance: d5f988f2-765b-4ecf-bad5-4a70fc95d74c] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 618.352102] env[63515]: ERROR nova.compute.manager [instance: d5f988f2-765b-4ecf-bad5-4a70fc95d74c] created_port_ids = self._update_ports_for_instance( [ 618.352102] env[63515]: ERROR nova.compute.manager [instance: d5f988f2-765b-4ecf-bad5-4a70fc95d74c] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 618.352102] env[63515]: ERROR nova.compute.manager [instance: d5f988f2-765b-4ecf-bad5-4a70fc95d74c] with excutils.save_and_reraise_exception(): [ 618.352102] env[63515]: ERROR nova.compute.manager [instance: d5f988f2-765b-4ecf-bad5-4a70fc95d74c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 618.352460] env[63515]: ERROR nova.compute.manager [instance: d5f988f2-765b-4ecf-bad5-4a70fc95d74c] self.force_reraise() [ 618.352460] env[63515]: ERROR nova.compute.manager [instance: d5f988f2-765b-4ecf-bad5-4a70fc95d74c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 618.352460] env[63515]: ERROR nova.compute.manager [instance: d5f988f2-765b-4ecf-bad5-4a70fc95d74c] raise self.value [ 618.352460] env[63515]: ERROR nova.compute.manager [instance: d5f988f2-765b-4ecf-bad5-4a70fc95d74c] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 618.352460] env[63515]: ERROR nova.compute.manager [instance: d5f988f2-765b-4ecf-bad5-4a70fc95d74c] updated_port = self._update_port( [ 618.352460] env[63515]: ERROR nova.compute.manager [instance: d5f988f2-765b-4ecf-bad5-4a70fc95d74c] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 618.352460] env[63515]: ERROR nova.compute.manager [instance: d5f988f2-765b-4ecf-bad5-4a70fc95d74c] _ensure_no_port_binding_failure(port) [ 618.352460] env[63515]: ERROR nova.compute.manager [instance: d5f988f2-765b-4ecf-bad5-4a70fc95d74c] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 618.352460] env[63515]: ERROR nova.compute.manager [instance: d5f988f2-765b-4ecf-bad5-4a70fc95d74c] raise exception.PortBindingFailed(port_id=port['id']) [ 618.352460] env[63515]: ERROR nova.compute.manager [instance: d5f988f2-765b-4ecf-bad5-4a70fc95d74c] nova.exception.PortBindingFailed: Binding failed for port fca51b1b-a915-4365-a1b6-ccc0a5a43604, please check neutron logs for more information. [ 618.352460] env[63515]: ERROR nova.compute.manager [instance: d5f988f2-765b-4ecf-bad5-4a70fc95d74c] [ 618.352757] env[63515]: DEBUG nova.compute.utils [None req-22f02712-96d1-4031-8b8f-5bc880f34c33 tempest-ImagesNegativeTestJSON-2041328384 tempest-ImagesNegativeTestJSON-2041328384-project-member] [instance: d5f988f2-765b-4ecf-bad5-4a70fc95d74c] Binding failed for port fca51b1b-a915-4365-a1b6-ccc0a5a43604, please check neutron logs for more information. {{(pid=63515) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 618.353544] env[63515]: DEBUG oslo_concurrency.lockutils [None req-9e973647-d524-4c71-b05a-f1a253758bab tempest-ServerAddressesTestJSON-9863250 tempest-ServerAddressesTestJSON-9863250-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 19.253s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 618.356325] env[63515]: DEBUG nova.compute.manager [None req-22f02712-96d1-4031-8b8f-5bc880f34c33 tempest-ImagesNegativeTestJSON-2041328384 tempest-ImagesNegativeTestJSON-2041328384-project-member] [instance: d5f988f2-765b-4ecf-bad5-4a70fc95d74c] Build of instance d5f988f2-765b-4ecf-bad5-4a70fc95d74c was re-scheduled: Binding failed for port fca51b1b-a915-4365-a1b6-ccc0a5a43604, please check neutron logs for more information. {{(pid=63515) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 618.356787] env[63515]: DEBUG nova.compute.manager [None req-22f02712-96d1-4031-8b8f-5bc880f34c33 tempest-ImagesNegativeTestJSON-2041328384 tempest-ImagesNegativeTestJSON-2041328384-project-member] [instance: d5f988f2-765b-4ecf-bad5-4a70fc95d74c] Unplugging VIFs for instance {{(pid=63515) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 618.357034] env[63515]: DEBUG oslo_concurrency.lockutils [None req-22f02712-96d1-4031-8b8f-5bc880f34c33 tempest-ImagesNegativeTestJSON-2041328384 tempest-ImagesNegativeTestJSON-2041328384-project-member] Acquiring lock "refresh_cache-d5f988f2-765b-4ecf-bad5-4a70fc95d74c" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 618.357191] env[63515]: DEBUG oslo_concurrency.lockutils [None req-22f02712-96d1-4031-8b8f-5bc880f34c33 tempest-ImagesNegativeTestJSON-2041328384 tempest-ImagesNegativeTestJSON-2041328384-project-member] Acquired lock "refresh_cache-d5f988f2-765b-4ecf-bad5-4a70fc95d74c" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 618.357348] env[63515]: DEBUG nova.network.neutron [None req-22f02712-96d1-4031-8b8f-5bc880f34c33 tempest-ImagesNegativeTestJSON-2041328384 tempest-ImagesNegativeTestJSON-2041328384-project-member] [instance: d5f988f2-765b-4ecf-bad5-4a70fc95d74c] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 618.661364] env[63515]: DEBUG nova.compute.manager [None req-30030db8-77b3-4af4-8ae6-30653bd81e23 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: 02ffb723-d6df-42cb-93c5-d582705d1e03] Starting instance... {{(pid=63515) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 618.893013] env[63515]: DEBUG nova.network.neutron [None req-22f02712-96d1-4031-8b8f-5bc880f34c33 tempest-ImagesNegativeTestJSON-2041328384 tempest-ImagesNegativeTestJSON-2041328384-project-member] [instance: d5f988f2-765b-4ecf-bad5-4a70fc95d74c] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 619.050258] env[63515]: DEBUG nova.network.neutron [None req-22f02712-96d1-4031-8b8f-5bc880f34c33 tempest-ImagesNegativeTestJSON-2041328384 tempest-ImagesNegativeTestJSON-2041328384-project-member] [instance: d5f988f2-765b-4ecf-bad5-4a70fc95d74c] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 619.186889] env[63515]: DEBUG oslo_concurrency.lockutils [None req-30030db8-77b3-4af4-8ae6-30653bd81e23 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 619.268949] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e257ea3-16bd-4e71-bb61-e188b012aac8 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.276951] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3720dd16-6ed8-4d7e-a43b-7a8028a2fa6a {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.308783] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ecf51347-8191-48c6-a278-199a28c202cd {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.317491] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dea6a55e-6a48-4625-9aea-3c07ca7a5332 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.331097] env[63515]: DEBUG nova.compute.provider_tree [None req-9e973647-d524-4c71-b05a-f1a253758bab tempest-ServerAddressesTestJSON-9863250 tempest-ServerAddressesTestJSON-9863250-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 619.552617] env[63515]: DEBUG oslo_concurrency.lockutils [None req-22f02712-96d1-4031-8b8f-5bc880f34c33 tempest-ImagesNegativeTestJSON-2041328384 tempest-ImagesNegativeTestJSON-2041328384-project-member] Releasing lock "refresh_cache-d5f988f2-765b-4ecf-bad5-4a70fc95d74c" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 619.552896] env[63515]: DEBUG nova.compute.manager [None req-22f02712-96d1-4031-8b8f-5bc880f34c33 tempest-ImagesNegativeTestJSON-2041328384 tempest-ImagesNegativeTestJSON-2041328384-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63515) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 619.553089] env[63515]: DEBUG nova.compute.manager [None req-22f02712-96d1-4031-8b8f-5bc880f34c33 tempest-ImagesNegativeTestJSON-2041328384 tempest-ImagesNegativeTestJSON-2041328384-project-member] [instance: d5f988f2-765b-4ecf-bad5-4a70fc95d74c] Deallocating network for instance {{(pid=63515) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 619.553261] env[63515]: DEBUG nova.network.neutron [None req-22f02712-96d1-4031-8b8f-5bc880f34c33 tempest-ImagesNegativeTestJSON-2041328384 tempest-ImagesNegativeTestJSON-2041328384-project-member] [instance: d5f988f2-765b-4ecf-bad5-4a70fc95d74c] deallocate_for_instance() {{(pid=63515) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 619.571020] env[63515]: DEBUG nova.network.neutron [None req-22f02712-96d1-4031-8b8f-5bc880f34c33 tempest-ImagesNegativeTestJSON-2041328384 tempest-ImagesNegativeTestJSON-2041328384-project-member] [instance: d5f988f2-765b-4ecf-bad5-4a70fc95d74c] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 619.837222] env[63515]: DEBUG nova.scheduler.client.report [None req-9e973647-d524-4c71-b05a-f1a253758bab tempest-ServerAddressesTestJSON-9863250 tempest-ServerAddressesTestJSON-9863250-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 620.075991] env[63515]: DEBUG nova.network.neutron [None req-22f02712-96d1-4031-8b8f-5bc880f34c33 tempest-ImagesNegativeTestJSON-2041328384 tempest-ImagesNegativeTestJSON-2041328384-project-member] [instance: d5f988f2-765b-4ecf-bad5-4a70fc95d74c] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 620.343304] env[63515]: DEBUG oslo_concurrency.lockutils [None req-9e973647-d524-4c71-b05a-f1a253758bab tempest-ServerAddressesTestJSON-9863250 tempest-ServerAddressesTestJSON-9863250-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.986s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 620.343304] env[63515]: ERROR nova.compute.manager [None req-9e973647-d524-4c71-b05a-f1a253758bab tempest-ServerAddressesTestJSON-9863250 tempest-ServerAddressesTestJSON-9863250-project-member] [instance: 7fee0afb-88c9-4e93-a4d9-94312394f927] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 79794b46-d801-404a-94f2-a223e825fcfc, please check neutron logs for more information. [ 620.343304] env[63515]: ERROR nova.compute.manager [instance: 7fee0afb-88c9-4e93-a4d9-94312394f927] Traceback (most recent call last): [ 620.343304] env[63515]: ERROR nova.compute.manager [instance: 7fee0afb-88c9-4e93-a4d9-94312394f927] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 620.343304] env[63515]: ERROR nova.compute.manager [instance: 7fee0afb-88c9-4e93-a4d9-94312394f927] self.driver.spawn(context, instance, image_meta, [ 620.343304] env[63515]: ERROR nova.compute.manager [instance: 7fee0afb-88c9-4e93-a4d9-94312394f927] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 620.343304] env[63515]: ERROR nova.compute.manager [instance: 7fee0afb-88c9-4e93-a4d9-94312394f927] self._vmops.spawn(context, instance, image_meta, injected_files, [ 620.343304] env[63515]: ERROR nova.compute.manager [instance: 7fee0afb-88c9-4e93-a4d9-94312394f927] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 620.343304] env[63515]: ERROR nova.compute.manager [instance: 7fee0afb-88c9-4e93-a4d9-94312394f927] vm_ref = self.build_virtual_machine(instance, [ 620.343522] env[63515]: ERROR nova.compute.manager [instance: 7fee0afb-88c9-4e93-a4d9-94312394f927] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 620.343522] env[63515]: ERROR nova.compute.manager [instance: 7fee0afb-88c9-4e93-a4d9-94312394f927] vif_infos = vmwarevif.get_vif_info(self._session, [ 620.343522] env[63515]: ERROR nova.compute.manager [instance: 7fee0afb-88c9-4e93-a4d9-94312394f927] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 620.343522] env[63515]: ERROR nova.compute.manager [instance: 7fee0afb-88c9-4e93-a4d9-94312394f927] for vif in network_info: [ 620.343522] env[63515]: ERROR nova.compute.manager [instance: 7fee0afb-88c9-4e93-a4d9-94312394f927] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 620.343522] env[63515]: ERROR nova.compute.manager [instance: 7fee0afb-88c9-4e93-a4d9-94312394f927] return self._sync_wrapper(fn, *args, **kwargs) [ 620.343522] env[63515]: ERROR nova.compute.manager [instance: 7fee0afb-88c9-4e93-a4d9-94312394f927] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 620.343522] env[63515]: ERROR nova.compute.manager [instance: 7fee0afb-88c9-4e93-a4d9-94312394f927] self.wait() [ 620.343522] env[63515]: ERROR nova.compute.manager [instance: 7fee0afb-88c9-4e93-a4d9-94312394f927] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 620.343522] env[63515]: ERROR nova.compute.manager [instance: 7fee0afb-88c9-4e93-a4d9-94312394f927] self[:] = self._gt.wait() [ 620.343522] env[63515]: ERROR nova.compute.manager [instance: 7fee0afb-88c9-4e93-a4d9-94312394f927] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 620.343522] env[63515]: ERROR nova.compute.manager [instance: 7fee0afb-88c9-4e93-a4d9-94312394f927] return self._exit_event.wait() [ 620.343522] env[63515]: ERROR nova.compute.manager [instance: 7fee0afb-88c9-4e93-a4d9-94312394f927] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 620.343828] env[63515]: ERROR nova.compute.manager [instance: 7fee0afb-88c9-4e93-a4d9-94312394f927] result = hub.switch() [ 620.343828] env[63515]: ERROR nova.compute.manager [instance: 7fee0afb-88c9-4e93-a4d9-94312394f927] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 620.343828] env[63515]: ERROR nova.compute.manager [instance: 7fee0afb-88c9-4e93-a4d9-94312394f927] return self.greenlet.switch() [ 620.343828] env[63515]: ERROR nova.compute.manager [instance: 7fee0afb-88c9-4e93-a4d9-94312394f927] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 620.343828] env[63515]: ERROR nova.compute.manager [instance: 7fee0afb-88c9-4e93-a4d9-94312394f927] result = function(*args, **kwargs) [ 620.343828] env[63515]: ERROR nova.compute.manager [instance: 7fee0afb-88c9-4e93-a4d9-94312394f927] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 620.343828] env[63515]: ERROR nova.compute.manager [instance: 7fee0afb-88c9-4e93-a4d9-94312394f927] return func(*args, **kwargs) [ 620.343828] env[63515]: ERROR nova.compute.manager [instance: 7fee0afb-88c9-4e93-a4d9-94312394f927] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 620.343828] env[63515]: ERROR nova.compute.manager [instance: 7fee0afb-88c9-4e93-a4d9-94312394f927] raise e [ 620.343828] env[63515]: ERROR nova.compute.manager [instance: 7fee0afb-88c9-4e93-a4d9-94312394f927] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 620.343828] env[63515]: ERROR nova.compute.manager [instance: 7fee0afb-88c9-4e93-a4d9-94312394f927] nwinfo = self.network_api.allocate_for_instance( [ 620.343828] env[63515]: ERROR nova.compute.manager [instance: 7fee0afb-88c9-4e93-a4d9-94312394f927] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 620.343828] env[63515]: ERROR nova.compute.manager [instance: 7fee0afb-88c9-4e93-a4d9-94312394f927] created_port_ids = self._update_ports_for_instance( [ 620.344120] env[63515]: ERROR nova.compute.manager [instance: 7fee0afb-88c9-4e93-a4d9-94312394f927] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 620.344120] env[63515]: ERROR nova.compute.manager [instance: 7fee0afb-88c9-4e93-a4d9-94312394f927] with excutils.save_and_reraise_exception(): [ 620.344120] env[63515]: ERROR nova.compute.manager [instance: 7fee0afb-88c9-4e93-a4d9-94312394f927] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 620.344120] env[63515]: ERROR nova.compute.manager [instance: 7fee0afb-88c9-4e93-a4d9-94312394f927] self.force_reraise() [ 620.344120] env[63515]: ERROR nova.compute.manager [instance: 7fee0afb-88c9-4e93-a4d9-94312394f927] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 620.344120] env[63515]: ERROR nova.compute.manager [instance: 7fee0afb-88c9-4e93-a4d9-94312394f927] raise self.value [ 620.344120] env[63515]: ERROR nova.compute.manager [instance: 7fee0afb-88c9-4e93-a4d9-94312394f927] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 620.344120] env[63515]: ERROR nova.compute.manager [instance: 7fee0afb-88c9-4e93-a4d9-94312394f927] updated_port = self._update_port( [ 620.344120] env[63515]: ERROR nova.compute.manager [instance: 7fee0afb-88c9-4e93-a4d9-94312394f927] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 620.344120] env[63515]: ERROR nova.compute.manager [instance: 7fee0afb-88c9-4e93-a4d9-94312394f927] _ensure_no_port_binding_failure(port) [ 620.344120] env[63515]: ERROR nova.compute.manager [instance: 7fee0afb-88c9-4e93-a4d9-94312394f927] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 620.344120] env[63515]: ERROR nova.compute.manager [instance: 7fee0afb-88c9-4e93-a4d9-94312394f927] raise exception.PortBindingFailed(port_id=port['id']) [ 620.344387] env[63515]: ERROR nova.compute.manager [instance: 7fee0afb-88c9-4e93-a4d9-94312394f927] nova.exception.PortBindingFailed: Binding failed for port 79794b46-d801-404a-94f2-a223e825fcfc, please check neutron logs for more information. [ 620.344387] env[63515]: ERROR nova.compute.manager [instance: 7fee0afb-88c9-4e93-a4d9-94312394f927] [ 620.344387] env[63515]: DEBUG nova.compute.utils [None req-9e973647-d524-4c71-b05a-f1a253758bab tempest-ServerAddressesTestJSON-9863250 tempest-ServerAddressesTestJSON-9863250-project-member] [instance: 7fee0afb-88c9-4e93-a4d9-94312394f927] Binding failed for port 79794b46-d801-404a-94f2-a223e825fcfc, please check neutron logs for more information. {{(pid=63515) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 620.344387] env[63515]: DEBUG oslo_concurrency.lockutils [None req-9131a685-703c-44a1-81f8-7c157d95e037 tempest-ServerAddressesNegativeTestJSON-1262066083 tempest-ServerAddressesNegativeTestJSON-1262066083-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.320s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 620.348029] env[63515]: INFO nova.compute.claims [None req-9131a685-703c-44a1-81f8-7c157d95e037 tempest-ServerAddressesNegativeTestJSON-1262066083 tempest-ServerAddressesNegativeTestJSON-1262066083-project-member] [instance: 5b201ab6-c7ec-4a5c-a310-982649f34be4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 620.352021] env[63515]: DEBUG nova.compute.manager [None req-9e973647-d524-4c71-b05a-f1a253758bab tempest-ServerAddressesTestJSON-9863250 tempest-ServerAddressesTestJSON-9863250-project-member] [instance: 7fee0afb-88c9-4e93-a4d9-94312394f927] Build of instance 7fee0afb-88c9-4e93-a4d9-94312394f927 was re-scheduled: Binding failed for port 79794b46-d801-404a-94f2-a223e825fcfc, please check neutron logs for more information. {{(pid=63515) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 620.352021] env[63515]: DEBUG nova.compute.manager [None req-9e973647-d524-4c71-b05a-f1a253758bab tempest-ServerAddressesTestJSON-9863250 tempest-ServerAddressesTestJSON-9863250-project-member] [instance: 7fee0afb-88c9-4e93-a4d9-94312394f927] Unplugging VIFs for instance {{(pid=63515) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 620.352021] env[63515]: DEBUG oslo_concurrency.lockutils [None req-9e973647-d524-4c71-b05a-f1a253758bab tempest-ServerAddressesTestJSON-9863250 tempest-ServerAddressesTestJSON-9863250-project-member] Acquiring lock "refresh_cache-7fee0afb-88c9-4e93-a4d9-94312394f927" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 620.352021] env[63515]: DEBUG oslo_concurrency.lockutils [None req-9e973647-d524-4c71-b05a-f1a253758bab tempest-ServerAddressesTestJSON-9863250 tempest-ServerAddressesTestJSON-9863250-project-member] Acquired lock "refresh_cache-7fee0afb-88c9-4e93-a4d9-94312394f927" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 620.352271] env[63515]: DEBUG nova.network.neutron [None req-9e973647-d524-4c71-b05a-f1a253758bab tempest-ServerAddressesTestJSON-9863250 tempest-ServerAddressesTestJSON-9863250-project-member] [instance: 7fee0afb-88c9-4e93-a4d9-94312394f927] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 621.051041] env[63515]: INFO nova.compute.manager [None req-22f02712-96d1-4031-8b8f-5bc880f34c33 tempest-ImagesNegativeTestJSON-2041328384 tempest-ImagesNegativeTestJSON-2041328384-project-member] [instance: d5f988f2-765b-4ecf-bad5-4a70fc95d74c] Took 1.50 seconds to deallocate network for instance. [ 621.083888] env[63515]: DEBUG nova.network.neutron [None req-9e973647-d524-4c71-b05a-f1a253758bab tempest-ServerAddressesTestJSON-9863250 tempest-ServerAddressesTestJSON-9863250-project-member] [instance: 7fee0afb-88c9-4e93-a4d9-94312394f927] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 621.273571] env[63515]: DEBUG nova.network.neutron [None req-9e973647-d524-4c71-b05a-f1a253758bab tempest-ServerAddressesTestJSON-9863250 tempest-ServerAddressesTestJSON-9863250-project-member] [instance: 7fee0afb-88c9-4e93-a4d9-94312394f927] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 621.776390] env[63515]: DEBUG oslo_concurrency.lockutils [None req-9e973647-d524-4c71-b05a-f1a253758bab tempest-ServerAddressesTestJSON-9863250 tempest-ServerAddressesTestJSON-9863250-project-member] Releasing lock "refresh_cache-7fee0afb-88c9-4e93-a4d9-94312394f927" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 621.776639] env[63515]: DEBUG nova.compute.manager [None req-9e973647-d524-4c71-b05a-f1a253758bab tempest-ServerAddressesTestJSON-9863250 tempest-ServerAddressesTestJSON-9863250-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63515) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 621.776819] env[63515]: DEBUG nova.compute.manager [None req-9e973647-d524-4c71-b05a-f1a253758bab tempest-ServerAddressesTestJSON-9863250 tempest-ServerAddressesTestJSON-9863250-project-member] [instance: 7fee0afb-88c9-4e93-a4d9-94312394f927] Deallocating network for instance {{(pid=63515) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 621.776982] env[63515]: DEBUG nova.network.neutron [None req-9e973647-d524-4c71-b05a-f1a253758bab tempest-ServerAddressesTestJSON-9863250 tempest-ServerAddressesTestJSON-9863250-project-member] [instance: 7fee0afb-88c9-4e93-a4d9-94312394f927] deallocate_for_instance() {{(pid=63515) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 621.801199] env[63515]: DEBUG nova.network.neutron [None req-9e973647-d524-4c71-b05a-f1a253758bab tempest-ServerAddressesTestJSON-9863250 tempest-ServerAddressesTestJSON-9863250-project-member] [instance: 7fee0afb-88c9-4e93-a4d9-94312394f927] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 621.989042] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1e11ae4-66a7-4bfa-bb18-670e6e7db1c1 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.002503] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89b7897d-8639-41aa-bd9d-76cb0176f0f4 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.045942] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d06f36fd-173b-4d2c-865c-8ce16fd979a6 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.057608] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2fa1533d-d765-44fc-9789-bfadff435da3 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.073096] env[63515]: DEBUG nova.compute.provider_tree [None req-9131a685-703c-44a1-81f8-7c157d95e037 tempest-ServerAddressesNegativeTestJSON-1262066083 tempest-ServerAddressesNegativeTestJSON-1262066083-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 622.103944] env[63515]: INFO nova.scheduler.client.report [None req-22f02712-96d1-4031-8b8f-5bc880f34c33 tempest-ImagesNegativeTestJSON-2041328384 tempest-ImagesNegativeTestJSON-2041328384-project-member] Deleted allocations for instance d5f988f2-765b-4ecf-bad5-4a70fc95d74c [ 622.310667] env[63515]: DEBUG nova.network.neutron [None req-9e973647-d524-4c71-b05a-f1a253758bab tempest-ServerAddressesTestJSON-9863250 tempest-ServerAddressesTestJSON-9863250-project-member] [instance: 7fee0afb-88c9-4e93-a4d9-94312394f927] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 622.576716] env[63515]: DEBUG nova.scheduler.client.report [None req-9131a685-703c-44a1-81f8-7c157d95e037 tempest-ServerAddressesNegativeTestJSON-1262066083 tempest-ServerAddressesNegativeTestJSON-1262066083-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 622.617535] env[63515]: DEBUG oslo_concurrency.lockutils [None req-22f02712-96d1-4031-8b8f-5bc880f34c33 tempest-ImagesNegativeTestJSON-2041328384 tempest-ImagesNegativeTestJSON-2041328384-project-member] Lock "d5f988f2-765b-4ecf-bad5-4a70fc95d74c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 74.251s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 622.817242] env[63515]: INFO nova.compute.manager [None req-9e973647-d524-4c71-b05a-f1a253758bab tempest-ServerAddressesTestJSON-9863250 tempest-ServerAddressesTestJSON-9863250-project-member] [instance: 7fee0afb-88c9-4e93-a4d9-94312394f927] Took 1.04 seconds to deallocate network for instance. [ 623.086301] env[63515]: DEBUG oslo_concurrency.lockutils [None req-9131a685-703c-44a1-81f8-7c157d95e037 tempest-ServerAddressesNegativeTestJSON-1262066083 tempest-ServerAddressesNegativeTestJSON-1262066083-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.743s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 623.086630] env[63515]: DEBUG nova.compute.manager [None req-9131a685-703c-44a1-81f8-7c157d95e037 tempest-ServerAddressesNegativeTestJSON-1262066083 tempest-ServerAddressesNegativeTestJSON-1262066083-project-member] [instance: 5b201ab6-c7ec-4a5c-a310-982649f34be4] Start building networks asynchronously for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 623.089222] env[63515]: DEBUG oslo_concurrency.lockutils [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 19.489s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 623.089401] env[63515]: DEBUG oslo_concurrency.lockutils [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 623.089582] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=63515) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 623.089841] env[63515]: DEBUG oslo_concurrency.lockutils [None req-24636c84-3869-4a1b-acfb-9c10d33a9826 tempest-ServersWithSpecificFlavorTestJSON-691988195 tempest-ServersWithSpecificFlavorTestJSON-691988195-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 18.038s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 623.099014] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a31a478-23b5-462a-bb86-4879962c145e {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.105903] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e67651d6-4f66-43b7-b276-a3afee35c3ab {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.120227] env[63515]: DEBUG nova.compute.manager [None req-30987b61-ce89-4676-89be-fee7786462c9 tempest-ImagesOneServerNegativeTestJSON-52225474 tempest-ImagesOneServerNegativeTestJSON-52225474-project-member] [instance: bbce1137-d691-4633-87ff-f9f4ea257ed7] Starting instance... {{(pid=63515) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 623.123636] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d80a2061-abda-4ee1-a5ed-bdfc2efe2313 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.130717] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70e47407-638f-43fc-b149-65baeb7aa096 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.169124] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181585MB free_disk=171GB free_vcpus=48 pci_devices=None {{(pid=63515) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 623.169124] env[63515]: DEBUG oslo_concurrency.lockutils [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 623.236993] env[63515]: DEBUG oslo_concurrency.lockutils [None req-e71315b0-9587-4929-82a2-71bebc6fcbfc tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Acquiring lock "d7e8dcdb-41cb-46fb-8b61-d251e7c2d372" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 623.236993] env[63515]: DEBUG oslo_concurrency.lockutils [None req-e71315b0-9587-4929-82a2-71bebc6fcbfc tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Lock "d7e8dcdb-41cb-46fb-8b61-d251e7c2d372" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 623.596073] env[63515]: DEBUG nova.compute.utils [None req-9131a685-703c-44a1-81f8-7c157d95e037 tempest-ServerAddressesNegativeTestJSON-1262066083 tempest-ServerAddressesNegativeTestJSON-1262066083-project-member] Using /dev/sd instead of None {{(pid=63515) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 623.596656] env[63515]: DEBUG nova.compute.manager [None req-9131a685-703c-44a1-81f8-7c157d95e037 tempest-ServerAddressesNegativeTestJSON-1262066083 tempest-ServerAddressesNegativeTestJSON-1262066083-project-member] [instance: 5b201ab6-c7ec-4a5c-a310-982649f34be4] Allocating IP information in the background. {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 623.603018] env[63515]: DEBUG nova.network.neutron [None req-9131a685-703c-44a1-81f8-7c157d95e037 tempest-ServerAddressesNegativeTestJSON-1262066083 tempest-ServerAddressesNegativeTestJSON-1262066083-project-member] [instance: 5b201ab6-c7ec-4a5c-a310-982649f34be4] allocate_for_instance() {{(pid=63515) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 623.652892] env[63515]: DEBUG oslo_concurrency.lockutils [None req-30987b61-ce89-4676-89be-fee7786462c9 tempest-ImagesOneServerNegativeTestJSON-52225474 tempest-ImagesOneServerNegativeTestJSON-52225474-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 623.694183] env[63515]: DEBUG nova.policy [None req-9131a685-703c-44a1-81f8-7c157d95e037 tempest-ServerAddressesNegativeTestJSON-1262066083 tempest-ServerAddressesNegativeTestJSON-1262066083-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '37482b8eeb3d4e36b93f3b27d48a4207', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1873b3bcf9d24776a198aaebeb87ed4c', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63515) authorize /opt/stack/nova/nova/policy.py:201}} [ 623.850726] env[63515]: INFO nova.scheduler.client.report [None req-9e973647-d524-4c71-b05a-f1a253758bab tempest-ServerAddressesTestJSON-9863250 tempest-ServerAddressesTestJSON-9863250-project-member] Deleted allocations for instance 7fee0afb-88c9-4e93-a4d9-94312394f927 [ 624.039570] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-343bda48-bfb3-45aa-9be2-ee1ceeed616f {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.050102] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b04e84a8-2e3e-48c5-8381-5fe70314b0ce {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.085507] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b2dc270-7bdf-4f9d-9525-b4d56a3bcdef {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.093126] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8f75a0a-4788-427c-bc57-0adc4265defb {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.107575] env[63515]: DEBUG nova.compute.manager [None req-9131a685-703c-44a1-81f8-7c157d95e037 tempest-ServerAddressesNegativeTestJSON-1262066083 tempest-ServerAddressesNegativeTestJSON-1262066083-project-member] [instance: 5b201ab6-c7ec-4a5c-a310-982649f34be4] Start building block device mappings for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 624.110705] env[63515]: DEBUG nova.compute.provider_tree [None req-24636c84-3869-4a1b-acfb-9c10d33a9826 tempest-ServersWithSpecificFlavorTestJSON-691988195 tempest-ServersWithSpecificFlavorTestJSON-691988195-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 624.330621] env[63515]: DEBUG nova.network.neutron [None req-9131a685-703c-44a1-81f8-7c157d95e037 tempest-ServerAddressesNegativeTestJSON-1262066083 tempest-ServerAddressesNegativeTestJSON-1262066083-project-member] [instance: 5b201ab6-c7ec-4a5c-a310-982649f34be4] Successfully created port: 4d81feb5-11c1-415a-bdaf-cd39beb2e524 {{(pid=63515) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 624.364606] env[63515]: DEBUG oslo_concurrency.lockutils [None req-9e973647-d524-4c71-b05a-f1a253758bab tempest-ServerAddressesTestJSON-9863250 tempest-ServerAddressesTestJSON-9863250-project-member] Lock "7fee0afb-88c9-4e93-a4d9-94312394f927" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 75.076s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 624.619135] env[63515]: DEBUG nova.scheduler.client.report [None req-24636c84-3869-4a1b-acfb-9c10d33a9826 tempest-ServersWithSpecificFlavorTestJSON-691988195 tempest-ServersWithSpecificFlavorTestJSON-691988195-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 624.869030] env[63515]: DEBUG nova.compute.manager [None req-480b7d6d-2f7f-4da3-9eb1-818f0711a940 tempest-ServerRescueTestJSON-1961208277 tempest-ServerRescueTestJSON-1961208277-project-member] [instance: 7f49d79e-3bda-4949-a976-7e3e6513b2c7] Starting instance... {{(pid=63515) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 625.126716] env[63515]: DEBUG nova.compute.manager [None req-9131a685-703c-44a1-81f8-7c157d95e037 tempest-ServerAddressesNegativeTestJSON-1262066083 tempest-ServerAddressesNegativeTestJSON-1262066083-project-member] [instance: 5b201ab6-c7ec-4a5c-a310-982649f34be4] Start spawning the instance on the hypervisor. {{(pid=63515) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 625.132994] env[63515]: DEBUG oslo_concurrency.lockutils [None req-24636c84-3869-4a1b-acfb-9c10d33a9826 tempest-ServersWithSpecificFlavorTestJSON-691988195 tempest-ServersWithSpecificFlavorTestJSON-691988195-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.043s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 625.133970] env[63515]: ERROR nova.compute.manager [None req-24636c84-3869-4a1b-acfb-9c10d33a9826 tempest-ServersWithSpecificFlavorTestJSON-691988195 tempest-ServersWithSpecificFlavorTestJSON-691988195-project-member] [instance: ebc737a3-e6cd-4f4c-a4b6-fce5924de4c6] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 168387a2-8a0f-41c4-b727-bad7d761613f, please check neutron logs for more information. [ 625.133970] env[63515]: ERROR nova.compute.manager [instance: ebc737a3-e6cd-4f4c-a4b6-fce5924de4c6] Traceback (most recent call last): [ 625.133970] env[63515]: ERROR nova.compute.manager [instance: ebc737a3-e6cd-4f4c-a4b6-fce5924de4c6] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 625.133970] env[63515]: ERROR nova.compute.manager [instance: ebc737a3-e6cd-4f4c-a4b6-fce5924de4c6] self.driver.spawn(context, instance, image_meta, [ 625.133970] env[63515]: ERROR nova.compute.manager [instance: ebc737a3-e6cd-4f4c-a4b6-fce5924de4c6] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 625.133970] env[63515]: ERROR nova.compute.manager [instance: ebc737a3-e6cd-4f4c-a4b6-fce5924de4c6] self._vmops.spawn(context, instance, image_meta, injected_files, [ 625.133970] env[63515]: ERROR nova.compute.manager [instance: ebc737a3-e6cd-4f4c-a4b6-fce5924de4c6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 625.133970] env[63515]: ERROR nova.compute.manager [instance: ebc737a3-e6cd-4f4c-a4b6-fce5924de4c6] vm_ref = self.build_virtual_machine(instance, [ 625.133970] env[63515]: ERROR nova.compute.manager [instance: ebc737a3-e6cd-4f4c-a4b6-fce5924de4c6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 625.133970] env[63515]: ERROR nova.compute.manager [instance: ebc737a3-e6cd-4f4c-a4b6-fce5924de4c6] vif_infos = vmwarevif.get_vif_info(self._session, [ 625.133970] env[63515]: ERROR nova.compute.manager [instance: ebc737a3-e6cd-4f4c-a4b6-fce5924de4c6] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 625.134308] env[63515]: ERROR nova.compute.manager [instance: ebc737a3-e6cd-4f4c-a4b6-fce5924de4c6] for vif in network_info: [ 625.134308] env[63515]: ERROR nova.compute.manager [instance: ebc737a3-e6cd-4f4c-a4b6-fce5924de4c6] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 625.134308] env[63515]: ERROR nova.compute.manager [instance: ebc737a3-e6cd-4f4c-a4b6-fce5924de4c6] return self._sync_wrapper(fn, *args, **kwargs) [ 625.134308] env[63515]: ERROR nova.compute.manager [instance: ebc737a3-e6cd-4f4c-a4b6-fce5924de4c6] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 625.134308] env[63515]: ERROR nova.compute.manager [instance: ebc737a3-e6cd-4f4c-a4b6-fce5924de4c6] self.wait() [ 625.134308] env[63515]: ERROR nova.compute.manager [instance: ebc737a3-e6cd-4f4c-a4b6-fce5924de4c6] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 625.134308] env[63515]: ERROR nova.compute.manager [instance: ebc737a3-e6cd-4f4c-a4b6-fce5924de4c6] self[:] = self._gt.wait() [ 625.134308] env[63515]: ERROR nova.compute.manager [instance: ebc737a3-e6cd-4f4c-a4b6-fce5924de4c6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 625.134308] env[63515]: ERROR nova.compute.manager [instance: ebc737a3-e6cd-4f4c-a4b6-fce5924de4c6] return self._exit_event.wait() [ 625.134308] env[63515]: ERROR nova.compute.manager [instance: ebc737a3-e6cd-4f4c-a4b6-fce5924de4c6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 625.134308] env[63515]: ERROR nova.compute.manager [instance: ebc737a3-e6cd-4f4c-a4b6-fce5924de4c6] result = hub.switch() [ 625.134308] env[63515]: ERROR nova.compute.manager [instance: ebc737a3-e6cd-4f4c-a4b6-fce5924de4c6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 625.134308] env[63515]: ERROR nova.compute.manager [instance: ebc737a3-e6cd-4f4c-a4b6-fce5924de4c6] return self.greenlet.switch() [ 625.134624] env[63515]: ERROR nova.compute.manager [instance: ebc737a3-e6cd-4f4c-a4b6-fce5924de4c6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 625.134624] env[63515]: ERROR nova.compute.manager [instance: ebc737a3-e6cd-4f4c-a4b6-fce5924de4c6] result = function(*args, **kwargs) [ 625.134624] env[63515]: ERROR nova.compute.manager [instance: ebc737a3-e6cd-4f4c-a4b6-fce5924de4c6] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 625.134624] env[63515]: ERROR nova.compute.manager [instance: ebc737a3-e6cd-4f4c-a4b6-fce5924de4c6] return func(*args, **kwargs) [ 625.134624] env[63515]: ERROR nova.compute.manager [instance: ebc737a3-e6cd-4f4c-a4b6-fce5924de4c6] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 625.134624] env[63515]: ERROR nova.compute.manager [instance: ebc737a3-e6cd-4f4c-a4b6-fce5924de4c6] raise e [ 625.134624] env[63515]: ERROR nova.compute.manager [instance: ebc737a3-e6cd-4f4c-a4b6-fce5924de4c6] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 625.134624] env[63515]: ERROR nova.compute.manager [instance: ebc737a3-e6cd-4f4c-a4b6-fce5924de4c6] nwinfo = self.network_api.allocate_for_instance( [ 625.134624] env[63515]: ERROR nova.compute.manager [instance: ebc737a3-e6cd-4f4c-a4b6-fce5924de4c6] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 625.134624] env[63515]: ERROR nova.compute.manager [instance: ebc737a3-e6cd-4f4c-a4b6-fce5924de4c6] created_port_ids = self._update_ports_for_instance( [ 625.134624] env[63515]: ERROR nova.compute.manager [instance: ebc737a3-e6cd-4f4c-a4b6-fce5924de4c6] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 625.134624] env[63515]: ERROR nova.compute.manager [instance: ebc737a3-e6cd-4f4c-a4b6-fce5924de4c6] with excutils.save_and_reraise_exception(): [ 625.134624] env[63515]: ERROR nova.compute.manager [instance: ebc737a3-e6cd-4f4c-a4b6-fce5924de4c6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 625.134905] env[63515]: ERROR nova.compute.manager [instance: ebc737a3-e6cd-4f4c-a4b6-fce5924de4c6] self.force_reraise() [ 625.134905] env[63515]: ERROR nova.compute.manager [instance: ebc737a3-e6cd-4f4c-a4b6-fce5924de4c6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 625.134905] env[63515]: ERROR nova.compute.manager [instance: ebc737a3-e6cd-4f4c-a4b6-fce5924de4c6] raise self.value [ 625.134905] env[63515]: ERROR nova.compute.manager [instance: ebc737a3-e6cd-4f4c-a4b6-fce5924de4c6] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 625.134905] env[63515]: ERROR nova.compute.manager [instance: ebc737a3-e6cd-4f4c-a4b6-fce5924de4c6] updated_port = self._update_port( [ 625.134905] env[63515]: ERROR nova.compute.manager [instance: ebc737a3-e6cd-4f4c-a4b6-fce5924de4c6] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 625.134905] env[63515]: ERROR nova.compute.manager [instance: ebc737a3-e6cd-4f4c-a4b6-fce5924de4c6] _ensure_no_port_binding_failure(port) [ 625.134905] env[63515]: ERROR nova.compute.manager [instance: ebc737a3-e6cd-4f4c-a4b6-fce5924de4c6] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 625.134905] env[63515]: ERROR nova.compute.manager [instance: ebc737a3-e6cd-4f4c-a4b6-fce5924de4c6] raise exception.PortBindingFailed(port_id=port['id']) [ 625.134905] env[63515]: ERROR nova.compute.manager [instance: ebc737a3-e6cd-4f4c-a4b6-fce5924de4c6] nova.exception.PortBindingFailed: Binding failed for port 168387a2-8a0f-41c4-b727-bad7d761613f, please check neutron logs for more information. [ 625.134905] env[63515]: ERROR nova.compute.manager [instance: ebc737a3-e6cd-4f4c-a4b6-fce5924de4c6] [ 625.135159] env[63515]: DEBUG nova.compute.utils [None req-24636c84-3869-4a1b-acfb-9c10d33a9826 tempest-ServersWithSpecificFlavorTestJSON-691988195 tempest-ServersWithSpecificFlavorTestJSON-691988195-project-member] [instance: ebc737a3-e6cd-4f4c-a4b6-fce5924de4c6] Binding failed for port 168387a2-8a0f-41c4-b727-bad7d761613f, please check neutron logs for more information. {{(pid=63515) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 625.138194] env[63515]: DEBUG oslo_concurrency.lockutils [None req-292f2bc3-2911-4498-95fb-4e4a03681601 tempest-FloatingIPsAssociationTestJSON-629096942 tempest-FloatingIPsAssociationTestJSON-629096942-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.446s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 625.139733] env[63515]: INFO nova.compute.claims [None req-292f2bc3-2911-4498-95fb-4e4a03681601 tempest-FloatingIPsAssociationTestJSON-629096942 tempest-FloatingIPsAssociationTestJSON-629096942-project-member] [instance: 2728c9f9-0d03-4313-ba2a-3dc22aff8538] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 625.143415] env[63515]: DEBUG nova.compute.manager [None req-24636c84-3869-4a1b-acfb-9c10d33a9826 tempest-ServersWithSpecificFlavorTestJSON-691988195 tempest-ServersWithSpecificFlavorTestJSON-691988195-project-member] [instance: ebc737a3-e6cd-4f4c-a4b6-fce5924de4c6] Build of instance ebc737a3-e6cd-4f4c-a4b6-fce5924de4c6 was re-scheduled: Binding failed for port 168387a2-8a0f-41c4-b727-bad7d761613f, please check neutron logs for more information. {{(pid=63515) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 625.143945] env[63515]: DEBUG nova.compute.manager [None req-24636c84-3869-4a1b-acfb-9c10d33a9826 tempest-ServersWithSpecificFlavorTestJSON-691988195 tempest-ServersWithSpecificFlavorTestJSON-691988195-project-member] [instance: ebc737a3-e6cd-4f4c-a4b6-fce5924de4c6] Unplugging VIFs for instance {{(pid=63515) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 625.144548] env[63515]: DEBUG oslo_concurrency.lockutils [None req-24636c84-3869-4a1b-acfb-9c10d33a9826 tempest-ServersWithSpecificFlavorTestJSON-691988195 tempest-ServersWithSpecificFlavorTestJSON-691988195-project-member] Acquiring lock "refresh_cache-ebc737a3-e6cd-4f4c-a4b6-fce5924de4c6" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 625.144548] env[63515]: DEBUG oslo_concurrency.lockutils [None req-24636c84-3869-4a1b-acfb-9c10d33a9826 tempest-ServersWithSpecificFlavorTestJSON-691988195 tempest-ServersWithSpecificFlavorTestJSON-691988195-project-member] Acquired lock "refresh_cache-ebc737a3-e6cd-4f4c-a4b6-fce5924de4c6" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 625.144548] env[63515]: DEBUG nova.network.neutron [None req-24636c84-3869-4a1b-acfb-9c10d33a9826 tempest-ServersWithSpecificFlavorTestJSON-691988195 tempest-ServersWithSpecificFlavorTestJSON-691988195-project-member] [instance: ebc737a3-e6cd-4f4c-a4b6-fce5924de4c6] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 625.170685] env[63515]: DEBUG nova.virt.hardware [None req-9131a685-703c-44a1-81f8-7c157d95e037 tempest-ServerAddressesNegativeTestJSON-1262066083 tempest-ServerAddressesNegativeTestJSON-1262066083-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T02:52:14Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T02:51:56Z,direct_url=,disk_format='vmdk',id=8a120570-cb06-4099-b262-554ca0ad15c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b569255cc43e42e7ae3f2b4ad37c6ef4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T02:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 625.170934] env[63515]: DEBUG nova.virt.hardware [None req-9131a685-703c-44a1-81f8-7c157d95e037 tempest-ServerAddressesNegativeTestJSON-1262066083 tempest-ServerAddressesNegativeTestJSON-1262066083-project-member] Flavor limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 625.171810] env[63515]: DEBUG nova.virt.hardware [None req-9131a685-703c-44a1-81f8-7c157d95e037 tempest-ServerAddressesNegativeTestJSON-1262066083 tempest-ServerAddressesNegativeTestJSON-1262066083-project-member] Image limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 625.171810] env[63515]: DEBUG nova.virt.hardware [None req-9131a685-703c-44a1-81f8-7c157d95e037 tempest-ServerAddressesNegativeTestJSON-1262066083 tempest-ServerAddressesNegativeTestJSON-1262066083-project-member] Flavor pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 625.171810] env[63515]: DEBUG nova.virt.hardware [None req-9131a685-703c-44a1-81f8-7c157d95e037 tempest-ServerAddressesNegativeTestJSON-1262066083 tempest-ServerAddressesNegativeTestJSON-1262066083-project-member] Image pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 625.171810] env[63515]: DEBUG nova.virt.hardware [None req-9131a685-703c-44a1-81f8-7c157d95e037 tempest-ServerAddressesNegativeTestJSON-1262066083 tempest-ServerAddressesNegativeTestJSON-1262066083-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 625.173038] env[63515]: DEBUG nova.virt.hardware [None req-9131a685-703c-44a1-81f8-7c157d95e037 tempest-ServerAddressesNegativeTestJSON-1262066083 tempest-ServerAddressesNegativeTestJSON-1262066083-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 625.173038] env[63515]: DEBUG nova.virt.hardware [None req-9131a685-703c-44a1-81f8-7c157d95e037 tempest-ServerAddressesNegativeTestJSON-1262066083 tempest-ServerAddressesNegativeTestJSON-1262066083-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 625.173038] env[63515]: DEBUG nova.virt.hardware [None req-9131a685-703c-44a1-81f8-7c157d95e037 tempest-ServerAddressesNegativeTestJSON-1262066083 tempest-ServerAddressesNegativeTestJSON-1262066083-project-member] Got 1 possible topologies {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 625.173038] env[63515]: DEBUG nova.virt.hardware [None req-9131a685-703c-44a1-81f8-7c157d95e037 tempest-ServerAddressesNegativeTestJSON-1262066083 tempest-ServerAddressesNegativeTestJSON-1262066083-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 625.173595] env[63515]: DEBUG nova.virt.hardware [None req-9131a685-703c-44a1-81f8-7c157d95e037 tempest-ServerAddressesNegativeTestJSON-1262066083 tempest-ServerAddressesNegativeTestJSON-1262066083-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 625.174358] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b7ef179-544e-4ca3-a65f-5487bf84b887 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.182744] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6c82f4b-b5dd-4dd7-90e0-30e2528056d8 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.399020] env[63515]: DEBUG oslo_concurrency.lockutils [None req-480b7d6d-2f7f-4da3-9eb1-818f0711a940 tempest-ServerRescueTestJSON-1961208277 tempest-ServerRescueTestJSON-1961208277-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 625.685604] env[63515]: DEBUG nova.network.neutron [None req-24636c84-3869-4a1b-acfb-9c10d33a9826 tempest-ServersWithSpecificFlavorTestJSON-691988195 tempest-ServersWithSpecificFlavorTestJSON-691988195-project-member] [instance: ebc737a3-e6cd-4f4c-a4b6-fce5924de4c6] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 625.936261] env[63515]: DEBUG nova.network.neutron [None req-24636c84-3869-4a1b-acfb-9c10d33a9826 tempest-ServersWithSpecificFlavorTestJSON-691988195 tempest-ServersWithSpecificFlavorTestJSON-691988195-project-member] [instance: ebc737a3-e6cd-4f4c-a4b6-fce5924de4c6] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 626.422827] env[63515]: DEBUG nova.compute.manager [req-f40a5256-2ea8-4253-b8bd-3dc372cf5e84 req-446b5ee3-4ed1-4d43-a2db-cabb660b9b8e service nova] [instance: 5b201ab6-c7ec-4a5c-a310-982649f34be4] Received event network-changed-4d81feb5-11c1-415a-bdaf-cd39beb2e524 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 626.423870] env[63515]: DEBUG nova.compute.manager [req-f40a5256-2ea8-4253-b8bd-3dc372cf5e84 req-446b5ee3-4ed1-4d43-a2db-cabb660b9b8e service nova] [instance: 5b201ab6-c7ec-4a5c-a310-982649f34be4] Refreshing instance network info cache due to event network-changed-4d81feb5-11c1-415a-bdaf-cd39beb2e524. {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 626.423870] env[63515]: DEBUG oslo_concurrency.lockutils [req-f40a5256-2ea8-4253-b8bd-3dc372cf5e84 req-446b5ee3-4ed1-4d43-a2db-cabb660b9b8e service nova] Acquiring lock "refresh_cache-5b201ab6-c7ec-4a5c-a310-982649f34be4" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 626.423870] env[63515]: DEBUG oslo_concurrency.lockutils [req-f40a5256-2ea8-4253-b8bd-3dc372cf5e84 req-446b5ee3-4ed1-4d43-a2db-cabb660b9b8e service nova] Acquired lock "refresh_cache-5b201ab6-c7ec-4a5c-a310-982649f34be4" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 626.423870] env[63515]: DEBUG nova.network.neutron [req-f40a5256-2ea8-4253-b8bd-3dc372cf5e84 req-446b5ee3-4ed1-4d43-a2db-cabb660b9b8e service nova] [instance: 5b201ab6-c7ec-4a5c-a310-982649f34be4] Refreshing network info cache for port 4d81feb5-11c1-415a-bdaf-cd39beb2e524 {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 626.442033] env[63515]: DEBUG oslo_concurrency.lockutils [None req-24636c84-3869-4a1b-acfb-9c10d33a9826 tempest-ServersWithSpecificFlavorTestJSON-691988195 tempest-ServersWithSpecificFlavorTestJSON-691988195-project-member] Releasing lock "refresh_cache-ebc737a3-e6cd-4f4c-a4b6-fce5924de4c6" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 626.442033] env[63515]: DEBUG nova.compute.manager [None req-24636c84-3869-4a1b-acfb-9c10d33a9826 tempest-ServersWithSpecificFlavorTestJSON-691988195 tempest-ServersWithSpecificFlavorTestJSON-691988195-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63515) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 626.442033] env[63515]: DEBUG nova.compute.manager [None req-24636c84-3869-4a1b-acfb-9c10d33a9826 tempest-ServersWithSpecificFlavorTestJSON-691988195 tempest-ServersWithSpecificFlavorTestJSON-691988195-project-member] [instance: ebc737a3-e6cd-4f4c-a4b6-fce5924de4c6] Deallocating network for instance {{(pid=63515) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 626.442033] env[63515]: DEBUG nova.network.neutron [None req-24636c84-3869-4a1b-acfb-9c10d33a9826 tempest-ServersWithSpecificFlavorTestJSON-691988195 tempest-ServersWithSpecificFlavorTestJSON-691988195-project-member] [instance: ebc737a3-e6cd-4f4c-a4b6-fce5924de4c6] deallocate_for_instance() {{(pid=63515) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 626.469550] env[63515]: DEBUG nova.network.neutron [None req-24636c84-3869-4a1b-acfb-9c10d33a9826 tempest-ServersWithSpecificFlavorTestJSON-691988195 tempest-ServersWithSpecificFlavorTestJSON-691988195-project-member] [instance: ebc737a3-e6cd-4f4c-a4b6-fce5924de4c6] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 626.548797] env[63515]: ERROR nova.compute.manager [None req-9131a685-703c-44a1-81f8-7c157d95e037 tempest-ServerAddressesNegativeTestJSON-1262066083 tempest-ServerAddressesNegativeTestJSON-1262066083-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 4d81feb5-11c1-415a-bdaf-cd39beb2e524, please check neutron logs for more information. [ 626.548797] env[63515]: ERROR nova.compute.manager Traceback (most recent call last): [ 626.548797] env[63515]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 626.548797] env[63515]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 626.548797] env[63515]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 626.548797] env[63515]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 626.548797] env[63515]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 626.548797] env[63515]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 626.548797] env[63515]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 626.548797] env[63515]: ERROR nova.compute.manager self.force_reraise() [ 626.548797] env[63515]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 626.548797] env[63515]: ERROR nova.compute.manager raise self.value [ 626.548797] env[63515]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 626.548797] env[63515]: ERROR nova.compute.manager updated_port = self._update_port( [ 626.548797] env[63515]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 626.548797] env[63515]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 626.549285] env[63515]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 626.549285] env[63515]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 626.549285] env[63515]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 4d81feb5-11c1-415a-bdaf-cd39beb2e524, please check neutron logs for more information. [ 626.549285] env[63515]: ERROR nova.compute.manager [ 626.549285] env[63515]: Traceback (most recent call last): [ 626.549285] env[63515]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 626.549285] env[63515]: listener.cb(fileno) [ 626.549285] env[63515]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 626.549285] env[63515]: result = function(*args, **kwargs) [ 626.549285] env[63515]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 626.549285] env[63515]: return func(*args, **kwargs) [ 626.549285] env[63515]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 626.549285] env[63515]: raise e [ 626.549285] env[63515]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 626.549285] env[63515]: nwinfo = self.network_api.allocate_for_instance( [ 626.549285] env[63515]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 626.549285] env[63515]: created_port_ids = self._update_ports_for_instance( [ 626.549285] env[63515]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 626.549285] env[63515]: with excutils.save_and_reraise_exception(): [ 626.549285] env[63515]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 626.549285] env[63515]: self.force_reraise() [ 626.549285] env[63515]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 626.549285] env[63515]: raise self.value [ 626.549285] env[63515]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 626.549285] env[63515]: updated_port = self._update_port( [ 626.549285] env[63515]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 626.549285] env[63515]: _ensure_no_port_binding_failure(port) [ 626.549285] env[63515]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 626.549285] env[63515]: raise exception.PortBindingFailed(port_id=port['id']) [ 626.550093] env[63515]: nova.exception.PortBindingFailed: Binding failed for port 4d81feb5-11c1-415a-bdaf-cd39beb2e524, please check neutron logs for more information. [ 626.550093] env[63515]: Removing descriptor: 17 [ 626.550093] env[63515]: ERROR nova.compute.manager [None req-9131a685-703c-44a1-81f8-7c157d95e037 tempest-ServerAddressesNegativeTestJSON-1262066083 tempest-ServerAddressesNegativeTestJSON-1262066083-project-member] [instance: 5b201ab6-c7ec-4a5c-a310-982649f34be4] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 4d81feb5-11c1-415a-bdaf-cd39beb2e524, please check neutron logs for more information. [ 626.550093] env[63515]: ERROR nova.compute.manager [instance: 5b201ab6-c7ec-4a5c-a310-982649f34be4] Traceback (most recent call last): [ 626.550093] env[63515]: ERROR nova.compute.manager [instance: 5b201ab6-c7ec-4a5c-a310-982649f34be4] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 626.550093] env[63515]: ERROR nova.compute.manager [instance: 5b201ab6-c7ec-4a5c-a310-982649f34be4] yield resources [ 626.550093] env[63515]: ERROR nova.compute.manager [instance: 5b201ab6-c7ec-4a5c-a310-982649f34be4] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 626.550093] env[63515]: ERROR nova.compute.manager [instance: 5b201ab6-c7ec-4a5c-a310-982649f34be4] self.driver.spawn(context, instance, image_meta, [ 626.550093] env[63515]: ERROR nova.compute.manager [instance: 5b201ab6-c7ec-4a5c-a310-982649f34be4] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 626.550093] env[63515]: ERROR nova.compute.manager [instance: 5b201ab6-c7ec-4a5c-a310-982649f34be4] self._vmops.spawn(context, instance, image_meta, injected_files, [ 626.550093] env[63515]: ERROR nova.compute.manager [instance: 5b201ab6-c7ec-4a5c-a310-982649f34be4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 626.550093] env[63515]: ERROR nova.compute.manager [instance: 5b201ab6-c7ec-4a5c-a310-982649f34be4] vm_ref = self.build_virtual_machine(instance, [ 626.550483] env[63515]: ERROR nova.compute.manager [instance: 5b201ab6-c7ec-4a5c-a310-982649f34be4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 626.550483] env[63515]: ERROR nova.compute.manager [instance: 5b201ab6-c7ec-4a5c-a310-982649f34be4] vif_infos = vmwarevif.get_vif_info(self._session, [ 626.550483] env[63515]: ERROR nova.compute.manager [instance: 5b201ab6-c7ec-4a5c-a310-982649f34be4] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 626.550483] env[63515]: ERROR nova.compute.manager [instance: 5b201ab6-c7ec-4a5c-a310-982649f34be4] for vif in network_info: [ 626.550483] env[63515]: ERROR nova.compute.manager [instance: 5b201ab6-c7ec-4a5c-a310-982649f34be4] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 626.550483] env[63515]: ERROR nova.compute.manager [instance: 5b201ab6-c7ec-4a5c-a310-982649f34be4] return self._sync_wrapper(fn, *args, **kwargs) [ 626.550483] env[63515]: ERROR nova.compute.manager [instance: 5b201ab6-c7ec-4a5c-a310-982649f34be4] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 626.550483] env[63515]: ERROR nova.compute.manager [instance: 5b201ab6-c7ec-4a5c-a310-982649f34be4] self.wait() [ 626.550483] env[63515]: ERROR nova.compute.manager [instance: 5b201ab6-c7ec-4a5c-a310-982649f34be4] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 626.550483] env[63515]: ERROR nova.compute.manager [instance: 5b201ab6-c7ec-4a5c-a310-982649f34be4] self[:] = self._gt.wait() [ 626.550483] env[63515]: ERROR nova.compute.manager [instance: 5b201ab6-c7ec-4a5c-a310-982649f34be4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 626.550483] env[63515]: ERROR nova.compute.manager [instance: 5b201ab6-c7ec-4a5c-a310-982649f34be4] return self._exit_event.wait() [ 626.550483] env[63515]: ERROR nova.compute.manager [instance: 5b201ab6-c7ec-4a5c-a310-982649f34be4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 626.550833] env[63515]: ERROR nova.compute.manager [instance: 5b201ab6-c7ec-4a5c-a310-982649f34be4] result = hub.switch() [ 626.550833] env[63515]: ERROR nova.compute.manager [instance: 5b201ab6-c7ec-4a5c-a310-982649f34be4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 626.550833] env[63515]: ERROR nova.compute.manager [instance: 5b201ab6-c7ec-4a5c-a310-982649f34be4] return self.greenlet.switch() [ 626.550833] env[63515]: ERROR nova.compute.manager [instance: 5b201ab6-c7ec-4a5c-a310-982649f34be4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 626.550833] env[63515]: ERROR nova.compute.manager [instance: 5b201ab6-c7ec-4a5c-a310-982649f34be4] result = function(*args, **kwargs) [ 626.550833] env[63515]: ERROR nova.compute.manager [instance: 5b201ab6-c7ec-4a5c-a310-982649f34be4] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 626.550833] env[63515]: ERROR nova.compute.manager [instance: 5b201ab6-c7ec-4a5c-a310-982649f34be4] return func(*args, **kwargs) [ 626.550833] env[63515]: ERROR nova.compute.manager [instance: 5b201ab6-c7ec-4a5c-a310-982649f34be4] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 626.550833] env[63515]: ERROR nova.compute.manager [instance: 5b201ab6-c7ec-4a5c-a310-982649f34be4] raise e [ 626.550833] env[63515]: ERROR nova.compute.manager [instance: 5b201ab6-c7ec-4a5c-a310-982649f34be4] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 626.550833] env[63515]: ERROR nova.compute.manager [instance: 5b201ab6-c7ec-4a5c-a310-982649f34be4] nwinfo = self.network_api.allocate_for_instance( [ 626.550833] env[63515]: ERROR nova.compute.manager [instance: 5b201ab6-c7ec-4a5c-a310-982649f34be4] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 626.550833] env[63515]: ERROR nova.compute.manager [instance: 5b201ab6-c7ec-4a5c-a310-982649f34be4] created_port_ids = self._update_ports_for_instance( [ 626.551196] env[63515]: ERROR nova.compute.manager [instance: 5b201ab6-c7ec-4a5c-a310-982649f34be4] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 626.551196] env[63515]: ERROR nova.compute.manager [instance: 5b201ab6-c7ec-4a5c-a310-982649f34be4] with excutils.save_and_reraise_exception(): [ 626.551196] env[63515]: ERROR nova.compute.manager [instance: 5b201ab6-c7ec-4a5c-a310-982649f34be4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 626.551196] env[63515]: ERROR nova.compute.manager [instance: 5b201ab6-c7ec-4a5c-a310-982649f34be4] self.force_reraise() [ 626.551196] env[63515]: ERROR nova.compute.manager [instance: 5b201ab6-c7ec-4a5c-a310-982649f34be4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 626.551196] env[63515]: ERROR nova.compute.manager [instance: 5b201ab6-c7ec-4a5c-a310-982649f34be4] raise self.value [ 626.551196] env[63515]: ERROR nova.compute.manager [instance: 5b201ab6-c7ec-4a5c-a310-982649f34be4] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 626.551196] env[63515]: ERROR nova.compute.manager [instance: 5b201ab6-c7ec-4a5c-a310-982649f34be4] updated_port = self._update_port( [ 626.551196] env[63515]: ERROR nova.compute.manager [instance: 5b201ab6-c7ec-4a5c-a310-982649f34be4] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 626.551196] env[63515]: ERROR nova.compute.manager [instance: 5b201ab6-c7ec-4a5c-a310-982649f34be4] _ensure_no_port_binding_failure(port) [ 626.551196] env[63515]: ERROR nova.compute.manager [instance: 5b201ab6-c7ec-4a5c-a310-982649f34be4] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 626.551196] env[63515]: ERROR nova.compute.manager [instance: 5b201ab6-c7ec-4a5c-a310-982649f34be4] raise exception.PortBindingFailed(port_id=port['id']) [ 626.551601] env[63515]: ERROR nova.compute.manager [instance: 5b201ab6-c7ec-4a5c-a310-982649f34be4] nova.exception.PortBindingFailed: Binding failed for port 4d81feb5-11c1-415a-bdaf-cd39beb2e524, please check neutron logs for more information. [ 626.551601] env[63515]: ERROR nova.compute.manager [instance: 5b201ab6-c7ec-4a5c-a310-982649f34be4] [ 626.551601] env[63515]: INFO nova.compute.manager [None req-9131a685-703c-44a1-81f8-7c157d95e037 tempest-ServerAddressesNegativeTestJSON-1262066083 tempest-ServerAddressesNegativeTestJSON-1262066083-project-member] [instance: 5b201ab6-c7ec-4a5c-a310-982649f34be4] Terminating instance [ 626.557497] env[63515]: DEBUG oslo_concurrency.lockutils [None req-9131a685-703c-44a1-81f8-7c157d95e037 tempest-ServerAddressesNegativeTestJSON-1262066083 tempest-ServerAddressesNegativeTestJSON-1262066083-project-member] Acquiring lock "refresh_cache-5b201ab6-c7ec-4a5c-a310-982649f34be4" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 626.649835] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54ef29ce-1e16-4279-889f-14e4b44f4377 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.661018] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96f90c15-c25b-4aea-acf6-eb7addb459d1 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.697031] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e78fc3a-c9e1-4f12-a150-c09ee89b44fc {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.706544] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82411094-57fe-481d-b3e4-6dde85345411 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.724282] env[63515]: DEBUG nova.compute.provider_tree [None req-292f2bc3-2911-4498-95fb-4e4a03681601 tempest-FloatingIPsAssociationTestJSON-629096942 tempest-FloatingIPsAssociationTestJSON-629096942-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 626.730663] env[63515]: DEBUG oslo_concurrency.lockutils [None req-dcda8d78-de0d-44e9-b61f-e84b4a49533f tempest-ServersTestManualDisk-1114050206 tempest-ServersTestManualDisk-1114050206-project-member] Acquiring lock "e91aa479-1540-4950-851b-b2409e5f89f1" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 626.730896] env[63515]: DEBUG oslo_concurrency.lockutils [None req-dcda8d78-de0d-44e9-b61f-e84b4a49533f tempest-ServersTestManualDisk-1114050206 tempest-ServersTestManualDisk-1114050206-project-member] Lock "e91aa479-1540-4950-851b-b2409e5f89f1" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 626.949801] env[63515]: DEBUG nova.network.neutron [req-f40a5256-2ea8-4253-b8bd-3dc372cf5e84 req-446b5ee3-4ed1-4d43-a2db-cabb660b9b8e service nova] [instance: 5b201ab6-c7ec-4a5c-a310-982649f34be4] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 626.972479] env[63515]: DEBUG nova.network.neutron [None req-24636c84-3869-4a1b-acfb-9c10d33a9826 tempest-ServersWithSpecificFlavorTestJSON-691988195 tempest-ServersWithSpecificFlavorTestJSON-691988195-project-member] [instance: ebc737a3-e6cd-4f4c-a4b6-fce5924de4c6] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 627.106857] env[63515]: DEBUG nova.network.neutron [req-f40a5256-2ea8-4253-b8bd-3dc372cf5e84 req-446b5ee3-4ed1-4d43-a2db-cabb660b9b8e service nova] [instance: 5b201ab6-c7ec-4a5c-a310-982649f34be4] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 627.232067] env[63515]: DEBUG nova.scheduler.client.report [None req-292f2bc3-2911-4498-95fb-4e4a03681601 tempest-FloatingIPsAssociationTestJSON-629096942 tempest-FloatingIPsAssociationTestJSON-629096942-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 627.477126] env[63515]: INFO nova.compute.manager [None req-24636c84-3869-4a1b-acfb-9c10d33a9826 tempest-ServersWithSpecificFlavorTestJSON-691988195 tempest-ServersWithSpecificFlavorTestJSON-691988195-project-member] [instance: ebc737a3-e6cd-4f4c-a4b6-fce5924de4c6] Took 1.04 seconds to deallocate network for instance. [ 627.610045] env[63515]: DEBUG oslo_concurrency.lockutils [req-f40a5256-2ea8-4253-b8bd-3dc372cf5e84 req-446b5ee3-4ed1-4d43-a2db-cabb660b9b8e service nova] Releasing lock "refresh_cache-5b201ab6-c7ec-4a5c-a310-982649f34be4" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 627.610474] env[63515]: DEBUG oslo_concurrency.lockutils [None req-9131a685-703c-44a1-81f8-7c157d95e037 tempest-ServerAddressesNegativeTestJSON-1262066083 tempest-ServerAddressesNegativeTestJSON-1262066083-project-member] Acquired lock "refresh_cache-5b201ab6-c7ec-4a5c-a310-982649f34be4" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 627.610658] env[63515]: DEBUG nova.network.neutron [None req-9131a685-703c-44a1-81f8-7c157d95e037 tempest-ServerAddressesNegativeTestJSON-1262066083 tempest-ServerAddressesNegativeTestJSON-1262066083-project-member] [instance: 5b201ab6-c7ec-4a5c-a310-982649f34be4] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 627.742023] env[63515]: DEBUG oslo_concurrency.lockutils [None req-292f2bc3-2911-4498-95fb-4e4a03681601 tempest-FloatingIPsAssociationTestJSON-629096942 tempest-FloatingIPsAssociationTestJSON-629096942-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.601s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 627.742023] env[63515]: DEBUG nova.compute.manager [None req-292f2bc3-2911-4498-95fb-4e4a03681601 tempest-FloatingIPsAssociationTestJSON-629096942 tempest-FloatingIPsAssociationTestJSON-629096942-project-member] [instance: 2728c9f9-0d03-4313-ba2a-3dc22aff8538] Start building networks asynchronously for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 627.743898] env[63515]: DEBUG oslo_concurrency.lockutils [None req-d9ffa9d5-6856-47e5-8926-d090bbea2c09 tempest-DeleteServersAdminTestJSON-944542165 tempest-DeleteServersAdminTestJSON-944542165-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 18.262s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 628.129975] env[63515]: DEBUG nova.network.neutron [None req-9131a685-703c-44a1-81f8-7c157d95e037 tempest-ServerAddressesNegativeTestJSON-1262066083 tempest-ServerAddressesNegativeTestJSON-1262066083-project-member] [instance: 5b201ab6-c7ec-4a5c-a310-982649f34be4] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 628.221933] env[63515]: DEBUG nova.network.neutron [None req-9131a685-703c-44a1-81f8-7c157d95e037 tempest-ServerAddressesNegativeTestJSON-1262066083 tempest-ServerAddressesNegativeTestJSON-1262066083-project-member] [instance: 5b201ab6-c7ec-4a5c-a310-982649f34be4] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 628.246026] env[63515]: DEBUG nova.compute.utils [None req-292f2bc3-2911-4498-95fb-4e4a03681601 tempest-FloatingIPsAssociationTestJSON-629096942 tempest-FloatingIPsAssociationTestJSON-629096942-project-member] Using /dev/sd instead of None {{(pid=63515) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 628.247862] env[63515]: DEBUG nova.compute.manager [None req-292f2bc3-2911-4498-95fb-4e4a03681601 tempest-FloatingIPsAssociationTestJSON-629096942 tempest-FloatingIPsAssociationTestJSON-629096942-project-member] [instance: 2728c9f9-0d03-4313-ba2a-3dc22aff8538] Allocating IP information in the background. {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 628.253109] env[63515]: DEBUG nova.network.neutron [None req-292f2bc3-2911-4498-95fb-4e4a03681601 tempest-FloatingIPsAssociationTestJSON-629096942 tempest-FloatingIPsAssociationTestJSON-629096942-project-member] [instance: 2728c9f9-0d03-4313-ba2a-3dc22aff8538] allocate_for_instance() {{(pid=63515) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 628.304721] env[63515]: DEBUG nova.policy [None req-292f2bc3-2911-4498-95fb-4e4a03681601 tempest-FloatingIPsAssociationTestJSON-629096942 tempest-FloatingIPsAssociationTestJSON-629096942-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '1a52c473e6444c459e5f1d8b4d3df0b3', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd0e0035b06674732b65ef61dcdc61717', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63515) authorize /opt/stack/nova/nova/policy.py:201}} [ 628.523021] env[63515]: INFO nova.scheduler.client.report [None req-24636c84-3869-4a1b-acfb-9c10d33a9826 tempest-ServersWithSpecificFlavorTestJSON-691988195 tempest-ServersWithSpecificFlavorTestJSON-691988195-project-member] Deleted allocations for instance ebc737a3-e6cd-4f4c-a4b6-fce5924de4c6 [ 628.559842] env[63515]: DEBUG nova.compute.manager [req-24221c95-f10a-4de3-b9fe-d7dc06b5ea97 req-927ceced-d64a-43e0-a692-8b469a8738b6 service nova] [instance: 5b201ab6-c7ec-4a5c-a310-982649f34be4] Received event network-vif-deleted-4d81feb5-11c1-415a-bdaf-cd39beb2e524 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 628.680490] env[63515]: DEBUG nova.network.neutron [None req-292f2bc3-2911-4498-95fb-4e4a03681601 tempest-FloatingIPsAssociationTestJSON-629096942 tempest-FloatingIPsAssociationTestJSON-629096942-project-member] [instance: 2728c9f9-0d03-4313-ba2a-3dc22aff8538] Successfully created port: bfaff379-5880-4a38-86aa-c5d6331eb307 {{(pid=63515) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 628.702152] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0998f643-94ef-4c11-b2a7-7ad42ed2ea4c {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.711220] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e441005-d978-4f49-97af-3ee6dbd3658b {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.743551] env[63515]: DEBUG oslo_concurrency.lockutils [None req-9131a685-703c-44a1-81f8-7c157d95e037 tempest-ServerAddressesNegativeTestJSON-1262066083 tempest-ServerAddressesNegativeTestJSON-1262066083-project-member] Releasing lock "refresh_cache-5b201ab6-c7ec-4a5c-a310-982649f34be4" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 628.744223] env[63515]: DEBUG nova.compute.manager [None req-9131a685-703c-44a1-81f8-7c157d95e037 tempest-ServerAddressesNegativeTestJSON-1262066083 tempest-ServerAddressesNegativeTestJSON-1262066083-project-member] [instance: 5b201ab6-c7ec-4a5c-a310-982649f34be4] Start destroying the instance on the hypervisor. {{(pid=63515) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 628.744460] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-9131a685-703c-44a1-81f8-7c157d95e037 tempest-ServerAddressesNegativeTestJSON-1262066083 tempest-ServerAddressesNegativeTestJSON-1262066083-project-member] [instance: 5b201ab6-c7ec-4a5c-a310-982649f34be4] Destroying instance {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 628.744943] env[63515]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-43896654-9b96-42e2-99af-d709431e048e {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.747651] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4b85706-d8d8-4288-ae41-d5275b04d722 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.752356] env[63515]: DEBUG nova.compute.manager [None req-292f2bc3-2911-4498-95fb-4e4a03681601 tempest-FloatingIPsAssociationTestJSON-629096942 tempest-FloatingIPsAssociationTestJSON-629096942-project-member] [instance: 2728c9f9-0d03-4313-ba2a-3dc22aff8538] Start building block device mappings for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 628.762048] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a18914d-78dc-487f-9ddf-f7dd319ce241 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.775745] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27700a76-6be9-4284-acad-1f1d9b5ab9b2 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.790160] env[63515]: DEBUG nova.compute.provider_tree [None req-d9ffa9d5-6856-47e5-8926-d090bbea2c09 tempest-DeleteServersAdminTestJSON-944542165 tempest-DeleteServersAdminTestJSON-944542165-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 628.795564] env[63515]: WARNING nova.virt.vmwareapi.vmops [None req-9131a685-703c-44a1-81f8-7c157d95e037 tempest-ServerAddressesNegativeTestJSON-1262066083 tempest-ServerAddressesNegativeTestJSON-1262066083-project-member] [instance: 5b201ab6-c7ec-4a5c-a310-982649f34be4] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 5b201ab6-c7ec-4a5c-a310-982649f34be4 could not be found. [ 628.795779] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-9131a685-703c-44a1-81f8-7c157d95e037 tempest-ServerAddressesNegativeTestJSON-1262066083 tempest-ServerAddressesNegativeTestJSON-1262066083-project-member] [instance: 5b201ab6-c7ec-4a5c-a310-982649f34be4] Instance destroyed {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 628.795962] env[63515]: INFO nova.compute.manager [None req-9131a685-703c-44a1-81f8-7c157d95e037 tempest-ServerAddressesNegativeTestJSON-1262066083 tempest-ServerAddressesNegativeTestJSON-1262066083-project-member] [instance: 5b201ab6-c7ec-4a5c-a310-982649f34be4] Took 0.05 seconds to destroy the instance on the hypervisor. [ 628.796224] env[63515]: DEBUG oslo.service.loopingcall [None req-9131a685-703c-44a1-81f8-7c157d95e037 tempest-ServerAddressesNegativeTestJSON-1262066083 tempest-ServerAddressesNegativeTestJSON-1262066083-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 628.797042] env[63515]: DEBUG nova.compute.manager [-] [instance: 5b201ab6-c7ec-4a5c-a310-982649f34be4] Deallocating network for instance {{(pid=63515) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 628.797159] env[63515]: DEBUG nova.network.neutron [-] [instance: 5b201ab6-c7ec-4a5c-a310-982649f34be4] deallocate_for_instance() {{(pid=63515) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 628.821250] env[63515]: DEBUG nova.network.neutron [-] [instance: 5b201ab6-c7ec-4a5c-a310-982649f34be4] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 629.035253] env[63515]: DEBUG oslo_concurrency.lockutils [None req-24636c84-3869-4a1b-acfb-9c10d33a9826 tempest-ServersWithSpecificFlavorTestJSON-691988195 tempest-ServersWithSpecificFlavorTestJSON-691988195-project-member] Lock "ebc737a3-e6cd-4f4c-a4b6-fce5924de4c6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 79.302s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 629.301299] env[63515]: DEBUG nova.scheduler.client.report [None req-d9ffa9d5-6856-47e5-8926-d090bbea2c09 tempest-DeleteServersAdminTestJSON-944542165 tempest-DeleteServersAdminTestJSON-944542165-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 629.323247] env[63515]: DEBUG nova.network.neutron [-] [instance: 5b201ab6-c7ec-4a5c-a310-982649f34be4] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 629.538919] env[63515]: DEBUG nova.compute.manager [None req-b9c961a0-a559-4690-abb6-579359ed447c tempest-ServerPasswordTestJSON-433685770 tempest-ServerPasswordTestJSON-433685770-project-member] [instance: 07499601-62bc-4c31-b295-23f34a6e2e91] Starting instance... {{(pid=63515) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 629.764037] env[63515]: DEBUG nova.compute.manager [None req-292f2bc3-2911-4498-95fb-4e4a03681601 tempest-FloatingIPsAssociationTestJSON-629096942 tempest-FloatingIPsAssociationTestJSON-629096942-project-member] [instance: 2728c9f9-0d03-4313-ba2a-3dc22aff8538] Start spawning the instance on the hypervisor. {{(pid=63515) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 629.800008] env[63515]: DEBUG nova.virt.hardware [None req-292f2bc3-2911-4498-95fb-4e4a03681601 tempest-FloatingIPsAssociationTestJSON-629096942 tempest-FloatingIPsAssociationTestJSON-629096942-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T02:52:14Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T02:51:56Z,direct_url=,disk_format='vmdk',id=8a120570-cb06-4099-b262-554ca0ad15c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b569255cc43e42e7ae3f2b4ad37c6ef4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T02:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 629.800404] env[63515]: DEBUG nova.virt.hardware [None req-292f2bc3-2911-4498-95fb-4e4a03681601 tempest-FloatingIPsAssociationTestJSON-629096942 tempest-FloatingIPsAssociationTestJSON-629096942-project-member] Flavor limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 629.800652] env[63515]: DEBUG nova.virt.hardware [None req-292f2bc3-2911-4498-95fb-4e4a03681601 tempest-FloatingIPsAssociationTestJSON-629096942 tempest-FloatingIPsAssociationTestJSON-629096942-project-member] Image limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 629.800873] env[63515]: DEBUG nova.virt.hardware [None req-292f2bc3-2911-4498-95fb-4e4a03681601 tempest-FloatingIPsAssociationTestJSON-629096942 tempest-FloatingIPsAssociationTestJSON-629096942-project-member] Flavor pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 629.801707] env[63515]: DEBUG nova.virt.hardware [None req-292f2bc3-2911-4498-95fb-4e4a03681601 tempest-FloatingIPsAssociationTestJSON-629096942 tempest-FloatingIPsAssociationTestJSON-629096942-project-member] Image pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 629.801707] env[63515]: DEBUG nova.virt.hardware [None req-292f2bc3-2911-4498-95fb-4e4a03681601 tempest-FloatingIPsAssociationTestJSON-629096942 tempest-FloatingIPsAssociationTestJSON-629096942-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 629.801707] env[63515]: DEBUG nova.virt.hardware [None req-292f2bc3-2911-4498-95fb-4e4a03681601 tempest-FloatingIPsAssociationTestJSON-629096942 tempest-FloatingIPsAssociationTestJSON-629096942-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 629.801707] env[63515]: DEBUG nova.virt.hardware [None req-292f2bc3-2911-4498-95fb-4e4a03681601 tempest-FloatingIPsAssociationTestJSON-629096942 tempest-FloatingIPsAssociationTestJSON-629096942-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 629.801986] env[63515]: DEBUG nova.virt.hardware [None req-292f2bc3-2911-4498-95fb-4e4a03681601 tempest-FloatingIPsAssociationTestJSON-629096942 tempest-FloatingIPsAssociationTestJSON-629096942-project-member] Got 1 possible topologies {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 629.801986] env[63515]: DEBUG nova.virt.hardware [None req-292f2bc3-2911-4498-95fb-4e4a03681601 tempest-FloatingIPsAssociationTestJSON-629096942 tempest-FloatingIPsAssociationTestJSON-629096942-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 629.802489] env[63515]: DEBUG nova.virt.hardware [None req-292f2bc3-2911-4498-95fb-4e4a03681601 tempest-FloatingIPsAssociationTestJSON-629096942 tempest-FloatingIPsAssociationTestJSON-629096942-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 629.803213] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d1bdf06-db6c-429c-8d1f-2e583ee0c1d1 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.806894] env[63515]: DEBUG oslo_concurrency.lockutils [None req-d9ffa9d5-6856-47e5-8926-d090bbea2c09 tempest-DeleteServersAdminTestJSON-944542165 tempest-DeleteServersAdminTestJSON-944542165-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.063s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 629.807493] env[63515]: ERROR nova.compute.manager [None req-d9ffa9d5-6856-47e5-8926-d090bbea2c09 tempest-DeleteServersAdminTestJSON-944542165 tempest-DeleteServersAdminTestJSON-944542165-project-member] [instance: ae5c6e69-a885-48e3-9c9e-017f33430edd] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port c69d462c-5b89-4c1b-a518-3d862cdd355e, please check neutron logs for more information. [ 629.807493] env[63515]: ERROR nova.compute.manager [instance: ae5c6e69-a885-48e3-9c9e-017f33430edd] Traceback (most recent call last): [ 629.807493] env[63515]: ERROR nova.compute.manager [instance: ae5c6e69-a885-48e3-9c9e-017f33430edd] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 629.807493] env[63515]: ERROR nova.compute.manager [instance: ae5c6e69-a885-48e3-9c9e-017f33430edd] self.driver.spawn(context, instance, image_meta, [ 629.807493] env[63515]: ERROR nova.compute.manager [instance: ae5c6e69-a885-48e3-9c9e-017f33430edd] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 629.807493] env[63515]: ERROR nova.compute.manager [instance: ae5c6e69-a885-48e3-9c9e-017f33430edd] self._vmops.spawn(context, instance, image_meta, injected_files, [ 629.807493] env[63515]: ERROR nova.compute.manager [instance: ae5c6e69-a885-48e3-9c9e-017f33430edd] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 629.807493] env[63515]: ERROR nova.compute.manager [instance: ae5c6e69-a885-48e3-9c9e-017f33430edd] vm_ref = self.build_virtual_machine(instance, [ 629.807493] env[63515]: ERROR nova.compute.manager [instance: ae5c6e69-a885-48e3-9c9e-017f33430edd] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 629.807493] env[63515]: ERROR nova.compute.manager [instance: ae5c6e69-a885-48e3-9c9e-017f33430edd] vif_infos = vmwarevif.get_vif_info(self._session, [ 629.807493] env[63515]: ERROR nova.compute.manager [instance: ae5c6e69-a885-48e3-9c9e-017f33430edd] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 629.807867] env[63515]: ERROR nova.compute.manager [instance: ae5c6e69-a885-48e3-9c9e-017f33430edd] for vif in network_info: [ 629.807867] env[63515]: ERROR nova.compute.manager [instance: ae5c6e69-a885-48e3-9c9e-017f33430edd] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 629.807867] env[63515]: ERROR nova.compute.manager [instance: ae5c6e69-a885-48e3-9c9e-017f33430edd] return self._sync_wrapper(fn, *args, **kwargs) [ 629.807867] env[63515]: ERROR nova.compute.manager [instance: ae5c6e69-a885-48e3-9c9e-017f33430edd] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 629.807867] env[63515]: ERROR nova.compute.manager [instance: ae5c6e69-a885-48e3-9c9e-017f33430edd] self.wait() [ 629.807867] env[63515]: ERROR nova.compute.manager [instance: ae5c6e69-a885-48e3-9c9e-017f33430edd] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 629.807867] env[63515]: ERROR nova.compute.manager [instance: ae5c6e69-a885-48e3-9c9e-017f33430edd] self[:] = self._gt.wait() [ 629.807867] env[63515]: ERROR nova.compute.manager [instance: ae5c6e69-a885-48e3-9c9e-017f33430edd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 629.807867] env[63515]: ERROR nova.compute.manager [instance: ae5c6e69-a885-48e3-9c9e-017f33430edd] return self._exit_event.wait() [ 629.807867] env[63515]: ERROR nova.compute.manager [instance: ae5c6e69-a885-48e3-9c9e-017f33430edd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 629.807867] env[63515]: ERROR nova.compute.manager [instance: ae5c6e69-a885-48e3-9c9e-017f33430edd] current.throw(*self._exc) [ 629.807867] env[63515]: ERROR nova.compute.manager [instance: ae5c6e69-a885-48e3-9c9e-017f33430edd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 629.807867] env[63515]: ERROR nova.compute.manager [instance: ae5c6e69-a885-48e3-9c9e-017f33430edd] result = function(*args, **kwargs) [ 629.808199] env[63515]: ERROR nova.compute.manager [instance: ae5c6e69-a885-48e3-9c9e-017f33430edd] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 629.808199] env[63515]: ERROR nova.compute.manager [instance: ae5c6e69-a885-48e3-9c9e-017f33430edd] return func(*args, **kwargs) [ 629.808199] env[63515]: ERROR nova.compute.manager [instance: ae5c6e69-a885-48e3-9c9e-017f33430edd] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 629.808199] env[63515]: ERROR nova.compute.manager [instance: ae5c6e69-a885-48e3-9c9e-017f33430edd] raise e [ 629.808199] env[63515]: ERROR nova.compute.manager [instance: ae5c6e69-a885-48e3-9c9e-017f33430edd] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 629.808199] env[63515]: ERROR nova.compute.manager [instance: ae5c6e69-a885-48e3-9c9e-017f33430edd] nwinfo = self.network_api.allocate_for_instance( [ 629.808199] env[63515]: ERROR nova.compute.manager [instance: ae5c6e69-a885-48e3-9c9e-017f33430edd] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 629.808199] env[63515]: ERROR nova.compute.manager [instance: ae5c6e69-a885-48e3-9c9e-017f33430edd] created_port_ids = self._update_ports_for_instance( [ 629.808199] env[63515]: ERROR nova.compute.manager [instance: ae5c6e69-a885-48e3-9c9e-017f33430edd] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 629.808199] env[63515]: ERROR nova.compute.manager [instance: ae5c6e69-a885-48e3-9c9e-017f33430edd] with excutils.save_and_reraise_exception(): [ 629.808199] env[63515]: ERROR nova.compute.manager [instance: ae5c6e69-a885-48e3-9c9e-017f33430edd] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 629.808199] env[63515]: ERROR nova.compute.manager [instance: ae5c6e69-a885-48e3-9c9e-017f33430edd] self.force_reraise() [ 629.808199] env[63515]: ERROR nova.compute.manager [instance: ae5c6e69-a885-48e3-9c9e-017f33430edd] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 629.808512] env[63515]: ERROR nova.compute.manager [instance: ae5c6e69-a885-48e3-9c9e-017f33430edd] raise self.value [ 629.808512] env[63515]: ERROR nova.compute.manager [instance: ae5c6e69-a885-48e3-9c9e-017f33430edd] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 629.808512] env[63515]: ERROR nova.compute.manager [instance: ae5c6e69-a885-48e3-9c9e-017f33430edd] updated_port = self._update_port( [ 629.808512] env[63515]: ERROR nova.compute.manager [instance: ae5c6e69-a885-48e3-9c9e-017f33430edd] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 629.808512] env[63515]: ERROR nova.compute.manager [instance: ae5c6e69-a885-48e3-9c9e-017f33430edd] _ensure_no_port_binding_failure(port) [ 629.808512] env[63515]: ERROR nova.compute.manager [instance: ae5c6e69-a885-48e3-9c9e-017f33430edd] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 629.808512] env[63515]: ERROR nova.compute.manager [instance: ae5c6e69-a885-48e3-9c9e-017f33430edd] raise exception.PortBindingFailed(port_id=port['id']) [ 629.808512] env[63515]: ERROR nova.compute.manager [instance: ae5c6e69-a885-48e3-9c9e-017f33430edd] nova.exception.PortBindingFailed: Binding failed for port c69d462c-5b89-4c1b-a518-3d862cdd355e, please check neutron logs for more information. [ 629.808512] env[63515]: ERROR nova.compute.manager [instance: ae5c6e69-a885-48e3-9c9e-017f33430edd] [ 629.808744] env[63515]: DEBUG nova.compute.utils [None req-d9ffa9d5-6856-47e5-8926-d090bbea2c09 tempest-DeleteServersAdminTestJSON-944542165 tempest-DeleteServersAdminTestJSON-944542165-project-member] [instance: ae5c6e69-a885-48e3-9c9e-017f33430edd] Binding failed for port c69d462c-5b89-4c1b-a518-3d862cdd355e, please check neutron logs for more information. {{(pid=63515) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 629.809907] env[63515]: DEBUG oslo_concurrency.lockutils [None req-9f5c50b0-6bbe-4982-8818-c7255a8c23ab tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 18.649s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 629.814265] env[63515]: DEBUG nova.compute.manager [None req-d9ffa9d5-6856-47e5-8926-d090bbea2c09 tempest-DeleteServersAdminTestJSON-944542165 tempest-DeleteServersAdminTestJSON-944542165-project-member] [instance: ae5c6e69-a885-48e3-9c9e-017f33430edd] Build of instance ae5c6e69-a885-48e3-9c9e-017f33430edd was re-scheduled: Binding failed for port c69d462c-5b89-4c1b-a518-3d862cdd355e, please check neutron logs for more information. {{(pid=63515) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 629.815241] env[63515]: DEBUG nova.compute.manager [None req-d9ffa9d5-6856-47e5-8926-d090bbea2c09 tempest-DeleteServersAdminTestJSON-944542165 tempest-DeleteServersAdminTestJSON-944542165-project-member] [instance: ae5c6e69-a885-48e3-9c9e-017f33430edd] Unplugging VIFs for instance {{(pid=63515) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 629.815241] env[63515]: DEBUG oslo_concurrency.lockutils [None req-d9ffa9d5-6856-47e5-8926-d090bbea2c09 tempest-DeleteServersAdminTestJSON-944542165 tempest-DeleteServersAdminTestJSON-944542165-project-member] Acquiring lock "refresh_cache-ae5c6e69-a885-48e3-9c9e-017f33430edd" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 629.815241] env[63515]: DEBUG oslo_concurrency.lockutils [None req-d9ffa9d5-6856-47e5-8926-d090bbea2c09 tempest-DeleteServersAdminTestJSON-944542165 tempest-DeleteServersAdminTestJSON-944542165-project-member] Acquired lock "refresh_cache-ae5c6e69-a885-48e3-9c9e-017f33430edd" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 629.815241] env[63515]: DEBUG nova.network.neutron [None req-d9ffa9d5-6856-47e5-8926-d090bbea2c09 tempest-DeleteServersAdminTestJSON-944542165 tempest-DeleteServersAdminTestJSON-944542165-project-member] [instance: ae5c6e69-a885-48e3-9c9e-017f33430edd] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 629.822154] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c43dceb9-c787-4158-b180-c634295f093c {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.840382] env[63515]: INFO nova.compute.manager [-] [instance: 5b201ab6-c7ec-4a5c-a310-982649f34be4] Took 1.04 seconds to deallocate network for instance. [ 629.845649] env[63515]: DEBUG nova.compute.claims [None req-9131a685-703c-44a1-81f8-7c157d95e037 tempest-ServerAddressesNegativeTestJSON-1262066083 tempest-ServerAddressesNegativeTestJSON-1262066083-project-member] [instance: 5b201ab6-c7ec-4a5c-a310-982649f34be4] Aborting claim: {{(pid=63515) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 629.845858] env[63515]: DEBUG oslo_concurrency.lockutils [None req-9131a685-703c-44a1-81f8-7c157d95e037 tempest-ServerAddressesNegativeTestJSON-1262066083 tempest-ServerAddressesNegativeTestJSON-1262066083-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 630.069394] env[63515]: DEBUG oslo_concurrency.lockutils [None req-b9c961a0-a559-4690-abb6-579359ed447c tempest-ServerPasswordTestJSON-433685770 tempest-ServerPasswordTestJSON-433685770-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 630.222415] env[63515]: ERROR nova.compute.manager [None req-292f2bc3-2911-4498-95fb-4e4a03681601 tempest-FloatingIPsAssociationTestJSON-629096942 tempest-FloatingIPsAssociationTestJSON-629096942-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port bfaff379-5880-4a38-86aa-c5d6331eb307, please check neutron logs for more information. [ 630.222415] env[63515]: ERROR nova.compute.manager Traceback (most recent call last): [ 630.222415] env[63515]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 630.222415] env[63515]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 630.222415] env[63515]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 630.222415] env[63515]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 630.222415] env[63515]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 630.222415] env[63515]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 630.222415] env[63515]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 630.222415] env[63515]: ERROR nova.compute.manager self.force_reraise() [ 630.222415] env[63515]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 630.222415] env[63515]: ERROR nova.compute.manager raise self.value [ 630.222415] env[63515]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 630.222415] env[63515]: ERROR nova.compute.manager updated_port = self._update_port( [ 630.222415] env[63515]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 630.222415] env[63515]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 630.222845] env[63515]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 630.222845] env[63515]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 630.222845] env[63515]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port bfaff379-5880-4a38-86aa-c5d6331eb307, please check neutron logs for more information. [ 630.222845] env[63515]: ERROR nova.compute.manager [ 630.222845] env[63515]: Traceback (most recent call last): [ 630.222845] env[63515]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 630.222845] env[63515]: listener.cb(fileno) [ 630.222845] env[63515]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 630.222845] env[63515]: result = function(*args, **kwargs) [ 630.222845] env[63515]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 630.222845] env[63515]: return func(*args, **kwargs) [ 630.222845] env[63515]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 630.222845] env[63515]: raise e [ 630.222845] env[63515]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 630.222845] env[63515]: nwinfo = self.network_api.allocate_for_instance( [ 630.222845] env[63515]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 630.222845] env[63515]: created_port_ids = self._update_ports_for_instance( [ 630.222845] env[63515]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 630.222845] env[63515]: with excutils.save_and_reraise_exception(): [ 630.222845] env[63515]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 630.222845] env[63515]: self.force_reraise() [ 630.222845] env[63515]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 630.222845] env[63515]: raise self.value [ 630.222845] env[63515]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 630.222845] env[63515]: updated_port = self._update_port( [ 630.222845] env[63515]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 630.222845] env[63515]: _ensure_no_port_binding_failure(port) [ 630.222845] env[63515]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 630.222845] env[63515]: raise exception.PortBindingFailed(port_id=port['id']) [ 630.223515] env[63515]: nova.exception.PortBindingFailed: Binding failed for port bfaff379-5880-4a38-86aa-c5d6331eb307, please check neutron logs for more information. [ 630.223515] env[63515]: Removing descriptor: 17 [ 630.223515] env[63515]: ERROR nova.compute.manager [None req-292f2bc3-2911-4498-95fb-4e4a03681601 tempest-FloatingIPsAssociationTestJSON-629096942 tempest-FloatingIPsAssociationTestJSON-629096942-project-member] [instance: 2728c9f9-0d03-4313-ba2a-3dc22aff8538] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port bfaff379-5880-4a38-86aa-c5d6331eb307, please check neutron logs for more information. [ 630.223515] env[63515]: ERROR nova.compute.manager [instance: 2728c9f9-0d03-4313-ba2a-3dc22aff8538] Traceback (most recent call last): [ 630.223515] env[63515]: ERROR nova.compute.manager [instance: 2728c9f9-0d03-4313-ba2a-3dc22aff8538] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 630.223515] env[63515]: ERROR nova.compute.manager [instance: 2728c9f9-0d03-4313-ba2a-3dc22aff8538] yield resources [ 630.223515] env[63515]: ERROR nova.compute.manager [instance: 2728c9f9-0d03-4313-ba2a-3dc22aff8538] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 630.223515] env[63515]: ERROR nova.compute.manager [instance: 2728c9f9-0d03-4313-ba2a-3dc22aff8538] self.driver.spawn(context, instance, image_meta, [ 630.223515] env[63515]: ERROR nova.compute.manager [instance: 2728c9f9-0d03-4313-ba2a-3dc22aff8538] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 630.223515] env[63515]: ERROR nova.compute.manager [instance: 2728c9f9-0d03-4313-ba2a-3dc22aff8538] self._vmops.spawn(context, instance, image_meta, injected_files, [ 630.223515] env[63515]: ERROR nova.compute.manager [instance: 2728c9f9-0d03-4313-ba2a-3dc22aff8538] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 630.223515] env[63515]: ERROR nova.compute.manager [instance: 2728c9f9-0d03-4313-ba2a-3dc22aff8538] vm_ref = self.build_virtual_machine(instance, [ 630.223818] env[63515]: ERROR nova.compute.manager [instance: 2728c9f9-0d03-4313-ba2a-3dc22aff8538] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 630.223818] env[63515]: ERROR nova.compute.manager [instance: 2728c9f9-0d03-4313-ba2a-3dc22aff8538] vif_infos = vmwarevif.get_vif_info(self._session, [ 630.223818] env[63515]: ERROR nova.compute.manager [instance: 2728c9f9-0d03-4313-ba2a-3dc22aff8538] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 630.223818] env[63515]: ERROR nova.compute.manager [instance: 2728c9f9-0d03-4313-ba2a-3dc22aff8538] for vif in network_info: [ 630.223818] env[63515]: ERROR nova.compute.manager [instance: 2728c9f9-0d03-4313-ba2a-3dc22aff8538] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 630.223818] env[63515]: ERROR nova.compute.manager [instance: 2728c9f9-0d03-4313-ba2a-3dc22aff8538] return self._sync_wrapper(fn, *args, **kwargs) [ 630.223818] env[63515]: ERROR nova.compute.manager [instance: 2728c9f9-0d03-4313-ba2a-3dc22aff8538] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 630.223818] env[63515]: ERROR nova.compute.manager [instance: 2728c9f9-0d03-4313-ba2a-3dc22aff8538] self.wait() [ 630.223818] env[63515]: ERROR nova.compute.manager [instance: 2728c9f9-0d03-4313-ba2a-3dc22aff8538] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 630.223818] env[63515]: ERROR nova.compute.manager [instance: 2728c9f9-0d03-4313-ba2a-3dc22aff8538] self[:] = self._gt.wait() [ 630.223818] env[63515]: ERROR nova.compute.manager [instance: 2728c9f9-0d03-4313-ba2a-3dc22aff8538] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 630.223818] env[63515]: ERROR nova.compute.manager [instance: 2728c9f9-0d03-4313-ba2a-3dc22aff8538] return self._exit_event.wait() [ 630.223818] env[63515]: ERROR nova.compute.manager [instance: 2728c9f9-0d03-4313-ba2a-3dc22aff8538] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 630.224145] env[63515]: ERROR nova.compute.manager [instance: 2728c9f9-0d03-4313-ba2a-3dc22aff8538] result = hub.switch() [ 630.224145] env[63515]: ERROR nova.compute.manager [instance: 2728c9f9-0d03-4313-ba2a-3dc22aff8538] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 630.224145] env[63515]: ERROR nova.compute.manager [instance: 2728c9f9-0d03-4313-ba2a-3dc22aff8538] return self.greenlet.switch() [ 630.224145] env[63515]: ERROR nova.compute.manager [instance: 2728c9f9-0d03-4313-ba2a-3dc22aff8538] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 630.224145] env[63515]: ERROR nova.compute.manager [instance: 2728c9f9-0d03-4313-ba2a-3dc22aff8538] result = function(*args, **kwargs) [ 630.224145] env[63515]: ERROR nova.compute.manager [instance: 2728c9f9-0d03-4313-ba2a-3dc22aff8538] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 630.224145] env[63515]: ERROR nova.compute.manager [instance: 2728c9f9-0d03-4313-ba2a-3dc22aff8538] return func(*args, **kwargs) [ 630.224145] env[63515]: ERROR nova.compute.manager [instance: 2728c9f9-0d03-4313-ba2a-3dc22aff8538] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 630.224145] env[63515]: ERROR nova.compute.manager [instance: 2728c9f9-0d03-4313-ba2a-3dc22aff8538] raise e [ 630.224145] env[63515]: ERROR nova.compute.manager [instance: 2728c9f9-0d03-4313-ba2a-3dc22aff8538] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 630.224145] env[63515]: ERROR nova.compute.manager [instance: 2728c9f9-0d03-4313-ba2a-3dc22aff8538] nwinfo = self.network_api.allocate_for_instance( [ 630.224145] env[63515]: ERROR nova.compute.manager [instance: 2728c9f9-0d03-4313-ba2a-3dc22aff8538] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 630.224145] env[63515]: ERROR nova.compute.manager [instance: 2728c9f9-0d03-4313-ba2a-3dc22aff8538] created_port_ids = self._update_ports_for_instance( [ 630.224444] env[63515]: ERROR nova.compute.manager [instance: 2728c9f9-0d03-4313-ba2a-3dc22aff8538] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 630.224444] env[63515]: ERROR nova.compute.manager [instance: 2728c9f9-0d03-4313-ba2a-3dc22aff8538] with excutils.save_and_reraise_exception(): [ 630.224444] env[63515]: ERROR nova.compute.manager [instance: 2728c9f9-0d03-4313-ba2a-3dc22aff8538] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 630.224444] env[63515]: ERROR nova.compute.manager [instance: 2728c9f9-0d03-4313-ba2a-3dc22aff8538] self.force_reraise() [ 630.224444] env[63515]: ERROR nova.compute.manager [instance: 2728c9f9-0d03-4313-ba2a-3dc22aff8538] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 630.224444] env[63515]: ERROR nova.compute.manager [instance: 2728c9f9-0d03-4313-ba2a-3dc22aff8538] raise self.value [ 630.224444] env[63515]: ERROR nova.compute.manager [instance: 2728c9f9-0d03-4313-ba2a-3dc22aff8538] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 630.224444] env[63515]: ERROR nova.compute.manager [instance: 2728c9f9-0d03-4313-ba2a-3dc22aff8538] updated_port = self._update_port( [ 630.224444] env[63515]: ERROR nova.compute.manager [instance: 2728c9f9-0d03-4313-ba2a-3dc22aff8538] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 630.224444] env[63515]: ERROR nova.compute.manager [instance: 2728c9f9-0d03-4313-ba2a-3dc22aff8538] _ensure_no_port_binding_failure(port) [ 630.224444] env[63515]: ERROR nova.compute.manager [instance: 2728c9f9-0d03-4313-ba2a-3dc22aff8538] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 630.224444] env[63515]: ERROR nova.compute.manager [instance: 2728c9f9-0d03-4313-ba2a-3dc22aff8538] raise exception.PortBindingFailed(port_id=port['id']) [ 630.224736] env[63515]: ERROR nova.compute.manager [instance: 2728c9f9-0d03-4313-ba2a-3dc22aff8538] nova.exception.PortBindingFailed: Binding failed for port bfaff379-5880-4a38-86aa-c5d6331eb307, please check neutron logs for more information. [ 630.224736] env[63515]: ERROR nova.compute.manager [instance: 2728c9f9-0d03-4313-ba2a-3dc22aff8538] [ 630.224736] env[63515]: INFO nova.compute.manager [None req-292f2bc3-2911-4498-95fb-4e4a03681601 tempest-FloatingIPsAssociationTestJSON-629096942 tempest-FloatingIPsAssociationTestJSON-629096942-project-member] [instance: 2728c9f9-0d03-4313-ba2a-3dc22aff8538] Terminating instance [ 630.226322] env[63515]: DEBUG oslo_concurrency.lockutils [None req-292f2bc3-2911-4498-95fb-4e4a03681601 tempest-FloatingIPsAssociationTestJSON-629096942 tempest-FloatingIPsAssociationTestJSON-629096942-project-member] Acquiring lock "refresh_cache-2728c9f9-0d03-4313-ba2a-3dc22aff8538" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 630.226322] env[63515]: DEBUG oslo_concurrency.lockutils [None req-292f2bc3-2911-4498-95fb-4e4a03681601 tempest-FloatingIPsAssociationTestJSON-629096942 tempest-FloatingIPsAssociationTestJSON-629096942-project-member] Acquired lock "refresh_cache-2728c9f9-0d03-4313-ba2a-3dc22aff8538" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 630.226322] env[63515]: DEBUG nova.network.neutron [None req-292f2bc3-2911-4498-95fb-4e4a03681601 tempest-FloatingIPsAssociationTestJSON-629096942 tempest-FloatingIPsAssociationTestJSON-629096942-project-member] [instance: 2728c9f9-0d03-4313-ba2a-3dc22aff8538] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 630.348642] env[63515]: DEBUG nova.network.neutron [None req-d9ffa9d5-6856-47e5-8926-d090bbea2c09 tempest-DeleteServersAdminTestJSON-944542165 tempest-DeleteServersAdminTestJSON-944542165-project-member] [instance: ae5c6e69-a885-48e3-9c9e-017f33430edd] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 630.619230] env[63515]: DEBUG nova.compute.manager [req-7c4cee6c-59c3-4733-b459-35047bff205a req-210ee58f-9bd7-49f9-bea1-bf40e692754b service nova] [instance: 2728c9f9-0d03-4313-ba2a-3dc22aff8538] Received event network-changed-bfaff379-5880-4a38-86aa-c5d6331eb307 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 630.619230] env[63515]: DEBUG nova.compute.manager [req-7c4cee6c-59c3-4733-b459-35047bff205a req-210ee58f-9bd7-49f9-bea1-bf40e692754b service nova] [instance: 2728c9f9-0d03-4313-ba2a-3dc22aff8538] Refreshing instance network info cache due to event network-changed-bfaff379-5880-4a38-86aa-c5d6331eb307. {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 630.619230] env[63515]: DEBUG oslo_concurrency.lockutils [req-7c4cee6c-59c3-4733-b459-35047bff205a req-210ee58f-9bd7-49f9-bea1-bf40e692754b service nova] Acquiring lock "refresh_cache-2728c9f9-0d03-4313-ba2a-3dc22aff8538" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 630.649566] env[63515]: DEBUG nova.network.neutron [None req-d9ffa9d5-6856-47e5-8926-d090bbea2c09 tempest-DeleteServersAdminTestJSON-944542165 tempest-DeleteServersAdminTestJSON-944542165-project-member] [instance: ae5c6e69-a885-48e3-9c9e-017f33430edd] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 630.747998] env[63515]: DEBUG nova.network.neutron [None req-292f2bc3-2911-4498-95fb-4e4a03681601 tempest-FloatingIPsAssociationTestJSON-629096942 tempest-FloatingIPsAssociationTestJSON-629096942-project-member] [instance: 2728c9f9-0d03-4313-ba2a-3dc22aff8538] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 630.762739] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bbf295f9-44af-4d80-bd8f-85a79b678bc7 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.770179] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b6384e0-534b-499d-a802-108b5219ada7 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.801020] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f99f6e7-4974-4ab4-8406-5c2263f554bf {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.808733] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f911a3cb-f93b-41f7-8e2b-6e1847ca3f62 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.822567] env[63515]: DEBUG nova.compute.provider_tree [None req-9f5c50b0-6bbe-4982-8818-c7255a8c23ab tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 630.968043] env[63515]: DEBUG nova.network.neutron [None req-292f2bc3-2911-4498-95fb-4e4a03681601 tempest-FloatingIPsAssociationTestJSON-629096942 tempest-FloatingIPsAssociationTestJSON-629096942-project-member] [instance: 2728c9f9-0d03-4313-ba2a-3dc22aff8538] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 631.156514] env[63515]: DEBUG oslo_concurrency.lockutils [None req-d9ffa9d5-6856-47e5-8926-d090bbea2c09 tempest-DeleteServersAdminTestJSON-944542165 tempest-DeleteServersAdminTestJSON-944542165-project-member] Releasing lock "refresh_cache-ae5c6e69-a885-48e3-9c9e-017f33430edd" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 631.157663] env[63515]: DEBUG nova.compute.manager [None req-d9ffa9d5-6856-47e5-8926-d090bbea2c09 tempest-DeleteServersAdminTestJSON-944542165 tempest-DeleteServersAdminTestJSON-944542165-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63515) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 631.157663] env[63515]: DEBUG nova.compute.manager [None req-d9ffa9d5-6856-47e5-8926-d090bbea2c09 tempest-DeleteServersAdminTestJSON-944542165 tempest-DeleteServersAdminTestJSON-944542165-project-member] [instance: ae5c6e69-a885-48e3-9c9e-017f33430edd] Deallocating network for instance {{(pid=63515) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 631.157663] env[63515]: DEBUG nova.network.neutron [None req-d9ffa9d5-6856-47e5-8926-d090bbea2c09 tempest-DeleteServersAdminTestJSON-944542165 tempest-DeleteServersAdminTestJSON-944542165-project-member] [instance: ae5c6e69-a885-48e3-9c9e-017f33430edd] deallocate_for_instance() {{(pid=63515) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 631.180749] env[63515]: DEBUG nova.network.neutron [None req-d9ffa9d5-6856-47e5-8926-d090bbea2c09 tempest-DeleteServersAdminTestJSON-944542165 tempest-DeleteServersAdminTestJSON-944542165-project-member] [instance: ae5c6e69-a885-48e3-9c9e-017f33430edd] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 631.326956] env[63515]: DEBUG nova.scheduler.client.report [None req-9f5c50b0-6bbe-4982-8818-c7255a8c23ab tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 631.471223] env[63515]: DEBUG oslo_concurrency.lockutils [None req-292f2bc3-2911-4498-95fb-4e4a03681601 tempest-FloatingIPsAssociationTestJSON-629096942 tempest-FloatingIPsAssociationTestJSON-629096942-project-member] Releasing lock "refresh_cache-2728c9f9-0d03-4313-ba2a-3dc22aff8538" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 631.474063] env[63515]: DEBUG nova.compute.manager [None req-292f2bc3-2911-4498-95fb-4e4a03681601 tempest-FloatingIPsAssociationTestJSON-629096942 tempest-FloatingIPsAssociationTestJSON-629096942-project-member] [instance: 2728c9f9-0d03-4313-ba2a-3dc22aff8538] Start destroying the instance on the hypervisor. {{(pid=63515) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 631.474063] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-292f2bc3-2911-4498-95fb-4e4a03681601 tempest-FloatingIPsAssociationTestJSON-629096942 tempest-FloatingIPsAssociationTestJSON-629096942-project-member] [instance: 2728c9f9-0d03-4313-ba2a-3dc22aff8538] Destroying instance {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 631.474063] env[63515]: DEBUG oslo_concurrency.lockutils [req-7c4cee6c-59c3-4733-b459-35047bff205a req-210ee58f-9bd7-49f9-bea1-bf40e692754b service nova] Acquired lock "refresh_cache-2728c9f9-0d03-4313-ba2a-3dc22aff8538" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 631.474063] env[63515]: DEBUG nova.network.neutron [req-7c4cee6c-59c3-4733-b459-35047bff205a req-210ee58f-9bd7-49f9-bea1-bf40e692754b service nova] [instance: 2728c9f9-0d03-4313-ba2a-3dc22aff8538] Refreshing network info cache for port bfaff379-5880-4a38-86aa-c5d6331eb307 {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 631.475918] env[63515]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c47a1ed2-256a-400e-8093-97b933d53a91 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.487870] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48187efa-346d-455a-84e2-5c9f8c4aa2f7 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.517775] env[63515]: WARNING nova.virt.vmwareapi.vmops [None req-292f2bc3-2911-4498-95fb-4e4a03681601 tempest-FloatingIPsAssociationTestJSON-629096942 tempest-FloatingIPsAssociationTestJSON-629096942-project-member] [instance: 2728c9f9-0d03-4313-ba2a-3dc22aff8538] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 2728c9f9-0d03-4313-ba2a-3dc22aff8538 could not be found. [ 631.518435] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-292f2bc3-2911-4498-95fb-4e4a03681601 tempest-FloatingIPsAssociationTestJSON-629096942 tempest-FloatingIPsAssociationTestJSON-629096942-project-member] [instance: 2728c9f9-0d03-4313-ba2a-3dc22aff8538] Instance destroyed {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 631.518861] env[63515]: INFO nova.compute.manager [None req-292f2bc3-2911-4498-95fb-4e4a03681601 tempest-FloatingIPsAssociationTestJSON-629096942 tempest-FloatingIPsAssociationTestJSON-629096942-project-member] [instance: 2728c9f9-0d03-4313-ba2a-3dc22aff8538] Took 0.05 seconds to destroy the instance on the hypervisor. [ 631.519961] env[63515]: DEBUG oslo.service.loopingcall [None req-292f2bc3-2911-4498-95fb-4e4a03681601 tempest-FloatingIPsAssociationTestJSON-629096942 tempest-FloatingIPsAssociationTestJSON-629096942-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 631.519961] env[63515]: DEBUG nova.compute.manager [-] [instance: 2728c9f9-0d03-4313-ba2a-3dc22aff8538] Deallocating network for instance {{(pid=63515) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 631.519961] env[63515]: DEBUG nova.network.neutron [-] [instance: 2728c9f9-0d03-4313-ba2a-3dc22aff8538] deallocate_for_instance() {{(pid=63515) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 631.543196] env[63515]: DEBUG nova.network.neutron [-] [instance: 2728c9f9-0d03-4313-ba2a-3dc22aff8538] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 631.685142] env[63515]: DEBUG nova.network.neutron [None req-d9ffa9d5-6856-47e5-8926-d090bbea2c09 tempest-DeleteServersAdminTestJSON-944542165 tempest-DeleteServersAdminTestJSON-944542165-project-member] [instance: ae5c6e69-a885-48e3-9c9e-017f33430edd] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 631.835228] env[63515]: DEBUG oslo_concurrency.lockutils [None req-9f5c50b0-6bbe-4982-8818-c7255a8c23ab tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.022s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 631.835228] env[63515]: ERROR nova.compute.manager [None req-9f5c50b0-6bbe-4982-8818-c7255a8c23ab tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: b53754b1-ec31-4dc3-a839-771f71affe1e] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 46784f84-e839-4fa3-88ea-2005d2779b07, please check neutron logs for more information. [ 631.835228] env[63515]: ERROR nova.compute.manager [instance: b53754b1-ec31-4dc3-a839-771f71affe1e] Traceback (most recent call last): [ 631.835228] env[63515]: ERROR nova.compute.manager [instance: b53754b1-ec31-4dc3-a839-771f71affe1e] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 631.835228] env[63515]: ERROR nova.compute.manager [instance: b53754b1-ec31-4dc3-a839-771f71affe1e] self.driver.spawn(context, instance, image_meta, [ 631.835228] env[63515]: ERROR nova.compute.manager [instance: b53754b1-ec31-4dc3-a839-771f71affe1e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 631.835228] env[63515]: ERROR nova.compute.manager [instance: b53754b1-ec31-4dc3-a839-771f71affe1e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 631.835228] env[63515]: ERROR nova.compute.manager [instance: b53754b1-ec31-4dc3-a839-771f71affe1e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 631.835228] env[63515]: ERROR nova.compute.manager [instance: b53754b1-ec31-4dc3-a839-771f71affe1e] vm_ref = self.build_virtual_machine(instance, [ 631.835533] env[63515]: ERROR nova.compute.manager [instance: b53754b1-ec31-4dc3-a839-771f71affe1e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 631.835533] env[63515]: ERROR nova.compute.manager [instance: b53754b1-ec31-4dc3-a839-771f71affe1e] vif_infos = vmwarevif.get_vif_info(self._session, [ 631.835533] env[63515]: ERROR nova.compute.manager [instance: b53754b1-ec31-4dc3-a839-771f71affe1e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 631.835533] env[63515]: ERROR nova.compute.manager [instance: b53754b1-ec31-4dc3-a839-771f71affe1e] for vif in network_info: [ 631.835533] env[63515]: ERROR nova.compute.manager [instance: b53754b1-ec31-4dc3-a839-771f71affe1e] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 631.835533] env[63515]: ERROR nova.compute.manager [instance: b53754b1-ec31-4dc3-a839-771f71affe1e] return self._sync_wrapper(fn, *args, **kwargs) [ 631.835533] env[63515]: ERROR nova.compute.manager [instance: b53754b1-ec31-4dc3-a839-771f71affe1e] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 631.835533] env[63515]: ERROR nova.compute.manager [instance: b53754b1-ec31-4dc3-a839-771f71affe1e] self.wait() [ 631.835533] env[63515]: ERROR nova.compute.manager [instance: b53754b1-ec31-4dc3-a839-771f71affe1e] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 631.835533] env[63515]: ERROR nova.compute.manager [instance: b53754b1-ec31-4dc3-a839-771f71affe1e] self[:] = self._gt.wait() [ 631.835533] env[63515]: ERROR nova.compute.manager [instance: b53754b1-ec31-4dc3-a839-771f71affe1e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 631.835533] env[63515]: ERROR nova.compute.manager [instance: b53754b1-ec31-4dc3-a839-771f71affe1e] return self._exit_event.wait() [ 631.835533] env[63515]: ERROR nova.compute.manager [instance: b53754b1-ec31-4dc3-a839-771f71affe1e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 631.835899] env[63515]: ERROR nova.compute.manager [instance: b53754b1-ec31-4dc3-a839-771f71affe1e] current.throw(*self._exc) [ 631.835899] env[63515]: ERROR nova.compute.manager [instance: b53754b1-ec31-4dc3-a839-771f71affe1e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 631.835899] env[63515]: ERROR nova.compute.manager [instance: b53754b1-ec31-4dc3-a839-771f71affe1e] result = function(*args, **kwargs) [ 631.835899] env[63515]: ERROR nova.compute.manager [instance: b53754b1-ec31-4dc3-a839-771f71affe1e] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 631.835899] env[63515]: ERROR nova.compute.manager [instance: b53754b1-ec31-4dc3-a839-771f71affe1e] return func(*args, **kwargs) [ 631.835899] env[63515]: ERROR nova.compute.manager [instance: b53754b1-ec31-4dc3-a839-771f71affe1e] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 631.835899] env[63515]: ERROR nova.compute.manager [instance: b53754b1-ec31-4dc3-a839-771f71affe1e] raise e [ 631.835899] env[63515]: ERROR nova.compute.manager [instance: b53754b1-ec31-4dc3-a839-771f71affe1e] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 631.835899] env[63515]: ERROR nova.compute.manager [instance: b53754b1-ec31-4dc3-a839-771f71affe1e] nwinfo = self.network_api.allocate_for_instance( [ 631.835899] env[63515]: ERROR nova.compute.manager [instance: b53754b1-ec31-4dc3-a839-771f71affe1e] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 631.835899] env[63515]: ERROR nova.compute.manager [instance: b53754b1-ec31-4dc3-a839-771f71affe1e] created_port_ids = self._update_ports_for_instance( [ 631.835899] env[63515]: ERROR nova.compute.manager [instance: b53754b1-ec31-4dc3-a839-771f71affe1e] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 631.835899] env[63515]: ERROR nova.compute.manager [instance: b53754b1-ec31-4dc3-a839-771f71affe1e] with excutils.save_and_reraise_exception(): [ 631.836289] env[63515]: ERROR nova.compute.manager [instance: b53754b1-ec31-4dc3-a839-771f71affe1e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 631.836289] env[63515]: ERROR nova.compute.manager [instance: b53754b1-ec31-4dc3-a839-771f71affe1e] self.force_reraise() [ 631.836289] env[63515]: ERROR nova.compute.manager [instance: b53754b1-ec31-4dc3-a839-771f71affe1e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 631.836289] env[63515]: ERROR nova.compute.manager [instance: b53754b1-ec31-4dc3-a839-771f71affe1e] raise self.value [ 631.836289] env[63515]: ERROR nova.compute.manager [instance: b53754b1-ec31-4dc3-a839-771f71affe1e] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 631.836289] env[63515]: ERROR nova.compute.manager [instance: b53754b1-ec31-4dc3-a839-771f71affe1e] updated_port = self._update_port( [ 631.836289] env[63515]: ERROR nova.compute.manager [instance: b53754b1-ec31-4dc3-a839-771f71affe1e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 631.836289] env[63515]: ERROR nova.compute.manager [instance: b53754b1-ec31-4dc3-a839-771f71affe1e] _ensure_no_port_binding_failure(port) [ 631.836289] env[63515]: ERROR nova.compute.manager [instance: b53754b1-ec31-4dc3-a839-771f71affe1e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 631.836289] env[63515]: ERROR nova.compute.manager [instance: b53754b1-ec31-4dc3-a839-771f71affe1e] raise exception.PortBindingFailed(port_id=port['id']) [ 631.836289] env[63515]: ERROR nova.compute.manager [instance: b53754b1-ec31-4dc3-a839-771f71affe1e] nova.exception.PortBindingFailed: Binding failed for port 46784f84-e839-4fa3-88ea-2005d2779b07, please check neutron logs for more information. [ 631.836289] env[63515]: ERROR nova.compute.manager [instance: b53754b1-ec31-4dc3-a839-771f71affe1e] [ 631.836839] env[63515]: DEBUG nova.compute.utils [None req-9f5c50b0-6bbe-4982-8818-c7255a8c23ab tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: b53754b1-ec31-4dc3-a839-771f71affe1e] Binding failed for port 46784f84-e839-4fa3-88ea-2005d2779b07, please check neutron logs for more information. {{(pid=63515) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 631.836839] env[63515]: DEBUG oslo_concurrency.lockutils [None req-4334f93e-6967-4c78-b6ec-145141dc41da tempest-AttachInterfacesV270Test-352099301 tempest-AttachInterfacesV270Test-352099301-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.186s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 631.845549] env[63515]: INFO nova.compute.claims [None req-4334f93e-6967-4c78-b6ec-145141dc41da tempest-AttachInterfacesV270Test-352099301 tempest-AttachInterfacesV270Test-352099301-project-member] [instance: 6876f61c-2315-4e49-9944-4a5caddede5b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 631.848237] env[63515]: DEBUG nova.compute.manager [None req-9f5c50b0-6bbe-4982-8818-c7255a8c23ab tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: b53754b1-ec31-4dc3-a839-771f71affe1e] Build of instance b53754b1-ec31-4dc3-a839-771f71affe1e was re-scheduled: Binding failed for port 46784f84-e839-4fa3-88ea-2005d2779b07, please check neutron logs for more information. {{(pid=63515) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 631.849110] env[63515]: DEBUG nova.compute.manager [None req-9f5c50b0-6bbe-4982-8818-c7255a8c23ab tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: b53754b1-ec31-4dc3-a839-771f71affe1e] Unplugging VIFs for instance {{(pid=63515) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 631.849785] env[63515]: DEBUG oslo_concurrency.lockutils [None req-9f5c50b0-6bbe-4982-8818-c7255a8c23ab tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Acquiring lock "refresh_cache-b53754b1-ec31-4dc3-a839-771f71affe1e" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 631.849946] env[63515]: DEBUG oslo_concurrency.lockutils [None req-9f5c50b0-6bbe-4982-8818-c7255a8c23ab tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Acquired lock "refresh_cache-b53754b1-ec31-4dc3-a839-771f71affe1e" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 631.850121] env[63515]: DEBUG nova.network.neutron [None req-9f5c50b0-6bbe-4982-8818-c7255a8c23ab tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: b53754b1-ec31-4dc3-a839-771f71affe1e] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 632.002169] env[63515]: DEBUG nova.network.neutron [req-7c4cee6c-59c3-4733-b459-35047bff205a req-210ee58f-9bd7-49f9-bea1-bf40e692754b service nova] [instance: 2728c9f9-0d03-4313-ba2a-3dc22aff8538] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 632.046984] env[63515]: DEBUG nova.network.neutron [-] [instance: 2728c9f9-0d03-4313-ba2a-3dc22aff8538] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 632.144376] env[63515]: DEBUG nova.network.neutron [req-7c4cee6c-59c3-4733-b459-35047bff205a req-210ee58f-9bd7-49f9-bea1-bf40e692754b service nova] [instance: 2728c9f9-0d03-4313-ba2a-3dc22aff8538] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 632.187244] env[63515]: INFO nova.compute.manager [None req-d9ffa9d5-6856-47e5-8926-d090bbea2c09 tempest-DeleteServersAdminTestJSON-944542165 tempest-DeleteServersAdminTestJSON-944542165-project-member] [instance: ae5c6e69-a885-48e3-9c9e-017f33430edd] Took 1.03 seconds to deallocate network for instance. [ 632.383547] env[63515]: DEBUG nova.network.neutron [None req-9f5c50b0-6bbe-4982-8818-c7255a8c23ab tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: b53754b1-ec31-4dc3-a839-771f71affe1e] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 632.498192] env[63515]: DEBUG oslo_concurrency.lockutils [None req-4bc5b4e9-701c-413d-abd5-d300a7447cb8 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Acquiring lock "ed550b10-d58f-45b8-b766-198f431c3788" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 632.498461] env[63515]: DEBUG oslo_concurrency.lockutils [None req-4bc5b4e9-701c-413d-abd5-d300a7447cb8 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Lock "ed550b10-d58f-45b8-b766-198f431c3788" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 632.506738] env[63515]: DEBUG nova.network.neutron [None req-9f5c50b0-6bbe-4982-8818-c7255a8c23ab tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: b53754b1-ec31-4dc3-a839-771f71affe1e] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 632.551736] env[63515]: INFO nova.compute.manager [-] [instance: 2728c9f9-0d03-4313-ba2a-3dc22aff8538] Took 1.03 seconds to deallocate network for instance. [ 632.554184] env[63515]: DEBUG nova.compute.claims [None req-292f2bc3-2911-4498-95fb-4e4a03681601 tempest-FloatingIPsAssociationTestJSON-629096942 tempest-FloatingIPsAssociationTestJSON-629096942-project-member] [instance: 2728c9f9-0d03-4313-ba2a-3dc22aff8538] Aborting claim: {{(pid=63515) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 632.554356] env[63515]: DEBUG oslo_concurrency.lockutils [None req-292f2bc3-2911-4498-95fb-4e4a03681601 tempest-FloatingIPsAssociationTestJSON-629096942 tempest-FloatingIPsAssociationTestJSON-629096942-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 632.647666] env[63515]: DEBUG oslo_concurrency.lockutils [req-7c4cee6c-59c3-4733-b459-35047bff205a req-210ee58f-9bd7-49f9-bea1-bf40e692754b service nova] Releasing lock "refresh_cache-2728c9f9-0d03-4313-ba2a-3dc22aff8538" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 632.648066] env[63515]: DEBUG nova.compute.manager [req-7c4cee6c-59c3-4733-b459-35047bff205a req-210ee58f-9bd7-49f9-bea1-bf40e692754b service nova] [instance: 2728c9f9-0d03-4313-ba2a-3dc22aff8538] Received event network-vif-deleted-bfaff379-5880-4a38-86aa-c5d6331eb307 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 632.951521] env[63515]: DEBUG oslo_concurrency.lockutils [None req-3b658824-6e25-4b06-878c-99b3be1f1d31 tempest-ServerTagsTestJSON-1170819896 tempest-ServerTagsTestJSON-1170819896-project-member] Acquiring lock "de210780-5c0f-4fba-883c-13707566a2e1" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 632.951769] env[63515]: DEBUG oslo_concurrency.lockutils [None req-3b658824-6e25-4b06-878c-99b3be1f1d31 tempest-ServerTagsTestJSON-1170819896 tempest-ServerTagsTestJSON-1170819896-project-member] Lock "de210780-5c0f-4fba-883c-13707566a2e1" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 633.009330] env[63515]: DEBUG oslo_concurrency.lockutils [None req-9f5c50b0-6bbe-4982-8818-c7255a8c23ab tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Releasing lock "refresh_cache-b53754b1-ec31-4dc3-a839-771f71affe1e" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 633.009584] env[63515]: DEBUG nova.compute.manager [None req-9f5c50b0-6bbe-4982-8818-c7255a8c23ab tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63515) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 633.009735] env[63515]: DEBUG nova.compute.manager [None req-9f5c50b0-6bbe-4982-8818-c7255a8c23ab tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: b53754b1-ec31-4dc3-a839-771f71affe1e] Deallocating network for instance {{(pid=63515) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 633.009903] env[63515]: DEBUG nova.network.neutron [None req-9f5c50b0-6bbe-4982-8818-c7255a8c23ab tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: b53754b1-ec31-4dc3-a839-771f71affe1e] deallocate_for_instance() {{(pid=63515) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 633.031739] env[63515]: DEBUG nova.network.neutron [None req-9f5c50b0-6bbe-4982-8818-c7255a8c23ab tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: b53754b1-ec31-4dc3-a839-771f71affe1e] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 633.226110] env[63515]: INFO nova.scheduler.client.report [None req-d9ffa9d5-6856-47e5-8926-d090bbea2c09 tempest-DeleteServersAdminTestJSON-944542165 tempest-DeleteServersAdminTestJSON-944542165-project-member] Deleted allocations for instance ae5c6e69-a885-48e3-9c9e-017f33430edd [ 633.274757] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-278738d3-8be1-44f3-b99b-9ed2d22d1764 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.287365] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c72d43f3-639c-4fb2-8823-a74ce1359ef4 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.319667] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60cccb4e-b679-4bad-a154-0b9fa61e3eea {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.327342] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7c61f18-714a-4a2f-80fa-751833f0741b {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.341049] env[63515]: DEBUG nova.compute.provider_tree [None req-4334f93e-6967-4c78-b6ec-145141dc41da tempest-AttachInterfacesV270Test-352099301 tempest-AttachInterfacesV270Test-352099301-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 633.535171] env[63515]: DEBUG nova.network.neutron [None req-9f5c50b0-6bbe-4982-8818-c7255a8c23ab tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: b53754b1-ec31-4dc3-a839-771f71affe1e] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 633.738643] env[63515]: DEBUG oslo_concurrency.lockutils [None req-d9ffa9d5-6856-47e5-8926-d090bbea2c09 tempest-DeleteServersAdminTestJSON-944542165 tempest-DeleteServersAdminTestJSON-944542165-project-member] Lock "ae5c6e69-a885-48e3-9c9e-017f33430edd" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 80.045s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 633.846826] env[63515]: DEBUG nova.scheduler.client.report [None req-4334f93e-6967-4c78-b6ec-145141dc41da tempest-AttachInterfacesV270Test-352099301 tempest-AttachInterfacesV270Test-352099301-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 634.041508] env[63515]: INFO nova.compute.manager [None req-9f5c50b0-6bbe-4982-8818-c7255a8c23ab tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: b53754b1-ec31-4dc3-a839-771f71affe1e] Took 1.03 seconds to deallocate network for instance. [ 634.241213] env[63515]: DEBUG nova.compute.manager [None req-fa933c74-4a5b-4cc9-a499-9ea16a38d8b7 tempest-ServerActionsTestJSON-1591637270 tempest-ServerActionsTestJSON-1591637270-project-member] [instance: 005f45b3-630d-400f-9605-100f6aad88e6] Starting instance... {{(pid=63515) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 634.357042] env[63515]: DEBUG oslo_concurrency.lockutils [None req-4334f93e-6967-4c78-b6ec-145141dc41da tempest-AttachInterfacesV270Test-352099301 tempest-AttachInterfacesV270Test-352099301-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.522s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 634.357600] env[63515]: DEBUG nova.compute.manager [None req-4334f93e-6967-4c78-b6ec-145141dc41da tempest-AttachInterfacesV270Test-352099301 tempest-AttachInterfacesV270Test-352099301-project-member] [instance: 6876f61c-2315-4e49-9944-4a5caddede5b] Start building networks asynchronously for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 634.361129] env[63515]: DEBUG oslo_concurrency.lockutils [None req-886adfc0-007d-4f83-bc30-b755d00d04c7 tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.760s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 634.773193] env[63515]: DEBUG oslo_concurrency.lockutils [None req-fa933c74-4a5b-4cc9-a499-9ea16a38d8b7 tempest-ServerActionsTestJSON-1591637270 tempest-ServerActionsTestJSON-1591637270-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 634.868948] env[63515]: DEBUG nova.compute.utils [None req-4334f93e-6967-4c78-b6ec-145141dc41da tempest-AttachInterfacesV270Test-352099301 tempest-AttachInterfacesV270Test-352099301-project-member] Using /dev/sd instead of None {{(pid=63515) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 634.870650] env[63515]: DEBUG nova.compute.manager [None req-4334f93e-6967-4c78-b6ec-145141dc41da tempest-AttachInterfacesV270Test-352099301 tempest-AttachInterfacesV270Test-352099301-project-member] [instance: 6876f61c-2315-4e49-9944-4a5caddede5b] Allocating IP information in the background. {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 634.870855] env[63515]: DEBUG nova.network.neutron [None req-4334f93e-6967-4c78-b6ec-145141dc41da tempest-AttachInterfacesV270Test-352099301 tempest-AttachInterfacesV270Test-352099301-project-member] [instance: 6876f61c-2315-4e49-9944-4a5caddede5b] allocate_for_instance() {{(pid=63515) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 634.918685] env[63515]: DEBUG nova.policy [None req-4334f93e-6967-4c78-b6ec-145141dc41da tempest-AttachInterfacesV270Test-352099301 tempest-AttachInterfacesV270Test-352099301-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b8ac8eca9c3844f9980e11af28cc3d55', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5540dca5c21640438d96b95e7c7b27b6', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63515) authorize /opt/stack/nova/nova/policy.py:201}} [ 635.096808] env[63515]: INFO nova.scheduler.client.report [None req-9f5c50b0-6bbe-4982-8818-c7255a8c23ab tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Deleted allocations for instance b53754b1-ec31-4dc3-a839-771f71affe1e [ 635.273563] env[63515]: DEBUG nova.network.neutron [None req-4334f93e-6967-4c78-b6ec-145141dc41da tempest-AttachInterfacesV270Test-352099301 tempest-AttachInterfacesV270Test-352099301-project-member] [instance: 6876f61c-2315-4e49-9944-4a5caddede5b] Successfully created port: d382abb3-d2ce-4738-ab1a-4fd2da50fa06 {{(pid=63515) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 635.281328] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9779e463-ea43-4f21-9973-377018710446 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.288867] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7214f85-1154-4359-835c-db3b3e8ce41e {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.319578] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88ae172b-5813-41d3-93df-688cf4de1e7a {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.326835] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6de222a6-3e0a-4754-946d-e8c8d8db7561 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.340240] env[63515]: DEBUG nova.compute.provider_tree [None req-886adfc0-007d-4f83-bc30-b755d00d04c7 tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 635.374032] env[63515]: DEBUG nova.compute.manager [None req-4334f93e-6967-4c78-b6ec-145141dc41da tempest-AttachInterfacesV270Test-352099301 tempest-AttachInterfacesV270Test-352099301-project-member] [instance: 6876f61c-2315-4e49-9944-4a5caddede5b] Start building block device mappings for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 635.609849] env[63515]: DEBUG oslo_concurrency.lockutils [None req-9f5c50b0-6bbe-4982-8818-c7255a8c23ab tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Lock "b53754b1-ec31-4dc3-a839-771f71affe1e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 80.185s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 635.844689] env[63515]: DEBUG nova.scheduler.client.report [None req-886adfc0-007d-4f83-bc30-b755d00d04c7 tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 636.112560] env[63515]: DEBUG nova.compute.manager [None req-45f6ec8d-6d74-4e46-8fdf-6b0f7c107bf4 tempest-ServerDiagnosticsTest-1160321831 tempest-ServerDiagnosticsTest-1160321831-project-member] [instance: 4ad5440f-cd18-44c5-8836-0aa39824cf03] Starting instance... {{(pid=63515) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 636.192248] env[63515]: DEBUG nova.compute.manager [req-f4051ef5-c8ba-47d4-bfe7-fc43bbada7b7 req-7365fc3f-7a6c-467e-b7b1-175910fb69e6 service nova] [instance: 6876f61c-2315-4e49-9944-4a5caddede5b] Received event network-changed-d382abb3-d2ce-4738-ab1a-4fd2da50fa06 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 636.192421] env[63515]: DEBUG nova.compute.manager [req-f4051ef5-c8ba-47d4-bfe7-fc43bbada7b7 req-7365fc3f-7a6c-467e-b7b1-175910fb69e6 service nova] [instance: 6876f61c-2315-4e49-9944-4a5caddede5b] Refreshing instance network info cache due to event network-changed-d382abb3-d2ce-4738-ab1a-4fd2da50fa06. {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 636.192643] env[63515]: DEBUG oslo_concurrency.lockutils [req-f4051ef5-c8ba-47d4-bfe7-fc43bbada7b7 req-7365fc3f-7a6c-467e-b7b1-175910fb69e6 service nova] Acquiring lock "refresh_cache-6876f61c-2315-4e49-9944-4a5caddede5b" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 636.192788] env[63515]: DEBUG oslo_concurrency.lockutils [req-f4051ef5-c8ba-47d4-bfe7-fc43bbada7b7 req-7365fc3f-7a6c-467e-b7b1-175910fb69e6 service nova] Acquired lock "refresh_cache-6876f61c-2315-4e49-9944-4a5caddede5b" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 636.192949] env[63515]: DEBUG nova.network.neutron [req-f4051ef5-c8ba-47d4-bfe7-fc43bbada7b7 req-7365fc3f-7a6c-467e-b7b1-175910fb69e6 service nova] [instance: 6876f61c-2315-4e49-9944-4a5caddede5b] Refreshing network info cache for port d382abb3-d2ce-4738-ab1a-4fd2da50fa06 {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 636.328788] env[63515]: ERROR nova.compute.manager [None req-4334f93e-6967-4c78-b6ec-145141dc41da tempest-AttachInterfacesV270Test-352099301 tempest-AttachInterfacesV270Test-352099301-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port d382abb3-d2ce-4738-ab1a-4fd2da50fa06, please check neutron logs for more information. [ 636.328788] env[63515]: ERROR nova.compute.manager Traceback (most recent call last): [ 636.328788] env[63515]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 636.328788] env[63515]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 636.328788] env[63515]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 636.328788] env[63515]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 636.328788] env[63515]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 636.328788] env[63515]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 636.328788] env[63515]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 636.328788] env[63515]: ERROR nova.compute.manager self.force_reraise() [ 636.328788] env[63515]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 636.328788] env[63515]: ERROR nova.compute.manager raise self.value [ 636.328788] env[63515]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 636.328788] env[63515]: ERROR nova.compute.manager updated_port = self._update_port( [ 636.328788] env[63515]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 636.328788] env[63515]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 636.329441] env[63515]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 636.329441] env[63515]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 636.329441] env[63515]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port d382abb3-d2ce-4738-ab1a-4fd2da50fa06, please check neutron logs for more information. [ 636.329441] env[63515]: ERROR nova.compute.manager [ 636.329441] env[63515]: Traceback (most recent call last): [ 636.329441] env[63515]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 636.329441] env[63515]: listener.cb(fileno) [ 636.329441] env[63515]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 636.329441] env[63515]: result = function(*args, **kwargs) [ 636.329441] env[63515]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 636.329441] env[63515]: return func(*args, **kwargs) [ 636.329441] env[63515]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 636.329441] env[63515]: raise e [ 636.329441] env[63515]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 636.329441] env[63515]: nwinfo = self.network_api.allocate_for_instance( [ 636.329441] env[63515]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 636.329441] env[63515]: created_port_ids = self._update_ports_for_instance( [ 636.329441] env[63515]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 636.329441] env[63515]: with excutils.save_and_reraise_exception(): [ 636.329441] env[63515]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 636.329441] env[63515]: self.force_reraise() [ 636.329441] env[63515]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 636.329441] env[63515]: raise self.value [ 636.329441] env[63515]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 636.329441] env[63515]: updated_port = self._update_port( [ 636.329441] env[63515]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 636.329441] env[63515]: _ensure_no_port_binding_failure(port) [ 636.329441] env[63515]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 636.329441] env[63515]: raise exception.PortBindingFailed(port_id=port['id']) [ 636.330617] env[63515]: nova.exception.PortBindingFailed: Binding failed for port d382abb3-d2ce-4738-ab1a-4fd2da50fa06, please check neutron logs for more information. [ 636.330617] env[63515]: Removing descriptor: 17 [ 636.359020] env[63515]: DEBUG oslo_concurrency.lockutils [None req-886adfc0-007d-4f83-bc30-b755d00d04c7 tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.995s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 636.359020] env[63515]: ERROR nova.compute.manager [None req-886adfc0-007d-4f83-bc30-b755d00d04c7 tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] [instance: 8b65c223-03a8-41e0-917e-45dd76b6e57f] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port b7d27a8e-e6d6-42b4-a4bf-50e668e77788, please check neutron logs for more information. [ 636.359020] env[63515]: ERROR nova.compute.manager [instance: 8b65c223-03a8-41e0-917e-45dd76b6e57f] Traceback (most recent call last): [ 636.359020] env[63515]: ERROR nova.compute.manager [instance: 8b65c223-03a8-41e0-917e-45dd76b6e57f] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 636.359020] env[63515]: ERROR nova.compute.manager [instance: 8b65c223-03a8-41e0-917e-45dd76b6e57f] self.driver.spawn(context, instance, image_meta, [ 636.359020] env[63515]: ERROR nova.compute.manager [instance: 8b65c223-03a8-41e0-917e-45dd76b6e57f] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 636.359020] env[63515]: ERROR nova.compute.manager [instance: 8b65c223-03a8-41e0-917e-45dd76b6e57f] self._vmops.spawn(context, instance, image_meta, injected_files, [ 636.359020] env[63515]: ERROR nova.compute.manager [instance: 8b65c223-03a8-41e0-917e-45dd76b6e57f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 636.359020] env[63515]: ERROR nova.compute.manager [instance: 8b65c223-03a8-41e0-917e-45dd76b6e57f] vm_ref = self.build_virtual_machine(instance, [ 636.359554] env[63515]: ERROR nova.compute.manager [instance: 8b65c223-03a8-41e0-917e-45dd76b6e57f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 636.359554] env[63515]: ERROR nova.compute.manager [instance: 8b65c223-03a8-41e0-917e-45dd76b6e57f] vif_infos = vmwarevif.get_vif_info(self._session, [ 636.359554] env[63515]: ERROR nova.compute.manager [instance: 8b65c223-03a8-41e0-917e-45dd76b6e57f] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 636.359554] env[63515]: ERROR nova.compute.manager [instance: 8b65c223-03a8-41e0-917e-45dd76b6e57f] for vif in network_info: [ 636.359554] env[63515]: ERROR nova.compute.manager [instance: 8b65c223-03a8-41e0-917e-45dd76b6e57f] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 636.359554] env[63515]: ERROR nova.compute.manager [instance: 8b65c223-03a8-41e0-917e-45dd76b6e57f] return self._sync_wrapper(fn, *args, **kwargs) [ 636.359554] env[63515]: ERROR nova.compute.manager [instance: 8b65c223-03a8-41e0-917e-45dd76b6e57f] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 636.359554] env[63515]: ERROR nova.compute.manager [instance: 8b65c223-03a8-41e0-917e-45dd76b6e57f] self.wait() [ 636.359554] env[63515]: ERROR nova.compute.manager [instance: 8b65c223-03a8-41e0-917e-45dd76b6e57f] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 636.359554] env[63515]: ERROR nova.compute.manager [instance: 8b65c223-03a8-41e0-917e-45dd76b6e57f] self[:] = self._gt.wait() [ 636.359554] env[63515]: ERROR nova.compute.manager [instance: 8b65c223-03a8-41e0-917e-45dd76b6e57f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 636.359554] env[63515]: ERROR nova.compute.manager [instance: 8b65c223-03a8-41e0-917e-45dd76b6e57f] return self._exit_event.wait() [ 636.359554] env[63515]: ERROR nova.compute.manager [instance: 8b65c223-03a8-41e0-917e-45dd76b6e57f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 636.359969] env[63515]: ERROR nova.compute.manager [instance: 8b65c223-03a8-41e0-917e-45dd76b6e57f] current.throw(*self._exc) [ 636.359969] env[63515]: ERROR nova.compute.manager [instance: 8b65c223-03a8-41e0-917e-45dd76b6e57f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 636.359969] env[63515]: ERROR nova.compute.manager [instance: 8b65c223-03a8-41e0-917e-45dd76b6e57f] result = function(*args, **kwargs) [ 636.359969] env[63515]: ERROR nova.compute.manager [instance: 8b65c223-03a8-41e0-917e-45dd76b6e57f] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 636.359969] env[63515]: ERROR nova.compute.manager [instance: 8b65c223-03a8-41e0-917e-45dd76b6e57f] return func(*args, **kwargs) [ 636.359969] env[63515]: ERROR nova.compute.manager [instance: 8b65c223-03a8-41e0-917e-45dd76b6e57f] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 636.359969] env[63515]: ERROR nova.compute.manager [instance: 8b65c223-03a8-41e0-917e-45dd76b6e57f] raise e [ 636.359969] env[63515]: ERROR nova.compute.manager [instance: 8b65c223-03a8-41e0-917e-45dd76b6e57f] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 636.359969] env[63515]: ERROR nova.compute.manager [instance: 8b65c223-03a8-41e0-917e-45dd76b6e57f] nwinfo = self.network_api.allocate_for_instance( [ 636.359969] env[63515]: ERROR nova.compute.manager [instance: 8b65c223-03a8-41e0-917e-45dd76b6e57f] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 636.359969] env[63515]: ERROR nova.compute.manager [instance: 8b65c223-03a8-41e0-917e-45dd76b6e57f] created_port_ids = self._update_ports_for_instance( [ 636.359969] env[63515]: ERROR nova.compute.manager [instance: 8b65c223-03a8-41e0-917e-45dd76b6e57f] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 636.359969] env[63515]: ERROR nova.compute.manager [instance: 8b65c223-03a8-41e0-917e-45dd76b6e57f] with excutils.save_and_reraise_exception(): [ 636.360303] env[63515]: ERROR nova.compute.manager [instance: 8b65c223-03a8-41e0-917e-45dd76b6e57f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 636.360303] env[63515]: ERROR nova.compute.manager [instance: 8b65c223-03a8-41e0-917e-45dd76b6e57f] self.force_reraise() [ 636.360303] env[63515]: ERROR nova.compute.manager [instance: 8b65c223-03a8-41e0-917e-45dd76b6e57f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 636.360303] env[63515]: ERROR nova.compute.manager [instance: 8b65c223-03a8-41e0-917e-45dd76b6e57f] raise self.value [ 636.360303] env[63515]: ERROR nova.compute.manager [instance: 8b65c223-03a8-41e0-917e-45dd76b6e57f] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 636.360303] env[63515]: ERROR nova.compute.manager [instance: 8b65c223-03a8-41e0-917e-45dd76b6e57f] updated_port = self._update_port( [ 636.360303] env[63515]: ERROR nova.compute.manager [instance: 8b65c223-03a8-41e0-917e-45dd76b6e57f] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 636.360303] env[63515]: ERROR nova.compute.manager [instance: 8b65c223-03a8-41e0-917e-45dd76b6e57f] _ensure_no_port_binding_failure(port) [ 636.360303] env[63515]: ERROR nova.compute.manager [instance: 8b65c223-03a8-41e0-917e-45dd76b6e57f] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 636.360303] env[63515]: ERROR nova.compute.manager [instance: 8b65c223-03a8-41e0-917e-45dd76b6e57f] raise exception.PortBindingFailed(port_id=port['id']) [ 636.360303] env[63515]: ERROR nova.compute.manager [instance: 8b65c223-03a8-41e0-917e-45dd76b6e57f] nova.exception.PortBindingFailed: Binding failed for port b7d27a8e-e6d6-42b4-a4bf-50e668e77788, please check neutron logs for more information. [ 636.360303] env[63515]: ERROR nova.compute.manager [instance: 8b65c223-03a8-41e0-917e-45dd76b6e57f] [ 636.360585] env[63515]: DEBUG nova.compute.utils [None req-886adfc0-007d-4f83-bc30-b755d00d04c7 tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] [instance: 8b65c223-03a8-41e0-917e-45dd76b6e57f] Binding failed for port b7d27a8e-e6d6-42b4-a4bf-50e668e77788, please check neutron logs for more information. {{(pid=63515) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 636.360585] env[63515]: DEBUG oslo_concurrency.lockutils [None req-30030db8-77b3-4af4-8ae6-30653bd81e23 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.172s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 636.361962] env[63515]: INFO nova.compute.claims [None req-30030db8-77b3-4af4-8ae6-30653bd81e23 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: 02ffb723-d6df-42cb-93c5-d582705d1e03] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 636.366910] env[63515]: DEBUG nova.compute.manager [None req-886adfc0-007d-4f83-bc30-b755d00d04c7 tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] [instance: 8b65c223-03a8-41e0-917e-45dd76b6e57f] Build of instance 8b65c223-03a8-41e0-917e-45dd76b6e57f was re-scheduled: Binding failed for port b7d27a8e-e6d6-42b4-a4bf-50e668e77788, please check neutron logs for more information. {{(pid=63515) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 636.366910] env[63515]: DEBUG nova.compute.manager [None req-886adfc0-007d-4f83-bc30-b755d00d04c7 tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] [instance: 8b65c223-03a8-41e0-917e-45dd76b6e57f] Unplugging VIFs for instance {{(pid=63515) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 636.366910] env[63515]: DEBUG oslo_concurrency.lockutils [None req-886adfc0-007d-4f83-bc30-b755d00d04c7 tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] Acquiring lock "refresh_cache-8b65c223-03a8-41e0-917e-45dd76b6e57f" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 636.366910] env[63515]: DEBUG oslo_concurrency.lockutils [None req-886adfc0-007d-4f83-bc30-b755d00d04c7 tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] Acquired lock "refresh_cache-8b65c223-03a8-41e0-917e-45dd76b6e57f" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 636.367073] env[63515]: DEBUG nova.network.neutron [None req-886adfc0-007d-4f83-bc30-b755d00d04c7 tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] [instance: 8b65c223-03a8-41e0-917e-45dd76b6e57f] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 636.388995] env[63515]: DEBUG nova.compute.manager [None req-4334f93e-6967-4c78-b6ec-145141dc41da tempest-AttachInterfacesV270Test-352099301 tempest-AttachInterfacesV270Test-352099301-project-member] [instance: 6876f61c-2315-4e49-9944-4a5caddede5b] Start spawning the instance on the hypervisor. {{(pid=63515) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 636.415337] env[63515]: DEBUG nova.virt.hardware [None req-4334f93e-6967-4c78-b6ec-145141dc41da tempest-AttachInterfacesV270Test-352099301 tempest-AttachInterfacesV270Test-352099301-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T02:52:14Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T02:51:56Z,direct_url=,disk_format='vmdk',id=8a120570-cb06-4099-b262-554ca0ad15c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b569255cc43e42e7ae3f2b4ad37c6ef4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T02:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 636.415598] env[63515]: DEBUG nova.virt.hardware [None req-4334f93e-6967-4c78-b6ec-145141dc41da tempest-AttachInterfacesV270Test-352099301 tempest-AttachInterfacesV270Test-352099301-project-member] Flavor limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 636.415756] env[63515]: DEBUG nova.virt.hardware [None req-4334f93e-6967-4c78-b6ec-145141dc41da tempest-AttachInterfacesV270Test-352099301 tempest-AttachInterfacesV270Test-352099301-project-member] Image limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 636.415937] env[63515]: DEBUG nova.virt.hardware [None req-4334f93e-6967-4c78-b6ec-145141dc41da tempest-AttachInterfacesV270Test-352099301 tempest-AttachInterfacesV270Test-352099301-project-member] Flavor pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 636.416096] env[63515]: DEBUG nova.virt.hardware [None req-4334f93e-6967-4c78-b6ec-145141dc41da tempest-AttachInterfacesV270Test-352099301 tempest-AttachInterfacesV270Test-352099301-project-member] Image pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 636.416243] env[63515]: DEBUG nova.virt.hardware [None req-4334f93e-6967-4c78-b6ec-145141dc41da tempest-AttachInterfacesV270Test-352099301 tempest-AttachInterfacesV270Test-352099301-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 636.416444] env[63515]: DEBUG nova.virt.hardware [None req-4334f93e-6967-4c78-b6ec-145141dc41da tempest-AttachInterfacesV270Test-352099301 tempest-AttachInterfacesV270Test-352099301-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 636.416599] env[63515]: DEBUG nova.virt.hardware [None req-4334f93e-6967-4c78-b6ec-145141dc41da tempest-AttachInterfacesV270Test-352099301 tempest-AttachInterfacesV270Test-352099301-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 636.416761] env[63515]: DEBUG nova.virt.hardware [None req-4334f93e-6967-4c78-b6ec-145141dc41da tempest-AttachInterfacesV270Test-352099301 tempest-AttachInterfacesV270Test-352099301-project-member] Got 1 possible topologies {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 636.416920] env[63515]: DEBUG nova.virt.hardware [None req-4334f93e-6967-4c78-b6ec-145141dc41da tempest-AttachInterfacesV270Test-352099301 tempest-AttachInterfacesV270Test-352099301-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 636.417382] env[63515]: DEBUG nova.virt.hardware [None req-4334f93e-6967-4c78-b6ec-145141dc41da tempest-AttachInterfacesV270Test-352099301 tempest-AttachInterfacesV270Test-352099301-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 636.418639] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5bdeb26-5c02-4653-b9d4-224001cd21e8 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.427810] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06db9c90-b0f0-484d-bebf-db527bee8f7d {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.441803] env[63515]: ERROR nova.compute.manager [None req-4334f93e-6967-4c78-b6ec-145141dc41da tempest-AttachInterfacesV270Test-352099301 tempest-AttachInterfacesV270Test-352099301-project-member] [instance: 6876f61c-2315-4e49-9944-4a5caddede5b] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port d382abb3-d2ce-4738-ab1a-4fd2da50fa06, please check neutron logs for more information. [ 636.441803] env[63515]: ERROR nova.compute.manager [instance: 6876f61c-2315-4e49-9944-4a5caddede5b] Traceback (most recent call last): [ 636.441803] env[63515]: ERROR nova.compute.manager [instance: 6876f61c-2315-4e49-9944-4a5caddede5b] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 636.441803] env[63515]: ERROR nova.compute.manager [instance: 6876f61c-2315-4e49-9944-4a5caddede5b] yield resources [ 636.441803] env[63515]: ERROR nova.compute.manager [instance: 6876f61c-2315-4e49-9944-4a5caddede5b] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 636.441803] env[63515]: ERROR nova.compute.manager [instance: 6876f61c-2315-4e49-9944-4a5caddede5b] self.driver.spawn(context, instance, image_meta, [ 636.441803] env[63515]: ERROR nova.compute.manager [instance: 6876f61c-2315-4e49-9944-4a5caddede5b] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 636.441803] env[63515]: ERROR nova.compute.manager [instance: 6876f61c-2315-4e49-9944-4a5caddede5b] self._vmops.spawn(context, instance, image_meta, injected_files, [ 636.441803] env[63515]: ERROR nova.compute.manager [instance: 6876f61c-2315-4e49-9944-4a5caddede5b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 636.441803] env[63515]: ERROR nova.compute.manager [instance: 6876f61c-2315-4e49-9944-4a5caddede5b] vm_ref = self.build_virtual_machine(instance, [ 636.441803] env[63515]: ERROR nova.compute.manager [instance: 6876f61c-2315-4e49-9944-4a5caddede5b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 636.442237] env[63515]: ERROR nova.compute.manager [instance: 6876f61c-2315-4e49-9944-4a5caddede5b] vif_infos = vmwarevif.get_vif_info(self._session, [ 636.442237] env[63515]: ERROR nova.compute.manager [instance: 6876f61c-2315-4e49-9944-4a5caddede5b] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 636.442237] env[63515]: ERROR nova.compute.manager [instance: 6876f61c-2315-4e49-9944-4a5caddede5b] for vif in network_info: [ 636.442237] env[63515]: ERROR nova.compute.manager [instance: 6876f61c-2315-4e49-9944-4a5caddede5b] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 636.442237] env[63515]: ERROR nova.compute.manager [instance: 6876f61c-2315-4e49-9944-4a5caddede5b] return self._sync_wrapper(fn, *args, **kwargs) [ 636.442237] env[63515]: ERROR nova.compute.manager [instance: 6876f61c-2315-4e49-9944-4a5caddede5b] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 636.442237] env[63515]: ERROR nova.compute.manager [instance: 6876f61c-2315-4e49-9944-4a5caddede5b] self.wait() [ 636.442237] env[63515]: ERROR nova.compute.manager [instance: 6876f61c-2315-4e49-9944-4a5caddede5b] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 636.442237] env[63515]: ERROR nova.compute.manager [instance: 6876f61c-2315-4e49-9944-4a5caddede5b] self[:] = self._gt.wait() [ 636.442237] env[63515]: ERROR nova.compute.manager [instance: 6876f61c-2315-4e49-9944-4a5caddede5b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 636.442237] env[63515]: ERROR nova.compute.manager [instance: 6876f61c-2315-4e49-9944-4a5caddede5b] return self._exit_event.wait() [ 636.442237] env[63515]: ERROR nova.compute.manager [instance: 6876f61c-2315-4e49-9944-4a5caddede5b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 636.442237] env[63515]: ERROR nova.compute.manager [instance: 6876f61c-2315-4e49-9944-4a5caddede5b] current.throw(*self._exc) [ 636.442620] env[63515]: ERROR nova.compute.manager [instance: 6876f61c-2315-4e49-9944-4a5caddede5b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 636.442620] env[63515]: ERROR nova.compute.manager [instance: 6876f61c-2315-4e49-9944-4a5caddede5b] result = function(*args, **kwargs) [ 636.442620] env[63515]: ERROR nova.compute.manager [instance: 6876f61c-2315-4e49-9944-4a5caddede5b] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 636.442620] env[63515]: ERROR nova.compute.manager [instance: 6876f61c-2315-4e49-9944-4a5caddede5b] return func(*args, **kwargs) [ 636.442620] env[63515]: ERROR nova.compute.manager [instance: 6876f61c-2315-4e49-9944-4a5caddede5b] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 636.442620] env[63515]: ERROR nova.compute.manager [instance: 6876f61c-2315-4e49-9944-4a5caddede5b] raise e [ 636.442620] env[63515]: ERROR nova.compute.manager [instance: 6876f61c-2315-4e49-9944-4a5caddede5b] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 636.442620] env[63515]: ERROR nova.compute.manager [instance: 6876f61c-2315-4e49-9944-4a5caddede5b] nwinfo = self.network_api.allocate_for_instance( [ 636.442620] env[63515]: ERROR nova.compute.manager [instance: 6876f61c-2315-4e49-9944-4a5caddede5b] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 636.442620] env[63515]: ERROR nova.compute.manager [instance: 6876f61c-2315-4e49-9944-4a5caddede5b] created_port_ids = self._update_ports_for_instance( [ 636.442620] env[63515]: ERROR nova.compute.manager [instance: 6876f61c-2315-4e49-9944-4a5caddede5b] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 636.442620] env[63515]: ERROR nova.compute.manager [instance: 6876f61c-2315-4e49-9944-4a5caddede5b] with excutils.save_and_reraise_exception(): [ 636.442620] env[63515]: ERROR nova.compute.manager [instance: 6876f61c-2315-4e49-9944-4a5caddede5b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 636.443059] env[63515]: ERROR nova.compute.manager [instance: 6876f61c-2315-4e49-9944-4a5caddede5b] self.force_reraise() [ 636.443059] env[63515]: ERROR nova.compute.manager [instance: 6876f61c-2315-4e49-9944-4a5caddede5b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 636.443059] env[63515]: ERROR nova.compute.manager [instance: 6876f61c-2315-4e49-9944-4a5caddede5b] raise self.value [ 636.443059] env[63515]: ERROR nova.compute.manager [instance: 6876f61c-2315-4e49-9944-4a5caddede5b] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 636.443059] env[63515]: ERROR nova.compute.manager [instance: 6876f61c-2315-4e49-9944-4a5caddede5b] updated_port = self._update_port( [ 636.443059] env[63515]: ERROR nova.compute.manager [instance: 6876f61c-2315-4e49-9944-4a5caddede5b] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 636.443059] env[63515]: ERROR nova.compute.manager [instance: 6876f61c-2315-4e49-9944-4a5caddede5b] _ensure_no_port_binding_failure(port) [ 636.443059] env[63515]: ERROR nova.compute.manager [instance: 6876f61c-2315-4e49-9944-4a5caddede5b] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 636.443059] env[63515]: ERROR nova.compute.manager [instance: 6876f61c-2315-4e49-9944-4a5caddede5b] raise exception.PortBindingFailed(port_id=port['id']) [ 636.443059] env[63515]: ERROR nova.compute.manager [instance: 6876f61c-2315-4e49-9944-4a5caddede5b] nova.exception.PortBindingFailed: Binding failed for port d382abb3-d2ce-4738-ab1a-4fd2da50fa06, please check neutron logs for more information. [ 636.443059] env[63515]: ERROR nova.compute.manager [instance: 6876f61c-2315-4e49-9944-4a5caddede5b] [ 636.443059] env[63515]: INFO nova.compute.manager [None req-4334f93e-6967-4c78-b6ec-145141dc41da tempest-AttachInterfacesV270Test-352099301 tempest-AttachInterfacesV270Test-352099301-project-member] [instance: 6876f61c-2315-4e49-9944-4a5caddede5b] Terminating instance [ 636.444135] env[63515]: DEBUG oslo_concurrency.lockutils [None req-4334f93e-6967-4c78-b6ec-145141dc41da tempest-AttachInterfacesV270Test-352099301 tempest-AttachInterfacesV270Test-352099301-project-member] Acquiring lock "refresh_cache-6876f61c-2315-4e49-9944-4a5caddede5b" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 636.632406] env[63515]: DEBUG oslo_concurrency.lockutils [None req-45f6ec8d-6d74-4e46-8fdf-6b0f7c107bf4 tempest-ServerDiagnosticsTest-1160321831 tempest-ServerDiagnosticsTest-1160321831-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 636.715663] env[63515]: DEBUG nova.network.neutron [req-f4051ef5-c8ba-47d4-bfe7-fc43bbada7b7 req-7365fc3f-7a6c-467e-b7b1-175910fb69e6 service nova] [instance: 6876f61c-2315-4e49-9944-4a5caddede5b] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 636.830696] env[63515]: DEBUG nova.network.neutron [req-f4051ef5-c8ba-47d4-bfe7-fc43bbada7b7 req-7365fc3f-7a6c-467e-b7b1-175910fb69e6 service nova] [instance: 6876f61c-2315-4e49-9944-4a5caddede5b] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 636.887691] env[63515]: DEBUG nova.network.neutron [None req-886adfc0-007d-4f83-bc30-b755d00d04c7 tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] [instance: 8b65c223-03a8-41e0-917e-45dd76b6e57f] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 636.956319] env[63515]: DEBUG nova.network.neutron [None req-886adfc0-007d-4f83-bc30-b755d00d04c7 tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] [instance: 8b65c223-03a8-41e0-917e-45dd76b6e57f] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 637.164445] env[63515]: DEBUG oslo_concurrency.lockutils [None req-fd4a0cad-edc5-4c70-94f6-406aa5060c1e tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Acquiring lock "001a48e1-b237-41f8-a6f4-2fe98d173512" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 637.164761] env[63515]: DEBUG oslo_concurrency.lockutils [None req-fd4a0cad-edc5-4c70-94f6-406aa5060c1e tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Lock "001a48e1-b237-41f8-a6f4-2fe98d173512" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 637.332877] env[63515]: DEBUG oslo_concurrency.lockutils [req-f4051ef5-c8ba-47d4-bfe7-fc43bbada7b7 req-7365fc3f-7a6c-467e-b7b1-175910fb69e6 service nova] Releasing lock "refresh_cache-6876f61c-2315-4e49-9944-4a5caddede5b" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 637.333335] env[63515]: DEBUG oslo_concurrency.lockutils [None req-4334f93e-6967-4c78-b6ec-145141dc41da tempest-AttachInterfacesV270Test-352099301 tempest-AttachInterfacesV270Test-352099301-project-member] Acquired lock "refresh_cache-6876f61c-2315-4e49-9944-4a5caddede5b" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 637.333523] env[63515]: DEBUG nova.network.neutron [None req-4334f93e-6967-4c78-b6ec-145141dc41da tempest-AttachInterfacesV270Test-352099301 tempest-AttachInterfacesV270Test-352099301-project-member] [instance: 6876f61c-2315-4e49-9944-4a5caddede5b] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 637.461572] env[63515]: DEBUG oslo_concurrency.lockutils [None req-886adfc0-007d-4f83-bc30-b755d00d04c7 tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] Releasing lock "refresh_cache-8b65c223-03a8-41e0-917e-45dd76b6e57f" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 637.461818] env[63515]: DEBUG nova.compute.manager [None req-886adfc0-007d-4f83-bc30-b755d00d04c7 tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63515) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 637.462037] env[63515]: DEBUG nova.compute.manager [None req-886adfc0-007d-4f83-bc30-b755d00d04c7 tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] [instance: 8b65c223-03a8-41e0-917e-45dd76b6e57f] Deallocating network for instance {{(pid=63515) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 637.462214] env[63515]: DEBUG nova.network.neutron [None req-886adfc0-007d-4f83-bc30-b755d00d04c7 tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] [instance: 8b65c223-03a8-41e0-917e-45dd76b6e57f] deallocate_for_instance() {{(pid=63515) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 637.539334] env[63515]: DEBUG nova.network.neutron [None req-886adfc0-007d-4f83-bc30-b755d00d04c7 tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] [instance: 8b65c223-03a8-41e0-917e-45dd76b6e57f] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 637.792590] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93c957da-3709-4ba6-b4ab-3369e9bdadbb {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.799929] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-106164d3-6199-495d-92df-ec27c764908e {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.831234] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1b55956-e195-4e50-830b-2f2befe72da3 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.840677] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-628bdbb9-f4f6-4569-9c76-0b8e62d2f506 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.856385] env[63515]: DEBUG nova.compute.provider_tree [None req-30030db8-77b3-4af4-8ae6-30653bd81e23 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 637.858960] env[63515]: DEBUG nova.network.neutron [None req-4334f93e-6967-4c78-b6ec-145141dc41da tempest-AttachInterfacesV270Test-352099301 tempest-AttachInterfacesV270Test-352099301-project-member] [instance: 6876f61c-2315-4e49-9944-4a5caddede5b] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 637.940216] env[63515]: DEBUG nova.network.neutron [None req-4334f93e-6967-4c78-b6ec-145141dc41da tempest-AttachInterfacesV270Test-352099301 tempest-AttachInterfacesV270Test-352099301-project-member] [instance: 6876f61c-2315-4e49-9944-4a5caddede5b] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 638.041590] env[63515]: DEBUG nova.network.neutron [None req-886adfc0-007d-4f83-bc30-b755d00d04c7 tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] [instance: 8b65c223-03a8-41e0-917e-45dd76b6e57f] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 638.301109] env[63515]: DEBUG nova.compute.manager [req-c172fe58-1274-4110-ac7c-591ef7e1e26c req-aa30844e-9fae-44aa-825d-530e7cf87786 service nova] [instance: 6876f61c-2315-4e49-9944-4a5caddede5b] Received event network-vif-deleted-d382abb3-d2ce-4738-ab1a-4fd2da50fa06 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 638.359782] env[63515]: DEBUG nova.scheduler.client.report [None req-30030db8-77b3-4af4-8ae6-30653bd81e23 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 638.443488] env[63515]: DEBUG oslo_concurrency.lockutils [None req-4334f93e-6967-4c78-b6ec-145141dc41da tempest-AttachInterfacesV270Test-352099301 tempest-AttachInterfacesV270Test-352099301-project-member] Releasing lock "refresh_cache-6876f61c-2315-4e49-9944-4a5caddede5b" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 638.444177] env[63515]: DEBUG nova.compute.manager [None req-4334f93e-6967-4c78-b6ec-145141dc41da tempest-AttachInterfacesV270Test-352099301 tempest-AttachInterfacesV270Test-352099301-project-member] [instance: 6876f61c-2315-4e49-9944-4a5caddede5b] Start destroying the instance on the hypervisor. {{(pid=63515) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 638.444768] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-4334f93e-6967-4c78-b6ec-145141dc41da tempest-AttachInterfacesV270Test-352099301 tempest-AttachInterfacesV270Test-352099301-project-member] [instance: 6876f61c-2315-4e49-9944-4a5caddede5b] Destroying instance {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 638.445735] env[63515]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a81a4545-0d14-4a41-944a-b342fc757236 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.455735] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e865976-3e37-46cf-82e6-03d1efa9ef39 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.478026] env[63515]: WARNING nova.virt.vmwareapi.vmops [None req-4334f93e-6967-4c78-b6ec-145141dc41da tempest-AttachInterfacesV270Test-352099301 tempest-AttachInterfacesV270Test-352099301-project-member] [instance: 6876f61c-2315-4e49-9944-4a5caddede5b] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 6876f61c-2315-4e49-9944-4a5caddede5b could not be found. [ 638.478147] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-4334f93e-6967-4c78-b6ec-145141dc41da tempest-AttachInterfacesV270Test-352099301 tempest-AttachInterfacesV270Test-352099301-project-member] [instance: 6876f61c-2315-4e49-9944-4a5caddede5b] Instance destroyed {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 638.478339] env[63515]: INFO nova.compute.manager [None req-4334f93e-6967-4c78-b6ec-145141dc41da tempest-AttachInterfacesV270Test-352099301 tempest-AttachInterfacesV270Test-352099301-project-member] [instance: 6876f61c-2315-4e49-9944-4a5caddede5b] Took 0.03 seconds to destroy the instance on the hypervisor. [ 638.478592] env[63515]: DEBUG oslo.service.loopingcall [None req-4334f93e-6967-4c78-b6ec-145141dc41da tempest-AttachInterfacesV270Test-352099301 tempest-AttachInterfacesV270Test-352099301-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 638.478822] env[63515]: DEBUG nova.compute.manager [-] [instance: 6876f61c-2315-4e49-9944-4a5caddede5b] Deallocating network for instance {{(pid=63515) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 638.478943] env[63515]: DEBUG nova.network.neutron [-] [instance: 6876f61c-2315-4e49-9944-4a5caddede5b] deallocate_for_instance() {{(pid=63515) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 638.497481] env[63515]: DEBUG nova.network.neutron [-] [instance: 6876f61c-2315-4e49-9944-4a5caddede5b] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 638.544097] env[63515]: INFO nova.compute.manager [None req-886adfc0-007d-4f83-bc30-b755d00d04c7 tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] [instance: 8b65c223-03a8-41e0-917e-45dd76b6e57f] Took 1.08 seconds to deallocate network for instance. [ 638.866224] env[63515]: DEBUG oslo_concurrency.lockutils [None req-30030db8-77b3-4af4-8ae6-30653bd81e23 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.507s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 638.866689] env[63515]: DEBUG nova.compute.manager [None req-30030db8-77b3-4af4-8ae6-30653bd81e23 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: 02ffb723-d6df-42cb-93c5-d582705d1e03] Start building networks asynchronously for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 638.870948] env[63515]: DEBUG oslo_concurrency.lockutils [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 15.702s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 638.999692] env[63515]: DEBUG nova.network.neutron [-] [instance: 6876f61c-2315-4e49-9944-4a5caddede5b] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 639.015483] env[63515]: DEBUG oslo_concurrency.lockutils [None req-99ae4d48-1fae-4ace-913c-d249ca72bbeb tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Acquiring lock "d2a4f66a-f1a2-4f0d-826d-0b7c4f04bd23" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 639.015483] env[63515]: DEBUG oslo_concurrency.lockutils [None req-99ae4d48-1fae-4ace-913c-d249ca72bbeb tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Lock "d2a4f66a-f1a2-4f0d-826d-0b7c4f04bd23" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 639.337876] env[63515]: DEBUG oslo_concurrency.lockutils [None req-5bd41396-218d-4ce7-9765-053dcc695301 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Acquiring lock "6688cdaa-29ba-413a-8131-4f834cdb70e4" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 639.339061] env[63515]: DEBUG oslo_concurrency.lockutils [None req-5bd41396-218d-4ce7-9765-053dcc695301 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Lock "6688cdaa-29ba-413a-8131-4f834cdb70e4" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 639.376229] env[63515]: DEBUG nova.compute.utils [None req-30030db8-77b3-4af4-8ae6-30653bd81e23 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Using /dev/sd instead of None {{(pid=63515) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 639.378369] env[63515]: DEBUG nova.compute.manager [None req-30030db8-77b3-4af4-8ae6-30653bd81e23 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: 02ffb723-d6df-42cb-93c5-d582705d1e03] Allocating IP information in the background. {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 639.378369] env[63515]: DEBUG nova.network.neutron [None req-30030db8-77b3-4af4-8ae6-30653bd81e23 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: 02ffb723-d6df-42cb-93c5-d582705d1e03] allocate_for_instance() {{(pid=63515) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 639.425851] env[63515]: DEBUG nova.policy [None req-30030db8-77b3-4af4-8ae6-30653bd81e23 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '002c1a9364184b82b649fc18d017d97f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2259667ddc0c42639ea1af3b6e6fa0e8', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63515) authorize /opt/stack/nova/nova/policy.py:201}} [ 639.503677] env[63515]: INFO nova.compute.manager [-] [instance: 6876f61c-2315-4e49-9944-4a5caddede5b] Took 1.02 seconds to deallocate network for instance. [ 639.509795] env[63515]: DEBUG nova.compute.claims [None req-4334f93e-6967-4c78-b6ec-145141dc41da tempest-AttachInterfacesV270Test-352099301 tempest-AttachInterfacesV270Test-352099301-project-member] [instance: 6876f61c-2315-4e49-9944-4a5caddede5b] Aborting claim: {{(pid=63515) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 639.510011] env[63515]: DEBUG oslo_concurrency.lockutils [None req-4334f93e-6967-4c78-b6ec-145141dc41da tempest-AttachInterfacesV270Test-352099301 tempest-AttachInterfacesV270Test-352099301-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 639.574305] env[63515]: INFO nova.scheduler.client.report [None req-886adfc0-007d-4f83-bc30-b755d00d04c7 tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] Deleted allocations for instance 8b65c223-03a8-41e0-917e-45dd76b6e57f [ 639.788144] env[63515]: DEBUG nova.network.neutron [None req-30030db8-77b3-4af4-8ae6-30653bd81e23 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: 02ffb723-d6df-42cb-93c5-d582705d1e03] Successfully created port: a2188979-d321-41f1-8cbe-faa875445836 {{(pid=63515) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 639.835911] env[63515]: DEBUG oslo_concurrency.lockutils [None req-dfa187f2-adf0-4e24-9cf4-7df6d0016e12 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Acquiring lock "38d6d030-06b0-4185-904d-44a038b3a752" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 639.836203] env[63515]: DEBUG oslo_concurrency.lockutils [None req-dfa187f2-adf0-4e24-9cf4-7df6d0016e12 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Lock "38d6d030-06b0-4185-904d-44a038b3a752" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 639.883873] env[63515]: DEBUG nova.compute.manager [None req-30030db8-77b3-4af4-8ae6-30653bd81e23 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: 02ffb723-d6df-42cb-93c5-d582705d1e03] Start building block device mappings for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 639.909541] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Instance 5b201ab6-c7ec-4a5c-a310-982649f34be4 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63515) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 639.909693] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Instance 2728c9f9-0d03-4313-ba2a-3dc22aff8538 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63515) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 639.909813] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Instance 6876f61c-2315-4e49-9944-4a5caddede5b actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63515) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 639.909929] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Instance 02ffb723-d6df-42cb-93c5-d582705d1e03 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63515) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 640.087802] env[63515]: DEBUG oslo_concurrency.lockutils [None req-886adfc0-007d-4f83-bc30-b755d00d04c7 tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] Lock "8b65c223-03a8-41e0-917e-45dd76b6e57f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 84.206s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 640.414186] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Instance bbce1137-d691-4633-87ff-f9f4ea257ed7 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63515) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 640.592507] env[63515]: DEBUG nova.compute.manager [None req-9ecc0501-dc42-4f19-a331-8b337de58eaf tempest-AttachInterfacesUnderV243Test-1943594058 tempest-AttachInterfacesUnderV243Test-1943594058-project-member] [instance: b911a5b5-9617-4fb3-9b5e-fb8c492e4931] Starting instance... {{(pid=63515) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 640.814177] env[63515]: DEBUG nova.compute.manager [req-84d844d2-cf8e-4323-90b6-6f790f8bb43e req-dd1548d7-4ca5-4194-823e-452198db37d9 service nova] [instance: 02ffb723-d6df-42cb-93c5-d582705d1e03] Received event network-changed-a2188979-d321-41f1-8cbe-faa875445836 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 640.814650] env[63515]: DEBUG nova.compute.manager [req-84d844d2-cf8e-4323-90b6-6f790f8bb43e req-dd1548d7-4ca5-4194-823e-452198db37d9 service nova] [instance: 02ffb723-d6df-42cb-93c5-d582705d1e03] Refreshing instance network info cache due to event network-changed-a2188979-d321-41f1-8cbe-faa875445836. {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 640.817059] env[63515]: DEBUG oslo_concurrency.lockutils [req-84d844d2-cf8e-4323-90b6-6f790f8bb43e req-dd1548d7-4ca5-4194-823e-452198db37d9 service nova] Acquiring lock "refresh_cache-02ffb723-d6df-42cb-93c5-d582705d1e03" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 640.817295] env[63515]: DEBUG oslo_concurrency.lockutils [req-84d844d2-cf8e-4323-90b6-6f790f8bb43e req-dd1548d7-4ca5-4194-823e-452198db37d9 service nova] Acquired lock "refresh_cache-02ffb723-d6df-42cb-93c5-d582705d1e03" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 640.817515] env[63515]: DEBUG nova.network.neutron [req-84d844d2-cf8e-4323-90b6-6f790f8bb43e req-dd1548d7-4ca5-4194-823e-452198db37d9 service nova] [instance: 02ffb723-d6df-42cb-93c5-d582705d1e03] Refreshing network info cache for port a2188979-d321-41f1-8cbe-faa875445836 {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 640.842982] env[63515]: ERROR nova.compute.manager [None req-30030db8-77b3-4af4-8ae6-30653bd81e23 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port a2188979-d321-41f1-8cbe-faa875445836, please check neutron logs for more information. [ 640.842982] env[63515]: ERROR nova.compute.manager Traceback (most recent call last): [ 640.842982] env[63515]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 640.842982] env[63515]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 640.842982] env[63515]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 640.842982] env[63515]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 640.842982] env[63515]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 640.842982] env[63515]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 640.842982] env[63515]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 640.842982] env[63515]: ERROR nova.compute.manager self.force_reraise() [ 640.842982] env[63515]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 640.842982] env[63515]: ERROR nova.compute.manager raise self.value [ 640.842982] env[63515]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 640.842982] env[63515]: ERROR nova.compute.manager updated_port = self._update_port( [ 640.842982] env[63515]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 640.842982] env[63515]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 640.843452] env[63515]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 640.843452] env[63515]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 640.843452] env[63515]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port a2188979-d321-41f1-8cbe-faa875445836, please check neutron logs for more information. [ 640.843452] env[63515]: ERROR nova.compute.manager [ 640.843452] env[63515]: Traceback (most recent call last): [ 640.843452] env[63515]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 640.843452] env[63515]: listener.cb(fileno) [ 640.843452] env[63515]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 640.843452] env[63515]: result = function(*args, **kwargs) [ 640.843452] env[63515]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 640.843452] env[63515]: return func(*args, **kwargs) [ 640.843452] env[63515]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 640.843452] env[63515]: raise e [ 640.843452] env[63515]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 640.843452] env[63515]: nwinfo = self.network_api.allocate_for_instance( [ 640.843452] env[63515]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 640.843452] env[63515]: created_port_ids = self._update_ports_for_instance( [ 640.843452] env[63515]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 640.843452] env[63515]: with excutils.save_and_reraise_exception(): [ 640.843452] env[63515]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 640.843452] env[63515]: self.force_reraise() [ 640.843452] env[63515]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 640.843452] env[63515]: raise self.value [ 640.843452] env[63515]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 640.843452] env[63515]: updated_port = self._update_port( [ 640.843452] env[63515]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 640.843452] env[63515]: _ensure_no_port_binding_failure(port) [ 640.843452] env[63515]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 640.843452] env[63515]: raise exception.PortBindingFailed(port_id=port['id']) [ 640.844244] env[63515]: nova.exception.PortBindingFailed: Binding failed for port a2188979-d321-41f1-8cbe-faa875445836, please check neutron logs for more information. [ 640.844244] env[63515]: Removing descriptor: 17 [ 640.899622] env[63515]: DEBUG nova.compute.manager [None req-30030db8-77b3-4af4-8ae6-30653bd81e23 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: 02ffb723-d6df-42cb-93c5-d582705d1e03] Start spawning the instance on the hypervisor. {{(pid=63515) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 640.921406] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Instance 7f49d79e-3bda-4949-a976-7e3e6513b2c7 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63515) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 640.924773] env[63515]: DEBUG nova.virt.hardware [None req-30030db8-77b3-4af4-8ae6-30653bd81e23 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T02:52:14Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T02:51:56Z,direct_url=,disk_format='vmdk',id=8a120570-cb06-4099-b262-554ca0ad15c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b569255cc43e42e7ae3f2b4ad37c6ef4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T02:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 640.925113] env[63515]: DEBUG nova.virt.hardware [None req-30030db8-77b3-4af4-8ae6-30653bd81e23 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Flavor limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 640.925253] env[63515]: DEBUG nova.virt.hardware [None req-30030db8-77b3-4af4-8ae6-30653bd81e23 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Image limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 640.925436] env[63515]: DEBUG nova.virt.hardware [None req-30030db8-77b3-4af4-8ae6-30653bd81e23 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Flavor pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 640.925579] env[63515]: DEBUG nova.virt.hardware [None req-30030db8-77b3-4af4-8ae6-30653bd81e23 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Image pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 640.925723] env[63515]: DEBUG nova.virt.hardware [None req-30030db8-77b3-4af4-8ae6-30653bd81e23 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 640.927030] env[63515]: DEBUG nova.virt.hardware [None req-30030db8-77b3-4af4-8ae6-30653bd81e23 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 640.927030] env[63515]: DEBUG nova.virt.hardware [None req-30030db8-77b3-4af4-8ae6-30653bd81e23 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 640.927030] env[63515]: DEBUG nova.virt.hardware [None req-30030db8-77b3-4af4-8ae6-30653bd81e23 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Got 1 possible topologies {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 640.927030] env[63515]: DEBUG nova.virt.hardware [None req-30030db8-77b3-4af4-8ae6-30653bd81e23 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 640.927030] env[63515]: DEBUG nova.virt.hardware [None req-30030db8-77b3-4af4-8ae6-30653bd81e23 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 640.927471] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7458d3d3-990f-43d4-99f7-037326d3da7d {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.936248] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3da6ecaa-20ec-4d42-8e37-3a76639686f9 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.950356] env[63515]: ERROR nova.compute.manager [None req-30030db8-77b3-4af4-8ae6-30653bd81e23 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: 02ffb723-d6df-42cb-93c5-d582705d1e03] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port a2188979-d321-41f1-8cbe-faa875445836, please check neutron logs for more information. [ 640.950356] env[63515]: ERROR nova.compute.manager [instance: 02ffb723-d6df-42cb-93c5-d582705d1e03] Traceback (most recent call last): [ 640.950356] env[63515]: ERROR nova.compute.manager [instance: 02ffb723-d6df-42cb-93c5-d582705d1e03] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 640.950356] env[63515]: ERROR nova.compute.manager [instance: 02ffb723-d6df-42cb-93c5-d582705d1e03] yield resources [ 640.950356] env[63515]: ERROR nova.compute.manager [instance: 02ffb723-d6df-42cb-93c5-d582705d1e03] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 640.950356] env[63515]: ERROR nova.compute.manager [instance: 02ffb723-d6df-42cb-93c5-d582705d1e03] self.driver.spawn(context, instance, image_meta, [ 640.950356] env[63515]: ERROR nova.compute.manager [instance: 02ffb723-d6df-42cb-93c5-d582705d1e03] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 640.950356] env[63515]: ERROR nova.compute.manager [instance: 02ffb723-d6df-42cb-93c5-d582705d1e03] self._vmops.spawn(context, instance, image_meta, injected_files, [ 640.950356] env[63515]: ERROR nova.compute.manager [instance: 02ffb723-d6df-42cb-93c5-d582705d1e03] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 640.950356] env[63515]: ERROR nova.compute.manager [instance: 02ffb723-d6df-42cb-93c5-d582705d1e03] vm_ref = self.build_virtual_machine(instance, [ 640.950356] env[63515]: ERROR nova.compute.manager [instance: 02ffb723-d6df-42cb-93c5-d582705d1e03] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 640.950733] env[63515]: ERROR nova.compute.manager [instance: 02ffb723-d6df-42cb-93c5-d582705d1e03] vif_infos = vmwarevif.get_vif_info(self._session, [ 640.950733] env[63515]: ERROR nova.compute.manager [instance: 02ffb723-d6df-42cb-93c5-d582705d1e03] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 640.950733] env[63515]: ERROR nova.compute.manager [instance: 02ffb723-d6df-42cb-93c5-d582705d1e03] for vif in network_info: [ 640.950733] env[63515]: ERROR nova.compute.manager [instance: 02ffb723-d6df-42cb-93c5-d582705d1e03] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 640.950733] env[63515]: ERROR nova.compute.manager [instance: 02ffb723-d6df-42cb-93c5-d582705d1e03] return self._sync_wrapper(fn, *args, **kwargs) [ 640.950733] env[63515]: ERROR nova.compute.manager [instance: 02ffb723-d6df-42cb-93c5-d582705d1e03] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 640.950733] env[63515]: ERROR nova.compute.manager [instance: 02ffb723-d6df-42cb-93c5-d582705d1e03] self.wait() [ 640.950733] env[63515]: ERROR nova.compute.manager [instance: 02ffb723-d6df-42cb-93c5-d582705d1e03] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 640.950733] env[63515]: ERROR nova.compute.manager [instance: 02ffb723-d6df-42cb-93c5-d582705d1e03] self[:] = self._gt.wait() [ 640.950733] env[63515]: ERROR nova.compute.manager [instance: 02ffb723-d6df-42cb-93c5-d582705d1e03] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 640.950733] env[63515]: ERROR nova.compute.manager [instance: 02ffb723-d6df-42cb-93c5-d582705d1e03] return self._exit_event.wait() [ 640.950733] env[63515]: ERROR nova.compute.manager [instance: 02ffb723-d6df-42cb-93c5-d582705d1e03] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 640.950733] env[63515]: ERROR nova.compute.manager [instance: 02ffb723-d6df-42cb-93c5-d582705d1e03] current.throw(*self._exc) [ 640.951051] env[63515]: ERROR nova.compute.manager [instance: 02ffb723-d6df-42cb-93c5-d582705d1e03] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 640.951051] env[63515]: ERROR nova.compute.manager [instance: 02ffb723-d6df-42cb-93c5-d582705d1e03] result = function(*args, **kwargs) [ 640.951051] env[63515]: ERROR nova.compute.manager [instance: 02ffb723-d6df-42cb-93c5-d582705d1e03] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 640.951051] env[63515]: ERROR nova.compute.manager [instance: 02ffb723-d6df-42cb-93c5-d582705d1e03] return func(*args, **kwargs) [ 640.951051] env[63515]: ERROR nova.compute.manager [instance: 02ffb723-d6df-42cb-93c5-d582705d1e03] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 640.951051] env[63515]: ERROR nova.compute.manager [instance: 02ffb723-d6df-42cb-93c5-d582705d1e03] raise e [ 640.951051] env[63515]: ERROR nova.compute.manager [instance: 02ffb723-d6df-42cb-93c5-d582705d1e03] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 640.951051] env[63515]: ERROR nova.compute.manager [instance: 02ffb723-d6df-42cb-93c5-d582705d1e03] nwinfo = self.network_api.allocate_for_instance( [ 640.951051] env[63515]: ERROR nova.compute.manager [instance: 02ffb723-d6df-42cb-93c5-d582705d1e03] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 640.951051] env[63515]: ERROR nova.compute.manager [instance: 02ffb723-d6df-42cb-93c5-d582705d1e03] created_port_ids = self._update_ports_for_instance( [ 640.951051] env[63515]: ERROR nova.compute.manager [instance: 02ffb723-d6df-42cb-93c5-d582705d1e03] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 640.951051] env[63515]: ERROR nova.compute.manager [instance: 02ffb723-d6df-42cb-93c5-d582705d1e03] with excutils.save_and_reraise_exception(): [ 640.951051] env[63515]: ERROR nova.compute.manager [instance: 02ffb723-d6df-42cb-93c5-d582705d1e03] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 640.951360] env[63515]: ERROR nova.compute.manager [instance: 02ffb723-d6df-42cb-93c5-d582705d1e03] self.force_reraise() [ 640.951360] env[63515]: ERROR nova.compute.manager [instance: 02ffb723-d6df-42cb-93c5-d582705d1e03] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 640.951360] env[63515]: ERROR nova.compute.manager [instance: 02ffb723-d6df-42cb-93c5-d582705d1e03] raise self.value [ 640.951360] env[63515]: ERROR nova.compute.manager [instance: 02ffb723-d6df-42cb-93c5-d582705d1e03] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 640.951360] env[63515]: ERROR nova.compute.manager [instance: 02ffb723-d6df-42cb-93c5-d582705d1e03] updated_port = self._update_port( [ 640.951360] env[63515]: ERROR nova.compute.manager [instance: 02ffb723-d6df-42cb-93c5-d582705d1e03] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 640.951360] env[63515]: ERROR nova.compute.manager [instance: 02ffb723-d6df-42cb-93c5-d582705d1e03] _ensure_no_port_binding_failure(port) [ 640.951360] env[63515]: ERROR nova.compute.manager [instance: 02ffb723-d6df-42cb-93c5-d582705d1e03] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 640.951360] env[63515]: ERROR nova.compute.manager [instance: 02ffb723-d6df-42cb-93c5-d582705d1e03] raise exception.PortBindingFailed(port_id=port['id']) [ 640.951360] env[63515]: ERROR nova.compute.manager [instance: 02ffb723-d6df-42cb-93c5-d582705d1e03] nova.exception.PortBindingFailed: Binding failed for port a2188979-d321-41f1-8cbe-faa875445836, please check neutron logs for more information. [ 640.951360] env[63515]: ERROR nova.compute.manager [instance: 02ffb723-d6df-42cb-93c5-d582705d1e03] [ 640.951360] env[63515]: INFO nova.compute.manager [None req-30030db8-77b3-4af4-8ae6-30653bd81e23 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: 02ffb723-d6df-42cb-93c5-d582705d1e03] Terminating instance [ 640.952672] env[63515]: DEBUG oslo_concurrency.lockutils [None req-30030db8-77b3-4af4-8ae6-30653bd81e23 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Acquiring lock "refresh_cache-02ffb723-d6df-42cb-93c5-d582705d1e03" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 641.116128] env[63515]: DEBUG oslo_concurrency.lockutils [None req-9ecc0501-dc42-4f19-a331-8b337de58eaf tempest-AttachInterfacesUnderV243Test-1943594058 tempest-AttachInterfacesUnderV243Test-1943594058-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 641.341273] env[63515]: DEBUG nova.network.neutron [req-84d844d2-cf8e-4323-90b6-6f790f8bb43e req-dd1548d7-4ca5-4194-823e-452198db37d9 service nova] [instance: 02ffb723-d6df-42cb-93c5-d582705d1e03] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 641.352948] env[63515]: DEBUG oslo_concurrency.lockutils [None req-750d4038-2a95-4b14-b9d7-c9c6803c64b6 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] Acquiring lock "8eb17506-ff93-4d25-b9af-ec5886569e65" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 641.353159] env[63515]: DEBUG oslo_concurrency.lockutils [None req-750d4038-2a95-4b14-b9d7-c9c6803c64b6 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] Lock "8eb17506-ff93-4d25-b9af-ec5886569e65" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 641.431074] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Instance 07499601-62bc-4c31-b295-23f34a6e2e91 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63515) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 641.460923] env[63515]: DEBUG nova.network.neutron [req-84d844d2-cf8e-4323-90b6-6f790f8bb43e req-dd1548d7-4ca5-4194-823e-452198db37d9 service nova] [instance: 02ffb723-d6df-42cb-93c5-d582705d1e03] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 641.941822] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Instance 005f45b3-630d-400f-9605-100f6aad88e6 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63515) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 641.963716] env[63515]: DEBUG oslo_concurrency.lockutils [req-84d844d2-cf8e-4323-90b6-6f790f8bb43e req-dd1548d7-4ca5-4194-823e-452198db37d9 service nova] Releasing lock "refresh_cache-02ffb723-d6df-42cb-93c5-d582705d1e03" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 641.964260] env[63515]: DEBUG oslo_concurrency.lockutils [None req-30030db8-77b3-4af4-8ae6-30653bd81e23 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Acquired lock "refresh_cache-02ffb723-d6df-42cb-93c5-d582705d1e03" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 641.964447] env[63515]: DEBUG nova.network.neutron [None req-30030db8-77b3-4af4-8ae6-30653bd81e23 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: 02ffb723-d6df-42cb-93c5-d582705d1e03] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 642.445988] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Instance 4ad5440f-cd18-44c5-8836-0aa39824cf03 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63515) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 642.482527] env[63515]: DEBUG nova.network.neutron [None req-30030db8-77b3-4af4-8ae6-30653bd81e23 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: 02ffb723-d6df-42cb-93c5-d582705d1e03] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 642.584549] env[63515]: DEBUG nova.network.neutron [None req-30030db8-77b3-4af4-8ae6-30653bd81e23 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: 02ffb723-d6df-42cb-93c5-d582705d1e03] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 642.891268] env[63515]: DEBUG nova.compute.manager [req-3759aaff-0784-4df6-901f-246a4c67802b req-97cda126-8428-41ff-8751-f2aae6a80aa4 service nova] [instance: 02ffb723-d6df-42cb-93c5-d582705d1e03] Received event network-vif-deleted-a2188979-d321-41f1-8cbe-faa875445836 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 642.949244] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Instance b911a5b5-9617-4fb3-9b5e-fb8c492e4931 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63515) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 643.087132] env[63515]: DEBUG oslo_concurrency.lockutils [None req-30030db8-77b3-4af4-8ae6-30653bd81e23 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Releasing lock "refresh_cache-02ffb723-d6df-42cb-93c5-d582705d1e03" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 643.087572] env[63515]: DEBUG nova.compute.manager [None req-30030db8-77b3-4af4-8ae6-30653bd81e23 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: 02ffb723-d6df-42cb-93c5-d582705d1e03] Start destroying the instance on the hypervisor. {{(pid=63515) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 643.087765] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-30030db8-77b3-4af4-8ae6-30653bd81e23 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: 02ffb723-d6df-42cb-93c5-d582705d1e03] Destroying instance {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 643.088104] env[63515]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6151cb82-35b2-4aa1-bf70-c3ba24259822 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.097384] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6f2391b-7812-404a-92fe-8e06dbcd2f98 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.118025] env[63515]: WARNING nova.virt.vmwareapi.vmops [None req-30030db8-77b3-4af4-8ae6-30653bd81e23 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: 02ffb723-d6df-42cb-93c5-d582705d1e03] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 02ffb723-d6df-42cb-93c5-d582705d1e03 could not be found. [ 643.118251] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-30030db8-77b3-4af4-8ae6-30653bd81e23 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: 02ffb723-d6df-42cb-93c5-d582705d1e03] Instance destroyed {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 643.118433] env[63515]: INFO nova.compute.manager [None req-30030db8-77b3-4af4-8ae6-30653bd81e23 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: 02ffb723-d6df-42cb-93c5-d582705d1e03] Took 0.03 seconds to destroy the instance on the hypervisor. [ 643.118674] env[63515]: DEBUG oslo.service.loopingcall [None req-30030db8-77b3-4af4-8ae6-30653bd81e23 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 643.118889] env[63515]: DEBUG nova.compute.manager [-] [instance: 02ffb723-d6df-42cb-93c5-d582705d1e03] Deallocating network for instance {{(pid=63515) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 643.118987] env[63515]: DEBUG nova.network.neutron [-] [instance: 02ffb723-d6df-42cb-93c5-d582705d1e03] deallocate_for_instance() {{(pid=63515) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 643.132809] env[63515]: DEBUG nova.network.neutron [-] [instance: 02ffb723-d6df-42cb-93c5-d582705d1e03] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 643.399241] env[63515]: DEBUG oslo_concurrency.lockutils [None req-b8b2a7f6-808d-40d4-8de9-c8863e5b55c5 tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] Acquiring lock "30b8b872-40ca-4297-b98e-a64c83a75483" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 643.399476] env[63515]: DEBUG oslo_concurrency.lockutils [None req-b8b2a7f6-808d-40d4-8de9-c8863e5b55c5 tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] Lock "30b8b872-40ca-4297-b98e-a64c83a75483" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 643.451882] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Instance 3ad04a9f-62cf-4b1c-aab2-7dc4a219edc6 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63515) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 643.634808] env[63515]: DEBUG nova.network.neutron [-] [instance: 02ffb723-d6df-42cb-93c5-d582705d1e03] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 643.955719] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Instance 5b42f744-fdd6-45b1-8563-896869648c23 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63515) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 644.137646] env[63515]: INFO nova.compute.manager [-] [instance: 02ffb723-d6df-42cb-93c5-d582705d1e03] Took 1.02 seconds to deallocate network for instance. [ 644.140102] env[63515]: DEBUG nova.compute.claims [None req-30030db8-77b3-4af4-8ae6-30653bd81e23 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: 02ffb723-d6df-42cb-93c5-d582705d1e03] Aborting claim: {{(pid=63515) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 644.140285] env[63515]: DEBUG oslo_concurrency.lockutils [None req-30030db8-77b3-4af4-8ae6-30653bd81e23 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 644.459075] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Instance f1d01b75-ac9d-458d-8cc2-ae64cffca4e8 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63515) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 644.963770] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Instance 90f4930b-aaa0-4c4b-9ab8-92aed45e200b has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63515) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 645.466557] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Instance 396e49dd-48c0-496b-a1ec-190c33a22c5e has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63515) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 645.969723] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Instance 4087b2c4-6ed1-4b68-8b78-a36e34d935b1 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63515) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 646.474515] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Instance 87c468d9-9594-4804-b461-527f01f6118f has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63515) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 646.979031] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Instance 48668736-df27-4f2a-94d9-132f5b49701b has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63515) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 647.481721] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Instance 452f3ca2-6141-43b2-a77a-c9ab5754192d has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63515) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 647.986098] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Instance 37e6c27e-317b-45d2-bd55-2fd78ccf009f has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63515) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 648.490628] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Instance b4477e66-ae12-4929-90ed-b7b652e0f207 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63515) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 648.993231] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Instance b85e9a70-7f5b-4d32-b616-f2a97e3186c8 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63515) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 649.497058] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Instance d7e8dcdb-41cb-46fb-8b61-d251e7c2d372 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63515) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 649.999014] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Instance e91aa479-1540-4950-851b-b2409e5f89f1 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63515) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 650.502175] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Instance ed550b10-d58f-45b8-b766-198f431c3788 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63515) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 651.005726] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Instance de210780-5c0f-4fba-883c-13707566a2e1 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63515) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 651.509064] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Instance 001a48e1-b237-41f8-a6f4-2fe98d173512 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63515) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 652.012092] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Instance d2a4f66a-f1a2-4f0d-826d-0b7c4f04bd23 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63515) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 652.515814] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Instance 6688cdaa-29ba-413a-8131-4f834cdb70e4 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63515) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 653.019773] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Instance 38d6d030-06b0-4185-904d-44a038b3a752 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=63515) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 653.019941] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Total usable vcpus: 48, total allocated vcpus: 4 {{(pid=63515) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 653.020142] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1280MB phys_disk=200GB used_disk=4GB total_vcpus=48 used_vcpus=4 pci_stats=[] {{(pid=63515) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 653.364216] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5da5d53a-4d68-44ab-9de6-41973acd4336 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.372136] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef08b30a-f509-425e-a5ef-85ae01d65839 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.404359] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0814b829-5d91-452e-af63-89fc1094bc17 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.412948] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2b3e37a-bdc2-4331-a557-57b5bfb6cbc1 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.426011] env[63515]: DEBUG nova.compute.provider_tree [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 653.929624] env[63515]: DEBUG nova.scheduler.client.report [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 654.434818] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=63515) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 654.435164] env[63515]: DEBUG oslo_concurrency.lockutils [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 15.566s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 654.435543] env[63515]: DEBUG oslo_concurrency.lockutils [None req-30987b61-ce89-4676-89be-fee7786462c9 tempest-ImagesOneServerNegativeTestJSON-52225474 tempest-ImagesOneServerNegativeTestJSON-52225474-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 30.783s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 654.436928] env[63515]: INFO nova.compute.claims [None req-30987b61-ce89-4676-89be-fee7786462c9 tempest-ImagesOneServerNegativeTestJSON-52225474 tempest-ImagesOneServerNegativeTestJSON-52225474-project-member] [instance: bbce1137-d691-4633-87ff-f9f4ea257ed7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 655.868404] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2815c1b6-19c2-4e4b-892a-5f06529c5c7f {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.875946] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1aca1ab6-af93-4f5e-a2dc-cb62773ec142 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.905667] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e56f3208-b8a4-456e-b7ff-3aad1d2554e2 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.912453] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28bac0c7-6adb-4e16-9eb7-945680e73c58 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.925152] env[63515]: DEBUG nova.compute.provider_tree [None req-30987b61-ce89-4676-89be-fee7786462c9 tempest-ImagesOneServerNegativeTestJSON-52225474 tempest-ImagesOneServerNegativeTestJSON-52225474-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 656.428319] env[63515]: DEBUG nova.scheduler.client.report [None req-30987b61-ce89-4676-89be-fee7786462c9 tempest-ImagesOneServerNegativeTestJSON-52225474 tempest-ImagesOneServerNegativeTestJSON-52225474-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 656.933367] env[63515]: DEBUG oslo_concurrency.lockutils [None req-30987b61-ce89-4676-89be-fee7786462c9 tempest-ImagesOneServerNegativeTestJSON-52225474 tempest-ImagesOneServerNegativeTestJSON-52225474-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.498s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 656.933901] env[63515]: DEBUG nova.compute.manager [None req-30987b61-ce89-4676-89be-fee7786462c9 tempest-ImagesOneServerNegativeTestJSON-52225474 tempest-ImagesOneServerNegativeTestJSON-52225474-project-member] [instance: bbce1137-d691-4633-87ff-f9f4ea257ed7] Start building networks asynchronously for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 656.936547] env[63515]: DEBUG oslo_concurrency.lockutils [None req-480b7d6d-2f7f-4da3-9eb1-818f0711a940 tempest-ServerRescueTestJSON-1961208277 tempest-ServerRescueTestJSON-1961208277-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 31.541s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 656.938237] env[63515]: INFO nova.compute.claims [None req-480b7d6d-2f7f-4da3-9eb1-818f0711a940 tempest-ServerRescueTestJSON-1961208277 tempest-ServerRescueTestJSON-1961208277-project-member] [instance: 7f49d79e-3bda-4949-a976-7e3e6513b2c7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 657.442995] env[63515]: DEBUG nova.compute.utils [None req-30987b61-ce89-4676-89be-fee7786462c9 tempest-ImagesOneServerNegativeTestJSON-52225474 tempest-ImagesOneServerNegativeTestJSON-52225474-project-member] Using /dev/sd instead of None {{(pid=63515) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 657.446368] env[63515]: DEBUG nova.compute.manager [None req-30987b61-ce89-4676-89be-fee7786462c9 tempest-ImagesOneServerNegativeTestJSON-52225474 tempest-ImagesOneServerNegativeTestJSON-52225474-project-member] [instance: bbce1137-d691-4633-87ff-f9f4ea257ed7] Allocating IP information in the background. {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 657.446580] env[63515]: DEBUG nova.network.neutron [None req-30987b61-ce89-4676-89be-fee7786462c9 tempest-ImagesOneServerNegativeTestJSON-52225474 tempest-ImagesOneServerNegativeTestJSON-52225474-project-member] [instance: bbce1137-d691-4633-87ff-f9f4ea257ed7] allocate_for_instance() {{(pid=63515) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 657.499371] env[63515]: DEBUG nova.policy [None req-30987b61-ce89-4676-89be-fee7786462c9 tempest-ImagesOneServerNegativeTestJSON-52225474 tempest-ImagesOneServerNegativeTestJSON-52225474-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6d4dd869fdd44e4f9a457781bb5c8ce0', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c43f7616b8df4b2ca7bef1597bdcd30b', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63515) authorize /opt/stack/nova/nova/policy.py:201}} [ 657.780244] env[63515]: DEBUG nova.network.neutron [None req-30987b61-ce89-4676-89be-fee7786462c9 tempest-ImagesOneServerNegativeTestJSON-52225474 tempest-ImagesOneServerNegativeTestJSON-52225474-project-member] [instance: bbce1137-d691-4633-87ff-f9f4ea257ed7] Successfully created port: fb0a1841-27e9-4610-abba-889cf0217607 {{(pid=63515) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 657.946849] env[63515]: DEBUG nova.compute.manager [None req-30987b61-ce89-4676-89be-fee7786462c9 tempest-ImagesOneServerNegativeTestJSON-52225474 tempest-ImagesOneServerNegativeTestJSON-52225474-project-member] [instance: bbce1137-d691-4633-87ff-f9f4ea257ed7] Start building block device mappings for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 658.354047] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed26096c-6af8-42c4-8314-55278e9a3f85 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.361256] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a11bbd0-104c-462d-835d-b0bbe58053e0 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.393325] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b5ed127-4d21-462e-8e46-f0de87fa1a64 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.400902] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-502df863-ce63-495e-85d0-7c52e6b7599b {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.415295] env[63515]: DEBUG nova.compute.provider_tree [None req-480b7d6d-2f7f-4da3-9eb1-818f0711a940 tempest-ServerRescueTestJSON-1961208277 tempest-ServerRescueTestJSON-1961208277-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 658.654463] env[63515]: DEBUG nova.compute.manager [req-bb3207d2-ca06-4e5d-bd64-427741c5bdc7 req-3ca56192-01aa-43ef-94c4-3b8538d908b5 service nova] [instance: bbce1137-d691-4633-87ff-f9f4ea257ed7] Received event network-changed-fb0a1841-27e9-4610-abba-889cf0217607 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 658.656080] env[63515]: DEBUG nova.compute.manager [req-bb3207d2-ca06-4e5d-bd64-427741c5bdc7 req-3ca56192-01aa-43ef-94c4-3b8538d908b5 service nova] [instance: bbce1137-d691-4633-87ff-f9f4ea257ed7] Refreshing instance network info cache due to event network-changed-fb0a1841-27e9-4610-abba-889cf0217607. {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 658.656780] env[63515]: DEBUG oslo_concurrency.lockutils [req-bb3207d2-ca06-4e5d-bd64-427741c5bdc7 req-3ca56192-01aa-43ef-94c4-3b8538d908b5 service nova] Acquiring lock "refresh_cache-bbce1137-d691-4633-87ff-f9f4ea257ed7" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 658.656780] env[63515]: DEBUG oslo_concurrency.lockutils [req-bb3207d2-ca06-4e5d-bd64-427741c5bdc7 req-3ca56192-01aa-43ef-94c4-3b8538d908b5 service nova] Acquired lock "refresh_cache-bbce1137-d691-4633-87ff-f9f4ea257ed7" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 658.656780] env[63515]: DEBUG nova.network.neutron [req-bb3207d2-ca06-4e5d-bd64-427741c5bdc7 req-3ca56192-01aa-43ef-94c4-3b8538d908b5 service nova] [instance: bbce1137-d691-4633-87ff-f9f4ea257ed7] Refreshing network info cache for port fb0a1841-27e9-4610-abba-889cf0217607 {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 658.821175] env[63515]: ERROR nova.compute.manager [None req-30987b61-ce89-4676-89be-fee7786462c9 tempest-ImagesOneServerNegativeTestJSON-52225474 tempest-ImagesOneServerNegativeTestJSON-52225474-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port fb0a1841-27e9-4610-abba-889cf0217607, please check neutron logs for more information. [ 658.821175] env[63515]: ERROR nova.compute.manager Traceback (most recent call last): [ 658.821175] env[63515]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 658.821175] env[63515]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 658.821175] env[63515]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 658.821175] env[63515]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 658.821175] env[63515]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 658.821175] env[63515]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 658.821175] env[63515]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 658.821175] env[63515]: ERROR nova.compute.manager self.force_reraise() [ 658.821175] env[63515]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 658.821175] env[63515]: ERROR nova.compute.manager raise self.value [ 658.821175] env[63515]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 658.821175] env[63515]: ERROR nova.compute.manager updated_port = self._update_port( [ 658.821175] env[63515]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 658.821175] env[63515]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 658.821651] env[63515]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 658.821651] env[63515]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 658.821651] env[63515]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port fb0a1841-27e9-4610-abba-889cf0217607, please check neutron logs for more information. [ 658.821651] env[63515]: ERROR nova.compute.manager [ 658.821651] env[63515]: Traceback (most recent call last): [ 658.821651] env[63515]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 658.821651] env[63515]: listener.cb(fileno) [ 658.821651] env[63515]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 658.821651] env[63515]: result = function(*args, **kwargs) [ 658.821651] env[63515]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 658.821651] env[63515]: return func(*args, **kwargs) [ 658.821651] env[63515]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 658.821651] env[63515]: raise e [ 658.821651] env[63515]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 658.821651] env[63515]: nwinfo = self.network_api.allocate_for_instance( [ 658.821651] env[63515]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 658.821651] env[63515]: created_port_ids = self._update_ports_for_instance( [ 658.821651] env[63515]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 658.821651] env[63515]: with excutils.save_and_reraise_exception(): [ 658.821651] env[63515]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 658.821651] env[63515]: self.force_reraise() [ 658.821651] env[63515]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 658.821651] env[63515]: raise self.value [ 658.821651] env[63515]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 658.821651] env[63515]: updated_port = self._update_port( [ 658.821651] env[63515]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 658.821651] env[63515]: _ensure_no_port_binding_failure(port) [ 658.821651] env[63515]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 658.821651] env[63515]: raise exception.PortBindingFailed(port_id=port['id']) [ 658.822394] env[63515]: nova.exception.PortBindingFailed: Binding failed for port fb0a1841-27e9-4610-abba-889cf0217607, please check neutron logs for more information. [ 658.822394] env[63515]: Removing descriptor: 17 [ 658.922668] env[63515]: DEBUG nova.scheduler.client.report [None req-480b7d6d-2f7f-4da3-9eb1-818f0711a940 tempest-ServerRescueTestJSON-1961208277 tempest-ServerRescueTestJSON-1961208277-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 658.958026] env[63515]: DEBUG nova.compute.manager [None req-30987b61-ce89-4676-89be-fee7786462c9 tempest-ImagesOneServerNegativeTestJSON-52225474 tempest-ImagesOneServerNegativeTestJSON-52225474-project-member] [instance: bbce1137-d691-4633-87ff-f9f4ea257ed7] Start spawning the instance on the hypervisor. {{(pid=63515) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 658.986945] env[63515]: DEBUG nova.virt.hardware [None req-30987b61-ce89-4676-89be-fee7786462c9 tempest-ImagesOneServerNegativeTestJSON-52225474 tempest-ImagesOneServerNegativeTestJSON-52225474-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T02:52:14Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T02:51:56Z,direct_url=,disk_format='vmdk',id=8a120570-cb06-4099-b262-554ca0ad15c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b569255cc43e42e7ae3f2b4ad37c6ef4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T02:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 658.987248] env[63515]: DEBUG nova.virt.hardware [None req-30987b61-ce89-4676-89be-fee7786462c9 tempest-ImagesOneServerNegativeTestJSON-52225474 tempest-ImagesOneServerNegativeTestJSON-52225474-project-member] Flavor limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 658.987415] env[63515]: DEBUG nova.virt.hardware [None req-30987b61-ce89-4676-89be-fee7786462c9 tempest-ImagesOneServerNegativeTestJSON-52225474 tempest-ImagesOneServerNegativeTestJSON-52225474-project-member] Image limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 658.987596] env[63515]: DEBUG nova.virt.hardware [None req-30987b61-ce89-4676-89be-fee7786462c9 tempest-ImagesOneServerNegativeTestJSON-52225474 tempest-ImagesOneServerNegativeTestJSON-52225474-project-member] Flavor pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 658.987741] env[63515]: DEBUG nova.virt.hardware [None req-30987b61-ce89-4676-89be-fee7786462c9 tempest-ImagesOneServerNegativeTestJSON-52225474 tempest-ImagesOneServerNegativeTestJSON-52225474-project-member] Image pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 658.987885] env[63515]: DEBUG nova.virt.hardware [None req-30987b61-ce89-4676-89be-fee7786462c9 tempest-ImagesOneServerNegativeTestJSON-52225474 tempest-ImagesOneServerNegativeTestJSON-52225474-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 658.988105] env[63515]: DEBUG nova.virt.hardware [None req-30987b61-ce89-4676-89be-fee7786462c9 tempest-ImagesOneServerNegativeTestJSON-52225474 tempest-ImagesOneServerNegativeTestJSON-52225474-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 658.988268] env[63515]: DEBUG nova.virt.hardware [None req-30987b61-ce89-4676-89be-fee7786462c9 tempest-ImagesOneServerNegativeTestJSON-52225474 tempest-ImagesOneServerNegativeTestJSON-52225474-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 658.988469] env[63515]: DEBUG nova.virt.hardware [None req-30987b61-ce89-4676-89be-fee7786462c9 tempest-ImagesOneServerNegativeTestJSON-52225474 tempest-ImagesOneServerNegativeTestJSON-52225474-project-member] Got 1 possible topologies {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 658.988598] env[63515]: DEBUG nova.virt.hardware [None req-30987b61-ce89-4676-89be-fee7786462c9 tempest-ImagesOneServerNegativeTestJSON-52225474 tempest-ImagesOneServerNegativeTestJSON-52225474-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 658.988769] env[63515]: DEBUG nova.virt.hardware [None req-30987b61-ce89-4676-89be-fee7786462c9 tempest-ImagesOneServerNegativeTestJSON-52225474 tempest-ImagesOneServerNegativeTestJSON-52225474-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 658.989695] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c8fc91a-c17a-4f1a-b458-d6bb8531fa91 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.998936] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d086c973-ac94-447d-99a0-c643a48257df {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.012230] env[63515]: ERROR nova.compute.manager [None req-30987b61-ce89-4676-89be-fee7786462c9 tempest-ImagesOneServerNegativeTestJSON-52225474 tempest-ImagesOneServerNegativeTestJSON-52225474-project-member] [instance: bbce1137-d691-4633-87ff-f9f4ea257ed7] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port fb0a1841-27e9-4610-abba-889cf0217607, please check neutron logs for more information. [ 659.012230] env[63515]: ERROR nova.compute.manager [instance: bbce1137-d691-4633-87ff-f9f4ea257ed7] Traceback (most recent call last): [ 659.012230] env[63515]: ERROR nova.compute.manager [instance: bbce1137-d691-4633-87ff-f9f4ea257ed7] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 659.012230] env[63515]: ERROR nova.compute.manager [instance: bbce1137-d691-4633-87ff-f9f4ea257ed7] yield resources [ 659.012230] env[63515]: ERROR nova.compute.manager [instance: bbce1137-d691-4633-87ff-f9f4ea257ed7] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 659.012230] env[63515]: ERROR nova.compute.manager [instance: bbce1137-d691-4633-87ff-f9f4ea257ed7] self.driver.spawn(context, instance, image_meta, [ 659.012230] env[63515]: ERROR nova.compute.manager [instance: bbce1137-d691-4633-87ff-f9f4ea257ed7] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 659.012230] env[63515]: ERROR nova.compute.manager [instance: bbce1137-d691-4633-87ff-f9f4ea257ed7] self._vmops.spawn(context, instance, image_meta, injected_files, [ 659.012230] env[63515]: ERROR nova.compute.manager [instance: bbce1137-d691-4633-87ff-f9f4ea257ed7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 659.012230] env[63515]: ERROR nova.compute.manager [instance: bbce1137-d691-4633-87ff-f9f4ea257ed7] vm_ref = self.build_virtual_machine(instance, [ 659.012230] env[63515]: ERROR nova.compute.manager [instance: bbce1137-d691-4633-87ff-f9f4ea257ed7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 659.012634] env[63515]: ERROR nova.compute.manager [instance: bbce1137-d691-4633-87ff-f9f4ea257ed7] vif_infos = vmwarevif.get_vif_info(self._session, [ 659.012634] env[63515]: ERROR nova.compute.manager [instance: bbce1137-d691-4633-87ff-f9f4ea257ed7] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 659.012634] env[63515]: ERROR nova.compute.manager [instance: bbce1137-d691-4633-87ff-f9f4ea257ed7] for vif in network_info: [ 659.012634] env[63515]: ERROR nova.compute.manager [instance: bbce1137-d691-4633-87ff-f9f4ea257ed7] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 659.012634] env[63515]: ERROR nova.compute.manager [instance: bbce1137-d691-4633-87ff-f9f4ea257ed7] return self._sync_wrapper(fn, *args, **kwargs) [ 659.012634] env[63515]: ERROR nova.compute.manager [instance: bbce1137-d691-4633-87ff-f9f4ea257ed7] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 659.012634] env[63515]: ERROR nova.compute.manager [instance: bbce1137-d691-4633-87ff-f9f4ea257ed7] self.wait() [ 659.012634] env[63515]: ERROR nova.compute.manager [instance: bbce1137-d691-4633-87ff-f9f4ea257ed7] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 659.012634] env[63515]: ERROR nova.compute.manager [instance: bbce1137-d691-4633-87ff-f9f4ea257ed7] self[:] = self._gt.wait() [ 659.012634] env[63515]: ERROR nova.compute.manager [instance: bbce1137-d691-4633-87ff-f9f4ea257ed7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 659.012634] env[63515]: ERROR nova.compute.manager [instance: bbce1137-d691-4633-87ff-f9f4ea257ed7] return self._exit_event.wait() [ 659.012634] env[63515]: ERROR nova.compute.manager [instance: bbce1137-d691-4633-87ff-f9f4ea257ed7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 659.012634] env[63515]: ERROR nova.compute.manager [instance: bbce1137-d691-4633-87ff-f9f4ea257ed7] current.throw(*self._exc) [ 659.013012] env[63515]: ERROR nova.compute.manager [instance: bbce1137-d691-4633-87ff-f9f4ea257ed7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 659.013012] env[63515]: ERROR nova.compute.manager [instance: bbce1137-d691-4633-87ff-f9f4ea257ed7] result = function(*args, **kwargs) [ 659.013012] env[63515]: ERROR nova.compute.manager [instance: bbce1137-d691-4633-87ff-f9f4ea257ed7] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 659.013012] env[63515]: ERROR nova.compute.manager [instance: bbce1137-d691-4633-87ff-f9f4ea257ed7] return func(*args, **kwargs) [ 659.013012] env[63515]: ERROR nova.compute.manager [instance: bbce1137-d691-4633-87ff-f9f4ea257ed7] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 659.013012] env[63515]: ERROR nova.compute.manager [instance: bbce1137-d691-4633-87ff-f9f4ea257ed7] raise e [ 659.013012] env[63515]: ERROR nova.compute.manager [instance: bbce1137-d691-4633-87ff-f9f4ea257ed7] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 659.013012] env[63515]: ERROR nova.compute.manager [instance: bbce1137-d691-4633-87ff-f9f4ea257ed7] nwinfo = self.network_api.allocate_for_instance( [ 659.013012] env[63515]: ERROR nova.compute.manager [instance: bbce1137-d691-4633-87ff-f9f4ea257ed7] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 659.013012] env[63515]: ERROR nova.compute.manager [instance: bbce1137-d691-4633-87ff-f9f4ea257ed7] created_port_ids = self._update_ports_for_instance( [ 659.013012] env[63515]: ERROR nova.compute.manager [instance: bbce1137-d691-4633-87ff-f9f4ea257ed7] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 659.013012] env[63515]: ERROR nova.compute.manager [instance: bbce1137-d691-4633-87ff-f9f4ea257ed7] with excutils.save_and_reraise_exception(): [ 659.013012] env[63515]: ERROR nova.compute.manager [instance: bbce1137-d691-4633-87ff-f9f4ea257ed7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 659.013403] env[63515]: ERROR nova.compute.manager [instance: bbce1137-d691-4633-87ff-f9f4ea257ed7] self.force_reraise() [ 659.013403] env[63515]: ERROR nova.compute.manager [instance: bbce1137-d691-4633-87ff-f9f4ea257ed7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 659.013403] env[63515]: ERROR nova.compute.manager [instance: bbce1137-d691-4633-87ff-f9f4ea257ed7] raise self.value [ 659.013403] env[63515]: ERROR nova.compute.manager [instance: bbce1137-d691-4633-87ff-f9f4ea257ed7] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 659.013403] env[63515]: ERROR nova.compute.manager [instance: bbce1137-d691-4633-87ff-f9f4ea257ed7] updated_port = self._update_port( [ 659.013403] env[63515]: ERROR nova.compute.manager [instance: bbce1137-d691-4633-87ff-f9f4ea257ed7] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 659.013403] env[63515]: ERROR nova.compute.manager [instance: bbce1137-d691-4633-87ff-f9f4ea257ed7] _ensure_no_port_binding_failure(port) [ 659.013403] env[63515]: ERROR nova.compute.manager [instance: bbce1137-d691-4633-87ff-f9f4ea257ed7] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 659.013403] env[63515]: ERROR nova.compute.manager [instance: bbce1137-d691-4633-87ff-f9f4ea257ed7] raise exception.PortBindingFailed(port_id=port['id']) [ 659.013403] env[63515]: ERROR nova.compute.manager [instance: bbce1137-d691-4633-87ff-f9f4ea257ed7] nova.exception.PortBindingFailed: Binding failed for port fb0a1841-27e9-4610-abba-889cf0217607, please check neutron logs for more information. [ 659.013403] env[63515]: ERROR nova.compute.manager [instance: bbce1137-d691-4633-87ff-f9f4ea257ed7] [ 659.013403] env[63515]: INFO nova.compute.manager [None req-30987b61-ce89-4676-89be-fee7786462c9 tempest-ImagesOneServerNegativeTestJSON-52225474 tempest-ImagesOneServerNegativeTestJSON-52225474-project-member] [instance: bbce1137-d691-4633-87ff-f9f4ea257ed7] Terminating instance [ 659.015089] env[63515]: DEBUG oslo_concurrency.lockutils [None req-30987b61-ce89-4676-89be-fee7786462c9 tempest-ImagesOneServerNegativeTestJSON-52225474 tempest-ImagesOneServerNegativeTestJSON-52225474-project-member] Acquiring lock "refresh_cache-bbce1137-d691-4633-87ff-f9f4ea257ed7" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 659.178550] env[63515]: DEBUG nova.network.neutron [req-bb3207d2-ca06-4e5d-bd64-427741c5bdc7 req-3ca56192-01aa-43ef-94c4-3b8538d908b5 service nova] [instance: bbce1137-d691-4633-87ff-f9f4ea257ed7] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 659.260648] env[63515]: DEBUG nova.network.neutron [req-bb3207d2-ca06-4e5d-bd64-427741c5bdc7 req-3ca56192-01aa-43ef-94c4-3b8538d908b5 service nova] [instance: bbce1137-d691-4633-87ff-f9f4ea257ed7] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 659.428996] env[63515]: DEBUG oslo_concurrency.lockutils [None req-480b7d6d-2f7f-4da3-9eb1-818f0711a940 tempest-ServerRescueTestJSON-1961208277 tempest-ServerRescueTestJSON-1961208277-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.492s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 659.428996] env[63515]: DEBUG nova.compute.manager [None req-480b7d6d-2f7f-4da3-9eb1-818f0711a940 tempest-ServerRescueTestJSON-1961208277 tempest-ServerRescueTestJSON-1961208277-project-member] [instance: 7f49d79e-3bda-4949-a976-7e3e6513b2c7] Start building networks asynchronously for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 659.431519] env[63515]: DEBUG oslo_concurrency.lockutils [None req-9131a685-703c-44a1-81f8-7c157d95e037 tempest-ServerAddressesNegativeTestJSON-1262066083 tempest-ServerAddressesNegativeTestJSON-1262066083-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 29.586s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 659.762326] env[63515]: DEBUG oslo_concurrency.lockutils [req-bb3207d2-ca06-4e5d-bd64-427741c5bdc7 req-3ca56192-01aa-43ef-94c4-3b8538d908b5 service nova] Releasing lock "refresh_cache-bbce1137-d691-4633-87ff-f9f4ea257ed7" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 659.763807] env[63515]: DEBUG oslo_concurrency.lockutils [None req-30987b61-ce89-4676-89be-fee7786462c9 tempest-ImagesOneServerNegativeTestJSON-52225474 tempest-ImagesOneServerNegativeTestJSON-52225474-project-member] Acquired lock "refresh_cache-bbce1137-d691-4633-87ff-f9f4ea257ed7" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 659.763807] env[63515]: DEBUG nova.network.neutron [None req-30987b61-ce89-4676-89be-fee7786462c9 tempest-ImagesOneServerNegativeTestJSON-52225474 tempest-ImagesOneServerNegativeTestJSON-52225474-project-member] [instance: bbce1137-d691-4633-87ff-f9f4ea257ed7] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 659.936638] env[63515]: DEBUG nova.compute.utils [None req-480b7d6d-2f7f-4da3-9eb1-818f0711a940 tempest-ServerRescueTestJSON-1961208277 tempest-ServerRescueTestJSON-1961208277-project-member] Using /dev/sd instead of None {{(pid=63515) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 659.941072] env[63515]: DEBUG nova.compute.manager [None req-480b7d6d-2f7f-4da3-9eb1-818f0711a940 tempest-ServerRescueTestJSON-1961208277 tempest-ServerRescueTestJSON-1961208277-project-member] [instance: 7f49d79e-3bda-4949-a976-7e3e6513b2c7] Allocating IP information in the background. {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 659.941238] env[63515]: DEBUG nova.network.neutron [None req-480b7d6d-2f7f-4da3-9eb1-818f0711a940 tempest-ServerRescueTestJSON-1961208277 tempest-ServerRescueTestJSON-1961208277-project-member] [instance: 7f49d79e-3bda-4949-a976-7e3e6513b2c7] allocate_for_instance() {{(pid=63515) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 659.991889] env[63515]: DEBUG nova.policy [None req-480b7d6d-2f7f-4da3-9eb1-818f0711a940 tempest-ServerRescueTestJSON-1961208277 tempest-ServerRescueTestJSON-1961208277-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '80b60813fefe49b9a9734617f3d99072', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '002ab6c947c341f69b5401936658fc51', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63515) authorize /opt/stack/nova/nova/policy.py:201}} [ 660.268322] env[63515]: DEBUG nova.network.neutron [None req-480b7d6d-2f7f-4da3-9eb1-818f0711a940 tempest-ServerRescueTestJSON-1961208277 tempest-ServerRescueTestJSON-1961208277-project-member] [instance: 7f49d79e-3bda-4949-a976-7e3e6513b2c7] Successfully created port: cc888fc6-b889-48e3-bb61-bfa980ef91fd {{(pid=63515) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 660.288250] env[63515]: DEBUG nova.network.neutron [None req-30987b61-ce89-4676-89be-fee7786462c9 tempest-ImagesOneServerNegativeTestJSON-52225474 tempest-ImagesOneServerNegativeTestJSON-52225474-project-member] [instance: bbce1137-d691-4633-87ff-f9f4ea257ed7] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 660.332174] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b72823a2-74a2-4710-abad-81076e17d802 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.341205] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57917250-4058-4fbc-abdc-f6121a6a9557 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.376651] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c880540b-83cc-4a84-846a-0ead0d3a3c8a {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.384246] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b136febb-ac02-4d97-b460-f15dcc6490bc {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.397442] env[63515]: DEBUG nova.compute.provider_tree [None req-9131a685-703c-44a1-81f8-7c157d95e037 tempest-ServerAddressesNegativeTestJSON-1262066083 tempest-ServerAddressesNegativeTestJSON-1262066083-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 660.444080] env[63515]: DEBUG nova.compute.manager [None req-480b7d6d-2f7f-4da3-9eb1-818f0711a940 tempest-ServerRescueTestJSON-1961208277 tempest-ServerRescueTestJSON-1961208277-project-member] [instance: 7f49d79e-3bda-4949-a976-7e3e6513b2c7] Start building block device mappings for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 660.448656] env[63515]: DEBUG nova.network.neutron [None req-30987b61-ce89-4676-89be-fee7786462c9 tempest-ImagesOneServerNegativeTestJSON-52225474 tempest-ImagesOneServerNegativeTestJSON-52225474-project-member] [instance: bbce1137-d691-4633-87ff-f9f4ea257ed7] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 660.683014] env[63515]: DEBUG nova.compute.manager [req-fc2edd22-eed3-4e1e-8e52-1da3e14415e8 req-5141914d-76f9-4a2e-92ad-a2b2af237e79 service nova] [instance: bbce1137-d691-4633-87ff-f9f4ea257ed7] Received event network-vif-deleted-fb0a1841-27e9-4610-abba-889cf0217607 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 660.899907] env[63515]: DEBUG nova.scheduler.client.report [None req-9131a685-703c-44a1-81f8-7c157d95e037 tempest-ServerAddressesNegativeTestJSON-1262066083 tempest-ServerAddressesNegativeTestJSON-1262066083-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 660.958316] env[63515]: DEBUG oslo_concurrency.lockutils [None req-30987b61-ce89-4676-89be-fee7786462c9 tempest-ImagesOneServerNegativeTestJSON-52225474 tempest-ImagesOneServerNegativeTestJSON-52225474-project-member] Releasing lock "refresh_cache-bbce1137-d691-4633-87ff-f9f4ea257ed7" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 660.958316] env[63515]: DEBUG nova.compute.manager [None req-30987b61-ce89-4676-89be-fee7786462c9 tempest-ImagesOneServerNegativeTestJSON-52225474 tempest-ImagesOneServerNegativeTestJSON-52225474-project-member] [instance: bbce1137-d691-4633-87ff-f9f4ea257ed7] Start destroying the instance on the hypervisor. {{(pid=63515) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 660.958316] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-30987b61-ce89-4676-89be-fee7786462c9 tempest-ImagesOneServerNegativeTestJSON-52225474 tempest-ImagesOneServerNegativeTestJSON-52225474-project-member] [instance: bbce1137-d691-4633-87ff-f9f4ea257ed7] Destroying instance {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 660.958316] env[63515]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6d0733e4-7357-4c2a-b627-a6ef2ff33baa {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.964930] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3539819-11fc-4534-8904-9bc9d95f4620 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.988205] env[63515]: WARNING nova.virt.vmwareapi.vmops [None req-30987b61-ce89-4676-89be-fee7786462c9 tempest-ImagesOneServerNegativeTestJSON-52225474 tempest-ImagesOneServerNegativeTestJSON-52225474-project-member] [instance: bbce1137-d691-4633-87ff-f9f4ea257ed7] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance bbce1137-d691-4633-87ff-f9f4ea257ed7 could not be found. [ 660.988205] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-30987b61-ce89-4676-89be-fee7786462c9 tempest-ImagesOneServerNegativeTestJSON-52225474 tempest-ImagesOneServerNegativeTestJSON-52225474-project-member] [instance: bbce1137-d691-4633-87ff-f9f4ea257ed7] Instance destroyed {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 660.988205] env[63515]: INFO nova.compute.manager [None req-30987b61-ce89-4676-89be-fee7786462c9 tempest-ImagesOneServerNegativeTestJSON-52225474 tempest-ImagesOneServerNegativeTestJSON-52225474-project-member] [instance: bbce1137-d691-4633-87ff-f9f4ea257ed7] Took 0.03 seconds to destroy the instance on the hypervisor. [ 660.988205] env[63515]: DEBUG oslo.service.loopingcall [None req-30987b61-ce89-4676-89be-fee7786462c9 tempest-ImagesOneServerNegativeTestJSON-52225474 tempest-ImagesOneServerNegativeTestJSON-52225474-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 660.988205] env[63515]: DEBUG nova.compute.manager [-] [instance: bbce1137-d691-4633-87ff-f9f4ea257ed7] Deallocating network for instance {{(pid=63515) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 660.988205] env[63515]: DEBUG nova.network.neutron [-] [instance: bbce1137-d691-4633-87ff-f9f4ea257ed7] deallocate_for_instance() {{(pid=63515) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 661.009271] env[63515]: DEBUG nova.network.neutron [-] [instance: bbce1137-d691-4633-87ff-f9f4ea257ed7] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 661.342709] env[63515]: ERROR nova.compute.manager [None req-480b7d6d-2f7f-4da3-9eb1-818f0711a940 tempest-ServerRescueTestJSON-1961208277 tempest-ServerRescueTestJSON-1961208277-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port cc888fc6-b889-48e3-bb61-bfa980ef91fd, please check neutron logs for more information. [ 661.342709] env[63515]: ERROR nova.compute.manager Traceback (most recent call last): [ 661.342709] env[63515]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 661.342709] env[63515]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 661.342709] env[63515]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 661.342709] env[63515]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 661.342709] env[63515]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 661.342709] env[63515]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 661.342709] env[63515]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 661.342709] env[63515]: ERROR nova.compute.manager self.force_reraise() [ 661.342709] env[63515]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 661.342709] env[63515]: ERROR nova.compute.manager raise self.value [ 661.342709] env[63515]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 661.342709] env[63515]: ERROR nova.compute.manager updated_port = self._update_port( [ 661.342709] env[63515]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 661.342709] env[63515]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 661.343251] env[63515]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 661.343251] env[63515]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 661.343251] env[63515]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port cc888fc6-b889-48e3-bb61-bfa980ef91fd, please check neutron logs for more information. [ 661.343251] env[63515]: ERROR nova.compute.manager [ 661.343251] env[63515]: Traceback (most recent call last): [ 661.343251] env[63515]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 661.343251] env[63515]: listener.cb(fileno) [ 661.343251] env[63515]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 661.343251] env[63515]: result = function(*args, **kwargs) [ 661.343251] env[63515]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 661.343251] env[63515]: return func(*args, **kwargs) [ 661.343251] env[63515]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 661.343251] env[63515]: raise e [ 661.343251] env[63515]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 661.343251] env[63515]: nwinfo = self.network_api.allocate_for_instance( [ 661.343251] env[63515]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 661.343251] env[63515]: created_port_ids = self._update_ports_for_instance( [ 661.343251] env[63515]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 661.343251] env[63515]: with excutils.save_and_reraise_exception(): [ 661.343251] env[63515]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 661.343251] env[63515]: self.force_reraise() [ 661.343251] env[63515]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 661.343251] env[63515]: raise self.value [ 661.343251] env[63515]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 661.343251] env[63515]: updated_port = self._update_port( [ 661.343251] env[63515]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 661.343251] env[63515]: _ensure_no_port_binding_failure(port) [ 661.343251] env[63515]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 661.343251] env[63515]: raise exception.PortBindingFailed(port_id=port['id']) [ 661.344459] env[63515]: nova.exception.PortBindingFailed: Binding failed for port cc888fc6-b889-48e3-bb61-bfa980ef91fd, please check neutron logs for more information. [ 661.344459] env[63515]: Removing descriptor: 17 [ 661.405958] env[63515]: DEBUG oslo_concurrency.lockutils [None req-9131a685-703c-44a1-81f8-7c157d95e037 tempest-ServerAddressesNegativeTestJSON-1262066083 tempest-ServerAddressesNegativeTestJSON-1262066083-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.974s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 661.406766] env[63515]: ERROR nova.compute.manager [None req-9131a685-703c-44a1-81f8-7c157d95e037 tempest-ServerAddressesNegativeTestJSON-1262066083 tempest-ServerAddressesNegativeTestJSON-1262066083-project-member] [instance: 5b201ab6-c7ec-4a5c-a310-982649f34be4] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 4d81feb5-11c1-415a-bdaf-cd39beb2e524, please check neutron logs for more information. [ 661.406766] env[63515]: ERROR nova.compute.manager [instance: 5b201ab6-c7ec-4a5c-a310-982649f34be4] Traceback (most recent call last): [ 661.406766] env[63515]: ERROR nova.compute.manager [instance: 5b201ab6-c7ec-4a5c-a310-982649f34be4] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 661.406766] env[63515]: ERROR nova.compute.manager [instance: 5b201ab6-c7ec-4a5c-a310-982649f34be4] self.driver.spawn(context, instance, image_meta, [ 661.406766] env[63515]: ERROR nova.compute.manager [instance: 5b201ab6-c7ec-4a5c-a310-982649f34be4] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 661.406766] env[63515]: ERROR nova.compute.manager [instance: 5b201ab6-c7ec-4a5c-a310-982649f34be4] self._vmops.spawn(context, instance, image_meta, injected_files, [ 661.406766] env[63515]: ERROR nova.compute.manager [instance: 5b201ab6-c7ec-4a5c-a310-982649f34be4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 661.406766] env[63515]: ERROR nova.compute.manager [instance: 5b201ab6-c7ec-4a5c-a310-982649f34be4] vm_ref = self.build_virtual_machine(instance, [ 661.406766] env[63515]: ERROR nova.compute.manager [instance: 5b201ab6-c7ec-4a5c-a310-982649f34be4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 661.406766] env[63515]: ERROR nova.compute.manager [instance: 5b201ab6-c7ec-4a5c-a310-982649f34be4] vif_infos = vmwarevif.get_vif_info(self._session, [ 661.406766] env[63515]: ERROR nova.compute.manager [instance: 5b201ab6-c7ec-4a5c-a310-982649f34be4] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 661.407073] env[63515]: ERROR nova.compute.manager [instance: 5b201ab6-c7ec-4a5c-a310-982649f34be4] for vif in network_info: [ 661.407073] env[63515]: ERROR nova.compute.manager [instance: 5b201ab6-c7ec-4a5c-a310-982649f34be4] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 661.407073] env[63515]: ERROR nova.compute.manager [instance: 5b201ab6-c7ec-4a5c-a310-982649f34be4] return self._sync_wrapper(fn, *args, **kwargs) [ 661.407073] env[63515]: ERROR nova.compute.manager [instance: 5b201ab6-c7ec-4a5c-a310-982649f34be4] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 661.407073] env[63515]: ERROR nova.compute.manager [instance: 5b201ab6-c7ec-4a5c-a310-982649f34be4] self.wait() [ 661.407073] env[63515]: ERROR nova.compute.manager [instance: 5b201ab6-c7ec-4a5c-a310-982649f34be4] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 661.407073] env[63515]: ERROR nova.compute.manager [instance: 5b201ab6-c7ec-4a5c-a310-982649f34be4] self[:] = self._gt.wait() [ 661.407073] env[63515]: ERROR nova.compute.manager [instance: 5b201ab6-c7ec-4a5c-a310-982649f34be4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 661.407073] env[63515]: ERROR nova.compute.manager [instance: 5b201ab6-c7ec-4a5c-a310-982649f34be4] return self._exit_event.wait() [ 661.407073] env[63515]: ERROR nova.compute.manager [instance: 5b201ab6-c7ec-4a5c-a310-982649f34be4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 661.407073] env[63515]: ERROR nova.compute.manager [instance: 5b201ab6-c7ec-4a5c-a310-982649f34be4] result = hub.switch() [ 661.407073] env[63515]: ERROR nova.compute.manager [instance: 5b201ab6-c7ec-4a5c-a310-982649f34be4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 661.407073] env[63515]: ERROR nova.compute.manager [instance: 5b201ab6-c7ec-4a5c-a310-982649f34be4] return self.greenlet.switch() [ 661.407412] env[63515]: ERROR nova.compute.manager [instance: 5b201ab6-c7ec-4a5c-a310-982649f34be4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 661.407412] env[63515]: ERROR nova.compute.manager [instance: 5b201ab6-c7ec-4a5c-a310-982649f34be4] result = function(*args, **kwargs) [ 661.407412] env[63515]: ERROR nova.compute.manager [instance: 5b201ab6-c7ec-4a5c-a310-982649f34be4] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 661.407412] env[63515]: ERROR nova.compute.manager [instance: 5b201ab6-c7ec-4a5c-a310-982649f34be4] return func(*args, **kwargs) [ 661.407412] env[63515]: ERROR nova.compute.manager [instance: 5b201ab6-c7ec-4a5c-a310-982649f34be4] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 661.407412] env[63515]: ERROR nova.compute.manager [instance: 5b201ab6-c7ec-4a5c-a310-982649f34be4] raise e [ 661.407412] env[63515]: ERROR nova.compute.manager [instance: 5b201ab6-c7ec-4a5c-a310-982649f34be4] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 661.407412] env[63515]: ERROR nova.compute.manager [instance: 5b201ab6-c7ec-4a5c-a310-982649f34be4] nwinfo = self.network_api.allocate_for_instance( [ 661.407412] env[63515]: ERROR nova.compute.manager [instance: 5b201ab6-c7ec-4a5c-a310-982649f34be4] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 661.407412] env[63515]: ERROR nova.compute.manager [instance: 5b201ab6-c7ec-4a5c-a310-982649f34be4] created_port_ids = self._update_ports_for_instance( [ 661.407412] env[63515]: ERROR nova.compute.manager [instance: 5b201ab6-c7ec-4a5c-a310-982649f34be4] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 661.407412] env[63515]: ERROR nova.compute.manager [instance: 5b201ab6-c7ec-4a5c-a310-982649f34be4] with excutils.save_and_reraise_exception(): [ 661.407412] env[63515]: ERROR nova.compute.manager [instance: 5b201ab6-c7ec-4a5c-a310-982649f34be4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 661.407830] env[63515]: ERROR nova.compute.manager [instance: 5b201ab6-c7ec-4a5c-a310-982649f34be4] self.force_reraise() [ 661.407830] env[63515]: ERROR nova.compute.manager [instance: 5b201ab6-c7ec-4a5c-a310-982649f34be4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 661.407830] env[63515]: ERROR nova.compute.manager [instance: 5b201ab6-c7ec-4a5c-a310-982649f34be4] raise self.value [ 661.407830] env[63515]: ERROR nova.compute.manager [instance: 5b201ab6-c7ec-4a5c-a310-982649f34be4] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 661.407830] env[63515]: ERROR nova.compute.manager [instance: 5b201ab6-c7ec-4a5c-a310-982649f34be4] updated_port = self._update_port( [ 661.407830] env[63515]: ERROR nova.compute.manager [instance: 5b201ab6-c7ec-4a5c-a310-982649f34be4] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 661.407830] env[63515]: ERROR nova.compute.manager [instance: 5b201ab6-c7ec-4a5c-a310-982649f34be4] _ensure_no_port_binding_failure(port) [ 661.407830] env[63515]: ERROR nova.compute.manager [instance: 5b201ab6-c7ec-4a5c-a310-982649f34be4] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 661.407830] env[63515]: ERROR nova.compute.manager [instance: 5b201ab6-c7ec-4a5c-a310-982649f34be4] raise exception.PortBindingFailed(port_id=port['id']) [ 661.407830] env[63515]: ERROR nova.compute.manager [instance: 5b201ab6-c7ec-4a5c-a310-982649f34be4] nova.exception.PortBindingFailed: Binding failed for port 4d81feb5-11c1-415a-bdaf-cd39beb2e524, please check neutron logs for more information. [ 661.407830] env[63515]: ERROR nova.compute.manager [instance: 5b201ab6-c7ec-4a5c-a310-982649f34be4] [ 661.408129] env[63515]: DEBUG nova.compute.utils [None req-9131a685-703c-44a1-81f8-7c157d95e037 tempest-ServerAddressesNegativeTestJSON-1262066083 tempest-ServerAddressesNegativeTestJSON-1262066083-project-member] [instance: 5b201ab6-c7ec-4a5c-a310-982649f34be4] Binding failed for port 4d81feb5-11c1-415a-bdaf-cd39beb2e524, please check neutron logs for more information. {{(pid=63515) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 661.409576] env[63515]: DEBUG nova.compute.manager [None req-9131a685-703c-44a1-81f8-7c157d95e037 tempest-ServerAddressesNegativeTestJSON-1262066083 tempest-ServerAddressesNegativeTestJSON-1262066083-project-member] [instance: 5b201ab6-c7ec-4a5c-a310-982649f34be4] Build of instance 5b201ab6-c7ec-4a5c-a310-982649f34be4 was re-scheduled: Binding failed for port 4d81feb5-11c1-415a-bdaf-cd39beb2e524, please check neutron logs for more information. {{(pid=63515) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 661.410013] env[63515]: DEBUG nova.compute.manager [None req-9131a685-703c-44a1-81f8-7c157d95e037 tempest-ServerAddressesNegativeTestJSON-1262066083 tempest-ServerAddressesNegativeTestJSON-1262066083-project-member] [instance: 5b201ab6-c7ec-4a5c-a310-982649f34be4] Unplugging VIFs for instance {{(pid=63515) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 661.410255] env[63515]: DEBUG oslo_concurrency.lockutils [None req-9131a685-703c-44a1-81f8-7c157d95e037 tempest-ServerAddressesNegativeTestJSON-1262066083 tempest-ServerAddressesNegativeTestJSON-1262066083-project-member] Acquiring lock "refresh_cache-5b201ab6-c7ec-4a5c-a310-982649f34be4" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 661.410408] env[63515]: DEBUG oslo_concurrency.lockutils [None req-9131a685-703c-44a1-81f8-7c157d95e037 tempest-ServerAddressesNegativeTestJSON-1262066083 tempest-ServerAddressesNegativeTestJSON-1262066083-project-member] Acquired lock "refresh_cache-5b201ab6-c7ec-4a5c-a310-982649f34be4" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 661.410565] env[63515]: DEBUG nova.network.neutron [None req-9131a685-703c-44a1-81f8-7c157d95e037 tempest-ServerAddressesNegativeTestJSON-1262066083 tempest-ServerAddressesNegativeTestJSON-1262066083-project-member] [instance: 5b201ab6-c7ec-4a5c-a310-982649f34be4] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 661.411965] env[63515]: DEBUG oslo_concurrency.lockutils [None req-b9c961a0-a559-4690-abb6-579359ed447c tempest-ServerPasswordTestJSON-433685770 tempest-ServerPasswordTestJSON-433685770-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 31.343s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 661.413523] env[63515]: INFO nova.compute.claims [None req-b9c961a0-a559-4690-abb6-579359ed447c tempest-ServerPasswordTestJSON-433685770 tempest-ServerPasswordTestJSON-433685770-project-member] [instance: 07499601-62bc-4c31-b295-23f34a6e2e91] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 661.456136] env[63515]: DEBUG nova.compute.manager [None req-480b7d6d-2f7f-4da3-9eb1-818f0711a940 tempest-ServerRescueTestJSON-1961208277 tempest-ServerRescueTestJSON-1961208277-project-member] [instance: 7f49d79e-3bda-4949-a976-7e3e6513b2c7] Start spawning the instance on the hypervisor. {{(pid=63515) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 661.481734] env[63515]: DEBUG nova.virt.hardware [None req-480b7d6d-2f7f-4da3-9eb1-818f0711a940 tempest-ServerRescueTestJSON-1961208277 tempest-ServerRescueTestJSON-1961208277-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T02:52:14Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T02:51:56Z,direct_url=,disk_format='vmdk',id=8a120570-cb06-4099-b262-554ca0ad15c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b569255cc43e42e7ae3f2b4ad37c6ef4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T02:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 661.481983] env[63515]: DEBUG nova.virt.hardware [None req-480b7d6d-2f7f-4da3-9eb1-818f0711a940 tempest-ServerRescueTestJSON-1961208277 tempest-ServerRescueTestJSON-1961208277-project-member] Flavor limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 661.482158] env[63515]: DEBUG nova.virt.hardware [None req-480b7d6d-2f7f-4da3-9eb1-818f0711a940 tempest-ServerRescueTestJSON-1961208277 tempest-ServerRescueTestJSON-1961208277-project-member] Image limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 661.482343] env[63515]: DEBUG nova.virt.hardware [None req-480b7d6d-2f7f-4da3-9eb1-818f0711a940 tempest-ServerRescueTestJSON-1961208277 tempest-ServerRescueTestJSON-1961208277-project-member] Flavor pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 661.482490] env[63515]: DEBUG nova.virt.hardware [None req-480b7d6d-2f7f-4da3-9eb1-818f0711a940 tempest-ServerRescueTestJSON-1961208277 tempest-ServerRescueTestJSON-1961208277-project-member] Image pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 661.482634] env[63515]: DEBUG nova.virt.hardware [None req-480b7d6d-2f7f-4da3-9eb1-818f0711a940 tempest-ServerRescueTestJSON-1961208277 tempest-ServerRescueTestJSON-1961208277-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 661.482836] env[63515]: DEBUG nova.virt.hardware [None req-480b7d6d-2f7f-4da3-9eb1-818f0711a940 tempest-ServerRescueTestJSON-1961208277 tempest-ServerRescueTestJSON-1961208277-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 661.482991] env[63515]: DEBUG nova.virt.hardware [None req-480b7d6d-2f7f-4da3-9eb1-818f0711a940 tempest-ServerRescueTestJSON-1961208277 tempest-ServerRescueTestJSON-1961208277-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 661.483180] env[63515]: DEBUG nova.virt.hardware [None req-480b7d6d-2f7f-4da3-9eb1-818f0711a940 tempest-ServerRescueTestJSON-1961208277 tempest-ServerRescueTestJSON-1961208277-project-member] Got 1 possible topologies {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 661.483342] env[63515]: DEBUG nova.virt.hardware [None req-480b7d6d-2f7f-4da3-9eb1-818f0711a940 tempest-ServerRescueTestJSON-1961208277 tempest-ServerRescueTestJSON-1961208277-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 661.483509] env[63515]: DEBUG nova.virt.hardware [None req-480b7d6d-2f7f-4da3-9eb1-818f0711a940 tempest-ServerRescueTestJSON-1961208277 tempest-ServerRescueTestJSON-1961208277-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 661.484710] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41b1af14-23df-4145-b9f8-8a36bfb6f291 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.492550] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b0d42ec-33b7-4750-b5b8-4f04054c7cd2 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.507458] env[63515]: ERROR nova.compute.manager [None req-480b7d6d-2f7f-4da3-9eb1-818f0711a940 tempest-ServerRescueTestJSON-1961208277 tempest-ServerRescueTestJSON-1961208277-project-member] [instance: 7f49d79e-3bda-4949-a976-7e3e6513b2c7] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port cc888fc6-b889-48e3-bb61-bfa980ef91fd, please check neutron logs for more information. [ 661.507458] env[63515]: ERROR nova.compute.manager [instance: 7f49d79e-3bda-4949-a976-7e3e6513b2c7] Traceback (most recent call last): [ 661.507458] env[63515]: ERROR nova.compute.manager [instance: 7f49d79e-3bda-4949-a976-7e3e6513b2c7] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 661.507458] env[63515]: ERROR nova.compute.manager [instance: 7f49d79e-3bda-4949-a976-7e3e6513b2c7] yield resources [ 661.507458] env[63515]: ERROR nova.compute.manager [instance: 7f49d79e-3bda-4949-a976-7e3e6513b2c7] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 661.507458] env[63515]: ERROR nova.compute.manager [instance: 7f49d79e-3bda-4949-a976-7e3e6513b2c7] self.driver.spawn(context, instance, image_meta, [ 661.507458] env[63515]: ERROR nova.compute.manager [instance: 7f49d79e-3bda-4949-a976-7e3e6513b2c7] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 661.507458] env[63515]: ERROR nova.compute.manager [instance: 7f49d79e-3bda-4949-a976-7e3e6513b2c7] self._vmops.spawn(context, instance, image_meta, injected_files, [ 661.507458] env[63515]: ERROR nova.compute.manager [instance: 7f49d79e-3bda-4949-a976-7e3e6513b2c7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 661.507458] env[63515]: ERROR nova.compute.manager [instance: 7f49d79e-3bda-4949-a976-7e3e6513b2c7] vm_ref = self.build_virtual_machine(instance, [ 661.507458] env[63515]: ERROR nova.compute.manager [instance: 7f49d79e-3bda-4949-a976-7e3e6513b2c7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 661.508030] env[63515]: ERROR nova.compute.manager [instance: 7f49d79e-3bda-4949-a976-7e3e6513b2c7] vif_infos = vmwarevif.get_vif_info(self._session, [ 661.508030] env[63515]: ERROR nova.compute.manager [instance: 7f49d79e-3bda-4949-a976-7e3e6513b2c7] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 661.508030] env[63515]: ERROR nova.compute.manager [instance: 7f49d79e-3bda-4949-a976-7e3e6513b2c7] for vif in network_info: [ 661.508030] env[63515]: ERROR nova.compute.manager [instance: 7f49d79e-3bda-4949-a976-7e3e6513b2c7] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 661.508030] env[63515]: ERROR nova.compute.manager [instance: 7f49d79e-3bda-4949-a976-7e3e6513b2c7] return self._sync_wrapper(fn, *args, **kwargs) [ 661.508030] env[63515]: ERROR nova.compute.manager [instance: 7f49d79e-3bda-4949-a976-7e3e6513b2c7] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 661.508030] env[63515]: ERROR nova.compute.manager [instance: 7f49d79e-3bda-4949-a976-7e3e6513b2c7] self.wait() [ 661.508030] env[63515]: ERROR nova.compute.manager [instance: 7f49d79e-3bda-4949-a976-7e3e6513b2c7] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 661.508030] env[63515]: ERROR nova.compute.manager [instance: 7f49d79e-3bda-4949-a976-7e3e6513b2c7] self[:] = self._gt.wait() [ 661.508030] env[63515]: ERROR nova.compute.manager [instance: 7f49d79e-3bda-4949-a976-7e3e6513b2c7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 661.508030] env[63515]: ERROR nova.compute.manager [instance: 7f49d79e-3bda-4949-a976-7e3e6513b2c7] return self._exit_event.wait() [ 661.508030] env[63515]: ERROR nova.compute.manager [instance: 7f49d79e-3bda-4949-a976-7e3e6513b2c7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 661.508030] env[63515]: ERROR nova.compute.manager [instance: 7f49d79e-3bda-4949-a976-7e3e6513b2c7] current.throw(*self._exc) [ 661.508421] env[63515]: ERROR nova.compute.manager [instance: 7f49d79e-3bda-4949-a976-7e3e6513b2c7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 661.508421] env[63515]: ERROR nova.compute.manager [instance: 7f49d79e-3bda-4949-a976-7e3e6513b2c7] result = function(*args, **kwargs) [ 661.508421] env[63515]: ERROR nova.compute.manager [instance: 7f49d79e-3bda-4949-a976-7e3e6513b2c7] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 661.508421] env[63515]: ERROR nova.compute.manager [instance: 7f49d79e-3bda-4949-a976-7e3e6513b2c7] return func(*args, **kwargs) [ 661.508421] env[63515]: ERROR nova.compute.manager [instance: 7f49d79e-3bda-4949-a976-7e3e6513b2c7] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 661.508421] env[63515]: ERROR nova.compute.manager [instance: 7f49d79e-3bda-4949-a976-7e3e6513b2c7] raise e [ 661.508421] env[63515]: ERROR nova.compute.manager [instance: 7f49d79e-3bda-4949-a976-7e3e6513b2c7] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 661.508421] env[63515]: ERROR nova.compute.manager [instance: 7f49d79e-3bda-4949-a976-7e3e6513b2c7] nwinfo = self.network_api.allocate_for_instance( [ 661.508421] env[63515]: ERROR nova.compute.manager [instance: 7f49d79e-3bda-4949-a976-7e3e6513b2c7] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 661.508421] env[63515]: ERROR nova.compute.manager [instance: 7f49d79e-3bda-4949-a976-7e3e6513b2c7] created_port_ids = self._update_ports_for_instance( [ 661.508421] env[63515]: ERROR nova.compute.manager [instance: 7f49d79e-3bda-4949-a976-7e3e6513b2c7] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 661.508421] env[63515]: ERROR nova.compute.manager [instance: 7f49d79e-3bda-4949-a976-7e3e6513b2c7] with excutils.save_and_reraise_exception(): [ 661.508421] env[63515]: ERROR nova.compute.manager [instance: 7f49d79e-3bda-4949-a976-7e3e6513b2c7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 661.508731] env[63515]: ERROR nova.compute.manager [instance: 7f49d79e-3bda-4949-a976-7e3e6513b2c7] self.force_reraise() [ 661.508731] env[63515]: ERROR nova.compute.manager [instance: 7f49d79e-3bda-4949-a976-7e3e6513b2c7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 661.508731] env[63515]: ERROR nova.compute.manager [instance: 7f49d79e-3bda-4949-a976-7e3e6513b2c7] raise self.value [ 661.508731] env[63515]: ERROR nova.compute.manager [instance: 7f49d79e-3bda-4949-a976-7e3e6513b2c7] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 661.508731] env[63515]: ERROR nova.compute.manager [instance: 7f49d79e-3bda-4949-a976-7e3e6513b2c7] updated_port = self._update_port( [ 661.508731] env[63515]: ERROR nova.compute.manager [instance: 7f49d79e-3bda-4949-a976-7e3e6513b2c7] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 661.508731] env[63515]: ERROR nova.compute.manager [instance: 7f49d79e-3bda-4949-a976-7e3e6513b2c7] _ensure_no_port_binding_failure(port) [ 661.508731] env[63515]: ERROR nova.compute.manager [instance: 7f49d79e-3bda-4949-a976-7e3e6513b2c7] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 661.508731] env[63515]: ERROR nova.compute.manager [instance: 7f49d79e-3bda-4949-a976-7e3e6513b2c7] raise exception.PortBindingFailed(port_id=port['id']) [ 661.508731] env[63515]: ERROR nova.compute.manager [instance: 7f49d79e-3bda-4949-a976-7e3e6513b2c7] nova.exception.PortBindingFailed: Binding failed for port cc888fc6-b889-48e3-bb61-bfa980ef91fd, please check neutron logs for more information. [ 661.508731] env[63515]: ERROR nova.compute.manager [instance: 7f49d79e-3bda-4949-a976-7e3e6513b2c7] [ 661.508731] env[63515]: INFO nova.compute.manager [None req-480b7d6d-2f7f-4da3-9eb1-818f0711a940 tempest-ServerRescueTestJSON-1961208277 tempest-ServerRescueTestJSON-1961208277-project-member] [instance: 7f49d79e-3bda-4949-a976-7e3e6513b2c7] Terminating instance [ 661.509918] env[63515]: DEBUG oslo_concurrency.lockutils [None req-480b7d6d-2f7f-4da3-9eb1-818f0711a940 tempest-ServerRescueTestJSON-1961208277 tempest-ServerRescueTestJSON-1961208277-project-member] Acquiring lock "refresh_cache-7f49d79e-3bda-4949-a976-7e3e6513b2c7" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 661.510090] env[63515]: DEBUG oslo_concurrency.lockutils [None req-480b7d6d-2f7f-4da3-9eb1-818f0711a940 tempest-ServerRescueTestJSON-1961208277 tempest-ServerRescueTestJSON-1961208277-project-member] Acquired lock "refresh_cache-7f49d79e-3bda-4949-a976-7e3e6513b2c7" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 661.510259] env[63515]: DEBUG nova.network.neutron [None req-480b7d6d-2f7f-4da3-9eb1-818f0711a940 tempest-ServerRescueTestJSON-1961208277 tempest-ServerRescueTestJSON-1961208277-project-member] [instance: 7f49d79e-3bda-4949-a976-7e3e6513b2c7] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 661.511554] env[63515]: DEBUG nova.network.neutron [-] [instance: bbce1137-d691-4633-87ff-f9f4ea257ed7] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 661.926804] env[63515]: DEBUG nova.network.neutron [None req-9131a685-703c-44a1-81f8-7c157d95e037 tempest-ServerAddressesNegativeTestJSON-1262066083 tempest-ServerAddressesNegativeTestJSON-1262066083-project-member] [instance: 5b201ab6-c7ec-4a5c-a310-982649f34be4] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 662.011926] env[63515]: DEBUG nova.network.neutron [None req-9131a685-703c-44a1-81f8-7c157d95e037 tempest-ServerAddressesNegativeTestJSON-1262066083 tempest-ServerAddressesNegativeTestJSON-1262066083-project-member] [instance: 5b201ab6-c7ec-4a5c-a310-982649f34be4] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 662.015943] env[63515]: INFO nova.compute.manager [-] [instance: bbce1137-d691-4633-87ff-f9f4ea257ed7] Took 1.03 seconds to deallocate network for instance. [ 662.017667] env[63515]: DEBUG nova.compute.claims [None req-30987b61-ce89-4676-89be-fee7786462c9 tempest-ImagesOneServerNegativeTestJSON-52225474 tempest-ImagesOneServerNegativeTestJSON-52225474-project-member] [instance: bbce1137-d691-4633-87ff-f9f4ea257ed7] Aborting claim: {{(pid=63515) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 662.017848] env[63515]: DEBUG oslo_concurrency.lockutils [None req-30987b61-ce89-4676-89be-fee7786462c9 tempest-ImagesOneServerNegativeTestJSON-52225474 tempest-ImagesOneServerNegativeTestJSON-52225474-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 662.029716] env[63515]: DEBUG nova.network.neutron [None req-480b7d6d-2f7f-4da3-9eb1-818f0711a940 tempest-ServerRescueTestJSON-1961208277 tempest-ServerRescueTestJSON-1961208277-project-member] [instance: 7f49d79e-3bda-4949-a976-7e3e6513b2c7] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 662.123225] env[63515]: DEBUG nova.network.neutron [None req-480b7d6d-2f7f-4da3-9eb1-818f0711a940 tempest-ServerRescueTestJSON-1961208277 tempest-ServerRescueTestJSON-1961208277-project-member] [instance: 7f49d79e-3bda-4949-a976-7e3e6513b2c7] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 662.517135] env[63515]: DEBUG oslo_concurrency.lockutils [None req-9131a685-703c-44a1-81f8-7c157d95e037 tempest-ServerAddressesNegativeTestJSON-1262066083 tempest-ServerAddressesNegativeTestJSON-1262066083-project-member] Releasing lock "refresh_cache-5b201ab6-c7ec-4a5c-a310-982649f34be4" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 662.517335] env[63515]: DEBUG nova.compute.manager [None req-9131a685-703c-44a1-81f8-7c157d95e037 tempest-ServerAddressesNegativeTestJSON-1262066083 tempest-ServerAddressesNegativeTestJSON-1262066083-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63515) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 662.517520] env[63515]: DEBUG nova.compute.manager [None req-9131a685-703c-44a1-81f8-7c157d95e037 tempest-ServerAddressesNegativeTestJSON-1262066083 tempest-ServerAddressesNegativeTestJSON-1262066083-project-member] [instance: 5b201ab6-c7ec-4a5c-a310-982649f34be4] Deallocating network for instance {{(pid=63515) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 662.517697] env[63515]: DEBUG nova.network.neutron [None req-9131a685-703c-44a1-81f8-7c157d95e037 tempest-ServerAddressesNegativeTestJSON-1262066083 tempest-ServerAddressesNegativeTestJSON-1262066083-project-member] [instance: 5b201ab6-c7ec-4a5c-a310-982649f34be4] deallocate_for_instance() {{(pid=63515) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 662.533334] env[63515]: DEBUG nova.network.neutron [None req-9131a685-703c-44a1-81f8-7c157d95e037 tempest-ServerAddressesNegativeTestJSON-1262066083 tempest-ServerAddressesNegativeTestJSON-1262066083-project-member] [instance: 5b201ab6-c7ec-4a5c-a310-982649f34be4] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 662.626965] env[63515]: DEBUG oslo_concurrency.lockutils [None req-480b7d6d-2f7f-4da3-9eb1-818f0711a940 tempest-ServerRescueTestJSON-1961208277 tempest-ServerRescueTestJSON-1961208277-project-member] Releasing lock "refresh_cache-7f49d79e-3bda-4949-a976-7e3e6513b2c7" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 662.627431] env[63515]: DEBUG nova.compute.manager [None req-480b7d6d-2f7f-4da3-9eb1-818f0711a940 tempest-ServerRescueTestJSON-1961208277 tempest-ServerRescueTestJSON-1961208277-project-member] [instance: 7f49d79e-3bda-4949-a976-7e3e6513b2c7] Start destroying the instance on the hypervisor. {{(pid=63515) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 662.627627] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-480b7d6d-2f7f-4da3-9eb1-818f0711a940 tempest-ServerRescueTestJSON-1961208277 tempest-ServerRescueTestJSON-1961208277-project-member] [instance: 7f49d79e-3bda-4949-a976-7e3e6513b2c7] Destroying instance {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 662.627907] env[63515]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d07c4951-3949-4a19-bf8f-d3f0d388f2b7 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.640885] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-523e2c6d-be7c-4918-81d6-0a2de9750dca {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.662016] env[63515]: WARNING nova.virt.vmwareapi.vmops [None req-480b7d6d-2f7f-4da3-9eb1-818f0711a940 tempest-ServerRescueTestJSON-1961208277 tempest-ServerRescueTestJSON-1961208277-project-member] [instance: 7f49d79e-3bda-4949-a976-7e3e6513b2c7] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 7f49d79e-3bda-4949-a976-7e3e6513b2c7 could not be found. [ 662.662277] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-480b7d6d-2f7f-4da3-9eb1-818f0711a940 tempest-ServerRescueTestJSON-1961208277 tempest-ServerRescueTestJSON-1961208277-project-member] [instance: 7f49d79e-3bda-4949-a976-7e3e6513b2c7] Instance destroyed {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 662.662467] env[63515]: INFO nova.compute.manager [None req-480b7d6d-2f7f-4da3-9eb1-818f0711a940 tempest-ServerRescueTestJSON-1961208277 tempest-ServerRescueTestJSON-1961208277-project-member] [instance: 7f49d79e-3bda-4949-a976-7e3e6513b2c7] Took 0.03 seconds to destroy the instance on the hypervisor. [ 662.662698] env[63515]: DEBUG oslo.service.loopingcall [None req-480b7d6d-2f7f-4da3-9eb1-818f0711a940 tempest-ServerRescueTestJSON-1961208277 tempest-ServerRescueTestJSON-1961208277-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 662.665168] env[63515]: DEBUG nova.compute.manager [-] [instance: 7f49d79e-3bda-4949-a976-7e3e6513b2c7] Deallocating network for instance {{(pid=63515) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 662.665273] env[63515]: DEBUG nova.network.neutron [-] [instance: 7f49d79e-3bda-4949-a976-7e3e6513b2c7] deallocate_for_instance() {{(pid=63515) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 662.681146] env[63515]: DEBUG nova.network.neutron [-] [instance: 7f49d79e-3bda-4949-a976-7e3e6513b2c7] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 662.710271] env[63515]: DEBUG nova.compute.manager [req-dc96c3f8-c1f7-4045-bba4-ec34e71cbac9 req-51f1a9ae-8867-4b47-b575-1c261bf04106 service nova] [instance: 7f49d79e-3bda-4949-a976-7e3e6513b2c7] Received event network-changed-cc888fc6-b889-48e3-bb61-bfa980ef91fd {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 662.710271] env[63515]: DEBUG nova.compute.manager [req-dc96c3f8-c1f7-4045-bba4-ec34e71cbac9 req-51f1a9ae-8867-4b47-b575-1c261bf04106 service nova] [instance: 7f49d79e-3bda-4949-a976-7e3e6513b2c7] Refreshing instance network info cache due to event network-changed-cc888fc6-b889-48e3-bb61-bfa980ef91fd. {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 662.710271] env[63515]: DEBUG oslo_concurrency.lockutils [req-dc96c3f8-c1f7-4045-bba4-ec34e71cbac9 req-51f1a9ae-8867-4b47-b575-1c261bf04106 service nova] Acquiring lock "refresh_cache-7f49d79e-3bda-4949-a976-7e3e6513b2c7" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 662.710271] env[63515]: DEBUG oslo_concurrency.lockutils [req-dc96c3f8-c1f7-4045-bba4-ec34e71cbac9 req-51f1a9ae-8867-4b47-b575-1c261bf04106 service nova] Acquired lock "refresh_cache-7f49d79e-3bda-4949-a976-7e3e6513b2c7" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 662.710271] env[63515]: DEBUG nova.network.neutron [req-dc96c3f8-c1f7-4045-bba4-ec34e71cbac9 req-51f1a9ae-8867-4b47-b575-1c261bf04106 service nova] [instance: 7f49d79e-3bda-4949-a976-7e3e6513b2c7] Refreshing network info cache for port cc888fc6-b889-48e3-bb61-bfa980ef91fd {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 662.789661] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d78f7cf6-7471-4893-81de-a5723cb542af {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.797471] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-efe8844f-6d97-4032-8e79-6d61c8568ab2 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.827909] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f97e07c-39d5-4a7a-a8c6-372b88a41a2e {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.835139] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26e37d67-3854-4ef8-9554-050e3278b0b3 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.848852] env[63515]: DEBUG nova.compute.provider_tree [None req-b9c961a0-a559-4690-abb6-579359ed447c tempest-ServerPasswordTestJSON-433685770 tempest-ServerPasswordTestJSON-433685770-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 663.037047] env[63515]: DEBUG nova.network.neutron [None req-9131a685-703c-44a1-81f8-7c157d95e037 tempest-ServerAddressesNegativeTestJSON-1262066083 tempest-ServerAddressesNegativeTestJSON-1262066083-project-member] [instance: 5b201ab6-c7ec-4a5c-a310-982649f34be4] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 663.184069] env[63515]: DEBUG nova.network.neutron [-] [instance: 7f49d79e-3bda-4949-a976-7e3e6513b2c7] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 663.229084] env[63515]: DEBUG nova.network.neutron [req-dc96c3f8-c1f7-4045-bba4-ec34e71cbac9 req-51f1a9ae-8867-4b47-b575-1c261bf04106 service nova] [instance: 7f49d79e-3bda-4949-a976-7e3e6513b2c7] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 663.306133] env[63515]: DEBUG nova.network.neutron [req-dc96c3f8-c1f7-4045-bba4-ec34e71cbac9 req-51f1a9ae-8867-4b47-b575-1c261bf04106 service nova] [instance: 7f49d79e-3bda-4949-a976-7e3e6513b2c7] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 663.355170] env[63515]: DEBUG nova.scheduler.client.report [None req-b9c961a0-a559-4690-abb6-579359ed447c tempest-ServerPasswordTestJSON-433685770 tempest-ServerPasswordTestJSON-433685770-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 663.538534] env[63515]: INFO nova.compute.manager [None req-9131a685-703c-44a1-81f8-7c157d95e037 tempest-ServerAddressesNegativeTestJSON-1262066083 tempest-ServerAddressesNegativeTestJSON-1262066083-project-member] [instance: 5b201ab6-c7ec-4a5c-a310-982649f34be4] Took 1.02 seconds to deallocate network for instance. [ 663.686899] env[63515]: INFO nova.compute.manager [-] [instance: 7f49d79e-3bda-4949-a976-7e3e6513b2c7] Took 1.02 seconds to deallocate network for instance. [ 663.689637] env[63515]: DEBUG nova.compute.claims [None req-480b7d6d-2f7f-4da3-9eb1-818f0711a940 tempest-ServerRescueTestJSON-1961208277 tempest-ServerRescueTestJSON-1961208277-project-member] [instance: 7f49d79e-3bda-4949-a976-7e3e6513b2c7] Aborting claim: {{(pid=63515) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 663.689637] env[63515]: DEBUG oslo_concurrency.lockutils [None req-480b7d6d-2f7f-4da3-9eb1-818f0711a940 tempest-ServerRescueTestJSON-1961208277 tempest-ServerRescueTestJSON-1961208277-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 663.807566] env[63515]: DEBUG oslo_concurrency.lockutils [req-dc96c3f8-c1f7-4045-bba4-ec34e71cbac9 req-51f1a9ae-8867-4b47-b575-1c261bf04106 service nova] Releasing lock "refresh_cache-7f49d79e-3bda-4949-a976-7e3e6513b2c7" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 663.807833] env[63515]: DEBUG nova.compute.manager [req-dc96c3f8-c1f7-4045-bba4-ec34e71cbac9 req-51f1a9ae-8867-4b47-b575-1c261bf04106 service nova] [instance: 7f49d79e-3bda-4949-a976-7e3e6513b2c7] Received event network-vif-deleted-cc888fc6-b889-48e3-bb61-bfa980ef91fd {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 663.861460] env[63515]: DEBUG oslo_concurrency.lockutils [None req-b9c961a0-a559-4690-abb6-579359ed447c tempest-ServerPasswordTestJSON-433685770 tempest-ServerPasswordTestJSON-433685770-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.449s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 663.861988] env[63515]: DEBUG nova.compute.manager [None req-b9c961a0-a559-4690-abb6-579359ed447c tempest-ServerPasswordTestJSON-433685770 tempest-ServerPasswordTestJSON-433685770-project-member] [instance: 07499601-62bc-4c31-b295-23f34a6e2e91] Start building networks asynchronously for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 663.864548] env[63515]: DEBUG oslo_concurrency.lockutils [None req-292f2bc3-2911-4498-95fb-4e4a03681601 tempest-FloatingIPsAssociationTestJSON-629096942 tempest-FloatingIPsAssociationTestJSON-629096942-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 31.310s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 664.369375] env[63515]: DEBUG nova.compute.utils [None req-b9c961a0-a559-4690-abb6-579359ed447c tempest-ServerPasswordTestJSON-433685770 tempest-ServerPasswordTestJSON-433685770-project-member] Using /dev/sd instead of None {{(pid=63515) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 664.373795] env[63515]: DEBUG nova.compute.manager [None req-b9c961a0-a559-4690-abb6-579359ed447c tempest-ServerPasswordTestJSON-433685770 tempest-ServerPasswordTestJSON-433685770-project-member] [instance: 07499601-62bc-4c31-b295-23f34a6e2e91] Allocating IP information in the background. {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 664.373967] env[63515]: DEBUG nova.network.neutron [None req-b9c961a0-a559-4690-abb6-579359ed447c tempest-ServerPasswordTestJSON-433685770 tempest-ServerPasswordTestJSON-433685770-project-member] [instance: 07499601-62bc-4c31-b295-23f34a6e2e91] allocate_for_instance() {{(pid=63515) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 664.421617] env[63515]: DEBUG nova.policy [None req-b9c961a0-a559-4690-abb6-579359ed447c tempest-ServerPasswordTestJSON-433685770 tempest-ServerPasswordTestJSON-433685770-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '265f012ea11d41eb868bdebf800f3763', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4d0cac1cb20f409eae261e91146ade77', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63515) authorize /opt/stack/nova/nova/policy.py:201}} [ 664.568926] env[63515]: INFO nova.scheduler.client.report [None req-9131a685-703c-44a1-81f8-7c157d95e037 tempest-ServerAddressesNegativeTestJSON-1262066083 tempest-ServerAddressesNegativeTestJSON-1262066083-project-member] Deleted allocations for instance 5b201ab6-c7ec-4a5c-a310-982649f34be4 [ 664.733827] env[63515]: DEBUG nova.network.neutron [None req-b9c961a0-a559-4690-abb6-579359ed447c tempest-ServerPasswordTestJSON-433685770 tempest-ServerPasswordTestJSON-433685770-project-member] [instance: 07499601-62bc-4c31-b295-23f34a6e2e91] Successfully created port: b497be91-987b-42ce-9f91-0850d502656f {{(pid=63515) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 664.875052] env[63515]: DEBUG nova.compute.manager [None req-b9c961a0-a559-4690-abb6-579359ed447c tempest-ServerPasswordTestJSON-433685770 tempest-ServerPasswordTestJSON-433685770-project-member] [instance: 07499601-62bc-4c31-b295-23f34a6e2e91] Start building block device mappings for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 664.901979] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b40c2b2-9cd4-4929-ae0e-2323e2a2db13 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 664.909840] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed2241e3-102e-4778-994a-083229084809 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 664.942539] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4765e73c-9b3a-4349-8259-18e8e6824252 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 664.950133] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c73a58a5-e36b-4e04-b12b-0bead48d70e9 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 664.963834] env[63515]: DEBUG nova.compute.provider_tree [None req-292f2bc3-2911-4498-95fb-4e4a03681601 tempest-FloatingIPsAssociationTestJSON-629096942 tempest-FloatingIPsAssociationTestJSON-629096942-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 665.080945] env[63515]: DEBUG oslo_concurrency.lockutils [None req-9131a685-703c-44a1-81f8-7c157d95e037 tempest-ServerAddressesNegativeTestJSON-1262066083 tempest-ServerAddressesNegativeTestJSON-1262066083-project-member] Lock "5b201ab6-c7ec-4a5c-a310-982649f34be4" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 107.986s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 665.466363] env[63515]: DEBUG nova.scheduler.client.report [None req-292f2bc3-2911-4498-95fb-4e4a03681601 tempest-FloatingIPsAssociationTestJSON-629096942 tempest-FloatingIPsAssociationTestJSON-629096942-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 665.582338] env[63515]: DEBUG nova.compute.manager [None req-95251dc9-cf22-4a89-a3b0-fce88f9e12db tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: 3ad04a9f-62cf-4b1c-aab2-7dc4a219edc6] Starting instance... {{(pid=63515) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 665.707177] env[63515]: DEBUG nova.compute.manager [req-8bcad23d-0438-437e-bb73-bc9bea4a97de req-f82f1131-a500-4e3e-a1d6-c22d82a0e2db service nova] [instance: 07499601-62bc-4c31-b295-23f34a6e2e91] Received event network-changed-b497be91-987b-42ce-9f91-0850d502656f {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 665.707177] env[63515]: DEBUG nova.compute.manager [req-8bcad23d-0438-437e-bb73-bc9bea4a97de req-f82f1131-a500-4e3e-a1d6-c22d82a0e2db service nova] [instance: 07499601-62bc-4c31-b295-23f34a6e2e91] Refreshing instance network info cache due to event network-changed-b497be91-987b-42ce-9f91-0850d502656f. {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 665.708299] env[63515]: DEBUG oslo_concurrency.lockutils [req-8bcad23d-0438-437e-bb73-bc9bea4a97de req-f82f1131-a500-4e3e-a1d6-c22d82a0e2db service nova] Acquiring lock "refresh_cache-07499601-62bc-4c31-b295-23f34a6e2e91" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 665.708710] env[63515]: DEBUG oslo_concurrency.lockutils [req-8bcad23d-0438-437e-bb73-bc9bea4a97de req-f82f1131-a500-4e3e-a1d6-c22d82a0e2db service nova] Acquired lock "refresh_cache-07499601-62bc-4c31-b295-23f34a6e2e91" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 665.709038] env[63515]: DEBUG nova.network.neutron [req-8bcad23d-0438-437e-bb73-bc9bea4a97de req-f82f1131-a500-4e3e-a1d6-c22d82a0e2db service nova] [instance: 07499601-62bc-4c31-b295-23f34a6e2e91] Refreshing network info cache for port b497be91-987b-42ce-9f91-0850d502656f {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 665.886988] env[63515]: DEBUG nova.compute.manager [None req-b9c961a0-a559-4690-abb6-579359ed447c tempest-ServerPasswordTestJSON-433685770 tempest-ServerPasswordTestJSON-433685770-project-member] [instance: 07499601-62bc-4c31-b295-23f34a6e2e91] Start spawning the instance on the hypervisor. {{(pid=63515) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 665.913181] env[63515]: ERROR nova.compute.manager [None req-b9c961a0-a559-4690-abb6-579359ed447c tempest-ServerPasswordTestJSON-433685770 tempest-ServerPasswordTestJSON-433685770-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port b497be91-987b-42ce-9f91-0850d502656f, please check neutron logs for more information. [ 665.913181] env[63515]: ERROR nova.compute.manager Traceback (most recent call last): [ 665.913181] env[63515]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 665.913181] env[63515]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 665.913181] env[63515]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 665.913181] env[63515]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 665.913181] env[63515]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 665.913181] env[63515]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 665.913181] env[63515]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 665.913181] env[63515]: ERROR nova.compute.manager self.force_reraise() [ 665.913181] env[63515]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 665.913181] env[63515]: ERROR nova.compute.manager raise self.value [ 665.913181] env[63515]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 665.913181] env[63515]: ERROR nova.compute.manager updated_port = self._update_port( [ 665.913181] env[63515]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 665.913181] env[63515]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 665.914143] env[63515]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 665.914143] env[63515]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 665.914143] env[63515]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port b497be91-987b-42ce-9f91-0850d502656f, please check neutron logs for more information. [ 665.914143] env[63515]: ERROR nova.compute.manager [ 665.914143] env[63515]: Traceback (most recent call last): [ 665.914143] env[63515]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 665.914143] env[63515]: listener.cb(fileno) [ 665.914143] env[63515]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 665.914143] env[63515]: result = function(*args, **kwargs) [ 665.914143] env[63515]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 665.914143] env[63515]: return func(*args, **kwargs) [ 665.914143] env[63515]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 665.914143] env[63515]: raise e [ 665.914143] env[63515]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 665.914143] env[63515]: nwinfo = self.network_api.allocate_for_instance( [ 665.914143] env[63515]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 665.914143] env[63515]: created_port_ids = self._update_ports_for_instance( [ 665.914143] env[63515]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 665.914143] env[63515]: with excutils.save_and_reraise_exception(): [ 665.914143] env[63515]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 665.914143] env[63515]: self.force_reraise() [ 665.914143] env[63515]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 665.914143] env[63515]: raise self.value [ 665.914143] env[63515]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 665.914143] env[63515]: updated_port = self._update_port( [ 665.914143] env[63515]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 665.914143] env[63515]: _ensure_no_port_binding_failure(port) [ 665.914143] env[63515]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 665.914143] env[63515]: raise exception.PortBindingFailed(port_id=port['id']) [ 665.915130] env[63515]: nova.exception.PortBindingFailed: Binding failed for port b497be91-987b-42ce-9f91-0850d502656f, please check neutron logs for more information. [ 665.915130] env[63515]: Removing descriptor: 17 [ 665.918154] env[63515]: DEBUG nova.virt.hardware [None req-b9c961a0-a559-4690-abb6-579359ed447c tempest-ServerPasswordTestJSON-433685770 tempest-ServerPasswordTestJSON-433685770-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T02:52:14Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T02:51:56Z,direct_url=,disk_format='vmdk',id=8a120570-cb06-4099-b262-554ca0ad15c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b569255cc43e42e7ae3f2b4ad37c6ef4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T02:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 665.918530] env[63515]: DEBUG nova.virt.hardware [None req-b9c961a0-a559-4690-abb6-579359ed447c tempest-ServerPasswordTestJSON-433685770 tempest-ServerPasswordTestJSON-433685770-project-member] Flavor limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 665.919033] env[63515]: DEBUG nova.virt.hardware [None req-b9c961a0-a559-4690-abb6-579359ed447c tempest-ServerPasswordTestJSON-433685770 tempest-ServerPasswordTestJSON-433685770-project-member] Image limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 665.919111] env[63515]: DEBUG nova.virt.hardware [None req-b9c961a0-a559-4690-abb6-579359ed447c tempest-ServerPasswordTestJSON-433685770 tempest-ServerPasswordTestJSON-433685770-project-member] Flavor pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 665.919241] env[63515]: DEBUG nova.virt.hardware [None req-b9c961a0-a559-4690-abb6-579359ed447c tempest-ServerPasswordTestJSON-433685770 tempest-ServerPasswordTestJSON-433685770-project-member] Image pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 665.922943] env[63515]: DEBUG nova.virt.hardware [None req-b9c961a0-a559-4690-abb6-579359ed447c tempest-ServerPasswordTestJSON-433685770 tempest-ServerPasswordTestJSON-433685770-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 665.926106] env[63515]: DEBUG nova.virt.hardware [None req-b9c961a0-a559-4690-abb6-579359ed447c tempest-ServerPasswordTestJSON-433685770 tempest-ServerPasswordTestJSON-433685770-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 665.926106] env[63515]: DEBUG nova.virt.hardware [None req-b9c961a0-a559-4690-abb6-579359ed447c tempest-ServerPasswordTestJSON-433685770 tempest-ServerPasswordTestJSON-433685770-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 665.926106] env[63515]: DEBUG nova.virt.hardware [None req-b9c961a0-a559-4690-abb6-579359ed447c tempest-ServerPasswordTestJSON-433685770 tempest-ServerPasswordTestJSON-433685770-project-member] Got 1 possible topologies {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 665.926106] env[63515]: DEBUG nova.virt.hardware [None req-b9c961a0-a559-4690-abb6-579359ed447c tempest-ServerPasswordTestJSON-433685770 tempest-ServerPasswordTestJSON-433685770-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 665.926106] env[63515]: DEBUG nova.virt.hardware [None req-b9c961a0-a559-4690-abb6-579359ed447c tempest-ServerPasswordTestJSON-433685770 tempest-ServerPasswordTestJSON-433685770-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 665.926281] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7458fca7-bf00-45f1-9770-2c7a89d98991 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 665.935186] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b29d5bab-b826-44ce-8f93-4db53f6ec2d2 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 665.949355] env[63515]: ERROR nova.compute.manager [None req-b9c961a0-a559-4690-abb6-579359ed447c tempest-ServerPasswordTestJSON-433685770 tempest-ServerPasswordTestJSON-433685770-project-member] [instance: 07499601-62bc-4c31-b295-23f34a6e2e91] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port b497be91-987b-42ce-9f91-0850d502656f, please check neutron logs for more information. [ 665.949355] env[63515]: ERROR nova.compute.manager [instance: 07499601-62bc-4c31-b295-23f34a6e2e91] Traceback (most recent call last): [ 665.949355] env[63515]: ERROR nova.compute.manager [instance: 07499601-62bc-4c31-b295-23f34a6e2e91] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 665.949355] env[63515]: ERROR nova.compute.manager [instance: 07499601-62bc-4c31-b295-23f34a6e2e91] yield resources [ 665.949355] env[63515]: ERROR nova.compute.manager [instance: 07499601-62bc-4c31-b295-23f34a6e2e91] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 665.949355] env[63515]: ERROR nova.compute.manager [instance: 07499601-62bc-4c31-b295-23f34a6e2e91] self.driver.spawn(context, instance, image_meta, [ 665.949355] env[63515]: ERROR nova.compute.manager [instance: 07499601-62bc-4c31-b295-23f34a6e2e91] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 665.949355] env[63515]: ERROR nova.compute.manager [instance: 07499601-62bc-4c31-b295-23f34a6e2e91] self._vmops.spawn(context, instance, image_meta, injected_files, [ 665.949355] env[63515]: ERROR nova.compute.manager [instance: 07499601-62bc-4c31-b295-23f34a6e2e91] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 665.949355] env[63515]: ERROR nova.compute.manager [instance: 07499601-62bc-4c31-b295-23f34a6e2e91] vm_ref = self.build_virtual_machine(instance, [ 665.949355] env[63515]: ERROR nova.compute.manager [instance: 07499601-62bc-4c31-b295-23f34a6e2e91] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 665.949812] env[63515]: ERROR nova.compute.manager [instance: 07499601-62bc-4c31-b295-23f34a6e2e91] vif_infos = vmwarevif.get_vif_info(self._session, [ 665.949812] env[63515]: ERROR nova.compute.manager [instance: 07499601-62bc-4c31-b295-23f34a6e2e91] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 665.949812] env[63515]: ERROR nova.compute.manager [instance: 07499601-62bc-4c31-b295-23f34a6e2e91] for vif in network_info: [ 665.949812] env[63515]: ERROR nova.compute.manager [instance: 07499601-62bc-4c31-b295-23f34a6e2e91] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 665.949812] env[63515]: ERROR nova.compute.manager [instance: 07499601-62bc-4c31-b295-23f34a6e2e91] return self._sync_wrapper(fn, *args, **kwargs) [ 665.949812] env[63515]: ERROR nova.compute.manager [instance: 07499601-62bc-4c31-b295-23f34a6e2e91] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 665.949812] env[63515]: ERROR nova.compute.manager [instance: 07499601-62bc-4c31-b295-23f34a6e2e91] self.wait() [ 665.949812] env[63515]: ERROR nova.compute.manager [instance: 07499601-62bc-4c31-b295-23f34a6e2e91] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 665.949812] env[63515]: ERROR nova.compute.manager [instance: 07499601-62bc-4c31-b295-23f34a6e2e91] self[:] = self._gt.wait() [ 665.949812] env[63515]: ERROR nova.compute.manager [instance: 07499601-62bc-4c31-b295-23f34a6e2e91] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 665.949812] env[63515]: ERROR nova.compute.manager [instance: 07499601-62bc-4c31-b295-23f34a6e2e91] return self._exit_event.wait() [ 665.949812] env[63515]: ERROR nova.compute.manager [instance: 07499601-62bc-4c31-b295-23f34a6e2e91] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 665.949812] env[63515]: ERROR nova.compute.manager [instance: 07499601-62bc-4c31-b295-23f34a6e2e91] current.throw(*self._exc) [ 665.950156] env[63515]: ERROR nova.compute.manager [instance: 07499601-62bc-4c31-b295-23f34a6e2e91] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 665.950156] env[63515]: ERROR nova.compute.manager [instance: 07499601-62bc-4c31-b295-23f34a6e2e91] result = function(*args, **kwargs) [ 665.950156] env[63515]: ERROR nova.compute.manager [instance: 07499601-62bc-4c31-b295-23f34a6e2e91] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 665.950156] env[63515]: ERROR nova.compute.manager [instance: 07499601-62bc-4c31-b295-23f34a6e2e91] return func(*args, **kwargs) [ 665.950156] env[63515]: ERROR nova.compute.manager [instance: 07499601-62bc-4c31-b295-23f34a6e2e91] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 665.950156] env[63515]: ERROR nova.compute.manager [instance: 07499601-62bc-4c31-b295-23f34a6e2e91] raise e [ 665.950156] env[63515]: ERROR nova.compute.manager [instance: 07499601-62bc-4c31-b295-23f34a6e2e91] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 665.950156] env[63515]: ERROR nova.compute.manager [instance: 07499601-62bc-4c31-b295-23f34a6e2e91] nwinfo = self.network_api.allocate_for_instance( [ 665.950156] env[63515]: ERROR nova.compute.manager [instance: 07499601-62bc-4c31-b295-23f34a6e2e91] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 665.950156] env[63515]: ERROR nova.compute.manager [instance: 07499601-62bc-4c31-b295-23f34a6e2e91] created_port_ids = self._update_ports_for_instance( [ 665.950156] env[63515]: ERROR nova.compute.manager [instance: 07499601-62bc-4c31-b295-23f34a6e2e91] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 665.950156] env[63515]: ERROR nova.compute.manager [instance: 07499601-62bc-4c31-b295-23f34a6e2e91] with excutils.save_and_reraise_exception(): [ 665.950156] env[63515]: ERROR nova.compute.manager [instance: 07499601-62bc-4c31-b295-23f34a6e2e91] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 665.950482] env[63515]: ERROR nova.compute.manager [instance: 07499601-62bc-4c31-b295-23f34a6e2e91] self.force_reraise() [ 665.950482] env[63515]: ERROR nova.compute.manager [instance: 07499601-62bc-4c31-b295-23f34a6e2e91] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 665.950482] env[63515]: ERROR nova.compute.manager [instance: 07499601-62bc-4c31-b295-23f34a6e2e91] raise self.value [ 665.950482] env[63515]: ERROR nova.compute.manager [instance: 07499601-62bc-4c31-b295-23f34a6e2e91] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 665.950482] env[63515]: ERROR nova.compute.manager [instance: 07499601-62bc-4c31-b295-23f34a6e2e91] updated_port = self._update_port( [ 665.950482] env[63515]: ERROR nova.compute.manager [instance: 07499601-62bc-4c31-b295-23f34a6e2e91] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 665.950482] env[63515]: ERROR nova.compute.manager [instance: 07499601-62bc-4c31-b295-23f34a6e2e91] _ensure_no_port_binding_failure(port) [ 665.950482] env[63515]: ERROR nova.compute.manager [instance: 07499601-62bc-4c31-b295-23f34a6e2e91] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 665.950482] env[63515]: ERROR nova.compute.manager [instance: 07499601-62bc-4c31-b295-23f34a6e2e91] raise exception.PortBindingFailed(port_id=port['id']) [ 665.950482] env[63515]: ERROR nova.compute.manager [instance: 07499601-62bc-4c31-b295-23f34a6e2e91] nova.exception.PortBindingFailed: Binding failed for port b497be91-987b-42ce-9f91-0850d502656f, please check neutron logs for more information. [ 665.950482] env[63515]: ERROR nova.compute.manager [instance: 07499601-62bc-4c31-b295-23f34a6e2e91] [ 665.950482] env[63515]: INFO nova.compute.manager [None req-b9c961a0-a559-4690-abb6-579359ed447c tempest-ServerPasswordTestJSON-433685770 tempest-ServerPasswordTestJSON-433685770-project-member] [instance: 07499601-62bc-4c31-b295-23f34a6e2e91] Terminating instance [ 665.951620] env[63515]: DEBUG oslo_concurrency.lockutils [None req-b9c961a0-a559-4690-abb6-579359ed447c tempest-ServerPasswordTestJSON-433685770 tempest-ServerPasswordTestJSON-433685770-project-member] Acquiring lock "refresh_cache-07499601-62bc-4c31-b295-23f34a6e2e91" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 665.971472] env[63515]: DEBUG oslo_concurrency.lockutils [None req-292f2bc3-2911-4498-95fb-4e4a03681601 tempest-FloatingIPsAssociationTestJSON-629096942 tempest-FloatingIPsAssociationTestJSON-629096942-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.107s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 665.972113] env[63515]: ERROR nova.compute.manager [None req-292f2bc3-2911-4498-95fb-4e4a03681601 tempest-FloatingIPsAssociationTestJSON-629096942 tempest-FloatingIPsAssociationTestJSON-629096942-project-member] [instance: 2728c9f9-0d03-4313-ba2a-3dc22aff8538] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port bfaff379-5880-4a38-86aa-c5d6331eb307, please check neutron logs for more information. [ 665.972113] env[63515]: ERROR nova.compute.manager [instance: 2728c9f9-0d03-4313-ba2a-3dc22aff8538] Traceback (most recent call last): [ 665.972113] env[63515]: ERROR nova.compute.manager [instance: 2728c9f9-0d03-4313-ba2a-3dc22aff8538] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 665.972113] env[63515]: ERROR nova.compute.manager [instance: 2728c9f9-0d03-4313-ba2a-3dc22aff8538] self.driver.spawn(context, instance, image_meta, [ 665.972113] env[63515]: ERROR nova.compute.manager [instance: 2728c9f9-0d03-4313-ba2a-3dc22aff8538] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 665.972113] env[63515]: ERROR nova.compute.manager [instance: 2728c9f9-0d03-4313-ba2a-3dc22aff8538] self._vmops.spawn(context, instance, image_meta, injected_files, [ 665.972113] env[63515]: ERROR nova.compute.manager [instance: 2728c9f9-0d03-4313-ba2a-3dc22aff8538] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 665.972113] env[63515]: ERROR nova.compute.manager [instance: 2728c9f9-0d03-4313-ba2a-3dc22aff8538] vm_ref = self.build_virtual_machine(instance, [ 665.972113] env[63515]: ERROR nova.compute.manager [instance: 2728c9f9-0d03-4313-ba2a-3dc22aff8538] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 665.972113] env[63515]: ERROR nova.compute.manager [instance: 2728c9f9-0d03-4313-ba2a-3dc22aff8538] vif_infos = vmwarevif.get_vif_info(self._session, [ 665.972113] env[63515]: ERROR nova.compute.manager [instance: 2728c9f9-0d03-4313-ba2a-3dc22aff8538] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 665.972401] env[63515]: ERROR nova.compute.manager [instance: 2728c9f9-0d03-4313-ba2a-3dc22aff8538] for vif in network_info: [ 665.972401] env[63515]: ERROR nova.compute.manager [instance: 2728c9f9-0d03-4313-ba2a-3dc22aff8538] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 665.972401] env[63515]: ERROR nova.compute.manager [instance: 2728c9f9-0d03-4313-ba2a-3dc22aff8538] return self._sync_wrapper(fn, *args, **kwargs) [ 665.972401] env[63515]: ERROR nova.compute.manager [instance: 2728c9f9-0d03-4313-ba2a-3dc22aff8538] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 665.972401] env[63515]: ERROR nova.compute.manager [instance: 2728c9f9-0d03-4313-ba2a-3dc22aff8538] self.wait() [ 665.972401] env[63515]: ERROR nova.compute.manager [instance: 2728c9f9-0d03-4313-ba2a-3dc22aff8538] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 665.972401] env[63515]: ERROR nova.compute.manager [instance: 2728c9f9-0d03-4313-ba2a-3dc22aff8538] self[:] = self._gt.wait() [ 665.972401] env[63515]: ERROR nova.compute.manager [instance: 2728c9f9-0d03-4313-ba2a-3dc22aff8538] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 665.972401] env[63515]: ERROR nova.compute.manager [instance: 2728c9f9-0d03-4313-ba2a-3dc22aff8538] return self._exit_event.wait() [ 665.972401] env[63515]: ERROR nova.compute.manager [instance: 2728c9f9-0d03-4313-ba2a-3dc22aff8538] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 665.972401] env[63515]: ERROR nova.compute.manager [instance: 2728c9f9-0d03-4313-ba2a-3dc22aff8538] result = hub.switch() [ 665.972401] env[63515]: ERROR nova.compute.manager [instance: 2728c9f9-0d03-4313-ba2a-3dc22aff8538] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 665.972401] env[63515]: ERROR nova.compute.manager [instance: 2728c9f9-0d03-4313-ba2a-3dc22aff8538] return self.greenlet.switch() [ 665.972705] env[63515]: ERROR nova.compute.manager [instance: 2728c9f9-0d03-4313-ba2a-3dc22aff8538] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 665.972705] env[63515]: ERROR nova.compute.manager [instance: 2728c9f9-0d03-4313-ba2a-3dc22aff8538] result = function(*args, **kwargs) [ 665.972705] env[63515]: ERROR nova.compute.manager [instance: 2728c9f9-0d03-4313-ba2a-3dc22aff8538] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 665.972705] env[63515]: ERROR nova.compute.manager [instance: 2728c9f9-0d03-4313-ba2a-3dc22aff8538] return func(*args, **kwargs) [ 665.972705] env[63515]: ERROR nova.compute.manager [instance: 2728c9f9-0d03-4313-ba2a-3dc22aff8538] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 665.972705] env[63515]: ERROR nova.compute.manager [instance: 2728c9f9-0d03-4313-ba2a-3dc22aff8538] raise e [ 665.972705] env[63515]: ERROR nova.compute.manager [instance: 2728c9f9-0d03-4313-ba2a-3dc22aff8538] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 665.972705] env[63515]: ERROR nova.compute.manager [instance: 2728c9f9-0d03-4313-ba2a-3dc22aff8538] nwinfo = self.network_api.allocate_for_instance( [ 665.972705] env[63515]: ERROR nova.compute.manager [instance: 2728c9f9-0d03-4313-ba2a-3dc22aff8538] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 665.972705] env[63515]: ERROR nova.compute.manager [instance: 2728c9f9-0d03-4313-ba2a-3dc22aff8538] created_port_ids = self._update_ports_for_instance( [ 665.972705] env[63515]: ERROR nova.compute.manager [instance: 2728c9f9-0d03-4313-ba2a-3dc22aff8538] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 665.972705] env[63515]: ERROR nova.compute.manager [instance: 2728c9f9-0d03-4313-ba2a-3dc22aff8538] with excutils.save_and_reraise_exception(): [ 665.972705] env[63515]: ERROR nova.compute.manager [instance: 2728c9f9-0d03-4313-ba2a-3dc22aff8538] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 665.973021] env[63515]: ERROR nova.compute.manager [instance: 2728c9f9-0d03-4313-ba2a-3dc22aff8538] self.force_reraise() [ 665.973021] env[63515]: ERROR nova.compute.manager [instance: 2728c9f9-0d03-4313-ba2a-3dc22aff8538] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 665.973021] env[63515]: ERROR nova.compute.manager [instance: 2728c9f9-0d03-4313-ba2a-3dc22aff8538] raise self.value [ 665.973021] env[63515]: ERROR nova.compute.manager [instance: 2728c9f9-0d03-4313-ba2a-3dc22aff8538] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 665.973021] env[63515]: ERROR nova.compute.manager [instance: 2728c9f9-0d03-4313-ba2a-3dc22aff8538] updated_port = self._update_port( [ 665.973021] env[63515]: ERROR nova.compute.manager [instance: 2728c9f9-0d03-4313-ba2a-3dc22aff8538] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 665.973021] env[63515]: ERROR nova.compute.manager [instance: 2728c9f9-0d03-4313-ba2a-3dc22aff8538] _ensure_no_port_binding_failure(port) [ 665.973021] env[63515]: ERROR nova.compute.manager [instance: 2728c9f9-0d03-4313-ba2a-3dc22aff8538] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 665.973021] env[63515]: ERROR nova.compute.manager [instance: 2728c9f9-0d03-4313-ba2a-3dc22aff8538] raise exception.PortBindingFailed(port_id=port['id']) [ 665.973021] env[63515]: ERROR nova.compute.manager [instance: 2728c9f9-0d03-4313-ba2a-3dc22aff8538] nova.exception.PortBindingFailed: Binding failed for port bfaff379-5880-4a38-86aa-c5d6331eb307, please check neutron logs for more information. [ 665.973021] env[63515]: ERROR nova.compute.manager [instance: 2728c9f9-0d03-4313-ba2a-3dc22aff8538] [ 665.973294] env[63515]: DEBUG nova.compute.utils [None req-292f2bc3-2911-4498-95fb-4e4a03681601 tempest-FloatingIPsAssociationTestJSON-629096942 tempest-FloatingIPsAssociationTestJSON-629096942-project-member] [instance: 2728c9f9-0d03-4313-ba2a-3dc22aff8538] Binding failed for port bfaff379-5880-4a38-86aa-c5d6331eb307, please check neutron logs for more information. {{(pid=63515) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 665.974211] env[63515]: DEBUG oslo_concurrency.lockutils [None req-fa933c74-4a5b-4cc9-a499-9ea16a38d8b7 tempest-ServerActionsTestJSON-1591637270 tempest-ServerActionsTestJSON-1591637270-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 31.204s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 665.975755] env[63515]: INFO nova.compute.claims [None req-fa933c74-4a5b-4cc9-a499-9ea16a38d8b7 tempest-ServerActionsTestJSON-1591637270 tempest-ServerActionsTestJSON-1591637270-project-member] [instance: 005f45b3-630d-400f-9605-100f6aad88e6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 665.978477] env[63515]: DEBUG nova.compute.manager [None req-292f2bc3-2911-4498-95fb-4e4a03681601 tempest-FloatingIPsAssociationTestJSON-629096942 tempest-FloatingIPsAssociationTestJSON-629096942-project-member] [instance: 2728c9f9-0d03-4313-ba2a-3dc22aff8538] Build of instance 2728c9f9-0d03-4313-ba2a-3dc22aff8538 was re-scheduled: Binding failed for port bfaff379-5880-4a38-86aa-c5d6331eb307, please check neutron logs for more information. {{(pid=63515) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 665.978909] env[63515]: DEBUG nova.compute.manager [None req-292f2bc3-2911-4498-95fb-4e4a03681601 tempest-FloatingIPsAssociationTestJSON-629096942 tempest-FloatingIPsAssociationTestJSON-629096942-project-member] [instance: 2728c9f9-0d03-4313-ba2a-3dc22aff8538] Unplugging VIFs for instance {{(pid=63515) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 665.979133] env[63515]: DEBUG oslo_concurrency.lockutils [None req-292f2bc3-2911-4498-95fb-4e4a03681601 tempest-FloatingIPsAssociationTestJSON-629096942 tempest-FloatingIPsAssociationTestJSON-629096942-project-member] Acquiring lock "refresh_cache-2728c9f9-0d03-4313-ba2a-3dc22aff8538" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 665.979299] env[63515]: DEBUG oslo_concurrency.lockutils [None req-292f2bc3-2911-4498-95fb-4e4a03681601 tempest-FloatingIPsAssociationTestJSON-629096942 tempest-FloatingIPsAssociationTestJSON-629096942-project-member] Acquired lock "refresh_cache-2728c9f9-0d03-4313-ba2a-3dc22aff8538" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 665.979567] env[63515]: DEBUG nova.network.neutron [None req-292f2bc3-2911-4498-95fb-4e4a03681601 tempest-FloatingIPsAssociationTestJSON-629096942 tempest-FloatingIPsAssociationTestJSON-629096942-project-member] [instance: 2728c9f9-0d03-4313-ba2a-3dc22aff8538] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 666.107832] env[63515]: DEBUG oslo_concurrency.lockutils [None req-95251dc9-cf22-4a89-a3b0-fce88f9e12db tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 666.232971] env[63515]: DEBUG nova.network.neutron [req-8bcad23d-0438-437e-bb73-bc9bea4a97de req-f82f1131-a500-4e3e-a1d6-c22d82a0e2db service nova] [instance: 07499601-62bc-4c31-b295-23f34a6e2e91] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 666.343144] env[63515]: DEBUG nova.network.neutron [req-8bcad23d-0438-437e-bb73-bc9bea4a97de req-f82f1131-a500-4e3e-a1d6-c22d82a0e2db service nova] [instance: 07499601-62bc-4c31-b295-23f34a6e2e91] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 666.508227] env[63515]: DEBUG nova.network.neutron [None req-292f2bc3-2911-4498-95fb-4e4a03681601 tempest-FloatingIPsAssociationTestJSON-629096942 tempest-FloatingIPsAssociationTestJSON-629096942-project-member] [instance: 2728c9f9-0d03-4313-ba2a-3dc22aff8538] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 666.618054] env[63515]: DEBUG nova.network.neutron [None req-292f2bc3-2911-4498-95fb-4e4a03681601 tempest-FloatingIPsAssociationTestJSON-629096942 tempest-FloatingIPsAssociationTestJSON-629096942-project-member] [instance: 2728c9f9-0d03-4313-ba2a-3dc22aff8538] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 666.846516] env[63515]: DEBUG oslo_concurrency.lockutils [req-8bcad23d-0438-437e-bb73-bc9bea4a97de req-f82f1131-a500-4e3e-a1d6-c22d82a0e2db service nova] Releasing lock "refresh_cache-07499601-62bc-4c31-b295-23f34a6e2e91" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 666.846606] env[63515]: DEBUG oslo_concurrency.lockutils [None req-b9c961a0-a559-4690-abb6-579359ed447c tempest-ServerPasswordTestJSON-433685770 tempest-ServerPasswordTestJSON-433685770-project-member] Acquired lock "refresh_cache-07499601-62bc-4c31-b295-23f34a6e2e91" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 666.847163] env[63515]: DEBUG nova.network.neutron [None req-b9c961a0-a559-4690-abb6-579359ed447c tempest-ServerPasswordTestJSON-433685770 tempest-ServerPasswordTestJSON-433685770-project-member] [instance: 07499601-62bc-4c31-b295-23f34a6e2e91] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 667.120559] env[63515]: DEBUG oslo_concurrency.lockutils [None req-292f2bc3-2911-4498-95fb-4e4a03681601 tempest-FloatingIPsAssociationTestJSON-629096942 tempest-FloatingIPsAssociationTestJSON-629096942-project-member] Releasing lock "refresh_cache-2728c9f9-0d03-4313-ba2a-3dc22aff8538" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 667.120790] env[63515]: DEBUG nova.compute.manager [None req-292f2bc3-2911-4498-95fb-4e4a03681601 tempest-FloatingIPsAssociationTestJSON-629096942 tempest-FloatingIPsAssociationTestJSON-629096942-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63515) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 667.120965] env[63515]: DEBUG nova.compute.manager [None req-292f2bc3-2911-4498-95fb-4e4a03681601 tempest-FloatingIPsAssociationTestJSON-629096942 tempest-FloatingIPsAssociationTestJSON-629096942-project-member] [instance: 2728c9f9-0d03-4313-ba2a-3dc22aff8538] Deallocating network for instance {{(pid=63515) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 667.121152] env[63515]: DEBUG nova.network.neutron [None req-292f2bc3-2911-4498-95fb-4e4a03681601 tempest-FloatingIPsAssociationTestJSON-629096942 tempest-FloatingIPsAssociationTestJSON-629096942-project-member] [instance: 2728c9f9-0d03-4313-ba2a-3dc22aff8538] deallocate_for_instance() {{(pid=63515) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 667.137316] env[63515]: DEBUG nova.network.neutron [None req-292f2bc3-2911-4498-95fb-4e4a03681601 tempest-FloatingIPsAssociationTestJSON-629096942 tempest-FloatingIPsAssociationTestJSON-629096942-project-member] [instance: 2728c9f9-0d03-4313-ba2a-3dc22aff8538] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 667.357577] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7cbd4791-03e5-4a53-984a-f60b067b8261 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.365217] env[63515]: DEBUG nova.network.neutron [None req-b9c961a0-a559-4690-abb6-579359ed447c tempest-ServerPasswordTestJSON-433685770 tempest-ServerPasswordTestJSON-433685770-project-member] [instance: 07499601-62bc-4c31-b295-23f34a6e2e91] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 667.367571] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ffdcff48-eafa-423e-95f0-b122ac3e7223 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.399142] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9344c7b-da4c-488e-a7e2-94a222a4afb9 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.406257] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0d753c2-5572-42fa-b58d-f1dc3ebe9b17 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.420082] env[63515]: DEBUG nova.compute.provider_tree [None req-fa933c74-4a5b-4cc9-a499-9ea16a38d8b7 tempest-ServerActionsTestJSON-1591637270 tempest-ServerActionsTestJSON-1591637270-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 667.464062] env[63515]: DEBUG nova.network.neutron [None req-b9c961a0-a559-4690-abb6-579359ed447c tempest-ServerPasswordTestJSON-433685770 tempest-ServerPasswordTestJSON-433685770-project-member] [instance: 07499601-62bc-4c31-b295-23f34a6e2e91] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 667.639972] env[63515]: DEBUG nova.network.neutron [None req-292f2bc3-2911-4498-95fb-4e4a03681601 tempest-FloatingIPsAssociationTestJSON-629096942 tempest-FloatingIPsAssociationTestJSON-629096942-project-member] [instance: 2728c9f9-0d03-4313-ba2a-3dc22aff8538] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 667.730162] env[63515]: DEBUG nova.compute.manager [req-37425b0b-0103-4101-abf4-1a3a67a29e5b req-70289c26-bb32-456a-90c7-a58472ef62d7 service nova] [instance: 07499601-62bc-4c31-b295-23f34a6e2e91] Received event network-vif-deleted-b497be91-987b-42ce-9f91-0850d502656f {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 667.925224] env[63515]: DEBUG nova.scheduler.client.report [None req-fa933c74-4a5b-4cc9-a499-9ea16a38d8b7 tempest-ServerActionsTestJSON-1591637270 tempest-ServerActionsTestJSON-1591637270-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 667.967220] env[63515]: DEBUG oslo_concurrency.lockutils [None req-b9c961a0-a559-4690-abb6-579359ed447c tempest-ServerPasswordTestJSON-433685770 tempest-ServerPasswordTestJSON-433685770-project-member] Releasing lock "refresh_cache-07499601-62bc-4c31-b295-23f34a6e2e91" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 667.967648] env[63515]: DEBUG nova.compute.manager [None req-b9c961a0-a559-4690-abb6-579359ed447c tempest-ServerPasswordTestJSON-433685770 tempest-ServerPasswordTestJSON-433685770-project-member] [instance: 07499601-62bc-4c31-b295-23f34a6e2e91] Start destroying the instance on the hypervisor. {{(pid=63515) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 667.967838] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-b9c961a0-a559-4690-abb6-579359ed447c tempest-ServerPasswordTestJSON-433685770 tempest-ServerPasswordTestJSON-433685770-project-member] [instance: 07499601-62bc-4c31-b295-23f34a6e2e91] Destroying instance {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 667.968144] env[63515]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d2f41f95-ef6a-439b-9b0d-fec92d02a155 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.976869] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5226a8e4-11f0-4940-8a37-af406e34699f {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.999119] env[63515]: WARNING nova.virt.vmwareapi.vmops [None req-b9c961a0-a559-4690-abb6-579359ed447c tempest-ServerPasswordTestJSON-433685770 tempest-ServerPasswordTestJSON-433685770-project-member] [instance: 07499601-62bc-4c31-b295-23f34a6e2e91] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 07499601-62bc-4c31-b295-23f34a6e2e91 could not be found. [ 667.999341] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-b9c961a0-a559-4690-abb6-579359ed447c tempest-ServerPasswordTestJSON-433685770 tempest-ServerPasswordTestJSON-433685770-project-member] [instance: 07499601-62bc-4c31-b295-23f34a6e2e91] Instance destroyed {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 667.999648] env[63515]: INFO nova.compute.manager [None req-b9c961a0-a559-4690-abb6-579359ed447c tempest-ServerPasswordTestJSON-433685770 tempest-ServerPasswordTestJSON-433685770-project-member] [instance: 07499601-62bc-4c31-b295-23f34a6e2e91] Took 0.03 seconds to destroy the instance on the hypervisor. [ 667.999754] env[63515]: DEBUG oslo.service.loopingcall [None req-b9c961a0-a559-4690-abb6-579359ed447c tempest-ServerPasswordTestJSON-433685770 tempest-ServerPasswordTestJSON-433685770-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 668.000678] env[63515]: DEBUG nova.compute.manager [-] [instance: 07499601-62bc-4c31-b295-23f34a6e2e91] Deallocating network for instance {{(pid=63515) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 668.000678] env[63515]: DEBUG nova.network.neutron [-] [instance: 07499601-62bc-4c31-b295-23f34a6e2e91] deallocate_for_instance() {{(pid=63515) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 668.022419] env[63515]: DEBUG nova.network.neutron [-] [instance: 07499601-62bc-4c31-b295-23f34a6e2e91] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 668.143767] env[63515]: INFO nova.compute.manager [None req-292f2bc3-2911-4498-95fb-4e4a03681601 tempest-FloatingIPsAssociationTestJSON-629096942 tempest-FloatingIPsAssociationTestJSON-629096942-project-member] [instance: 2728c9f9-0d03-4313-ba2a-3dc22aff8538] Took 1.02 seconds to deallocate network for instance. [ 668.431384] env[63515]: DEBUG oslo_concurrency.lockutils [None req-fa933c74-4a5b-4cc9-a499-9ea16a38d8b7 tempest-ServerActionsTestJSON-1591637270 tempest-ServerActionsTestJSON-1591637270-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.457s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 668.431491] env[63515]: DEBUG nova.compute.manager [None req-fa933c74-4a5b-4cc9-a499-9ea16a38d8b7 tempest-ServerActionsTestJSON-1591637270 tempest-ServerActionsTestJSON-1591637270-project-member] [instance: 005f45b3-630d-400f-9605-100f6aad88e6] Start building networks asynchronously for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 668.434699] env[63515]: DEBUG oslo_concurrency.lockutils [None req-45f6ec8d-6d74-4e46-8fdf-6b0f7c107bf4 tempest-ServerDiagnosticsTest-1160321831 tempest-ServerDiagnosticsTest-1160321831-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 31.802s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 668.436062] env[63515]: INFO nova.compute.claims [None req-45f6ec8d-6d74-4e46-8fdf-6b0f7c107bf4 tempest-ServerDiagnosticsTest-1160321831 tempest-ServerDiagnosticsTest-1160321831-project-member] [instance: 4ad5440f-cd18-44c5-8836-0aa39824cf03] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 668.524694] env[63515]: DEBUG nova.network.neutron [-] [instance: 07499601-62bc-4c31-b295-23f34a6e2e91] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 668.940651] env[63515]: DEBUG nova.compute.utils [None req-fa933c74-4a5b-4cc9-a499-9ea16a38d8b7 tempest-ServerActionsTestJSON-1591637270 tempest-ServerActionsTestJSON-1591637270-project-member] Using /dev/sd instead of None {{(pid=63515) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 668.943934] env[63515]: DEBUG nova.compute.manager [None req-fa933c74-4a5b-4cc9-a499-9ea16a38d8b7 tempest-ServerActionsTestJSON-1591637270 tempest-ServerActionsTestJSON-1591637270-project-member] [instance: 005f45b3-630d-400f-9605-100f6aad88e6] Allocating IP information in the background. {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 668.944178] env[63515]: DEBUG nova.network.neutron [None req-fa933c74-4a5b-4cc9-a499-9ea16a38d8b7 tempest-ServerActionsTestJSON-1591637270 tempest-ServerActionsTestJSON-1591637270-project-member] [instance: 005f45b3-630d-400f-9605-100f6aad88e6] allocate_for_instance() {{(pid=63515) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 669.002767] env[63515]: DEBUG nova.policy [None req-fa933c74-4a5b-4cc9-a499-9ea16a38d8b7 tempest-ServerActionsTestJSON-1591637270 tempest-ServerActionsTestJSON-1591637270-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5305b59f21f94215b2f601e73f8e1076', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4d245bcb1d984b0b86261aebafaa5e56', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63515) authorize /opt/stack/nova/nova/policy.py:201}} [ 669.027482] env[63515]: INFO nova.compute.manager [-] [instance: 07499601-62bc-4c31-b295-23f34a6e2e91] Took 1.03 seconds to deallocate network for instance. [ 669.029821] env[63515]: DEBUG nova.compute.claims [None req-b9c961a0-a559-4690-abb6-579359ed447c tempest-ServerPasswordTestJSON-433685770 tempest-ServerPasswordTestJSON-433685770-project-member] [instance: 07499601-62bc-4c31-b295-23f34a6e2e91] Aborting claim: {{(pid=63515) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 669.029998] env[63515]: DEBUG oslo_concurrency.lockutils [None req-b9c961a0-a559-4690-abb6-579359ed447c tempest-ServerPasswordTestJSON-433685770 tempest-ServerPasswordTestJSON-433685770-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 669.177920] env[63515]: INFO nova.scheduler.client.report [None req-292f2bc3-2911-4498-95fb-4e4a03681601 tempest-FloatingIPsAssociationTestJSON-629096942 tempest-FloatingIPsAssociationTestJSON-629096942-project-member] Deleted allocations for instance 2728c9f9-0d03-4313-ba2a-3dc22aff8538 [ 669.329205] env[63515]: DEBUG nova.network.neutron [None req-fa933c74-4a5b-4cc9-a499-9ea16a38d8b7 tempest-ServerActionsTestJSON-1591637270 tempest-ServerActionsTestJSON-1591637270-project-member] [instance: 005f45b3-630d-400f-9605-100f6aad88e6] Successfully created port: e7f7de8b-0bf6-48de-b304-c89a566c1cdc {{(pid=63515) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 669.444962] env[63515]: DEBUG nova.compute.manager [None req-fa933c74-4a5b-4cc9-a499-9ea16a38d8b7 tempest-ServerActionsTestJSON-1591637270 tempest-ServerActionsTestJSON-1591637270-project-member] [instance: 005f45b3-630d-400f-9605-100f6aad88e6] Start building block device mappings for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 669.692472] env[63515]: DEBUG oslo_concurrency.lockutils [None req-292f2bc3-2911-4498-95fb-4e4a03681601 tempest-FloatingIPsAssociationTestJSON-629096942 tempest-FloatingIPsAssociationTestJSON-629096942-project-member] Lock "2728c9f9-0d03-4313-ba2a-3dc22aff8538" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 112.363s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 669.840427] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-595ceaf0-f607-425f-960b-78a8dce2151c {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.850632] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1f662fc-1baa-441d-8150-d21853c0aa55 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.878635] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5ecff77-e27d-4dc9-92c7-c4055eeabc94 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.885928] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af7278ff-aad5-46fe-82b0-4404f47ae656 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.899051] env[63515]: DEBUG nova.compute.provider_tree [None req-45f6ec8d-6d74-4e46-8fdf-6b0f7c107bf4 tempest-ServerDiagnosticsTest-1160321831 tempest-ServerDiagnosticsTest-1160321831-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 670.176974] env[63515]: DEBUG nova.compute.manager [req-04fda937-7c86-48b7-a280-c2611463c703 req-c8a30f77-3cee-4554-a770-b30bfefe0ae4 service nova] [instance: 005f45b3-630d-400f-9605-100f6aad88e6] Received event network-changed-e7f7de8b-0bf6-48de-b304-c89a566c1cdc {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 670.177261] env[63515]: DEBUG nova.compute.manager [req-04fda937-7c86-48b7-a280-c2611463c703 req-c8a30f77-3cee-4554-a770-b30bfefe0ae4 service nova] [instance: 005f45b3-630d-400f-9605-100f6aad88e6] Refreshing instance network info cache due to event network-changed-e7f7de8b-0bf6-48de-b304-c89a566c1cdc. {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 670.177399] env[63515]: DEBUG oslo_concurrency.lockutils [req-04fda937-7c86-48b7-a280-c2611463c703 req-c8a30f77-3cee-4554-a770-b30bfefe0ae4 service nova] Acquiring lock "refresh_cache-005f45b3-630d-400f-9605-100f6aad88e6" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 670.177543] env[63515]: DEBUG oslo_concurrency.lockutils [req-04fda937-7c86-48b7-a280-c2611463c703 req-c8a30f77-3cee-4554-a770-b30bfefe0ae4 service nova] Acquired lock "refresh_cache-005f45b3-630d-400f-9605-100f6aad88e6" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 670.177701] env[63515]: DEBUG nova.network.neutron [req-04fda937-7c86-48b7-a280-c2611463c703 req-c8a30f77-3cee-4554-a770-b30bfefe0ae4 service nova] [instance: 005f45b3-630d-400f-9605-100f6aad88e6] Refreshing network info cache for port e7f7de8b-0bf6-48de-b304-c89a566c1cdc {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 670.195975] env[63515]: DEBUG nova.compute.manager [None req-c7572d27-5b85-4acb-89ec-c56462911d65 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] [instance: 5b42f744-fdd6-45b1-8563-896869648c23] Starting instance... {{(pid=63515) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 670.339577] env[63515]: ERROR nova.compute.manager [None req-fa933c74-4a5b-4cc9-a499-9ea16a38d8b7 tempest-ServerActionsTestJSON-1591637270 tempest-ServerActionsTestJSON-1591637270-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port e7f7de8b-0bf6-48de-b304-c89a566c1cdc, please check neutron logs for more information. [ 670.339577] env[63515]: ERROR nova.compute.manager Traceback (most recent call last): [ 670.339577] env[63515]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 670.339577] env[63515]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 670.339577] env[63515]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 670.339577] env[63515]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 670.339577] env[63515]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 670.339577] env[63515]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 670.339577] env[63515]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 670.339577] env[63515]: ERROR nova.compute.manager self.force_reraise() [ 670.339577] env[63515]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 670.339577] env[63515]: ERROR nova.compute.manager raise self.value [ 670.339577] env[63515]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 670.339577] env[63515]: ERROR nova.compute.manager updated_port = self._update_port( [ 670.339577] env[63515]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 670.339577] env[63515]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 670.340066] env[63515]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 670.340066] env[63515]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 670.340066] env[63515]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port e7f7de8b-0bf6-48de-b304-c89a566c1cdc, please check neutron logs for more information. [ 670.340066] env[63515]: ERROR nova.compute.manager [ 670.340066] env[63515]: Traceback (most recent call last): [ 670.340066] env[63515]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 670.340066] env[63515]: listener.cb(fileno) [ 670.340066] env[63515]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 670.340066] env[63515]: result = function(*args, **kwargs) [ 670.340066] env[63515]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 670.340066] env[63515]: return func(*args, **kwargs) [ 670.340066] env[63515]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 670.340066] env[63515]: raise e [ 670.340066] env[63515]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 670.340066] env[63515]: nwinfo = self.network_api.allocate_for_instance( [ 670.340066] env[63515]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 670.340066] env[63515]: created_port_ids = self._update_ports_for_instance( [ 670.340066] env[63515]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 670.340066] env[63515]: with excutils.save_and_reraise_exception(): [ 670.340066] env[63515]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 670.340066] env[63515]: self.force_reraise() [ 670.340066] env[63515]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 670.340066] env[63515]: raise self.value [ 670.340066] env[63515]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 670.340066] env[63515]: updated_port = self._update_port( [ 670.340066] env[63515]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 670.340066] env[63515]: _ensure_no_port_binding_failure(port) [ 670.340066] env[63515]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 670.340066] env[63515]: raise exception.PortBindingFailed(port_id=port['id']) [ 670.340867] env[63515]: nova.exception.PortBindingFailed: Binding failed for port e7f7de8b-0bf6-48de-b304-c89a566c1cdc, please check neutron logs for more information. [ 670.340867] env[63515]: Removing descriptor: 17 [ 670.402517] env[63515]: DEBUG nova.scheduler.client.report [None req-45f6ec8d-6d74-4e46-8fdf-6b0f7c107bf4 tempest-ServerDiagnosticsTest-1160321831 tempest-ServerDiagnosticsTest-1160321831-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 670.458040] env[63515]: DEBUG nova.compute.manager [None req-fa933c74-4a5b-4cc9-a499-9ea16a38d8b7 tempest-ServerActionsTestJSON-1591637270 tempest-ServerActionsTestJSON-1591637270-project-member] [instance: 005f45b3-630d-400f-9605-100f6aad88e6] Start spawning the instance on the hypervisor. {{(pid=63515) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 670.483558] env[63515]: DEBUG nova.virt.hardware [None req-fa933c74-4a5b-4cc9-a499-9ea16a38d8b7 tempest-ServerActionsTestJSON-1591637270 tempest-ServerActionsTestJSON-1591637270-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T02:52:14Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T02:51:56Z,direct_url=,disk_format='vmdk',id=8a120570-cb06-4099-b262-554ca0ad15c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b569255cc43e42e7ae3f2b4ad37c6ef4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T02:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 670.483967] env[63515]: DEBUG nova.virt.hardware [None req-fa933c74-4a5b-4cc9-a499-9ea16a38d8b7 tempest-ServerActionsTestJSON-1591637270 tempest-ServerActionsTestJSON-1591637270-project-member] Flavor limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 670.484248] env[63515]: DEBUG nova.virt.hardware [None req-fa933c74-4a5b-4cc9-a499-9ea16a38d8b7 tempest-ServerActionsTestJSON-1591637270 tempest-ServerActionsTestJSON-1591637270-project-member] Image limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 670.484520] env[63515]: DEBUG nova.virt.hardware [None req-fa933c74-4a5b-4cc9-a499-9ea16a38d8b7 tempest-ServerActionsTestJSON-1591637270 tempest-ServerActionsTestJSON-1591637270-project-member] Flavor pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 670.484712] env[63515]: DEBUG nova.virt.hardware [None req-fa933c74-4a5b-4cc9-a499-9ea16a38d8b7 tempest-ServerActionsTestJSON-1591637270 tempest-ServerActionsTestJSON-1591637270-project-member] Image pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 670.484869] env[63515]: DEBUG nova.virt.hardware [None req-fa933c74-4a5b-4cc9-a499-9ea16a38d8b7 tempest-ServerActionsTestJSON-1591637270 tempest-ServerActionsTestJSON-1591637270-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 670.485095] env[63515]: DEBUG nova.virt.hardware [None req-fa933c74-4a5b-4cc9-a499-9ea16a38d8b7 tempest-ServerActionsTestJSON-1591637270 tempest-ServerActionsTestJSON-1591637270-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 670.485309] env[63515]: DEBUG nova.virt.hardware [None req-fa933c74-4a5b-4cc9-a499-9ea16a38d8b7 tempest-ServerActionsTestJSON-1591637270 tempest-ServerActionsTestJSON-1591637270-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 670.485495] env[63515]: DEBUG nova.virt.hardware [None req-fa933c74-4a5b-4cc9-a499-9ea16a38d8b7 tempest-ServerActionsTestJSON-1591637270 tempest-ServerActionsTestJSON-1591637270-project-member] Got 1 possible topologies {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 670.485660] env[63515]: DEBUG nova.virt.hardware [None req-fa933c74-4a5b-4cc9-a499-9ea16a38d8b7 tempest-ServerActionsTestJSON-1591637270 tempest-ServerActionsTestJSON-1591637270-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 670.485831] env[63515]: DEBUG nova.virt.hardware [None req-fa933c74-4a5b-4cc9-a499-9ea16a38d8b7 tempest-ServerActionsTestJSON-1591637270 tempest-ServerActionsTestJSON-1591637270-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 670.487114] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9236d89-f92b-41e5-8303-d62c8f42650d {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.495508] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a95c3dad-262a-4f88-92cb-9050e82187fc {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.509669] env[63515]: ERROR nova.compute.manager [None req-fa933c74-4a5b-4cc9-a499-9ea16a38d8b7 tempest-ServerActionsTestJSON-1591637270 tempest-ServerActionsTestJSON-1591637270-project-member] [instance: 005f45b3-630d-400f-9605-100f6aad88e6] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port e7f7de8b-0bf6-48de-b304-c89a566c1cdc, please check neutron logs for more information. [ 670.509669] env[63515]: ERROR nova.compute.manager [instance: 005f45b3-630d-400f-9605-100f6aad88e6] Traceback (most recent call last): [ 670.509669] env[63515]: ERROR nova.compute.manager [instance: 005f45b3-630d-400f-9605-100f6aad88e6] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 670.509669] env[63515]: ERROR nova.compute.manager [instance: 005f45b3-630d-400f-9605-100f6aad88e6] yield resources [ 670.509669] env[63515]: ERROR nova.compute.manager [instance: 005f45b3-630d-400f-9605-100f6aad88e6] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 670.509669] env[63515]: ERROR nova.compute.manager [instance: 005f45b3-630d-400f-9605-100f6aad88e6] self.driver.spawn(context, instance, image_meta, [ 670.509669] env[63515]: ERROR nova.compute.manager [instance: 005f45b3-630d-400f-9605-100f6aad88e6] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 670.509669] env[63515]: ERROR nova.compute.manager [instance: 005f45b3-630d-400f-9605-100f6aad88e6] self._vmops.spawn(context, instance, image_meta, injected_files, [ 670.509669] env[63515]: ERROR nova.compute.manager [instance: 005f45b3-630d-400f-9605-100f6aad88e6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 670.509669] env[63515]: ERROR nova.compute.manager [instance: 005f45b3-630d-400f-9605-100f6aad88e6] vm_ref = self.build_virtual_machine(instance, [ 670.509669] env[63515]: ERROR nova.compute.manager [instance: 005f45b3-630d-400f-9605-100f6aad88e6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 670.510035] env[63515]: ERROR nova.compute.manager [instance: 005f45b3-630d-400f-9605-100f6aad88e6] vif_infos = vmwarevif.get_vif_info(self._session, [ 670.510035] env[63515]: ERROR nova.compute.manager [instance: 005f45b3-630d-400f-9605-100f6aad88e6] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 670.510035] env[63515]: ERROR nova.compute.manager [instance: 005f45b3-630d-400f-9605-100f6aad88e6] for vif in network_info: [ 670.510035] env[63515]: ERROR nova.compute.manager [instance: 005f45b3-630d-400f-9605-100f6aad88e6] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 670.510035] env[63515]: ERROR nova.compute.manager [instance: 005f45b3-630d-400f-9605-100f6aad88e6] return self._sync_wrapper(fn, *args, **kwargs) [ 670.510035] env[63515]: ERROR nova.compute.manager [instance: 005f45b3-630d-400f-9605-100f6aad88e6] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 670.510035] env[63515]: ERROR nova.compute.manager [instance: 005f45b3-630d-400f-9605-100f6aad88e6] self.wait() [ 670.510035] env[63515]: ERROR nova.compute.manager [instance: 005f45b3-630d-400f-9605-100f6aad88e6] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 670.510035] env[63515]: ERROR nova.compute.manager [instance: 005f45b3-630d-400f-9605-100f6aad88e6] self[:] = self._gt.wait() [ 670.510035] env[63515]: ERROR nova.compute.manager [instance: 005f45b3-630d-400f-9605-100f6aad88e6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 670.510035] env[63515]: ERROR nova.compute.manager [instance: 005f45b3-630d-400f-9605-100f6aad88e6] return self._exit_event.wait() [ 670.510035] env[63515]: ERROR nova.compute.manager [instance: 005f45b3-630d-400f-9605-100f6aad88e6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 670.510035] env[63515]: ERROR nova.compute.manager [instance: 005f45b3-630d-400f-9605-100f6aad88e6] current.throw(*self._exc) [ 670.510353] env[63515]: ERROR nova.compute.manager [instance: 005f45b3-630d-400f-9605-100f6aad88e6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 670.510353] env[63515]: ERROR nova.compute.manager [instance: 005f45b3-630d-400f-9605-100f6aad88e6] result = function(*args, **kwargs) [ 670.510353] env[63515]: ERROR nova.compute.manager [instance: 005f45b3-630d-400f-9605-100f6aad88e6] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 670.510353] env[63515]: ERROR nova.compute.manager [instance: 005f45b3-630d-400f-9605-100f6aad88e6] return func(*args, **kwargs) [ 670.510353] env[63515]: ERROR nova.compute.manager [instance: 005f45b3-630d-400f-9605-100f6aad88e6] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 670.510353] env[63515]: ERROR nova.compute.manager [instance: 005f45b3-630d-400f-9605-100f6aad88e6] raise e [ 670.510353] env[63515]: ERROR nova.compute.manager [instance: 005f45b3-630d-400f-9605-100f6aad88e6] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 670.510353] env[63515]: ERROR nova.compute.manager [instance: 005f45b3-630d-400f-9605-100f6aad88e6] nwinfo = self.network_api.allocate_for_instance( [ 670.510353] env[63515]: ERROR nova.compute.manager [instance: 005f45b3-630d-400f-9605-100f6aad88e6] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 670.510353] env[63515]: ERROR nova.compute.manager [instance: 005f45b3-630d-400f-9605-100f6aad88e6] created_port_ids = self._update_ports_for_instance( [ 670.510353] env[63515]: ERROR nova.compute.manager [instance: 005f45b3-630d-400f-9605-100f6aad88e6] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 670.510353] env[63515]: ERROR nova.compute.manager [instance: 005f45b3-630d-400f-9605-100f6aad88e6] with excutils.save_and_reraise_exception(): [ 670.510353] env[63515]: ERROR nova.compute.manager [instance: 005f45b3-630d-400f-9605-100f6aad88e6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 670.510685] env[63515]: ERROR nova.compute.manager [instance: 005f45b3-630d-400f-9605-100f6aad88e6] self.force_reraise() [ 670.510685] env[63515]: ERROR nova.compute.manager [instance: 005f45b3-630d-400f-9605-100f6aad88e6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 670.510685] env[63515]: ERROR nova.compute.manager [instance: 005f45b3-630d-400f-9605-100f6aad88e6] raise self.value [ 670.510685] env[63515]: ERROR nova.compute.manager [instance: 005f45b3-630d-400f-9605-100f6aad88e6] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 670.510685] env[63515]: ERROR nova.compute.manager [instance: 005f45b3-630d-400f-9605-100f6aad88e6] updated_port = self._update_port( [ 670.510685] env[63515]: ERROR nova.compute.manager [instance: 005f45b3-630d-400f-9605-100f6aad88e6] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 670.510685] env[63515]: ERROR nova.compute.manager [instance: 005f45b3-630d-400f-9605-100f6aad88e6] _ensure_no_port_binding_failure(port) [ 670.510685] env[63515]: ERROR nova.compute.manager [instance: 005f45b3-630d-400f-9605-100f6aad88e6] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 670.510685] env[63515]: ERROR nova.compute.manager [instance: 005f45b3-630d-400f-9605-100f6aad88e6] raise exception.PortBindingFailed(port_id=port['id']) [ 670.510685] env[63515]: ERROR nova.compute.manager [instance: 005f45b3-630d-400f-9605-100f6aad88e6] nova.exception.PortBindingFailed: Binding failed for port e7f7de8b-0bf6-48de-b304-c89a566c1cdc, please check neutron logs for more information. [ 670.510685] env[63515]: ERROR nova.compute.manager [instance: 005f45b3-630d-400f-9605-100f6aad88e6] [ 670.510685] env[63515]: INFO nova.compute.manager [None req-fa933c74-4a5b-4cc9-a499-9ea16a38d8b7 tempest-ServerActionsTestJSON-1591637270 tempest-ServerActionsTestJSON-1591637270-project-member] [instance: 005f45b3-630d-400f-9605-100f6aad88e6] Terminating instance [ 670.511995] env[63515]: DEBUG oslo_concurrency.lockutils [None req-fa933c74-4a5b-4cc9-a499-9ea16a38d8b7 tempest-ServerActionsTestJSON-1591637270 tempest-ServerActionsTestJSON-1591637270-project-member] Acquiring lock "refresh_cache-005f45b3-630d-400f-9605-100f6aad88e6" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 670.705438] env[63515]: DEBUG nova.network.neutron [req-04fda937-7c86-48b7-a280-c2611463c703 req-c8a30f77-3cee-4554-a770-b30bfefe0ae4 service nova] [instance: 005f45b3-630d-400f-9605-100f6aad88e6] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 670.721484] env[63515]: DEBUG oslo_concurrency.lockutils [None req-c7572d27-5b85-4acb-89ec-c56462911d65 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 670.854612] env[63515]: DEBUG nova.network.neutron [req-04fda937-7c86-48b7-a280-c2611463c703 req-c8a30f77-3cee-4554-a770-b30bfefe0ae4 service nova] [instance: 005f45b3-630d-400f-9605-100f6aad88e6] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 670.907141] env[63515]: DEBUG oslo_concurrency.lockutils [None req-45f6ec8d-6d74-4e46-8fdf-6b0f7c107bf4 tempest-ServerDiagnosticsTest-1160321831 tempest-ServerDiagnosticsTest-1160321831-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.472s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 670.907707] env[63515]: DEBUG nova.compute.manager [None req-45f6ec8d-6d74-4e46-8fdf-6b0f7c107bf4 tempest-ServerDiagnosticsTest-1160321831 tempest-ServerDiagnosticsTest-1160321831-project-member] [instance: 4ad5440f-cd18-44c5-8836-0aa39824cf03] Start building networks asynchronously for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 670.911758] env[63515]: DEBUG oslo_concurrency.lockutils [None req-4334f93e-6967-4c78-b6ec-145141dc41da tempest-AttachInterfacesV270Test-352099301 tempest-AttachInterfacesV270Test-352099301-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 31.402s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 671.359694] env[63515]: DEBUG oslo_concurrency.lockutils [req-04fda937-7c86-48b7-a280-c2611463c703 req-c8a30f77-3cee-4554-a770-b30bfefe0ae4 service nova] Releasing lock "refresh_cache-005f45b3-630d-400f-9605-100f6aad88e6" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 671.360119] env[63515]: DEBUG oslo_concurrency.lockutils [None req-fa933c74-4a5b-4cc9-a499-9ea16a38d8b7 tempest-ServerActionsTestJSON-1591637270 tempest-ServerActionsTestJSON-1591637270-project-member] Acquired lock "refresh_cache-005f45b3-630d-400f-9605-100f6aad88e6" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 671.360303] env[63515]: DEBUG nova.network.neutron [None req-fa933c74-4a5b-4cc9-a499-9ea16a38d8b7 tempest-ServerActionsTestJSON-1591637270 tempest-ServerActionsTestJSON-1591637270-project-member] [instance: 005f45b3-630d-400f-9605-100f6aad88e6] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 671.416018] env[63515]: DEBUG nova.compute.utils [None req-45f6ec8d-6d74-4e46-8fdf-6b0f7c107bf4 tempest-ServerDiagnosticsTest-1160321831 tempest-ServerDiagnosticsTest-1160321831-project-member] Using /dev/sd instead of None {{(pid=63515) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 671.421842] env[63515]: DEBUG nova.compute.manager [None req-45f6ec8d-6d74-4e46-8fdf-6b0f7c107bf4 tempest-ServerDiagnosticsTest-1160321831 tempest-ServerDiagnosticsTest-1160321831-project-member] [instance: 4ad5440f-cd18-44c5-8836-0aa39824cf03] Allocating IP information in the background. {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 671.421842] env[63515]: DEBUG nova.network.neutron [None req-45f6ec8d-6d74-4e46-8fdf-6b0f7c107bf4 tempest-ServerDiagnosticsTest-1160321831 tempest-ServerDiagnosticsTest-1160321831-project-member] [instance: 4ad5440f-cd18-44c5-8836-0aa39824cf03] allocate_for_instance() {{(pid=63515) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 671.515279] env[63515]: DEBUG nova.policy [None req-45f6ec8d-6d74-4e46-8fdf-6b0f7c107bf4 tempest-ServerDiagnosticsTest-1160321831 tempest-ServerDiagnosticsTest-1160321831-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'bc3b98ff5c97400a8be47202a7d2e73a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '03abf6ce2cb2400abdae862a82906553', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63515) authorize /opt/stack/nova/nova/policy.py:201}} [ 671.898203] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a9ea957-1da1-4674-b896-ab7259d25b5e {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.901252] env[63515]: DEBUG nova.network.neutron [None req-fa933c74-4a5b-4cc9-a499-9ea16a38d8b7 tempest-ServerActionsTestJSON-1591637270 tempest-ServerActionsTestJSON-1591637270-project-member] [instance: 005f45b3-630d-400f-9605-100f6aad88e6] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 671.908479] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17184303-785d-4131-8f2a-93b181f7c935 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.939710] env[63515]: DEBUG nova.compute.manager [None req-45f6ec8d-6d74-4e46-8fdf-6b0f7c107bf4 tempest-ServerDiagnosticsTest-1160321831 tempest-ServerDiagnosticsTest-1160321831-project-member] [instance: 4ad5440f-cd18-44c5-8836-0aa39824cf03] Start building block device mappings for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 671.942811] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c3bd801-b3d4-4f08-b446-4ddbea63dd59 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.954136] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25b64368-308e-4fd0-8309-6c0cec60f8cc {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.976899] env[63515]: DEBUG nova.compute.provider_tree [None req-4334f93e-6967-4c78-b6ec-145141dc41da tempest-AttachInterfacesV270Test-352099301 tempest-AttachInterfacesV270Test-352099301-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 672.043253] env[63515]: DEBUG nova.network.neutron [None req-fa933c74-4a5b-4cc9-a499-9ea16a38d8b7 tempest-ServerActionsTestJSON-1591637270 tempest-ServerActionsTestJSON-1591637270-project-member] [instance: 005f45b3-630d-400f-9605-100f6aad88e6] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 672.246194] env[63515]: DEBUG nova.network.neutron [None req-45f6ec8d-6d74-4e46-8fdf-6b0f7c107bf4 tempest-ServerDiagnosticsTest-1160321831 tempest-ServerDiagnosticsTest-1160321831-project-member] [instance: 4ad5440f-cd18-44c5-8836-0aa39824cf03] Successfully created port: c4df35f7-5206-4bec-ac71-0578da364e1a {{(pid=63515) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 672.293162] env[63515]: DEBUG nova.compute.manager [req-d92da337-dde1-4988-91bf-3d1236c4f0e1 req-467ffe61-016f-47a7-bdcb-3a204e153143 service nova] [instance: 005f45b3-630d-400f-9605-100f6aad88e6] Received event network-vif-deleted-e7f7de8b-0bf6-48de-b304-c89a566c1cdc {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 672.482524] env[63515]: DEBUG nova.scheduler.client.report [None req-4334f93e-6967-4c78-b6ec-145141dc41da tempest-AttachInterfacesV270Test-352099301 tempest-AttachInterfacesV270Test-352099301-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 672.547647] env[63515]: DEBUG oslo_concurrency.lockutils [None req-fa933c74-4a5b-4cc9-a499-9ea16a38d8b7 tempest-ServerActionsTestJSON-1591637270 tempest-ServerActionsTestJSON-1591637270-project-member] Releasing lock "refresh_cache-005f45b3-630d-400f-9605-100f6aad88e6" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 672.548861] env[63515]: DEBUG nova.compute.manager [None req-fa933c74-4a5b-4cc9-a499-9ea16a38d8b7 tempest-ServerActionsTestJSON-1591637270 tempest-ServerActionsTestJSON-1591637270-project-member] [instance: 005f45b3-630d-400f-9605-100f6aad88e6] Start destroying the instance on the hypervisor. {{(pid=63515) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 672.548861] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-fa933c74-4a5b-4cc9-a499-9ea16a38d8b7 tempest-ServerActionsTestJSON-1591637270 tempest-ServerActionsTestJSON-1591637270-project-member] [instance: 005f45b3-630d-400f-9605-100f6aad88e6] Destroying instance {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 672.548861] env[63515]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f9083261-ad9b-43c7-8875-728ab367e7df {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.558911] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-219d7733-aa4c-4fe5-87f8-791fffdc5775 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.582952] env[63515]: WARNING nova.virt.vmwareapi.vmops [None req-fa933c74-4a5b-4cc9-a499-9ea16a38d8b7 tempest-ServerActionsTestJSON-1591637270 tempest-ServerActionsTestJSON-1591637270-project-member] [instance: 005f45b3-630d-400f-9605-100f6aad88e6] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 005f45b3-630d-400f-9605-100f6aad88e6 could not be found. [ 672.583082] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-fa933c74-4a5b-4cc9-a499-9ea16a38d8b7 tempest-ServerActionsTestJSON-1591637270 tempest-ServerActionsTestJSON-1591637270-project-member] [instance: 005f45b3-630d-400f-9605-100f6aad88e6] Instance destroyed {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 672.583292] env[63515]: INFO nova.compute.manager [None req-fa933c74-4a5b-4cc9-a499-9ea16a38d8b7 tempest-ServerActionsTestJSON-1591637270 tempest-ServerActionsTestJSON-1591637270-project-member] [instance: 005f45b3-630d-400f-9605-100f6aad88e6] Took 0.03 seconds to destroy the instance on the hypervisor. [ 672.583552] env[63515]: DEBUG oslo.service.loopingcall [None req-fa933c74-4a5b-4cc9-a499-9ea16a38d8b7 tempest-ServerActionsTestJSON-1591637270 tempest-ServerActionsTestJSON-1591637270-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 672.583780] env[63515]: DEBUG nova.compute.manager [-] [instance: 005f45b3-630d-400f-9605-100f6aad88e6] Deallocating network for instance {{(pid=63515) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 672.583876] env[63515]: DEBUG nova.network.neutron [-] [instance: 005f45b3-630d-400f-9605-100f6aad88e6] deallocate_for_instance() {{(pid=63515) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 672.607671] env[63515]: DEBUG nova.network.neutron [-] [instance: 005f45b3-630d-400f-9605-100f6aad88e6] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 672.951388] env[63515]: DEBUG nova.compute.manager [None req-45f6ec8d-6d74-4e46-8fdf-6b0f7c107bf4 tempest-ServerDiagnosticsTest-1160321831 tempest-ServerDiagnosticsTest-1160321831-project-member] [instance: 4ad5440f-cd18-44c5-8836-0aa39824cf03] Start spawning the instance on the hypervisor. {{(pid=63515) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 672.972125] env[63515]: DEBUG nova.virt.hardware [None req-45f6ec8d-6d74-4e46-8fdf-6b0f7c107bf4 tempest-ServerDiagnosticsTest-1160321831 tempest-ServerDiagnosticsTest-1160321831-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T02:52:14Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T02:51:56Z,direct_url=,disk_format='vmdk',id=8a120570-cb06-4099-b262-554ca0ad15c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b569255cc43e42e7ae3f2b4ad37c6ef4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T02:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 672.972430] env[63515]: DEBUG nova.virt.hardware [None req-45f6ec8d-6d74-4e46-8fdf-6b0f7c107bf4 tempest-ServerDiagnosticsTest-1160321831 tempest-ServerDiagnosticsTest-1160321831-project-member] Flavor limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 672.972593] env[63515]: DEBUG nova.virt.hardware [None req-45f6ec8d-6d74-4e46-8fdf-6b0f7c107bf4 tempest-ServerDiagnosticsTest-1160321831 tempest-ServerDiagnosticsTest-1160321831-project-member] Image limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 672.973242] env[63515]: DEBUG nova.virt.hardware [None req-45f6ec8d-6d74-4e46-8fdf-6b0f7c107bf4 tempest-ServerDiagnosticsTest-1160321831 tempest-ServerDiagnosticsTest-1160321831-project-member] Flavor pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 672.973242] env[63515]: DEBUG nova.virt.hardware [None req-45f6ec8d-6d74-4e46-8fdf-6b0f7c107bf4 tempest-ServerDiagnosticsTest-1160321831 tempest-ServerDiagnosticsTest-1160321831-project-member] Image pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 672.973242] env[63515]: DEBUG nova.virt.hardware [None req-45f6ec8d-6d74-4e46-8fdf-6b0f7c107bf4 tempest-ServerDiagnosticsTest-1160321831 tempest-ServerDiagnosticsTest-1160321831-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 672.973242] env[63515]: DEBUG nova.virt.hardware [None req-45f6ec8d-6d74-4e46-8fdf-6b0f7c107bf4 tempest-ServerDiagnosticsTest-1160321831 tempest-ServerDiagnosticsTest-1160321831-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 672.973393] env[63515]: DEBUG nova.virt.hardware [None req-45f6ec8d-6d74-4e46-8fdf-6b0f7c107bf4 tempest-ServerDiagnosticsTest-1160321831 tempest-ServerDiagnosticsTest-1160321831-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 672.973552] env[63515]: DEBUG nova.virt.hardware [None req-45f6ec8d-6d74-4e46-8fdf-6b0f7c107bf4 tempest-ServerDiagnosticsTest-1160321831 tempest-ServerDiagnosticsTest-1160321831-project-member] Got 1 possible topologies {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 672.973710] env[63515]: DEBUG nova.virt.hardware [None req-45f6ec8d-6d74-4e46-8fdf-6b0f7c107bf4 tempest-ServerDiagnosticsTest-1160321831 tempest-ServerDiagnosticsTest-1160321831-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 672.973877] env[63515]: DEBUG nova.virt.hardware [None req-45f6ec8d-6d74-4e46-8fdf-6b0f7c107bf4 tempest-ServerDiagnosticsTest-1160321831 tempest-ServerDiagnosticsTest-1160321831-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 672.974810] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dee90671-f5ba-46b1-a9e1-a4efd0ec7501 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.982559] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95e0ce4e-40be-4c77-830d-385b71c2edd7 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.988732] env[63515]: DEBUG oslo_concurrency.lockutils [None req-4334f93e-6967-4c78-b6ec-145141dc41da tempest-AttachInterfacesV270Test-352099301 tempest-AttachInterfacesV270Test-352099301-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.077s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 672.989282] env[63515]: ERROR nova.compute.manager [None req-4334f93e-6967-4c78-b6ec-145141dc41da tempest-AttachInterfacesV270Test-352099301 tempest-AttachInterfacesV270Test-352099301-project-member] [instance: 6876f61c-2315-4e49-9944-4a5caddede5b] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port d382abb3-d2ce-4738-ab1a-4fd2da50fa06, please check neutron logs for more information. [ 672.989282] env[63515]: ERROR nova.compute.manager [instance: 6876f61c-2315-4e49-9944-4a5caddede5b] Traceback (most recent call last): [ 672.989282] env[63515]: ERROR nova.compute.manager [instance: 6876f61c-2315-4e49-9944-4a5caddede5b] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 672.989282] env[63515]: ERROR nova.compute.manager [instance: 6876f61c-2315-4e49-9944-4a5caddede5b] self.driver.spawn(context, instance, image_meta, [ 672.989282] env[63515]: ERROR nova.compute.manager [instance: 6876f61c-2315-4e49-9944-4a5caddede5b] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 672.989282] env[63515]: ERROR nova.compute.manager [instance: 6876f61c-2315-4e49-9944-4a5caddede5b] self._vmops.spawn(context, instance, image_meta, injected_files, [ 672.989282] env[63515]: ERROR nova.compute.manager [instance: 6876f61c-2315-4e49-9944-4a5caddede5b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 672.989282] env[63515]: ERROR nova.compute.manager [instance: 6876f61c-2315-4e49-9944-4a5caddede5b] vm_ref = self.build_virtual_machine(instance, [ 672.989282] env[63515]: ERROR nova.compute.manager [instance: 6876f61c-2315-4e49-9944-4a5caddede5b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 672.989282] env[63515]: ERROR nova.compute.manager [instance: 6876f61c-2315-4e49-9944-4a5caddede5b] vif_infos = vmwarevif.get_vif_info(self._session, [ 672.989282] env[63515]: ERROR nova.compute.manager [instance: 6876f61c-2315-4e49-9944-4a5caddede5b] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 672.989600] env[63515]: ERROR nova.compute.manager [instance: 6876f61c-2315-4e49-9944-4a5caddede5b] for vif in network_info: [ 672.989600] env[63515]: ERROR nova.compute.manager [instance: 6876f61c-2315-4e49-9944-4a5caddede5b] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 672.989600] env[63515]: ERROR nova.compute.manager [instance: 6876f61c-2315-4e49-9944-4a5caddede5b] return self._sync_wrapper(fn, *args, **kwargs) [ 672.989600] env[63515]: ERROR nova.compute.manager [instance: 6876f61c-2315-4e49-9944-4a5caddede5b] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 672.989600] env[63515]: ERROR nova.compute.manager [instance: 6876f61c-2315-4e49-9944-4a5caddede5b] self.wait() [ 672.989600] env[63515]: ERROR nova.compute.manager [instance: 6876f61c-2315-4e49-9944-4a5caddede5b] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 672.989600] env[63515]: ERROR nova.compute.manager [instance: 6876f61c-2315-4e49-9944-4a5caddede5b] self[:] = self._gt.wait() [ 672.989600] env[63515]: ERROR nova.compute.manager [instance: 6876f61c-2315-4e49-9944-4a5caddede5b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 672.989600] env[63515]: ERROR nova.compute.manager [instance: 6876f61c-2315-4e49-9944-4a5caddede5b] return self._exit_event.wait() [ 672.989600] env[63515]: ERROR nova.compute.manager [instance: 6876f61c-2315-4e49-9944-4a5caddede5b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 672.989600] env[63515]: ERROR nova.compute.manager [instance: 6876f61c-2315-4e49-9944-4a5caddede5b] current.throw(*self._exc) [ 672.989600] env[63515]: ERROR nova.compute.manager [instance: 6876f61c-2315-4e49-9944-4a5caddede5b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 672.989600] env[63515]: ERROR nova.compute.manager [instance: 6876f61c-2315-4e49-9944-4a5caddede5b] result = function(*args, **kwargs) [ 672.989906] env[63515]: ERROR nova.compute.manager [instance: 6876f61c-2315-4e49-9944-4a5caddede5b] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 672.989906] env[63515]: ERROR nova.compute.manager [instance: 6876f61c-2315-4e49-9944-4a5caddede5b] return func(*args, **kwargs) [ 672.989906] env[63515]: ERROR nova.compute.manager [instance: 6876f61c-2315-4e49-9944-4a5caddede5b] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 672.989906] env[63515]: ERROR nova.compute.manager [instance: 6876f61c-2315-4e49-9944-4a5caddede5b] raise e [ 672.989906] env[63515]: ERROR nova.compute.manager [instance: 6876f61c-2315-4e49-9944-4a5caddede5b] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 672.989906] env[63515]: ERROR nova.compute.manager [instance: 6876f61c-2315-4e49-9944-4a5caddede5b] nwinfo = self.network_api.allocate_for_instance( [ 672.989906] env[63515]: ERROR nova.compute.manager [instance: 6876f61c-2315-4e49-9944-4a5caddede5b] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 672.989906] env[63515]: ERROR nova.compute.manager [instance: 6876f61c-2315-4e49-9944-4a5caddede5b] created_port_ids = self._update_ports_for_instance( [ 672.989906] env[63515]: ERROR nova.compute.manager [instance: 6876f61c-2315-4e49-9944-4a5caddede5b] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 672.989906] env[63515]: ERROR nova.compute.manager [instance: 6876f61c-2315-4e49-9944-4a5caddede5b] with excutils.save_and_reraise_exception(): [ 672.989906] env[63515]: ERROR nova.compute.manager [instance: 6876f61c-2315-4e49-9944-4a5caddede5b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 672.989906] env[63515]: ERROR nova.compute.manager [instance: 6876f61c-2315-4e49-9944-4a5caddede5b] self.force_reraise() [ 672.989906] env[63515]: ERROR nova.compute.manager [instance: 6876f61c-2315-4e49-9944-4a5caddede5b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 672.990226] env[63515]: ERROR nova.compute.manager [instance: 6876f61c-2315-4e49-9944-4a5caddede5b] raise self.value [ 672.990226] env[63515]: ERROR nova.compute.manager [instance: 6876f61c-2315-4e49-9944-4a5caddede5b] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 672.990226] env[63515]: ERROR nova.compute.manager [instance: 6876f61c-2315-4e49-9944-4a5caddede5b] updated_port = self._update_port( [ 672.990226] env[63515]: ERROR nova.compute.manager [instance: 6876f61c-2315-4e49-9944-4a5caddede5b] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 672.990226] env[63515]: ERROR nova.compute.manager [instance: 6876f61c-2315-4e49-9944-4a5caddede5b] _ensure_no_port_binding_failure(port) [ 672.990226] env[63515]: ERROR nova.compute.manager [instance: 6876f61c-2315-4e49-9944-4a5caddede5b] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 672.990226] env[63515]: ERROR nova.compute.manager [instance: 6876f61c-2315-4e49-9944-4a5caddede5b] raise exception.PortBindingFailed(port_id=port['id']) [ 672.990226] env[63515]: ERROR nova.compute.manager [instance: 6876f61c-2315-4e49-9944-4a5caddede5b] nova.exception.PortBindingFailed: Binding failed for port d382abb3-d2ce-4738-ab1a-4fd2da50fa06, please check neutron logs for more information. [ 672.990226] env[63515]: ERROR nova.compute.manager [instance: 6876f61c-2315-4e49-9944-4a5caddede5b] [ 672.990226] env[63515]: DEBUG nova.compute.utils [None req-4334f93e-6967-4c78-b6ec-145141dc41da tempest-AttachInterfacesV270Test-352099301 tempest-AttachInterfacesV270Test-352099301-project-member] [instance: 6876f61c-2315-4e49-9944-4a5caddede5b] Binding failed for port d382abb3-d2ce-4738-ab1a-4fd2da50fa06, please check neutron logs for more information. {{(pid=63515) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 672.998759] env[63515]: DEBUG oslo_concurrency.lockutils [None req-9ecc0501-dc42-4f19-a331-8b337de58eaf tempest-AttachInterfacesUnderV243Test-1943594058 tempest-AttachInterfacesUnderV243Test-1943594058-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 31.883s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 673.000333] env[63515]: INFO nova.compute.claims [None req-9ecc0501-dc42-4f19-a331-8b337de58eaf tempest-AttachInterfacesUnderV243Test-1943594058 tempest-AttachInterfacesUnderV243Test-1943594058-project-member] [instance: b911a5b5-9617-4fb3-9b5e-fb8c492e4931] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 673.003746] env[63515]: DEBUG nova.compute.manager [None req-4334f93e-6967-4c78-b6ec-145141dc41da tempest-AttachInterfacesV270Test-352099301 tempest-AttachInterfacesV270Test-352099301-project-member] [instance: 6876f61c-2315-4e49-9944-4a5caddede5b] Build of instance 6876f61c-2315-4e49-9944-4a5caddede5b was re-scheduled: Binding failed for port d382abb3-d2ce-4738-ab1a-4fd2da50fa06, please check neutron logs for more information. {{(pid=63515) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 673.003869] env[63515]: DEBUG nova.compute.manager [None req-4334f93e-6967-4c78-b6ec-145141dc41da tempest-AttachInterfacesV270Test-352099301 tempest-AttachInterfacesV270Test-352099301-project-member] [instance: 6876f61c-2315-4e49-9944-4a5caddede5b] Unplugging VIFs for instance {{(pid=63515) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 673.004037] env[63515]: DEBUG oslo_concurrency.lockutils [None req-4334f93e-6967-4c78-b6ec-145141dc41da tempest-AttachInterfacesV270Test-352099301 tempest-AttachInterfacesV270Test-352099301-project-member] Acquiring lock "refresh_cache-6876f61c-2315-4e49-9944-4a5caddede5b" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 673.004208] env[63515]: DEBUG oslo_concurrency.lockutils [None req-4334f93e-6967-4c78-b6ec-145141dc41da tempest-AttachInterfacesV270Test-352099301 tempest-AttachInterfacesV270Test-352099301-project-member] Acquired lock "refresh_cache-6876f61c-2315-4e49-9944-4a5caddede5b" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 673.004381] env[63515]: DEBUG nova.network.neutron [None req-4334f93e-6967-4c78-b6ec-145141dc41da tempest-AttachInterfacesV270Test-352099301 tempest-AttachInterfacesV270Test-352099301-project-member] [instance: 6876f61c-2315-4e49-9944-4a5caddede5b] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 673.110204] env[63515]: DEBUG nova.network.neutron [-] [instance: 005f45b3-630d-400f-9605-100f6aad88e6] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 673.523992] env[63515]: DEBUG nova.network.neutron [None req-4334f93e-6967-4c78-b6ec-145141dc41da tempest-AttachInterfacesV270Test-352099301 tempest-AttachInterfacesV270Test-352099301-project-member] [instance: 6876f61c-2315-4e49-9944-4a5caddede5b] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 673.575920] env[63515]: DEBUG oslo_concurrency.lockutils [None req-8f7a1df0-5f0d-424a-ae8f-76bee2ee3dbb tempest-ServersAaction247Test-2019176257 tempest-ServersAaction247Test-2019176257-project-member] Acquiring lock "5d50fbd6-a7fb-422c-9dd7-df7140fd0aae" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 673.576220] env[63515]: DEBUG oslo_concurrency.lockutils [None req-8f7a1df0-5f0d-424a-ae8f-76bee2ee3dbb tempest-ServersAaction247Test-2019176257 tempest-ServersAaction247Test-2019176257-project-member] Lock "5d50fbd6-a7fb-422c-9dd7-df7140fd0aae" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 673.612697] env[63515]: INFO nova.compute.manager [-] [instance: 005f45b3-630d-400f-9605-100f6aad88e6] Took 1.03 seconds to deallocate network for instance. [ 673.614925] env[63515]: DEBUG nova.compute.claims [None req-fa933c74-4a5b-4cc9-a499-9ea16a38d8b7 tempest-ServerActionsTestJSON-1591637270 tempest-ServerActionsTestJSON-1591637270-project-member] [instance: 005f45b3-630d-400f-9605-100f6aad88e6] Aborting claim: {{(pid=63515) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 673.615114] env[63515]: DEBUG oslo_concurrency.lockutils [None req-fa933c74-4a5b-4cc9-a499-9ea16a38d8b7 tempest-ServerActionsTestJSON-1591637270 tempest-ServerActionsTestJSON-1591637270-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 673.674721] env[63515]: DEBUG nova.network.neutron [None req-4334f93e-6967-4c78-b6ec-145141dc41da tempest-AttachInterfacesV270Test-352099301 tempest-AttachInterfacesV270Test-352099301-project-member] [instance: 6876f61c-2315-4e49-9944-4a5caddede5b] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 673.971636] env[63515]: DEBUG nova.network.neutron [None req-45f6ec8d-6d74-4e46-8fdf-6b0f7c107bf4 tempest-ServerDiagnosticsTest-1160321831 tempest-ServerDiagnosticsTest-1160321831-project-member] [instance: 4ad5440f-cd18-44c5-8836-0aa39824cf03] Successfully updated port: c4df35f7-5206-4bec-ac71-0578da364e1a {{(pid=63515) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 674.178170] env[63515]: DEBUG oslo_concurrency.lockutils [None req-4334f93e-6967-4c78-b6ec-145141dc41da tempest-AttachInterfacesV270Test-352099301 tempest-AttachInterfacesV270Test-352099301-project-member] Releasing lock "refresh_cache-6876f61c-2315-4e49-9944-4a5caddede5b" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 674.178386] env[63515]: DEBUG nova.compute.manager [None req-4334f93e-6967-4c78-b6ec-145141dc41da tempest-AttachInterfacesV270Test-352099301 tempest-AttachInterfacesV270Test-352099301-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63515) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 674.178568] env[63515]: DEBUG nova.compute.manager [None req-4334f93e-6967-4c78-b6ec-145141dc41da tempest-AttachInterfacesV270Test-352099301 tempest-AttachInterfacesV270Test-352099301-project-member] [instance: 6876f61c-2315-4e49-9944-4a5caddede5b] Deallocating network for instance {{(pid=63515) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 674.178733] env[63515]: DEBUG nova.network.neutron [None req-4334f93e-6967-4c78-b6ec-145141dc41da tempest-AttachInterfacesV270Test-352099301 tempest-AttachInterfacesV270Test-352099301-project-member] [instance: 6876f61c-2315-4e49-9944-4a5caddede5b] deallocate_for_instance() {{(pid=63515) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 674.193305] env[63515]: DEBUG nova.network.neutron [None req-4334f93e-6967-4c78-b6ec-145141dc41da tempest-AttachInterfacesV270Test-352099301 tempest-AttachInterfacesV270Test-352099301-project-member] [instance: 6876f61c-2315-4e49-9944-4a5caddede5b] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 674.318583] env[63515]: DEBUG nova.compute.manager [req-b8cd7183-f2db-490c-8745-e2bfc72a234e req-007b16d4-70e9-48ca-8f7e-0b996ffd5ffc service nova] [instance: 4ad5440f-cd18-44c5-8836-0aa39824cf03] Received event network-vif-plugged-c4df35f7-5206-4bec-ac71-0578da364e1a {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 674.318863] env[63515]: DEBUG oslo_concurrency.lockutils [req-b8cd7183-f2db-490c-8745-e2bfc72a234e req-007b16d4-70e9-48ca-8f7e-0b996ffd5ffc service nova] Acquiring lock "4ad5440f-cd18-44c5-8836-0aa39824cf03-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 674.319148] env[63515]: DEBUG oslo_concurrency.lockutils [req-b8cd7183-f2db-490c-8745-e2bfc72a234e req-007b16d4-70e9-48ca-8f7e-0b996ffd5ffc service nova] Lock "4ad5440f-cd18-44c5-8836-0aa39824cf03-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 674.319328] env[63515]: DEBUG oslo_concurrency.lockutils [req-b8cd7183-f2db-490c-8745-e2bfc72a234e req-007b16d4-70e9-48ca-8f7e-0b996ffd5ffc service nova] Lock "4ad5440f-cd18-44c5-8836-0aa39824cf03-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 674.319500] env[63515]: DEBUG nova.compute.manager [req-b8cd7183-f2db-490c-8745-e2bfc72a234e req-007b16d4-70e9-48ca-8f7e-0b996ffd5ffc service nova] [instance: 4ad5440f-cd18-44c5-8836-0aa39824cf03] No waiting events found dispatching network-vif-plugged-c4df35f7-5206-4bec-ac71-0578da364e1a {{(pid=63515) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 674.319666] env[63515]: WARNING nova.compute.manager [req-b8cd7183-f2db-490c-8745-e2bfc72a234e req-007b16d4-70e9-48ca-8f7e-0b996ffd5ffc service nova] [instance: 4ad5440f-cd18-44c5-8836-0aa39824cf03] Received unexpected event network-vif-plugged-c4df35f7-5206-4bec-ac71-0578da364e1a for instance with vm_state building and task_state spawning. [ 674.319821] env[63515]: DEBUG nova.compute.manager [req-b8cd7183-f2db-490c-8745-e2bfc72a234e req-007b16d4-70e9-48ca-8f7e-0b996ffd5ffc service nova] [instance: 4ad5440f-cd18-44c5-8836-0aa39824cf03] Received event network-changed-c4df35f7-5206-4bec-ac71-0578da364e1a {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 674.319970] env[63515]: DEBUG nova.compute.manager [req-b8cd7183-f2db-490c-8745-e2bfc72a234e req-007b16d4-70e9-48ca-8f7e-0b996ffd5ffc service nova] [instance: 4ad5440f-cd18-44c5-8836-0aa39824cf03] Refreshing instance network info cache due to event network-changed-c4df35f7-5206-4bec-ac71-0578da364e1a. {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 674.320231] env[63515]: DEBUG oslo_concurrency.lockutils [req-b8cd7183-f2db-490c-8745-e2bfc72a234e req-007b16d4-70e9-48ca-8f7e-0b996ffd5ffc service nova] Acquiring lock "refresh_cache-4ad5440f-cd18-44c5-8836-0aa39824cf03" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 674.320473] env[63515]: DEBUG oslo_concurrency.lockutils [req-b8cd7183-f2db-490c-8745-e2bfc72a234e req-007b16d4-70e9-48ca-8f7e-0b996ffd5ffc service nova] Acquired lock "refresh_cache-4ad5440f-cd18-44c5-8836-0aa39824cf03" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 674.320601] env[63515]: DEBUG nova.network.neutron [req-b8cd7183-f2db-490c-8745-e2bfc72a234e req-007b16d4-70e9-48ca-8f7e-0b996ffd5ffc service nova] [instance: 4ad5440f-cd18-44c5-8836-0aa39824cf03] Refreshing network info cache for port c4df35f7-5206-4bec-ac71-0578da364e1a {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 674.350119] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3688994-de74-4b2f-9ef3-ce01c86c8a0a {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.359403] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58584ef3-c3d3-4d62-86a3-7789399bd3ad {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.390932] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44246623-758a-456c-b78f-0f03a61a39c5 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.399949] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a12bc77-ad76-4c91-a5f5-5e2281f79d97 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.413758] env[63515]: DEBUG nova.compute.provider_tree [None req-9ecc0501-dc42-4f19-a331-8b337de58eaf tempest-AttachInterfacesUnderV243Test-1943594058 tempest-AttachInterfacesUnderV243Test-1943594058-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 674.474646] env[63515]: DEBUG oslo_concurrency.lockutils [None req-45f6ec8d-6d74-4e46-8fdf-6b0f7c107bf4 tempest-ServerDiagnosticsTest-1160321831 tempest-ServerDiagnosticsTest-1160321831-project-member] Acquiring lock "refresh_cache-4ad5440f-cd18-44c5-8836-0aa39824cf03" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 674.697744] env[63515]: DEBUG nova.network.neutron [None req-4334f93e-6967-4c78-b6ec-145141dc41da tempest-AttachInterfacesV270Test-352099301 tempest-AttachInterfacesV270Test-352099301-project-member] [instance: 6876f61c-2315-4e49-9944-4a5caddede5b] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 674.858768] env[63515]: DEBUG nova.network.neutron [req-b8cd7183-f2db-490c-8745-e2bfc72a234e req-007b16d4-70e9-48ca-8f7e-0b996ffd5ffc service nova] [instance: 4ad5440f-cd18-44c5-8836-0aa39824cf03] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 674.911169] env[63515]: DEBUG nova.network.neutron [req-b8cd7183-f2db-490c-8745-e2bfc72a234e req-007b16d4-70e9-48ca-8f7e-0b996ffd5ffc service nova] [instance: 4ad5440f-cd18-44c5-8836-0aa39824cf03] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 674.915829] env[63515]: DEBUG nova.scheduler.client.report [None req-9ecc0501-dc42-4f19-a331-8b337de58eaf tempest-AttachInterfacesUnderV243Test-1943594058 tempest-AttachInterfacesUnderV243Test-1943594058-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 675.200615] env[63515]: INFO nova.compute.manager [None req-4334f93e-6967-4c78-b6ec-145141dc41da tempest-AttachInterfacesV270Test-352099301 tempest-AttachInterfacesV270Test-352099301-project-member] [instance: 6876f61c-2315-4e49-9944-4a5caddede5b] Took 1.02 seconds to deallocate network for instance. [ 675.413770] env[63515]: DEBUG oslo_concurrency.lockutils [req-b8cd7183-f2db-490c-8745-e2bfc72a234e req-007b16d4-70e9-48ca-8f7e-0b996ffd5ffc service nova] Releasing lock "refresh_cache-4ad5440f-cd18-44c5-8836-0aa39824cf03" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 675.414238] env[63515]: DEBUG oslo_concurrency.lockutils [None req-45f6ec8d-6d74-4e46-8fdf-6b0f7c107bf4 tempest-ServerDiagnosticsTest-1160321831 tempest-ServerDiagnosticsTest-1160321831-project-member] Acquired lock "refresh_cache-4ad5440f-cd18-44c5-8836-0aa39824cf03" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 675.414376] env[63515]: DEBUG nova.network.neutron [None req-45f6ec8d-6d74-4e46-8fdf-6b0f7c107bf4 tempest-ServerDiagnosticsTest-1160321831 tempest-ServerDiagnosticsTest-1160321831-project-member] [instance: 4ad5440f-cd18-44c5-8836-0aa39824cf03] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 675.420625] env[63515]: DEBUG oslo_concurrency.lockutils [None req-9ecc0501-dc42-4f19-a331-8b337de58eaf tempest-AttachInterfacesUnderV243Test-1943594058 tempest-AttachInterfacesUnderV243Test-1943594058-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.422s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 675.421177] env[63515]: DEBUG nova.compute.manager [None req-9ecc0501-dc42-4f19-a331-8b337de58eaf tempest-AttachInterfacesUnderV243Test-1943594058 tempest-AttachInterfacesUnderV243Test-1943594058-project-member] [instance: b911a5b5-9617-4fb3-9b5e-fb8c492e4931] Start building networks asynchronously for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 675.424180] env[63515]: DEBUG oslo_concurrency.lockutils [None req-30030db8-77b3-4af4-8ae6-30653bd81e23 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 31.284s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 675.928835] env[63515]: DEBUG nova.compute.utils [None req-9ecc0501-dc42-4f19-a331-8b337de58eaf tempest-AttachInterfacesUnderV243Test-1943594058 tempest-AttachInterfacesUnderV243Test-1943594058-project-member] Using /dev/sd instead of None {{(pid=63515) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 675.933538] env[63515]: DEBUG nova.compute.manager [None req-9ecc0501-dc42-4f19-a331-8b337de58eaf tempest-AttachInterfacesUnderV243Test-1943594058 tempest-AttachInterfacesUnderV243Test-1943594058-project-member] [instance: b911a5b5-9617-4fb3-9b5e-fb8c492e4931] Allocating IP information in the background. {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 675.933826] env[63515]: DEBUG nova.network.neutron [None req-9ecc0501-dc42-4f19-a331-8b337de58eaf tempest-AttachInterfacesUnderV243Test-1943594058 tempest-AttachInterfacesUnderV243Test-1943594058-project-member] [instance: b911a5b5-9617-4fb3-9b5e-fb8c492e4931] allocate_for_instance() {{(pid=63515) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 675.966291] env[63515]: DEBUG nova.network.neutron [None req-45f6ec8d-6d74-4e46-8fdf-6b0f7c107bf4 tempest-ServerDiagnosticsTest-1160321831 tempest-ServerDiagnosticsTest-1160321831-project-member] [instance: 4ad5440f-cd18-44c5-8836-0aa39824cf03] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 676.045408] env[63515]: DEBUG nova.policy [None req-9ecc0501-dc42-4f19-a331-8b337de58eaf tempest-AttachInterfacesUnderV243Test-1943594058 tempest-AttachInterfacesUnderV243Test-1943594058-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '814257bb6ddd41f181481a8236f08d26', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5086c79c531e4d039896a24a9ea29f73', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63515) authorize /opt/stack/nova/nova/policy.py:201}} [ 676.252044] env[63515]: INFO nova.scheduler.client.report [None req-4334f93e-6967-4c78-b6ec-145141dc41da tempest-AttachInterfacesV270Test-352099301 tempest-AttachInterfacesV270Test-352099301-project-member] Deleted allocations for instance 6876f61c-2315-4e49-9944-4a5caddede5b [ 676.336534] env[63515]: DEBUG nova.network.neutron [None req-45f6ec8d-6d74-4e46-8fdf-6b0f7c107bf4 tempest-ServerDiagnosticsTest-1160321831 tempest-ServerDiagnosticsTest-1160321831-project-member] [instance: 4ad5440f-cd18-44c5-8836-0aa39824cf03] Updating instance_info_cache with network_info: [{"id": "c4df35f7-5206-4bec-ac71-0578da364e1a", "address": "fa:16:3e:39:ce:cc", "network": {"id": "b4abcbd0-0eac-4bde-ad11-7f93afcedcef", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.208", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "b569255cc43e42e7ae3f2b4ad37c6ef4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0cd5d325-3053-407e-a4ee-f627e82a23f9", "external-id": "nsx-vlan-transportzone-809", "segmentation_id": 809, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc4df35f7-52", "ovs_interfaceid": "c4df35f7-5206-4bec-ac71-0578da364e1a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 676.388018] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7dcc831f-7894-4abe-adb1-8278a72c01fc {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.394498] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d1946de-f3a7-460b-ab8e-6997f90d960c {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.434022] env[63515]: DEBUG nova.compute.manager [None req-9ecc0501-dc42-4f19-a331-8b337de58eaf tempest-AttachInterfacesUnderV243Test-1943594058 tempest-AttachInterfacesUnderV243Test-1943594058-project-member] [instance: b911a5b5-9617-4fb3-9b5e-fb8c492e4931] Start building block device mappings for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 676.437430] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f03ec835-721b-4843-beaa-e89ffeeadc69 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.447662] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ada27004-38b3-47e3-ae15-26cd27c1b344 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.460941] env[63515]: DEBUG nova.compute.provider_tree [None req-30030db8-77b3-4af4-8ae6-30653bd81e23 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 676.631028] env[63515]: DEBUG nova.network.neutron [None req-9ecc0501-dc42-4f19-a331-8b337de58eaf tempest-AttachInterfacesUnderV243Test-1943594058 tempest-AttachInterfacesUnderV243Test-1943594058-project-member] [instance: b911a5b5-9617-4fb3-9b5e-fb8c492e4931] Successfully created port: 9bf84cb2-7053-4129-8d8a-dc09e80b384e {{(pid=63515) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 676.762787] env[63515]: DEBUG oslo_concurrency.lockutils [None req-4334f93e-6967-4c78-b6ec-145141dc41da tempest-AttachInterfacesV270Test-352099301 tempest-AttachInterfacesV270Test-352099301-project-member] Lock "6876f61c-2315-4e49-9944-4a5caddede5b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 117.758s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 676.840543] env[63515]: DEBUG oslo_concurrency.lockutils [None req-45f6ec8d-6d74-4e46-8fdf-6b0f7c107bf4 tempest-ServerDiagnosticsTest-1160321831 tempest-ServerDiagnosticsTest-1160321831-project-member] Releasing lock "refresh_cache-4ad5440f-cd18-44c5-8836-0aa39824cf03" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 676.840977] env[63515]: DEBUG nova.compute.manager [None req-45f6ec8d-6d74-4e46-8fdf-6b0f7c107bf4 tempest-ServerDiagnosticsTest-1160321831 tempest-ServerDiagnosticsTest-1160321831-project-member] [instance: 4ad5440f-cd18-44c5-8836-0aa39824cf03] Instance network_info: |[{"id": "c4df35f7-5206-4bec-ac71-0578da364e1a", "address": "fa:16:3e:39:ce:cc", "network": {"id": "b4abcbd0-0eac-4bde-ad11-7f93afcedcef", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.208", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "b569255cc43e42e7ae3f2b4ad37c6ef4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0cd5d325-3053-407e-a4ee-f627e82a23f9", "external-id": "nsx-vlan-transportzone-809", "segmentation_id": 809, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc4df35f7-52", "ovs_interfaceid": "c4df35f7-5206-4bec-ac71-0578da364e1a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 676.841529] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-45f6ec8d-6d74-4e46-8fdf-6b0f7c107bf4 tempest-ServerDiagnosticsTest-1160321831 tempest-ServerDiagnosticsTest-1160321831-project-member] [instance: 4ad5440f-cd18-44c5-8836-0aa39824cf03] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:39:ce:cc', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '0cd5d325-3053-407e-a4ee-f627e82a23f9', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c4df35f7-5206-4bec-ac71-0578da364e1a', 'vif_model': 'vmxnet3'}] {{(pid=63515) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 676.850824] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-45f6ec8d-6d74-4e46-8fdf-6b0f7c107bf4 tempest-ServerDiagnosticsTest-1160321831 tempest-ServerDiagnosticsTest-1160321831-project-member] Creating folder: Project (03abf6ce2cb2400abdae862a82906553). Parent ref: group-v243370. {{(pid=63515) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 676.851137] env[63515]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-2b452ef0-bf1f-48f3-885e-7793339bca9d {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.863375] env[63515]: INFO nova.virt.vmwareapi.vm_util [None req-45f6ec8d-6d74-4e46-8fdf-6b0f7c107bf4 tempest-ServerDiagnosticsTest-1160321831 tempest-ServerDiagnosticsTest-1160321831-project-member] Created folder: Project (03abf6ce2cb2400abdae862a82906553) in parent group-v243370. [ 676.863575] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-45f6ec8d-6d74-4e46-8fdf-6b0f7c107bf4 tempest-ServerDiagnosticsTest-1160321831 tempest-ServerDiagnosticsTest-1160321831-project-member] Creating folder: Instances. Parent ref: group-v243380. {{(pid=63515) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 676.864116] env[63515]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-652b3701-4de7-4861-aae3-29fcbb39cf10 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.874162] env[63515]: INFO nova.virt.vmwareapi.vm_util [None req-45f6ec8d-6d74-4e46-8fdf-6b0f7c107bf4 tempest-ServerDiagnosticsTest-1160321831 tempest-ServerDiagnosticsTest-1160321831-project-member] Created folder: Instances in parent group-v243380. [ 676.874523] env[63515]: DEBUG oslo.service.loopingcall [None req-45f6ec8d-6d74-4e46-8fdf-6b0f7c107bf4 tempest-ServerDiagnosticsTest-1160321831 tempest-ServerDiagnosticsTest-1160321831-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 676.874736] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4ad5440f-cd18-44c5-8836-0aa39824cf03] Creating VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 676.875711] env[63515]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-4032fd13-82d2-4264-9272-3e807ddab293 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.898574] env[63515]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 676.898574] env[63515]: value = "task-1110890" [ 676.898574] env[63515]: _type = "Task" [ 676.898574] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 676.906849] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1110890, 'name': CreateVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 676.964625] env[63515]: DEBUG nova.scheduler.client.report [None req-30030db8-77b3-4af4-8ae6-30653bd81e23 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 677.268464] env[63515]: DEBUG nova.compute.manager [None req-89884269-25d4-4981-bd3b-a346ece7b681 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] [instance: f1d01b75-ac9d-458d-8cc2-ae64cffca4e8] Starting instance... {{(pid=63515) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 677.407924] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1110890, 'name': CreateVM_Task, 'duration_secs': 0.334382} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 677.408149] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4ad5440f-cd18-44c5-8836-0aa39824cf03] Created VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 677.415230] env[63515]: DEBUG oslo_concurrency.lockutils [None req-45f6ec8d-6d74-4e46-8fdf-6b0f7c107bf4 tempest-ServerDiagnosticsTest-1160321831 tempest-ServerDiagnosticsTest-1160321831-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 677.415416] env[63515]: DEBUG oslo_concurrency.lockutils [None req-45f6ec8d-6d74-4e46-8fdf-6b0f7c107bf4 tempest-ServerDiagnosticsTest-1160321831 tempest-ServerDiagnosticsTest-1160321831-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 677.415743] env[63515]: DEBUG oslo_concurrency.lockutils [None req-45f6ec8d-6d74-4e46-8fdf-6b0f7c107bf4 tempest-ServerDiagnosticsTest-1160321831 tempest-ServerDiagnosticsTest-1160321831-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 677.415991] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-50461768-6823-485e-9c14-87758a1d570b {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.420629] env[63515]: DEBUG oslo_vmware.api [None req-45f6ec8d-6d74-4e46-8fdf-6b0f7c107bf4 tempest-ServerDiagnosticsTest-1160321831 tempest-ServerDiagnosticsTest-1160321831-project-member] Waiting for the task: (returnval){ [ 677.420629] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]52ddd6f3-7190-50a4-411b-526c798834c6" [ 677.420629] env[63515]: _type = "Task" [ 677.420629] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 677.428167] env[63515]: DEBUG oslo_vmware.api [None req-45f6ec8d-6d74-4e46-8fdf-6b0f7c107bf4 tempest-ServerDiagnosticsTest-1160321831 tempest-ServerDiagnosticsTest-1160321831-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52ddd6f3-7190-50a4-411b-526c798834c6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 677.451060] env[63515]: DEBUG nova.compute.manager [None req-9ecc0501-dc42-4f19-a331-8b337de58eaf tempest-AttachInterfacesUnderV243Test-1943594058 tempest-AttachInterfacesUnderV243Test-1943594058-project-member] [instance: b911a5b5-9617-4fb3-9b5e-fb8c492e4931] Start spawning the instance on the hypervisor. {{(pid=63515) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 677.470668] env[63515]: DEBUG oslo_concurrency.lockutils [None req-30030db8-77b3-4af4-8ae6-30653bd81e23 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.046s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 677.471468] env[63515]: ERROR nova.compute.manager [None req-30030db8-77b3-4af4-8ae6-30653bd81e23 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: 02ffb723-d6df-42cb-93c5-d582705d1e03] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port a2188979-d321-41f1-8cbe-faa875445836, please check neutron logs for more information. [ 677.471468] env[63515]: ERROR nova.compute.manager [instance: 02ffb723-d6df-42cb-93c5-d582705d1e03] Traceback (most recent call last): [ 677.471468] env[63515]: ERROR nova.compute.manager [instance: 02ffb723-d6df-42cb-93c5-d582705d1e03] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 677.471468] env[63515]: ERROR nova.compute.manager [instance: 02ffb723-d6df-42cb-93c5-d582705d1e03] self.driver.spawn(context, instance, image_meta, [ 677.471468] env[63515]: ERROR nova.compute.manager [instance: 02ffb723-d6df-42cb-93c5-d582705d1e03] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 677.471468] env[63515]: ERROR nova.compute.manager [instance: 02ffb723-d6df-42cb-93c5-d582705d1e03] self._vmops.spawn(context, instance, image_meta, injected_files, [ 677.471468] env[63515]: ERROR nova.compute.manager [instance: 02ffb723-d6df-42cb-93c5-d582705d1e03] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 677.471468] env[63515]: ERROR nova.compute.manager [instance: 02ffb723-d6df-42cb-93c5-d582705d1e03] vm_ref = self.build_virtual_machine(instance, [ 677.471468] env[63515]: ERROR nova.compute.manager [instance: 02ffb723-d6df-42cb-93c5-d582705d1e03] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 677.471468] env[63515]: ERROR nova.compute.manager [instance: 02ffb723-d6df-42cb-93c5-d582705d1e03] vif_infos = vmwarevif.get_vif_info(self._session, [ 677.471468] env[63515]: ERROR nova.compute.manager [instance: 02ffb723-d6df-42cb-93c5-d582705d1e03] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 677.471844] env[63515]: ERROR nova.compute.manager [instance: 02ffb723-d6df-42cb-93c5-d582705d1e03] for vif in network_info: [ 677.471844] env[63515]: ERROR nova.compute.manager [instance: 02ffb723-d6df-42cb-93c5-d582705d1e03] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 677.471844] env[63515]: ERROR nova.compute.manager [instance: 02ffb723-d6df-42cb-93c5-d582705d1e03] return self._sync_wrapper(fn, *args, **kwargs) [ 677.471844] env[63515]: ERROR nova.compute.manager [instance: 02ffb723-d6df-42cb-93c5-d582705d1e03] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 677.471844] env[63515]: ERROR nova.compute.manager [instance: 02ffb723-d6df-42cb-93c5-d582705d1e03] self.wait() [ 677.471844] env[63515]: ERROR nova.compute.manager [instance: 02ffb723-d6df-42cb-93c5-d582705d1e03] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 677.471844] env[63515]: ERROR nova.compute.manager [instance: 02ffb723-d6df-42cb-93c5-d582705d1e03] self[:] = self._gt.wait() [ 677.471844] env[63515]: ERROR nova.compute.manager [instance: 02ffb723-d6df-42cb-93c5-d582705d1e03] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 677.471844] env[63515]: ERROR nova.compute.manager [instance: 02ffb723-d6df-42cb-93c5-d582705d1e03] return self._exit_event.wait() [ 677.471844] env[63515]: ERROR nova.compute.manager [instance: 02ffb723-d6df-42cb-93c5-d582705d1e03] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 677.471844] env[63515]: ERROR nova.compute.manager [instance: 02ffb723-d6df-42cb-93c5-d582705d1e03] current.throw(*self._exc) [ 677.471844] env[63515]: ERROR nova.compute.manager [instance: 02ffb723-d6df-42cb-93c5-d582705d1e03] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 677.471844] env[63515]: ERROR nova.compute.manager [instance: 02ffb723-d6df-42cb-93c5-d582705d1e03] result = function(*args, **kwargs) [ 677.472316] env[63515]: ERROR nova.compute.manager [instance: 02ffb723-d6df-42cb-93c5-d582705d1e03] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 677.472316] env[63515]: ERROR nova.compute.manager [instance: 02ffb723-d6df-42cb-93c5-d582705d1e03] return func(*args, **kwargs) [ 677.472316] env[63515]: ERROR nova.compute.manager [instance: 02ffb723-d6df-42cb-93c5-d582705d1e03] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 677.472316] env[63515]: ERROR nova.compute.manager [instance: 02ffb723-d6df-42cb-93c5-d582705d1e03] raise e [ 677.472316] env[63515]: ERROR nova.compute.manager [instance: 02ffb723-d6df-42cb-93c5-d582705d1e03] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 677.472316] env[63515]: ERROR nova.compute.manager [instance: 02ffb723-d6df-42cb-93c5-d582705d1e03] nwinfo = self.network_api.allocate_for_instance( [ 677.472316] env[63515]: ERROR nova.compute.manager [instance: 02ffb723-d6df-42cb-93c5-d582705d1e03] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 677.472316] env[63515]: ERROR nova.compute.manager [instance: 02ffb723-d6df-42cb-93c5-d582705d1e03] created_port_ids = self._update_ports_for_instance( [ 677.472316] env[63515]: ERROR nova.compute.manager [instance: 02ffb723-d6df-42cb-93c5-d582705d1e03] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 677.472316] env[63515]: ERROR nova.compute.manager [instance: 02ffb723-d6df-42cb-93c5-d582705d1e03] with excutils.save_and_reraise_exception(): [ 677.472316] env[63515]: ERROR nova.compute.manager [instance: 02ffb723-d6df-42cb-93c5-d582705d1e03] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 677.472316] env[63515]: ERROR nova.compute.manager [instance: 02ffb723-d6df-42cb-93c5-d582705d1e03] self.force_reraise() [ 677.472316] env[63515]: ERROR nova.compute.manager [instance: 02ffb723-d6df-42cb-93c5-d582705d1e03] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 677.472679] env[63515]: ERROR nova.compute.manager [instance: 02ffb723-d6df-42cb-93c5-d582705d1e03] raise self.value [ 677.472679] env[63515]: ERROR nova.compute.manager [instance: 02ffb723-d6df-42cb-93c5-d582705d1e03] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 677.472679] env[63515]: ERROR nova.compute.manager [instance: 02ffb723-d6df-42cb-93c5-d582705d1e03] updated_port = self._update_port( [ 677.472679] env[63515]: ERROR nova.compute.manager [instance: 02ffb723-d6df-42cb-93c5-d582705d1e03] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 677.472679] env[63515]: ERROR nova.compute.manager [instance: 02ffb723-d6df-42cb-93c5-d582705d1e03] _ensure_no_port_binding_failure(port) [ 677.472679] env[63515]: ERROR nova.compute.manager [instance: 02ffb723-d6df-42cb-93c5-d582705d1e03] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 677.472679] env[63515]: ERROR nova.compute.manager [instance: 02ffb723-d6df-42cb-93c5-d582705d1e03] raise exception.PortBindingFailed(port_id=port['id']) [ 677.472679] env[63515]: ERROR nova.compute.manager [instance: 02ffb723-d6df-42cb-93c5-d582705d1e03] nova.exception.PortBindingFailed: Binding failed for port a2188979-d321-41f1-8cbe-faa875445836, please check neutron logs for more information. [ 677.472679] env[63515]: ERROR nova.compute.manager [instance: 02ffb723-d6df-42cb-93c5-d582705d1e03] [ 677.472679] env[63515]: DEBUG nova.compute.utils [None req-30030db8-77b3-4af4-8ae6-30653bd81e23 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: 02ffb723-d6df-42cb-93c5-d582705d1e03] Binding failed for port a2188979-d321-41f1-8cbe-faa875445836, please check neutron logs for more information. {{(pid=63515) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 677.473512] env[63515]: DEBUG oslo_concurrency.lockutils [None req-30987b61-ce89-4676-89be-fee7786462c9 tempest-ImagesOneServerNegativeTestJSON-52225474 tempest-ImagesOneServerNegativeTestJSON-52225474-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.456s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 677.476517] env[63515]: DEBUG nova.compute.manager [None req-30030db8-77b3-4af4-8ae6-30653bd81e23 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: 02ffb723-d6df-42cb-93c5-d582705d1e03] Build of instance 02ffb723-d6df-42cb-93c5-d582705d1e03 was re-scheduled: Binding failed for port a2188979-d321-41f1-8cbe-faa875445836, please check neutron logs for more information. {{(pid=63515) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 677.476950] env[63515]: DEBUG nova.compute.manager [None req-30030db8-77b3-4af4-8ae6-30653bd81e23 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: 02ffb723-d6df-42cb-93c5-d582705d1e03] Unplugging VIFs for instance {{(pid=63515) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 677.477195] env[63515]: DEBUG oslo_concurrency.lockutils [None req-30030db8-77b3-4af4-8ae6-30653bd81e23 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Acquiring lock "refresh_cache-02ffb723-d6df-42cb-93c5-d582705d1e03" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 677.477366] env[63515]: DEBUG oslo_concurrency.lockutils [None req-30030db8-77b3-4af4-8ae6-30653bd81e23 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Acquired lock "refresh_cache-02ffb723-d6df-42cb-93c5-d582705d1e03" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 677.481961] env[63515]: DEBUG nova.network.neutron [None req-30030db8-77b3-4af4-8ae6-30653bd81e23 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: 02ffb723-d6df-42cb-93c5-d582705d1e03] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 677.484060] env[63515]: DEBUG nova.virt.hardware [None req-9ecc0501-dc42-4f19-a331-8b337de58eaf tempest-AttachInterfacesUnderV243Test-1943594058 tempest-AttachInterfacesUnderV243Test-1943594058-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T02:52:14Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T02:51:56Z,direct_url=,disk_format='vmdk',id=8a120570-cb06-4099-b262-554ca0ad15c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b569255cc43e42e7ae3f2b4ad37c6ef4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T02:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 677.484319] env[63515]: DEBUG nova.virt.hardware [None req-9ecc0501-dc42-4f19-a331-8b337de58eaf tempest-AttachInterfacesUnderV243Test-1943594058 tempest-AttachInterfacesUnderV243Test-1943594058-project-member] Flavor limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 677.484479] env[63515]: DEBUG nova.virt.hardware [None req-9ecc0501-dc42-4f19-a331-8b337de58eaf tempest-AttachInterfacesUnderV243Test-1943594058 tempest-AttachInterfacesUnderV243Test-1943594058-project-member] Image limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 677.485530] env[63515]: DEBUG nova.virt.hardware [None req-9ecc0501-dc42-4f19-a331-8b337de58eaf tempest-AttachInterfacesUnderV243Test-1943594058 tempest-AttachInterfacesUnderV243Test-1943594058-project-member] Flavor pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 677.485530] env[63515]: DEBUG nova.virt.hardware [None req-9ecc0501-dc42-4f19-a331-8b337de58eaf tempest-AttachInterfacesUnderV243Test-1943594058 tempest-AttachInterfacesUnderV243Test-1943594058-project-member] Image pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 677.485530] env[63515]: DEBUG nova.virt.hardware [None req-9ecc0501-dc42-4f19-a331-8b337de58eaf tempest-AttachInterfacesUnderV243Test-1943594058 tempest-AttachInterfacesUnderV243Test-1943594058-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 677.485530] env[63515]: DEBUG nova.virt.hardware [None req-9ecc0501-dc42-4f19-a331-8b337de58eaf tempest-AttachInterfacesUnderV243Test-1943594058 tempest-AttachInterfacesUnderV243Test-1943594058-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 677.485530] env[63515]: DEBUG nova.virt.hardware [None req-9ecc0501-dc42-4f19-a331-8b337de58eaf tempest-AttachInterfacesUnderV243Test-1943594058 tempest-AttachInterfacesUnderV243Test-1943594058-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 677.485729] env[63515]: DEBUG nova.virt.hardware [None req-9ecc0501-dc42-4f19-a331-8b337de58eaf tempest-AttachInterfacesUnderV243Test-1943594058 tempest-AttachInterfacesUnderV243Test-1943594058-project-member] Got 1 possible topologies {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 677.485729] env[63515]: DEBUG nova.virt.hardware [None req-9ecc0501-dc42-4f19-a331-8b337de58eaf tempest-AttachInterfacesUnderV243Test-1943594058 tempest-AttachInterfacesUnderV243Test-1943594058-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 677.485857] env[63515]: DEBUG nova.virt.hardware [None req-9ecc0501-dc42-4f19-a331-8b337de58eaf tempest-AttachInterfacesUnderV243Test-1943594058 tempest-AttachInterfacesUnderV243Test-1943594058-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 677.486898] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c28143c-d179-4516-ad32-86926f3a26e8 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.498120] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c722ab36-9b1c-4a3c-b208-11ed8e2c9c1e {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.799153] env[63515]: DEBUG oslo_concurrency.lockutils [None req-89884269-25d4-4981-bd3b-a346ece7b681 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 677.933616] env[63515]: DEBUG oslo_vmware.api [None req-45f6ec8d-6d74-4e46-8fdf-6b0f7c107bf4 tempest-ServerDiagnosticsTest-1160321831 tempest-ServerDiagnosticsTest-1160321831-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52ddd6f3-7190-50a4-411b-526c798834c6, 'name': SearchDatastore_Task, 'duration_secs': 0.011254} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 677.933929] env[63515]: DEBUG oslo_concurrency.lockutils [None req-45f6ec8d-6d74-4e46-8fdf-6b0f7c107bf4 tempest-ServerDiagnosticsTest-1160321831 tempest-ServerDiagnosticsTest-1160321831-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 677.934175] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-45f6ec8d-6d74-4e46-8fdf-6b0f7c107bf4 tempest-ServerDiagnosticsTest-1160321831 tempest-ServerDiagnosticsTest-1160321831-project-member] [instance: 4ad5440f-cd18-44c5-8836-0aa39824cf03] Processing image 8a120570-cb06-4099-b262-554ca0ad15c5 {{(pid=63515) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 677.934451] env[63515]: DEBUG oslo_concurrency.lockutils [None req-45f6ec8d-6d74-4e46-8fdf-6b0f7c107bf4 tempest-ServerDiagnosticsTest-1160321831 tempest-ServerDiagnosticsTest-1160321831-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 677.934601] env[63515]: DEBUG oslo_concurrency.lockutils [None req-45f6ec8d-6d74-4e46-8fdf-6b0f7c107bf4 tempest-ServerDiagnosticsTest-1160321831 tempest-ServerDiagnosticsTest-1160321831-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 677.934881] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-45f6ec8d-6d74-4e46-8fdf-6b0f7c107bf4 tempest-ServerDiagnosticsTest-1160321831 tempest-ServerDiagnosticsTest-1160321831-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 677.935213] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-208f8433-dbe0-4006-a37d-c7cee235f445 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.943669] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-45f6ec8d-6d74-4e46-8fdf-6b0f7c107bf4 tempest-ServerDiagnosticsTest-1160321831 tempest-ServerDiagnosticsTest-1160321831-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 677.943843] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-45f6ec8d-6d74-4e46-8fdf-6b0f7c107bf4 tempest-ServerDiagnosticsTest-1160321831 tempest-ServerDiagnosticsTest-1160321831-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63515) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 677.944605] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7b634740-4466-46b2-85ba-94c8b7d6328d {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.949813] env[63515]: DEBUG oslo_vmware.api [None req-45f6ec8d-6d74-4e46-8fdf-6b0f7c107bf4 tempest-ServerDiagnosticsTest-1160321831 tempest-ServerDiagnosticsTest-1160321831-project-member] Waiting for the task: (returnval){ [ 677.949813] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]52695fb1-18e2-8d16-2f4a-af9e0c93a2b6" [ 677.949813] env[63515]: _type = "Task" [ 677.949813] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 677.958701] env[63515]: DEBUG oslo_vmware.api [None req-45f6ec8d-6d74-4e46-8fdf-6b0f7c107bf4 tempest-ServerDiagnosticsTest-1160321831 tempest-ServerDiagnosticsTest-1160321831-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52695fb1-18e2-8d16-2f4a-af9e0c93a2b6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 678.012240] env[63515]: DEBUG nova.network.neutron [None req-30030db8-77b3-4af4-8ae6-30653bd81e23 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: 02ffb723-d6df-42cb-93c5-d582705d1e03] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 678.148277] env[63515]: DEBUG nova.network.neutron [None req-30030db8-77b3-4af4-8ae6-30653bd81e23 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: 02ffb723-d6df-42cb-93c5-d582705d1e03] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 678.237249] env[63515]: DEBUG nova.compute.manager [req-64d239fd-1d34-4ded-914a-3d2e8994125d req-94b6e101-3a60-4a7b-9050-e37e5981e28a service nova] [instance: b911a5b5-9617-4fb3-9b5e-fb8c492e4931] Received event network-vif-plugged-9bf84cb2-7053-4129-8d8a-dc09e80b384e {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 678.237249] env[63515]: DEBUG oslo_concurrency.lockutils [req-64d239fd-1d34-4ded-914a-3d2e8994125d req-94b6e101-3a60-4a7b-9050-e37e5981e28a service nova] Acquiring lock "b911a5b5-9617-4fb3-9b5e-fb8c492e4931-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 678.237249] env[63515]: DEBUG oslo_concurrency.lockutils [req-64d239fd-1d34-4ded-914a-3d2e8994125d req-94b6e101-3a60-4a7b-9050-e37e5981e28a service nova] Lock "b911a5b5-9617-4fb3-9b5e-fb8c492e4931-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 678.237249] env[63515]: DEBUG oslo_concurrency.lockutils [req-64d239fd-1d34-4ded-914a-3d2e8994125d req-94b6e101-3a60-4a7b-9050-e37e5981e28a service nova] Lock "b911a5b5-9617-4fb3-9b5e-fb8c492e4931-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 678.237249] env[63515]: DEBUG nova.compute.manager [req-64d239fd-1d34-4ded-914a-3d2e8994125d req-94b6e101-3a60-4a7b-9050-e37e5981e28a service nova] [instance: b911a5b5-9617-4fb3-9b5e-fb8c492e4931] No waiting events found dispatching network-vif-plugged-9bf84cb2-7053-4129-8d8a-dc09e80b384e {{(pid=63515) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 678.237399] env[63515]: WARNING nova.compute.manager [req-64d239fd-1d34-4ded-914a-3d2e8994125d req-94b6e101-3a60-4a7b-9050-e37e5981e28a service nova] [instance: b911a5b5-9617-4fb3-9b5e-fb8c492e4931] Received unexpected event network-vif-plugged-9bf84cb2-7053-4129-8d8a-dc09e80b384e for instance with vm_state building and task_state spawning. [ 678.416710] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85fbd4ab-5eda-444e-bec0-a0237bb8ac8c {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.425466] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6cc657fa-6318-48dd-93d3-5e4977ca6c39 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.460517] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f3c4e24-70dc-4887-8289-8d40b9d7e84e {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.471018] env[63515]: DEBUG oslo_vmware.api [None req-45f6ec8d-6d74-4e46-8fdf-6b0f7c107bf4 tempest-ServerDiagnosticsTest-1160321831 tempest-ServerDiagnosticsTest-1160321831-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52695fb1-18e2-8d16-2f4a-af9e0c93a2b6, 'name': SearchDatastore_Task, 'duration_secs': 0.009757} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 678.471018] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2c552800-e7fd-4df6-bdda-703c1fd5c959 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.472931] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d42b968e-ec3c-487e-a1e5-f51dc8f68cf2 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.478279] env[63515]: DEBUG nova.network.neutron [None req-9ecc0501-dc42-4f19-a331-8b337de58eaf tempest-AttachInterfacesUnderV243Test-1943594058 tempest-AttachInterfacesUnderV243Test-1943594058-project-member] [instance: b911a5b5-9617-4fb3-9b5e-fb8c492e4931] Successfully updated port: 9bf84cb2-7053-4129-8d8a-dc09e80b384e {{(pid=63515) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 678.487818] env[63515]: DEBUG oslo_vmware.api [None req-45f6ec8d-6d74-4e46-8fdf-6b0f7c107bf4 tempest-ServerDiagnosticsTest-1160321831 tempest-ServerDiagnosticsTest-1160321831-project-member] Waiting for the task: (returnval){ [ 678.487818] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]52e61cbf-f014-98e3-4ac9-f9ab196961ab" [ 678.487818] env[63515]: _type = "Task" [ 678.487818] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 678.492029] env[63515]: DEBUG nova.compute.provider_tree [None req-30987b61-ce89-4676-89be-fee7786462c9 tempest-ImagesOneServerNegativeTestJSON-52225474 tempest-ImagesOneServerNegativeTestJSON-52225474-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 678.499058] env[63515]: DEBUG oslo_vmware.api [None req-45f6ec8d-6d74-4e46-8fdf-6b0f7c107bf4 tempest-ServerDiagnosticsTest-1160321831 tempest-ServerDiagnosticsTest-1160321831-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52e61cbf-f014-98e3-4ac9-f9ab196961ab, 'name': SearchDatastore_Task, 'duration_secs': 0.008743} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 678.500563] env[63515]: DEBUG oslo_concurrency.lockutils [None req-45f6ec8d-6d74-4e46-8fdf-6b0f7c107bf4 tempest-ServerDiagnosticsTest-1160321831 tempest-ServerDiagnosticsTest-1160321831-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 678.500563] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-45f6ec8d-6d74-4e46-8fdf-6b0f7c107bf4 tempest-ServerDiagnosticsTest-1160321831 tempest-ServerDiagnosticsTest-1160321831-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk to [datastore1] 4ad5440f-cd18-44c5-8836-0aa39824cf03/4ad5440f-cd18-44c5-8836-0aa39824cf03.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 678.500563] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ac9fc96d-4c8c-4e58-a254-8876cd7a17d9 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.508470] env[63515]: DEBUG oslo_vmware.api [None req-45f6ec8d-6d74-4e46-8fdf-6b0f7c107bf4 tempest-ServerDiagnosticsTest-1160321831 tempest-ServerDiagnosticsTest-1160321831-project-member] Waiting for the task: (returnval){ [ 678.508470] env[63515]: value = "task-1110891" [ 678.508470] env[63515]: _type = "Task" [ 678.508470] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 678.516574] env[63515]: DEBUG oslo_vmware.api [None req-45f6ec8d-6d74-4e46-8fdf-6b0f7c107bf4 tempest-ServerDiagnosticsTest-1160321831 tempest-ServerDiagnosticsTest-1160321831-project-member] Task: {'id': task-1110891, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 678.651137] env[63515]: DEBUG oslo_concurrency.lockutils [None req-30030db8-77b3-4af4-8ae6-30653bd81e23 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Releasing lock "refresh_cache-02ffb723-d6df-42cb-93c5-d582705d1e03" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 678.651137] env[63515]: DEBUG nova.compute.manager [None req-30030db8-77b3-4af4-8ae6-30653bd81e23 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63515) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 678.651289] env[63515]: DEBUG nova.compute.manager [None req-30030db8-77b3-4af4-8ae6-30653bd81e23 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: 02ffb723-d6df-42cb-93c5-d582705d1e03] Deallocating network for instance {{(pid=63515) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 678.651449] env[63515]: DEBUG nova.network.neutron [None req-30030db8-77b3-4af4-8ae6-30653bd81e23 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: 02ffb723-d6df-42cb-93c5-d582705d1e03] deallocate_for_instance() {{(pid=63515) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 678.769594] env[63515]: DEBUG nova.network.neutron [None req-30030db8-77b3-4af4-8ae6-30653bd81e23 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: 02ffb723-d6df-42cb-93c5-d582705d1e03] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 678.991191] env[63515]: DEBUG oslo_concurrency.lockutils [None req-9ecc0501-dc42-4f19-a331-8b337de58eaf tempest-AttachInterfacesUnderV243Test-1943594058 tempest-AttachInterfacesUnderV243Test-1943594058-project-member] Acquiring lock "refresh_cache-b911a5b5-9617-4fb3-9b5e-fb8c492e4931" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 678.991191] env[63515]: DEBUG oslo_concurrency.lockutils [None req-9ecc0501-dc42-4f19-a331-8b337de58eaf tempest-AttachInterfacesUnderV243Test-1943594058 tempest-AttachInterfacesUnderV243Test-1943594058-project-member] Acquired lock "refresh_cache-b911a5b5-9617-4fb3-9b5e-fb8c492e4931" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 678.991191] env[63515]: DEBUG nova.network.neutron [None req-9ecc0501-dc42-4f19-a331-8b337de58eaf tempest-AttachInterfacesUnderV243Test-1943594058 tempest-AttachInterfacesUnderV243Test-1943594058-project-member] [instance: b911a5b5-9617-4fb3-9b5e-fb8c492e4931] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 678.993320] env[63515]: DEBUG nova.scheduler.client.report [None req-30987b61-ce89-4676-89be-fee7786462c9 tempest-ImagesOneServerNegativeTestJSON-52225474 tempest-ImagesOneServerNegativeTestJSON-52225474-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 679.019331] env[63515]: DEBUG oslo_vmware.api [None req-45f6ec8d-6d74-4e46-8fdf-6b0f7c107bf4 tempest-ServerDiagnosticsTest-1160321831 tempest-ServerDiagnosticsTest-1160321831-project-member] Task: {'id': task-1110891, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.456153} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 679.020161] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-45f6ec8d-6d74-4e46-8fdf-6b0f7c107bf4 tempest-ServerDiagnosticsTest-1160321831 tempest-ServerDiagnosticsTest-1160321831-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk to [datastore1] 4ad5440f-cd18-44c5-8836-0aa39824cf03/4ad5440f-cd18-44c5-8836-0aa39824cf03.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 679.020497] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-45f6ec8d-6d74-4e46-8fdf-6b0f7c107bf4 tempest-ServerDiagnosticsTest-1160321831 tempest-ServerDiagnosticsTest-1160321831-project-member] [instance: 4ad5440f-cd18-44c5-8836-0aa39824cf03] Extending root virtual disk to 1048576 {{(pid=63515) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 679.020839] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a8f02724-a1a3-4c39-9477-b13ae9ee4b3a {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.028331] env[63515]: DEBUG oslo_vmware.api [None req-45f6ec8d-6d74-4e46-8fdf-6b0f7c107bf4 tempest-ServerDiagnosticsTest-1160321831 tempest-ServerDiagnosticsTest-1160321831-project-member] Waiting for the task: (returnval){ [ 679.028331] env[63515]: value = "task-1110892" [ 679.028331] env[63515]: _type = "Task" [ 679.028331] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 679.036873] env[63515]: DEBUG oslo_vmware.api [None req-45f6ec8d-6d74-4e46-8fdf-6b0f7c107bf4 tempest-ServerDiagnosticsTest-1160321831 tempest-ServerDiagnosticsTest-1160321831-project-member] Task: {'id': task-1110892, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 679.274019] env[63515]: DEBUG nova.network.neutron [None req-30030db8-77b3-4af4-8ae6-30653bd81e23 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: 02ffb723-d6df-42cb-93c5-d582705d1e03] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 679.501956] env[63515]: DEBUG oslo_concurrency.lockutils [None req-30987b61-ce89-4676-89be-fee7786462c9 tempest-ImagesOneServerNegativeTestJSON-52225474 tempest-ImagesOneServerNegativeTestJSON-52225474-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.028s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 679.502639] env[63515]: ERROR nova.compute.manager [None req-30987b61-ce89-4676-89be-fee7786462c9 tempest-ImagesOneServerNegativeTestJSON-52225474 tempest-ImagesOneServerNegativeTestJSON-52225474-project-member] [instance: bbce1137-d691-4633-87ff-f9f4ea257ed7] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port fb0a1841-27e9-4610-abba-889cf0217607, please check neutron logs for more information. [ 679.502639] env[63515]: ERROR nova.compute.manager [instance: bbce1137-d691-4633-87ff-f9f4ea257ed7] Traceback (most recent call last): [ 679.502639] env[63515]: ERROR nova.compute.manager [instance: bbce1137-d691-4633-87ff-f9f4ea257ed7] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 679.502639] env[63515]: ERROR nova.compute.manager [instance: bbce1137-d691-4633-87ff-f9f4ea257ed7] self.driver.spawn(context, instance, image_meta, [ 679.502639] env[63515]: ERROR nova.compute.manager [instance: bbce1137-d691-4633-87ff-f9f4ea257ed7] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 679.502639] env[63515]: ERROR nova.compute.manager [instance: bbce1137-d691-4633-87ff-f9f4ea257ed7] self._vmops.spawn(context, instance, image_meta, injected_files, [ 679.502639] env[63515]: ERROR nova.compute.manager [instance: bbce1137-d691-4633-87ff-f9f4ea257ed7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 679.502639] env[63515]: ERROR nova.compute.manager [instance: bbce1137-d691-4633-87ff-f9f4ea257ed7] vm_ref = self.build_virtual_machine(instance, [ 679.502639] env[63515]: ERROR nova.compute.manager [instance: bbce1137-d691-4633-87ff-f9f4ea257ed7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 679.502639] env[63515]: ERROR nova.compute.manager [instance: bbce1137-d691-4633-87ff-f9f4ea257ed7] vif_infos = vmwarevif.get_vif_info(self._session, [ 679.502639] env[63515]: ERROR nova.compute.manager [instance: bbce1137-d691-4633-87ff-f9f4ea257ed7] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 679.503073] env[63515]: ERROR nova.compute.manager [instance: bbce1137-d691-4633-87ff-f9f4ea257ed7] for vif in network_info: [ 679.503073] env[63515]: ERROR nova.compute.manager [instance: bbce1137-d691-4633-87ff-f9f4ea257ed7] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 679.503073] env[63515]: ERROR nova.compute.manager [instance: bbce1137-d691-4633-87ff-f9f4ea257ed7] return self._sync_wrapper(fn, *args, **kwargs) [ 679.503073] env[63515]: ERROR nova.compute.manager [instance: bbce1137-d691-4633-87ff-f9f4ea257ed7] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 679.503073] env[63515]: ERROR nova.compute.manager [instance: bbce1137-d691-4633-87ff-f9f4ea257ed7] self.wait() [ 679.503073] env[63515]: ERROR nova.compute.manager [instance: bbce1137-d691-4633-87ff-f9f4ea257ed7] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 679.503073] env[63515]: ERROR nova.compute.manager [instance: bbce1137-d691-4633-87ff-f9f4ea257ed7] self[:] = self._gt.wait() [ 679.503073] env[63515]: ERROR nova.compute.manager [instance: bbce1137-d691-4633-87ff-f9f4ea257ed7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 679.503073] env[63515]: ERROR nova.compute.manager [instance: bbce1137-d691-4633-87ff-f9f4ea257ed7] return self._exit_event.wait() [ 679.503073] env[63515]: ERROR nova.compute.manager [instance: bbce1137-d691-4633-87ff-f9f4ea257ed7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 679.503073] env[63515]: ERROR nova.compute.manager [instance: bbce1137-d691-4633-87ff-f9f4ea257ed7] current.throw(*self._exc) [ 679.503073] env[63515]: ERROR nova.compute.manager [instance: bbce1137-d691-4633-87ff-f9f4ea257ed7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 679.503073] env[63515]: ERROR nova.compute.manager [instance: bbce1137-d691-4633-87ff-f9f4ea257ed7] result = function(*args, **kwargs) [ 679.503529] env[63515]: ERROR nova.compute.manager [instance: bbce1137-d691-4633-87ff-f9f4ea257ed7] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 679.503529] env[63515]: ERROR nova.compute.manager [instance: bbce1137-d691-4633-87ff-f9f4ea257ed7] return func(*args, **kwargs) [ 679.503529] env[63515]: ERROR nova.compute.manager [instance: bbce1137-d691-4633-87ff-f9f4ea257ed7] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 679.503529] env[63515]: ERROR nova.compute.manager [instance: bbce1137-d691-4633-87ff-f9f4ea257ed7] raise e [ 679.503529] env[63515]: ERROR nova.compute.manager [instance: bbce1137-d691-4633-87ff-f9f4ea257ed7] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 679.503529] env[63515]: ERROR nova.compute.manager [instance: bbce1137-d691-4633-87ff-f9f4ea257ed7] nwinfo = self.network_api.allocate_for_instance( [ 679.503529] env[63515]: ERROR nova.compute.manager [instance: bbce1137-d691-4633-87ff-f9f4ea257ed7] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 679.503529] env[63515]: ERROR nova.compute.manager [instance: bbce1137-d691-4633-87ff-f9f4ea257ed7] created_port_ids = self._update_ports_for_instance( [ 679.503529] env[63515]: ERROR nova.compute.manager [instance: bbce1137-d691-4633-87ff-f9f4ea257ed7] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 679.503529] env[63515]: ERROR nova.compute.manager [instance: bbce1137-d691-4633-87ff-f9f4ea257ed7] with excutils.save_and_reraise_exception(): [ 679.503529] env[63515]: ERROR nova.compute.manager [instance: bbce1137-d691-4633-87ff-f9f4ea257ed7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 679.503529] env[63515]: ERROR nova.compute.manager [instance: bbce1137-d691-4633-87ff-f9f4ea257ed7] self.force_reraise() [ 679.503529] env[63515]: ERROR nova.compute.manager [instance: bbce1137-d691-4633-87ff-f9f4ea257ed7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 679.503873] env[63515]: ERROR nova.compute.manager [instance: bbce1137-d691-4633-87ff-f9f4ea257ed7] raise self.value [ 679.503873] env[63515]: ERROR nova.compute.manager [instance: bbce1137-d691-4633-87ff-f9f4ea257ed7] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 679.503873] env[63515]: ERROR nova.compute.manager [instance: bbce1137-d691-4633-87ff-f9f4ea257ed7] updated_port = self._update_port( [ 679.503873] env[63515]: ERROR nova.compute.manager [instance: bbce1137-d691-4633-87ff-f9f4ea257ed7] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 679.503873] env[63515]: ERROR nova.compute.manager [instance: bbce1137-d691-4633-87ff-f9f4ea257ed7] _ensure_no_port_binding_failure(port) [ 679.503873] env[63515]: ERROR nova.compute.manager [instance: bbce1137-d691-4633-87ff-f9f4ea257ed7] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 679.503873] env[63515]: ERROR nova.compute.manager [instance: bbce1137-d691-4633-87ff-f9f4ea257ed7] raise exception.PortBindingFailed(port_id=port['id']) [ 679.503873] env[63515]: ERROR nova.compute.manager [instance: bbce1137-d691-4633-87ff-f9f4ea257ed7] nova.exception.PortBindingFailed: Binding failed for port fb0a1841-27e9-4610-abba-889cf0217607, please check neutron logs for more information. [ 679.503873] env[63515]: ERROR nova.compute.manager [instance: bbce1137-d691-4633-87ff-f9f4ea257ed7] [ 679.503873] env[63515]: DEBUG nova.compute.utils [None req-30987b61-ce89-4676-89be-fee7786462c9 tempest-ImagesOneServerNegativeTestJSON-52225474 tempest-ImagesOneServerNegativeTestJSON-52225474-project-member] [instance: bbce1137-d691-4633-87ff-f9f4ea257ed7] Binding failed for port fb0a1841-27e9-4610-abba-889cf0217607, please check neutron logs for more information. {{(pid=63515) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 679.504791] env[63515]: DEBUG oslo_concurrency.lockutils [None req-480b7d6d-2f7f-4da3-9eb1-818f0711a940 tempest-ServerRescueTestJSON-1961208277 tempest-ServerRescueTestJSON-1961208277-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.815s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 679.507555] env[63515]: DEBUG nova.compute.manager [None req-30987b61-ce89-4676-89be-fee7786462c9 tempest-ImagesOneServerNegativeTestJSON-52225474 tempest-ImagesOneServerNegativeTestJSON-52225474-project-member] [instance: bbce1137-d691-4633-87ff-f9f4ea257ed7] Build of instance bbce1137-d691-4633-87ff-f9f4ea257ed7 was re-scheduled: Binding failed for port fb0a1841-27e9-4610-abba-889cf0217607, please check neutron logs for more information. {{(pid=63515) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 679.507997] env[63515]: DEBUG nova.compute.manager [None req-30987b61-ce89-4676-89be-fee7786462c9 tempest-ImagesOneServerNegativeTestJSON-52225474 tempest-ImagesOneServerNegativeTestJSON-52225474-project-member] [instance: bbce1137-d691-4633-87ff-f9f4ea257ed7] Unplugging VIFs for instance {{(pid=63515) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 679.508280] env[63515]: DEBUG oslo_concurrency.lockutils [None req-30987b61-ce89-4676-89be-fee7786462c9 tempest-ImagesOneServerNegativeTestJSON-52225474 tempest-ImagesOneServerNegativeTestJSON-52225474-project-member] Acquiring lock "refresh_cache-bbce1137-d691-4633-87ff-f9f4ea257ed7" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 679.508375] env[63515]: DEBUG oslo_concurrency.lockutils [None req-30987b61-ce89-4676-89be-fee7786462c9 tempest-ImagesOneServerNegativeTestJSON-52225474 tempest-ImagesOneServerNegativeTestJSON-52225474-project-member] Acquired lock "refresh_cache-bbce1137-d691-4633-87ff-f9f4ea257ed7" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 679.508531] env[63515]: DEBUG nova.network.neutron [None req-30987b61-ce89-4676-89be-fee7786462c9 tempest-ImagesOneServerNegativeTestJSON-52225474 tempest-ImagesOneServerNegativeTestJSON-52225474-project-member] [instance: bbce1137-d691-4633-87ff-f9f4ea257ed7] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 679.538226] env[63515]: DEBUG oslo_vmware.api [None req-45f6ec8d-6d74-4e46-8fdf-6b0f7c107bf4 tempest-ServerDiagnosticsTest-1160321831 tempest-ServerDiagnosticsTest-1160321831-project-member] Task: {'id': task-1110892, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.062375} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 679.538484] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-45f6ec8d-6d74-4e46-8fdf-6b0f7c107bf4 tempest-ServerDiagnosticsTest-1160321831 tempest-ServerDiagnosticsTest-1160321831-project-member] [instance: 4ad5440f-cd18-44c5-8836-0aa39824cf03] Extended root virtual disk {{(pid=63515) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 679.539251] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9cf81e38-a6a0-4416-9ffe-02a064eb0655 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.561970] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-45f6ec8d-6d74-4e46-8fdf-6b0f7c107bf4 tempest-ServerDiagnosticsTest-1160321831 tempest-ServerDiagnosticsTest-1160321831-project-member] [instance: 4ad5440f-cd18-44c5-8836-0aa39824cf03] Reconfiguring VM instance instance-0000001e to attach disk [datastore1] 4ad5440f-cd18-44c5-8836-0aa39824cf03/4ad5440f-cd18-44c5-8836-0aa39824cf03.vmdk or device None with type sparse {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 679.562848] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-18e9fac2-793d-44ab-9f76-1699080d5c89 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.577362] env[63515]: DEBUG nova.network.neutron [None req-9ecc0501-dc42-4f19-a331-8b337de58eaf tempest-AttachInterfacesUnderV243Test-1943594058 tempest-AttachInterfacesUnderV243Test-1943594058-project-member] [instance: b911a5b5-9617-4fb3-9b5e-fb8c492e4931] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 679.583767] env[63515]: DEBUG oslo_vmware.api [None req-45f6ec8d-6d74-4e46-8fdf-6b0f7c107bf4 tempest-ServerDiagnosticsTest-1160321831 tempest-ServerDiagnosticsTest-1160321831-project-member] Waiting for the task: (returnval){ [ 679.583767] env[63515]: value = "task-1110893" [ 679.583767] env[63515]: _type = "Task" [ 679.583767] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 679.591515] env[63515]: DEBUG oslo_vmware.api [None req-45f6ec8d-6d74-4e46-8fdf-6b0f7c107bf4 tempest-ServerDiagnosticsTest-1160321831 tempest-ServerDiagnosticsTest-1160321831-project-member] Task: {'id': task-1110893, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 679.776919] env[63515]: INFO nova.compute.manager [None req-30030db8-77b3-4af4-8ae6-30653bd81e23 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: 02ffb723-d6df-42cb-93c5-d582705d1e03] Took 1.13 seconds to deallocate network for instance. [ 679.873122] env[63515]: DEBUG nova.network.neutron [None req-9ecc0501-dc42-4f19-a331-8b337de58eaf tempest-AttachInterfacesUnderV243Test-1943594058 tempest-AttachInterfacesUnderV243Test-1943594058-project-member] [instance: b911a5b5-9617-4fb3-9b5e-fb8c492e4931] Updating instance_info_cache with network_info: [{"id": "9bf84cb2-7053-4129-8d8a-dc09e80b384e", "address": "fa:16:3e:c6:e8:2f", "network": {"id": "22a0cdee-4685-4b2b-9df7-2506b2373a19", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-878688764-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5086c79c531e4d039896a24a9ea29f73", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "78b49840-c3fc-455c-8491-a253ccd92bb5", "external-id": "nsx-vlan-transportzone-779", "segmentation_id": 779, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9bf84cb2-70", "ovs_interfaceid": "9bf84cb2-7053-4129-8d8a-dc09e80b384e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 680.037061] env[63515]: DEBUG nova.network.neutron [None req-30987b61-ce89-4676-89be-fee7786462c9 tempest-ImagesOneServerNegativeTestJSON-52225474 tempest-ImagesOneServerNegativeTestJSON-52225474-project-member] [instance: bbce1137-d691-4633-87ff-f9f4ea257ed7] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 680.097022] env[63515]: DEBUG oslo_vmware.api [None req-45f6ec8d-6d74-4e46-8fdf-6b0f7c107bf4 tempest-ServerDiagnosticsTest-1160321831 tempest-ServerDiagnosticsTest-1160321831-project-member] Task: {'id': task-1110893, 'name': ReconfigVM_Task, 'duration_secs': 0.301082} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 680.097022] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-45f6ec8d-6d74-4e46-8fdf-6b0f7c107bf4 tempest-ServerDiagnosticsTest-1160321831 tempest-ServerDiagnosticsTest-1160321831-project-member] [instance: 4ad5440f-cd18-44c5-8836-0aa39824cf03] Reconfigured VM instance instance-0000001e to attach disk [datastore1] 4ad5440f-cd18-44c5-8836-0aa39824cf03/4ad5440f-cd18-44c5-8836-0aa39824cf03.vmdk or device None with type sparse {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 680.097728] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-84ce5785-88cc-492e-93ff-5946cfef3c3b {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.105556] env[63515]: DEBUG oslo_vmware.api [None req-45f6ec8d-6d74-4e46-8fdf-6b0f7c107bf4 tempest-ServerDiagnosticsTest-1160321831 tempest-ServerDiagnosticsTest-1160321831-project-member] Waiting for the task: (returnval){ [ 680.105556] env[63515]: value = "task-1110894" [ 680.105556] env[63515]: _type = "Task" [ 680.105556] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 680.118092] env[63515]: DEBUG oslo_vmware.api [None req-45f6ec8d-6d74-4e46-8fdf-6b0f7c107bf4 tempest-ServerDiagnosticsTest-1160321831 tempest-ServerDiagnosticsTest-1160321831-project-member] Task: {'id': task-1110894, 'name': Rename_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 680.221099] env[63515]: DEBUG nova.network.neutron [None req-30987b61-ce89-4676-89be-fee7786462c9 tempest-ImagesOneServerNegativeTestJSON-52225474 tempest-ImagesOneServerNegativeTestJSON-52225474-project-member] [instance: bbce1137-d691-4633-87ff-f9f4ea257ed7] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 680.260534] env[63515]: DEBUG nova.compute.manager [req-11ce24a5-2b27-4009-963d-75a5915a2eab req-66c953a5-28da-4604-83a3-79180c422b32 service nova] [instance: b911a5b5-9617-4fb3-9b5e-fb8c492e4931] Received event network-changed-9bf84cb2-7053-4129-8d8a-dc09e80b384e {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 680.260534] env[63515]: DEBUG nova.compute.manager [req-11ce24a5-2b27-4009-963d-75a5915a2eab req-66c953a5-28da-4604-83a3-79180c422b32 service nova] [instance: b911a5b5-9617-4fb3-9b5e-fb8c492e4931] Refreshing instance network info cache due to event network-changed-9bf84cb2-7053-4129-8d8a-dc09e80b384e. {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 680.260707] env[63515]: DEBUG oslo_concurrency.lockutils [req-11ce24a5-2b27-4009-963d-75a5915a2eab req-66c953a5-28da-4604-83a3-79180c422b32 service nova] Acquiring lock "refresh_cache-b911a5b5-9617-4fb3-9b5e-fb8c492e4931" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 680.374502] env[63515]: DEBUG oslo_concurrency.lockutils [None req-9ecc0501-dc42-4f19-a331-8b337de58eaf tempest-AttachInterfacesUnderV243Test-1943594058 tempest-AttachInterfacesUnderV243Test-1943594058-project-member] Releasing lock "refresh_cache-b911a5b5-9617-4fb3-9b5e-fb8c492e4931" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 680.375265] env[63515]: DEBUG nova.compute.manager [None req-9ecc0501-dc42-4f19-a331-8b337de58eaf tempest-AttachInterfacesUnderV243Test-1943594058 tempest-AttachInterfacesUnderV243Test-1943594058-project-member] [instance: b911a5b5-9617-4fb3-9b5e-fb8c492e4931] Instance network_info: |[{"id": "9bf84cb2-7053-4129-8d8a-dc09e80b384e", "address": "fa:16:3e:c6:e8:2f", "network": {"id": "22a0cdee-4685-4b2b-9df7-2506b2373a19", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-878688764-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5086c79c531e4d039896a24a9ea29f73", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "78b49840-c3fc-455c-8491-a253ccd92bb5", "external-id": "nsx-vlan-transportzone-779", "segmentation_id": 779, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9bf84cb2-70", "ovs_interfaceid": "9bf84cb2-7053-4129-8d8a-dc09e80b384e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 680.375265] env[63515]: DEBUG oslo_concurrency.lockutils [req-11ce24a5-2b27-4009-963d-75a5915a2eab req-66c953a5-28da-4604-83a3-79180c422b32 service nova] Acquired lock "refresh_cache-b911a5b5-9617-4fb3-9b5e-fb8c492e4931" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 680.375440] env[63515]: DEBUG nova.network.neutron [req-11ce24a5-2b27-4009-963d-75a5915a2eab req-66c953a5-28da-4604-83a3-79180c422b32 service nova] [instance: b911a5b5-9617-4fb3-9b5e-fb8c492e4931] Refreshing network info cache for port 9bf84cb2-7053-4129-8d8a-dc09e80b384e {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 680.376906] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-9ecc0501-dc42-4f19-a331-8b337de58eaf tempest-AttachInterfacesUnderV243Test-1943594058 tempest-AttachInterfacesUnderV243Test-1943594058-project-member] [instance: b911a5b5-9617-4fb3-9b5e-fb8c492e4931] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c6:e8:2f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '78b49840-c3fc-455c-8491-a253ccd92bb5', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '9bf84cb2-7053-4129-8d8a-dc09e80b384e', 'vif_model': 'vmxnet3'}] {{(pid=63515) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 680.386482] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-9ecc0501-dc42-4f19-a331-8b337de58eaf tempest-AttachInterfacesUnderV243Test-1943594058 tempest-AttachInterfacesUnderV243Test-1943594058-project-member] Creating folder: Project (5086c79c531e4d039896a24a9ea29f73). Parent ref: group-v243370. {{(pid=63515) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 680.389498] env[63515]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-1a209c87-7cea-428c-a1d3-20cc6f4e4460 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.400857] env[63515]: INFO nova.virt.vmwareapi.vm_util [None req-9ecc0501-dc42-4f19-a331-8b337de58eaf tempest-AttachInterfacesUnderV243Test-1943594058 tempest-AttachInterfacesUnderV243Test-1943594058-project-member] Created folder: Project (5086c79c531e4d039896a24a9ea29f73) in parent group-v243370. [ 680.401083] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-9ecc0501-dc42-4f19-a331-8b337de58eaf tempest-AttachInterfacesUnderV243Test-1943594058 tempest-AttachInterfacesUnderV243Test-1943594058-project-member] Creating folder: Instances. Parent ref: group-v243383. {{(pid=63515) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 680.401332] env[63515]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-ffb9fec3-d193-437f-bdc1-dd1b035359b1 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.418020] env[63515]: INFO nova.virt.vmwareapi.vm_util [None req-9ecc0501-dc42-4f19-a331-8b337de58eaf tempest-AttachInterfacesUnderV243Test-1943594058 tempest-AttachInterfacesUnderV243Test-1943594058-project-member] Created folder: Instances in parent group-v243383. [ 680.418364] env[63515]: DEBUG oslo.service.loopingcall [None req-9ecc0501-dc42-4f19-a331-8b337de58eaf tempest-AttachInterfacesUnderV243Test-1943594058 tempest-AttachInterfacesUnderV243Test-1943594058-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 680.418568] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b911a5b5-9617-4fb3-9b5e-fb8c492e4931] Creating VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 680.419179] env[63515]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a857c364-71b5-4b86-8afc-0e0454ab0c7c {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.442270] env[63515]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 680.442270] env[63515]: value = "task-1110897" [ 680.442270] env[63515]: _type = "Task" [ 680.442270] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 680.452745] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1110897, 'name': CreateVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 680.507261] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6528aa25-eca2-43cd-a7de-47203ed604b7 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.514033] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07ba8b9d-08af-4e7d-a017-4eecb1ffd143 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.546104] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94901233-da26-49f3-9ff3-929f7cb52460 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.553342] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de8e076d-48a3-4234-8d33-886b2176ec43 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.567201] env[63515]: DEBUG nova.compute.provider_tree [None req-480b7d6d-2f7f-4da3-9eb1-818f0711a940 tempest-ServerRescueTestJSON-1961208277 tempest-ServerRescueTestJSON-1961208277-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 680.614613] env[63515]: DEBUG oslo_vmware.api [None req-45f6ec8d-6d74-4e46-8fdf-6b0f7c107bf4 tempest-ServerDiagnosticsTest-1160321831 tempest-ServerDiagnosticsTest-1160321831-project-member] Task: {'id': task-1110894, 'name': Rename_Task, 'duration_secs': 0.229644} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 680.615614] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-45f6ec8d-6d74-4e46-8fdf-6b0f7c107bf4 tempest-ServerDiagnosticsTest-1160321831 tempest-ServerDiagnosticsTest-1160321831-project-member] [instance: 4ad5440f-cd18-44c5-8836-0aa39824cf03] Powering on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 680.615614] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-96b35b04-fab6-46c8-ba60-95f75c5e905f {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.622065] env[63515]: DEBUG oslo_vmware.api [None req-45f6ec8d-6d74-4e46-8fdf-6b0f7c107bf4 tempest-ServerDiagnosticsTest-1160321831 tempest-ServerDiagnosticsTest-1160321831-project-member] Waiting for the task: (returnval){ [ 680.622065] env[63515]: value = "task-1110898" [ 680.622065] env[63515]: _type = "Task" [ 680.622065] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 680.629849] env[63515]: DEBUG oslo_vmware.api [None req-45f6ec8d-6d74-4e46-8fdf-6b0f7c107bf4 tempest-ServerDiagnosticsTest-1160321831 tempest-ServerDiagnosticsTest-1160321831-project-member] Task: {'id': task-1110898, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 680.723837] env[63515]: DEBUG oslo_concurrency.lockutils [None req-30987b61-ce89-4676-89be-fee7786462c9 tempest-ImagesOneServerNegativeTestJSON-52225474 tempest-ImagesOneServerNegativeTestJSON-52225474-project-member] Releasing lock "refresh_cache-bbce1137-d691-4633-87ff-f9f4ea257ed7" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 680.724101] env[63515]: DEBUG nova.compute.manager [None req-30987b61-ce89-4676-89be-fee7786462c9 tempest-ImagesOneServerNegativeTestJSON-52225474 tempest-ImagesOneServerNegativeTestJSON-52225474-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63515) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 680.724369] env[63515]: DEBUG nova.compute.manager [None req-30987b61-ce89-4676-89be-fee7786462c9 tempest-ImagesOneServerNegativeTestJSON-52225474 tempest-ImagesOneServerNegativeTestJSON-52225474-project-member] [instance: bbce1137-d691-4633-87ff-f9f4ea257ed7] Deallocating network for instance {{(pid=63515) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 680.724546] env[63515]: DEBUG nova.network.neutron [None req-30987b61-ce89-4676-89be-fee7786462c9 tempest-ImagesOneServerNegativeTestJSON-52225474 tempest-ImagesOneServerNegativeTestJSON-52225474-project-member] [instance: bbce1137-d691-4633-87ff-f9f4ea257ed7] deallocate_for_instance() {{(pid=63515) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 680.752942] env[63515]: DEBUG nova.network.neutron [None req-30987b61-ce89-4676-89be-fee7786462c9 tempest-ImagesOneServerNegativeTestJSON-52225474 tempest-ImagesOneServerNegativeTestJSON-52225474-project-member] [instance: bbce1137-d691-4633-87ff-f9f4ea257ed7] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 680.810091] env[63515]: INFO nova.scheduler.client.report [None req-30030db8-77b3-4af4-8ae6-30653bd81e23 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Deleted allocations for instance 02ffb723-d6df-42cb-93c5-d582705d1e03 [ 680.954023] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1110897, 'name': CreateVM_Task, 'duration_secs': 0.351882} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 680.954265] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b911a5b5-9617-4fb3-9b5e-fb8c492e4931] Created VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 680.955038] env[63515]: DEBUG oslo_concurrency.lockutils [None req-9ecc0501-dc42-4f19-a331-8b337de58eaf tempest-AttachInterfacesUnderV243Test-1943594058 tempest-AttachInterfacesUnderV243Test-1943594058-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 680.955226] env[63515]: DEBUG oslo_concurrency.lockutils [None req-9ecc0501-dc42-4f19-a331-8b337de58eaf tempest-AttachInterfacesUnderV243Test-1943594058 tempest-AttachInterfacesUnderV243Test-1943594058-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 680.955612] env[63515]: DEBUG oslo_concurrency.lockutils [None req-9ecc0501-dc42-4f19-a331-8b337de58eaf tempest-AttachInterfacesUnderV243Test-1943594058 tempest-AttachInterfacesUnderV243Test-1943594058-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 680.958067] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-23e1644d-2701-4c02-b369-ad8dda45ce18 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.963588] env[63515]: DEBUG oslo_vmware.api [None req-9ecc0501-dc42-4f19-a331-8b337de58eaf tempest-AttachInterfacesUnderV243Test-1943594058 tempest-AttachInterfacesUnderV243Test-1943594058-project-member] Waiting for the task: (returnval){ [ 680.963588] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]5214ff82-a62e-edb2-da41-4ce5ddeb2839" [ 680.963588] env[63515]: _type = "Task" [ 680.963588] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 680.977180] env[63515]: DEBUG oslo_vmware.api [None req-9ecc0501-dc42-4f19-a331-8b337de58eaf tempest-AttachInterfacesUnderV243Test-1943594058 tempest-AttachInterfacesUnderV243Test-1943594058-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]5214ff82-a62e-edb2-da41-4ce5ddeb2839, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 681.074028] env[63515]: DEBUG nova.scheduler.client.report [None req-480b7d6d-2f7f-4da3-9eb1-818f0711a940 tempest-ServerRescueTestJSON-1961208277 tempest-ServerRescueTestJSON-1961208277-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 681.133134] env[63515]: DEBUG oslo_vmware.api [None req-45f6ec8d-6d74-4e46-8fdf-6b0f7c107bf4 tempest-ServerDiagnosticsTest-1160321831 tempest-ServerDiagnosticsTest-1160321831-project-member] Task: {'id': task-1110898, 'name': PowerOnVM_Task, 'duration_secs': 0.475134} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 681.133413] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-45f6ec8d-6d74-4e46-8fdf-6b0f7c107bf4 tempest-ServerDiagnosticsTest-1160321831 tempest-ServerDiagnosticsTest-1160321831-project-member] [instance: 4ad5440f-cd18-44c5-8836-0aa39824cf03] Powered on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 681.133603] env[63515]: INFO nova.compute.manager [None req-45f6ec8d-6d74-4e46-8fdf-6b0f7c107bf4 tempest-ServerDiagnosticsTest-1160321831 tempest-ServerDiagnosticsTest-1160321831-project-member] [instance: 4ad5440f-cd18-44c5-8836-0aa39824cf03] Took 8.18 seconds to spawn the instance on the hypervisor. [ 681.133780] env[63515]: DEBUG nova.compute.manager [None req-45f6ec8d-6d74-4e46-8fdf-6b0f7c107bf4 tempest-ServerDiagnosticsTest-1160321831 tempest-ServerDiagnosticsTest-1160321831-project-member] [instance: 4ad5440f-cd18-44c5-8836-0aa39824cf03] Checking state {{(pid=63515) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 681.134600] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e05703b9-ff6a-4a89-8c96-a479f2a433b9 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.256623] env[63515]: DEBUG nova.network.neutron [None req-30987b61-ce89-4676-89be-fee7786462c9 tempest-ImagesOneServerNegativeTestJSON-52225474 tempest-ImagesOneServerNegativeTestJSON-52225474-project-member] [instance: bbce1137-d691-4633-87ff-f9f4ea257ed7] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 681.318201] env[63515]: DEBUG nova.network.neutron [req-11ce24a5-2b27-4009-963d-75a5915a2eab req-66c953a5-28da-4604-83a3-79180c422b32 service nova] [instance: b911a5b5-9617-4fb3-9b5e-fb8c492e4931] Updated VIF entry in instance network info cache for port 9bf84cb2-7053-4129-8d8a-dc09e80b384e. {{(pid=63515) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 681.318577] env[63515]: DEBUG nova.network.neutron [req-11ce24a5-2b27-4009-963d-75a5915a2eab req-66c953a5-28da-4604-83a3-79180c422b32 service nova] [instance: b911a5b5-9617-4fb3-9b5e-fb8c492e4931] Updating instance_info_cache with network_info: [{"id": "9bf84cb2-7053-4129-8d8a-dc09e80b384e", "address": "fa:16:3e:c6:e8:2f", "network": {"id": "22a0cdee-4685-4b2b-9df7-2506b2373a19", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-878688764-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5086c79c531e4d039896a24a9ea29f73", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "78b49840-c3fc-455c-8491-a253ccd92bb5", "external-id": "nsx-vlan-transportzone-779", "segmentation_id": 779, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9bf84cb2-70", "ovs_interfaceid": "9bf84cb2-7053-4129-8d8a-dc09e80b384e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 681.324438] env[63515]: DEBUG oslo_concurrency.lockutils [None req-30030db8-77b3-4af4-8ae6-30653bd81e23 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Lock "02ffb723-d6df-42cb-93c5-d582705d1e03" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 114.824s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 681.475054] env[63515]: DEBUG oslo_vmware.api [None req-9ecc0501-dc42-4f19-a331-8b337de58eaf tempest-AttachInterfacesUnderV243Test-1943594058 tempest-AttachInterfacesUnderV243Test-1943594058-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]5214ff82-a62e-edb2-da41-4ce5ddeb2839, 'name': SearchDatastore_Task, 'duration_secs': 0.008554} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 681.475374] env[63515]: DEBUG oslo_concurrency.lockutils [None req-9ecc0501-dc42-4f19-a331-8b337de58eaf tempest-AttachInterfacesUnderV243Test-1943594058 tempest-AttachInterfacesUnderV243Test-1943594058-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 681.475625] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-9ecc0501-dc42-4f19-a331-8b337de58eaf tempest-AttachInterfacesUnderV243Test-1943594058 tempest-AttachInterfacesUnderV243Test-1943594058-project-member] [instance: b911a5b5-9617-4fb3-9b5e-fb8c492e4931] Processing image 8a120570-cb06-4099-b262-554ca0ad15c5 {{(pid=63515) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 681.475862] env[63515]: DEBUG oslo_concurrency.lockutils [None req-9ecc0501-dc42-4f19-a331-8b337de58eaf tempest-AttachInterfacesUnderV243Test-1943594058 tempest-AttachInterfacesUnderV243Test-1943594058-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 681.476018] env[63515]: DEBUG oslo_concurrency.lockutils [None req-9ecc0501-dc42-4f19-a331-8b337de58eaf tempest-AttachInterfacesUnderV243Test-1943594058 tempest-AttachInterfacesUnderV243Test-1943594058-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 681.476201] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-9ecc0501-dc42-4f19-a331-8b337de58eaf tempest-AttachInterfacesUnderV243Test-1943594058 tempest-AttachInterfacesUnderV243Test-1943594058-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 681.476477] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-62919fd1-5b4b-4329-b1f6-049aaa2c3723 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.485687] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-9ecc0501-dc42-4f19-a331-8b337de58eaf tempest-AttachInterfacesUnderV243Test-1943594058 tempest-AttachInterfacesUnderV243Test-1943594058-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 681.485871] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-9ecc0501-dc42-4f19-a331-8b337de58eaf tempest-AttachInterfacesUnderV243Test-1943594058 tempest-AttachInterfacesUnderV243Test-1943594058-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63515) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 681.486599] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b7630a2c-6e31-44c6-b38b-845c3389706e {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.492074] env[63515]: DEBUG oslo_vmware.api [None req-9ecc0501-dc42-4f19-a331-8b337de58eaf tempest-AttachInterfacesUnderV243Test-1943594058 tempest-AttachInterfacesUnderV243Test-1943594058-project-member] Waiting for the task: (returnval){ [ 681.492074] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]52c08ffe-c806-3705-bec1-5908ee5151d9" [ 681.492074] env[63515]: _type = "Task" [ 681.492074] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 681.500667] env[63515]: DEBUG oslo_vmware.api [None req-9ecc0501-dc42-4f19-a331-8b337de58eaf tempest-AttachInterfacesUnderV243Test-1943594058 tempest-AttachInterfacesUnderV243Test-1943594058-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52c08ffe-c806-3705-bec1-5908ee5151d9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 681.578922] env[63515]: DEBUG oslo_concurrency.lockutils [None req-480b7d6d-2f7f-4da3-9eb1-818f0711a940 tempest-ServerRescueTestJSON-1961208277 tempest-ServerRescueTestJSON-1961208277-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.074s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 681.579685] env[63515]: ERROR nova.compute.manager [None req-480b7d6d-2f7f-4da3-9eb1-818f0711a940 tempest-ServerRescueTestJSON-1961208277 tempest-ServerRescueTestJSON-1961208277-project-member] [instance: 7f49d79e-3bda-4949-a976-7e3e6513b2c7] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port cc888fc6-b889-48e3-bb61-bfa980ef91fd, please check neutron logs for more information. [ 681.579685] env[63515]: ERROR nova.compute.manager [instance: 7f49d79e-3bda-4949-a976-7e3e6513b2c7] Traceback (most recent call last): [ 681.579685] env[63515]: ERROR nova.compute.manager [instance: 7f49d79e-3bda-4949-a976-7e3e6513b2c7] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 681.579685] env[63515]: ERROR nova.compute.manager [instance: 7f49d79e-3bda-4949-a976-7e3e6513b2c7] self.driver.spawn(context, instance, image_meta, [ 681.579685] env[63515]: ERROR nova.compute.manager [instance: 7f49d79e-3bda-4949-a976-7e3e6513b2c7] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 681.579685] env[63515]: ERROR nova.compute.manager [instance: 7f49d79e-3bda-4949-a976-7e3e6513b2c7] self._vmops.spawn(context, instance, image_meta, injected_files, [ 681.579685] env[63515]: ERROR nova.compute.manager [instance: 7f49d79e-3bda-4949-a976-7e3e6513b2c7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 681.579685] env[63515]: ERROR nova.compute.manager [instance: 7f49d79e-3bda-4949-a976-7e3e6513b2c7] vm_ref = self.build_virtual_machine(instance, [ 681.579685] env[63515]: ERROR nova.compute.manager [instance: 7f49d79e-3bda-4949-a976-7e3e6513b2c7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 681.579685] env[63515]: ERROR nova.compute.manager [instance: 7f49d79e-3bda-4949-a976-7e3e6513b2c7] vif_infos = vmwarevif.get_vif_info(self._session, [ 681.579685] env[63515]: ERROR nova.compute.manager [instance: 7f49d79e-3bda-4949-a976-7e3e6513b2c7] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 681.580183] env[63515]: ERROR nova.compute.manager [instance: 7f49d79e-3bda-4949-a976-7e3e6513b2c7] for vif in network_info: [ 681.580183] env[63515]: ERROR nova.compute.manager [instance: 7f49d79e-3bda-4949-a976-7e3e6513b2c7] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 681.580183] env[63515]: ERROR nova.compute.manager [instance: 7f49d79e-3bda-4949-a976-7e3e6513b2c7] return self._sync_wrapper(fn, *args, **kwargs) [ 681.580183] env[63515]: ERROR nova.compute.manager [instance: 7f49d79e-3bda-4949-a976-7e3e6513b2c7] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 681.580183] env[63515]: ERROR nova.compute.manager [instance: 7f49d79e-3bda-4949-a976-7e3e6513b2c7] self.wait() [ 681.580183] env[63515]: ERROR nova.compute.manager [instance: 7f49d79e-3bda-4949-a976-7e3e6513b2c7] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 681.580183] env[63515]: ERROR nova.compute.manager [instance: 7f49d79e-3bda-4949-a976-7e3e6513b2c7] self[:] = self._gt.wait() [ 681.580183] env[63515]: ERROR nova.compute.manager [instance: 7f49d79e-3bda-4949-a976-7e3e6513b2c7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 681.580183] env[63515]: ERROR nova.compute.manager [instance: 7f49d79e-3bda-4949-a976-7e3e6513b2c7] return self._exit_event.wait() [ 681.580183] env[63515]: ERROR nova.compute.manager [instance: 7f49d79e-3bda-4949-a976-7e3e6513b2c7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 681.580183] env[63515]: ERROR nova.compute.manager [instance: 7f49d79e-3bda-4949-a976-7e3e6513b2c7] current.throw(*self._exc) [ 681.580183] env[63515]: ERROR nova.compute.manager [instance: 7f49d79e-3bda-4949-a976-7e3e6513b2c7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 681.580183] env[63515]: ERROR nova.compute.manager [instance: 7f49d79e-3bda-4949-a976-7e3e6513b2c7] result = function(*args, **kwargs) [ 681.580523] env[63515]: ERROR nova.compute.manager [instance: 7f49d79e-3bda-4949-a976-7e3e6513b2c7] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 681.580523] env[63515]: ERROR nova.compute.manager [instance: 7f49d79e-3bda-4949-a976-7e3e6513b2c7] return func(*args, **kwargs) [ 681.580523] env[63515]: ERROR nova.compute.manager [instance: 7f49d79e-3bda-4949-a976-7e3e6513b2c7] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 681.580523] env[63515]: ERROR nova.compute.manager [instance: 7f49d79e-3bda-4949-a976-7e3e6513b2c7] raise e [ 681.580523] env[63515]: ERROR nova.compute.manager [instance: 7f49d79e-3bda-4949-a976-7e3e6513b2c7] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 681.580523] env[63515]: ERROR nova.compute.manager [instance: 7f49d79e-3bda-4949-a976-7e3e6513b2c7] nwinfo = self.network_api.allocate_for_instance( [ 681.580523] env[63515]: ERROR nova.compute.manager [instance: 7f49d79e-3bda-4949-a976-7e3e6513b2c7] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 681.580523] env[63515]: ERROR nova.compute.manager [instance: 7f49d79e-3bda-4949-a976-7e3e6513b2c7] created_port_ids = self._update_ports_for_instance( [ 681.580523] env[63515]: ERROR nova.compute.manager [instance: 7f49d79e-3bda-4949-a976-7e3e6513b2c7] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 681.580523] env[63515]: ERROR nova.compute.manager [instance: 7f49d79e-3bda-4949-a976-7e3e6513b2c7] with excutils.save_and_reraise_exception(): [ 681.580523] env[63515]: ERROR nova.compute.manager [instance: 7f49d79e-3bda-4949-a976-7e3e6513b2c7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 681.580523] env[63515]: ERROR nova.compute.manager [instance: 7f49d79e-3bda-4949-a976-7e3e6513b2c7] self.force_reraise() [ 681.580523] env[63515]: ERROR nova.compute.manager [instance: 7f49d79e-3bda-4949-a976-7e3e6513b2c7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 681.580828] env[63515]: ERROR nova.compute.manager [instance: 7f49d79e-3bda-4949-a976-7e3e6513b2c7] raise self.value [ 681.580828] env[63515]: ERROR nova.compute.manager [instance: 7f49d79e-3bda-4949-a976-7e3e6513b2c7] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 681.580828] env[63515]: ERROR nova.compute.manager [instance: 7f49d79e-3bda-4949-a976-7e3e6513b2c7] updated_port = self._update_port( [ 681.580828] env[63515]: ERROR nova.compute.manager [instance: 7f49d79e-3bda-4949-a976-7e3e6513b2c7] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 681.580828] env[63515]: ERROR nova.compute.manager [instance: 7f49d79e-3bda-4949-a976-7e3e6513b2c7] _ensure_no_port_binding_failure(port) [ 681.580828] env[63515]: ERROR nova.compute.manager [instance: 7f49d79e-3bda-4949-a976-7e3e6513b2c7] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 681.580828] env[63515]: ERROR nova.compute.manager [instance: 7f49d79e-3bda-4949-a976-7e3e6513b2c7] raise exception.PortBindingFailed(port_id=port['id']) [ 681.580828] env[63515]: ERROR nova.compute.manager [instance: 7f49d79e-3bda-4949-a976-7e3e6513b2c7] nova.exception.PortBindingFailed: Binding failed for port cc888fc6-b889-48e3-bb61-bfa980ef91fd, please check neutron logs for more information. [ 681.580828] env[63515]: ERROR nova.compute.manager [instance: 7f49d79e-3bda-4949-a976-7e3e6513b2c7] [ 681.580828] env[63515]: DEBUG nova.compute.utils [None req-480b7d6d-2f7f-4da3-9eb1-818f0711a940 tempest-ServerRescueTestJSON-1961208277 tempest-ServerRescueTestJSON-1961208277-project-member] [instance: 7f49d79e-3bda-4949-a976-7e3e6513b2c7] Binding failed for port cc888fc6-b889-48e3-bb61-bfa980ef91fd, please check neutron logs for more information. {{(pid=63515) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 681.582389] env[63515]: DEBUG oslo_concurrency.lockutils [None req-95251dc9-cf22-4a89-a3b0-fce88f9e12db tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.475s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 681.585295] env[63515]: INFO nova.compute.claims [None req-95251dc9-cf22-4a89-a3b0-fce88f9e12db tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: 3ad04a9f-62cf-4b1c-aab2-7dc4a219edc6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 681.588683] env[63515]: DEBUG nova.compute.manager [None req-480b7d6d-2f7f-4da3-9eb1-818f0711a940 tempest-ServerRescueTestJSON-1961208277 tempest-ServerRescueTestJSON-1961208277-project-member] [instance: 7f49d79e-3bda-4949-a976-7e3e6513b2c7] Build of instance 7f49d79e-3bda-4949-a976-7e3e6513b2c7 was re-scheduled: Binding failed for port cc888fc6-b889-48e3-bb61-bfa980ef91fd, please check neutron logs for more information. {{(pid=63515) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 681.589163] env[63515]: DEBUG nova.compute.manager [None req-480b7d6d-2f7f-4da3-9eb1-818f0711a940 tempest-ServerRescueTestJSON-1961208277 tempest-ServerRescueTestJSON-1961208277-project-member] [instance: 7f49d79e-3bda-4949-a976-7e3e6513b2c7] Unplugging VIFs for instance {{(pid=63515) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 681.589393] env[63515]: DEBUG oslo_concurrency.lockutils [None req-480b7d6d-2f7f-4da3-9eb1-818f0711a940 tempest-ServerRescueTestJSON-1961208277 tempest-ServerRescueTestJSON-1961208277-project-member] Acquiring lock "refresh_cache-7f49d79e-3bda-4949-a976-7e3e6513b2c7" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 681.589544] env[63515]: DEBUG oslo_concurrency.lockutils [None req-480b7d6d-2f7f-4da3-9eb1-818f0711a940 tempest-ServerRescueTestJSON-1961208277 tempest-ServerRescueTestJSON-1961208277-project-member] Acquired lock "refresh_cache-7f49d79e-3bda-4949-a976-7e3e6513b2c7" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 681.589701] env[63515]: DEBUG nova.network.neutron [None req-480b7d6d-2f7f-4da3-9eb1-818f0711a940 tempest-ServerRescueTestJSON-1961208277 tempest-ServerRescueTestJSON-1961208277-project-member] [instance: 7f49d79e-3bda-4949-a976-7e3e6513b2c7] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 681.651047] env[63515]: INFO nova.compute.manager [None req-45f6ec8d-6d74-4e46-8fdf-6b0f7c107bf4 tempest-ServerDiagnosticsTest-1160321831 tempest-ServerDiagnosticsTest-1160321831-project-member] [instance: 4ad5440f-cd18-44c5-8836-0aa39824cf03] Took 45.03 seconds to build instance. [ 681.761517] env[63515]: INFO nova.compute.manager [None req-30987b61-ce89-4676-89be-fee7786462c9 tempest-ImagesOneServerNegativeTestJSON-52225474 tempest-ImagesOneServerNegativeTestJSON-52225474-project-member] [instance: bbce1137-d691-4633-87ff-f9f4ea257ed7] Took 1.03 seconds to deallocate network for instance. [ 681.821391] env[63515]: DEBUG oslo_concurrency.lockutils [req-11ce24a5-2b27-4009-963d-75a5915a2eab req-66c953a5-28da-4604-83a3-79180c422b32 service nova] Releasing lock "refresh_cache-b911a5b5-9617-4fb3-9b5e-fb8c492e4931" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 681.832566] env[63515]: DEBUG nova.compute.manager [None req-c26675e2-4152-417b-99dd-0c18de4dc137 tempest-ServersTestBootFromVolume-2042560423 tempest-ServersTestBootFromVolume-2042560423-project-member] [instance: 90f4930b-aaa0-4c4b-9ab8-92aed45e200b] Starting instance... {{(pid=63515) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 682.007516] env[63515]: DEBUG oslo_vmware.api [None req-9ecc0501-dc42-4f19-a331-8b337de58eaf tempest-AttachInterfacesUnderV243Test-1943594058 tempest-AttachInterfacesUnderV243Test-1943594058-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52c08ffe-c806-3705-bec1-5908ee5151d9, 'name': SearchDatastore_Task, 'duration_secs': 0.008672} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 682.008348] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-eb192a82-6f15-4870-a4a0-2831a1cf4d6f {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.015585] env[63515]: DEBUG oslo_vmware.api [None req-9ecc0501-dc42-4f19-a331-8b337de58eaf tempest-AttachInterfacesUnderV243Test-1943594058 tempest-AttachInterfacesUnderV243Test-1943594058-project-member] Waiting for the task: (returnval){ [ 682.015585] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]52c2c14d-47d7-05bd-6730-d60d16d8f203" [ 682.015585] env[63515]: _type = "Task" [ 682.015585] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 682.024605] env[63515]: DEBUG oslo_vmware.api [None req-9ecc0501-dc42-4f19-a331-8b337de58eaf tempest-AttachInterfacesUnderV243Test-1943594058 tempest-AttachInterfacesUnderV243Test-1943594058-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52c2c14d-47d7-05bd-6730-d60d16d8f203, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 682.121177] env[63515]: DEBUG nova.network.neutron [None req-480b7d6d-2f7f-4da3-9eb1-818f0711a940 tempest-ServerRescueTestJSON-1961208277 tempest-ServerRescueTestJSON-1961208277-project-member] [instance: 7f49d79e-3bda-4949-a976-7e3e6513b2c7] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 682.155520] env[63515]: DEBUG oslo_concurrency.lockutils [None req-45f6ec8d-6d74-4e46-8fdf-6b0f7c107bf4 tempest-ServerDiagnosticsTest-1160321831 tempest-ServerDiagnosticsTest-1160321831-project-member] Lock "4ad5440f-cd18-44c5-8836-0aa39824cf03" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 102.214s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 682.246513] env[63515]: DEBUG nova.network.neutron [None req-480b7d6d-2f7f-4da3-9eb1-818f0711a940 tempest-ServerRescueTestJSON-1961208277 tempest-ServerRescueTestJSON-1961208277-project-member] [instance: 7f49d79e-3bda-4949-a976-7e3e6513b2c7] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 682.351655] env[63515]: DEBUG oslo_concurrency.lockutils [None req-c26675e2-4152-417b-99dd-0c18de4dc137 tempest-ServersTestBootFromVolume-2042560423 tempest-ServersTestBootFromVolume-2042560423-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 682.528078] env[63515]: DEBUG oslo_vmware.api [None req-9ecc0501-dc42-4f19-a331-8b337de58eaf tempest-AttachInterfacesUnderV243Test-1943594058 tempest-AttachInterfacesUnderV243Test-1943594058-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52c2c14d-47d7-05bd-6730-d60d16d8f203, 'name': SearchDatastore_Task, 'duration_secs': 0.008852} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 682.528078] env[63515]: DEBUG oslo_concurrency.lockutils [None req-9ecc0501-dc42-4f19-a331-8b337de58eaf tempest-AttachInterfacesUnderV243Test-1943594058 tempest-AttachInterfacesUnderV243Test-1943594058-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 682.528214] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-9ecc0501-dc42-4f19-a331-8b337de58eaf tempest-AttachInterfacesUnderV243Test-1943594058 tempest-AttachInterfacesUnderV243Test-1943594058-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk to [datastore1] b911a5b5-9617-4fb3-9b5e-fb8c492e4931/b911a5b5-9617-4fb3-9b5e-fb8c492e4931.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 682.528453] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-1d75117b-a978-4415-ae4b-5b4d26a65e3c {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.534605] env[63515]: DEBUG oslo_vmware.api [None req-9ecc0501-dc42-4f19-a331-8b337de58eaf tempest-AttachInterfacesUnderV243Test-1943594058 tempest-AttachInterfacesUnderV243Test-1943594058-project-member] Waiting for the task: (returnval){ [ 682.534605] env[63515]: value = "task-1110899" [ 682.534605] env[63515]: _type = "Task" [ 682.534605] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 682.541276] env[63515]: DEBUG nova.compute.manager [None req-795584a8-facd-4078-a7d4-e6ab73299a9d tempest-ServerDiagnosticsTest-2029970144 tempest-ServerDiagnosticsTest-2029970144-project-admin] [instance: 4ad5440f-cd18-44c5-8836-0aa39824cf03] Checking state {{(pid=63515) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 682.545855] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76005434-12b9-467c-bd0d-12ec62f1930a {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.551512] env[63515]: DEBUG oslo_vmware.api [None req-9ecc0501-dc42-4f19-a331-8b337de58eaf tempest-AttachInterfacesUnderV243Test-1943594058 tempest-AttachInterfacesUnderV243Test-1943594058-project-member] Task: {'id': task-1110899, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 682.552848] env[63515]: INFO nova.compute.manager [None req-795584a8-facd-4078-a7d4-e6ab73299a9d tempest-ServerDiagnosticsTest-2029970144 tempest-ServerDiagnosticsTest-2029970144-project-admin] [instance: 4ad5440f-cd18-44c5-8836-0aa39824cf03] Retrieving diagnostics [ 682.553597] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8e455f0-9142-4fd7-a1cb-6f7827431158 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.660846] env[63515]: DEBUG nova.compute.manager [None req-33f5b097-4c20-4377-b5be-7e558874e5d3 tempest-ServersV294TestFqdnHostnames-2112993702 tempest-ServersV294TestFqdnHostnames-2112993702-project-member] [instance: 396e49dd-48c0-496b-a1ec-190c33a22c5e] Starting instance... {{(pid=63515) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 682.749081] env[63515]: DEBUG oslo_concurrency.lockutils [None req-480b7d6d-2f7f-4da3-9eb1-818f0711a940 tempest-ServerRescueTestJSON-1961208277 tempest-ServerRescueTestJSON-1961208277-project-member] Releasing lock "refresh_cache-7f49d79e-3bda-4949-a976-7e3e6513b2c7" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 682.749410] env[63515]: DEBUG nova.compute.manager [None req-480b7d6d-2f7f-4da3-9eb1-818f0711a940 tempest-ServerRescueTestJSON-1961208277 tempest-ServerRescueTestJSON-1961208277-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63515) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 682.749644] env[63515]: DEBUG nova.compute.manager [None req-480b7d6d-2f7f-4da3-9eb1-818f0711a940 tempest-ServerRescueTestJSON-1961208277 tempest-ServerRescueTestJSON-1961208277-project-member] [instance: 7f49d79e-3bda-4949-a976-7e3e6513b2c7] Deallocating network for instance {{(pid=63515) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 682.749852] env[63515]: DEBUG nova.network.neutron [None req-480b7d6d-2f7f-4da3-9eb1-818f0711a940 tempest-ServerRescueTestJSON-1961208277 tempest-ServerRescueTestJSON-1961208277-project-member] [instance: 7f49d79e-3bda-4949-a976-7e3e6513b2c7] deallocate_for_instance() {{(pid=63515) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 682.778294] env[63515]: DEBUG nova.network.neutron [None req-480b7d6d-2f7f-4da3-9eb1-818f0711a940 tempest-ServerRescueTestJSON-1961208277 tempest-ServerRescueTestJSON-1961208277-project-member] [instance: 7f49d79e-3bda-4949-a976-7e3e6513b2c7] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 682.785277] env[63515]: DEBUG oslo_concurrency.lockutils [None req-549cc3f8-330f-4c6a-979d-35eedc6f8146 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Acquiring lock "84dbf321-2c70-4e08-b430-cb5a06fc6829" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 682.785505] env[63515]: DEBUG oslo_concurrency.lockutils [None req-549cc3f8-330f-4c6a-979d-35eedc6f8146 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Lock "84dbf321-2c70-4e08-b430-cb5a06fc6829" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 682.809548] env[63515]: INFO nova.scheduler.client.report [None req-30987b61-ce89-4676-89be-fee7786462c9 tempest-ImagesOneServerNegativeTestJSON-52225474 tempest-ImagesOneServerNegativeTestJSON-52225474-project-member] Deleted allocations for instance bbce1137-d691-4633-87ff-f9f4ea257ed7 [ 683.049325] env[63515]: DEBUG oslo_vmware.api [None req-9ecc0501-dc42-4f19-a331-8b337de58eaf tempest-AttachInterfacesUnderV243Test-1943594058 tempest-AttachInterfacesUnderV243Test-1943594058-project-member] Task: {'id': task-1110899, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.469205} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 683.049720] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-9ecc0501-dc42-4f19-a331-8b337de58eaf tempest-AttachInterfacesUnderV243Test-1943594058 tempest-AttachInterfacesUnderV243Test-1943594058-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk to [datastore1] b911a5b5-9617-4fb3-9b5e-fb8c492e4931/b911a5b5-9617-4fb3-9b5e-fb8c492e4931.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 683.049954] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-9ecc0501-dc42-4f19-a331-8b337de58eaf tempest-AttachInterfacesUnderV243Test-1943594058 tempest-AttachInterfacesUnderV243Test-1943594058-project-member] [instance: b911a5b5-9617-4fb3-9b5e-fb8c492e4931] Extending root virtual disk to 1048576 {{(pid=63515) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 683.050313] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-069c0696-3ca7-4f5a-9de0-46f3298d6131 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.057780] env[63515]: DEBUG oslo_vmware.api [None req-9ecc0501-dc42-4f19-a331-8b337de58eaf tempest-AttachInterfacesUnderV243Test-1943594058 tempest-AttachInterfacesUnderV243Test-1943594058-project-member] Waiting for the task: (returnval){ [ 683.057780] env[63515]: value = "task-1110900" [ 683.057780] env[63515]: _type = "Task" [ 683.057780] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 683.072521] env[63515]: DEBUG oslo_vmware.api [None req-9ecc0501-dc42-4f19-a331-8b337de58eaf tempest-AttachInterfacesUnderV243Test-1943594058 tempest-AttachInterfacesUnderV243Test-1943594058-project-member] Task: {'id': task-1110900, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 683.076360] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83f0870f-f2ce-4c78-9a3c-905181cad972 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.084636] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd502aad-f713-43cd-b4bd-70efd3131350 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.119639] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd54afc3-0d68-47f9-a05a-2a56a0d8a60a {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.128973] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f89bb5a5-255c-43dd-9b16-47b3134fcfbd {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.143325] env[63515]: DEBUG nova.compute.provider_tree [None req-95251dc9-cf22-4a89-a3b0-fce88f9e12db tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 683.183903] env[63515]: DEBUG oslo_concurrency.lockutils [None req-33f5b097-4c20-4377-b5be-7e558874e5d3 tempest-ServersV294TestFqdnHostnames-2112993702 tempest-ServersV294TestFqdnHostnames-2112993702-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 683.281593] env[63515]: DEBUG nova.network.neutron [None req-480b7d6d-2f7f-4da3-9eb1-818f0711a940 tempest-ServerRescueTestJSON-1961208277 tempest-ServerRescueTestJSON-1961208277-project-member] [instance: 7f49d79e-3bda-4949-a976-7e3e6513b2c7] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 683.321597] env[63515]: DEBUG oslo_concurrency.lockutils [None req-30987b61-ce89-4676-89be-fee7786462c9 tempest-ImagesOneServerNegativeTestJSON-52225474 tempest-ImagesOneServerNegativeTestJSON-52225474-project-member] Lock "bbce1137-d691-4633-87ff-f9f4ea257ed7" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 113.614s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 683.568462] env[63515]: DEBUG oslo_vmware.api [None req-9ecc0501-dc42-4f19-a331-8b337de58eaf tempest-AttachInterfacesUnderV243Test-1943594058 tempest-AttachInterfacesUnderV243Test-1943594058-project-member] Task: {'id': task-1110900, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.065017} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 683.568940] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-9ecc0501-dc42-4f19-a331-8b337de58eaf tempest-AttachInterfacesUnderV243Test-1943594058 tempest-AttachInterfacesUnderV243Test-1943594058-project-member] [instance: b911a5b5-9617-4fb3-9b5e-fb8c492e4931] Extended root virtual disk {{(pid=63515) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 683.569884] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c9eeace-b026-4094-9248-c2636ded8630 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.593076] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-9ecc0501-dc42-4f19-a331-8b337de58eaf tempest-AttachInterfacesUnderV243Test-1943594058 tempest-AttachInterfacesUnderV243Test-1943594058-project-member] [instance: b911a5b5-9617-4fb3-9b5e-fb8c492e4931] Reconfiguring VM instance instance-0000001f to attach disk [datastore1] b911a5b5-9617-4fb3-9b5e-fb8c492e4931/b911a5b5-9617-4fb3-9b5e-fb8c492e4931.vmdk or device None with type sparse {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 683.593571] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-27b3f147-659c-4b01-89f2-d1aabf7ef38b {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.615677] env[63515]: DEBUG oslo_vmware.api [None req-9ecc0501-dc42-4f19-a331-8b337de58eaf tempest-AttachInterfacesUnderV243Test-1943594058 tempest-AttachInterfacesUnderV243Test-1943594058-project-member] Waiting for the task: (returnval){ [ 683.615677] env[63515]: value = "task-1110901" [ 683.615677] env[63515]: _type = "Task" [ 683.615677] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 683.623647] env[63515]: DEBUG oslo_vmware.api [None req-9ecc0501-dc42-4f19-a331-8b337de58eaf tempest-AttachInterfacesUnderV243Test-1943594058 tempest-AttachInterfacesUnderV243Test-1943594058-project-member] Task: {'id': task-1110901, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 683.648373] env[63515]: DEBUG nova.scheduler.client.report [None req-95251dc9-cf22-4a89-a3b0-fce88f9e12db tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 683.784397] env[63515]: INFO nova.compute.manager [None req-480b7d6d-2f7f-4da3-9eb1-818f0711a940 tempest-ServerRescueTestJSON-1961208277 tempest-ServerRescueTestJSON-1961208277-project-member] [instance: 7f49d79e-3bda-4949-a976-7e3e6513b2c7] Took 1.03 seconds to deallocate network for instance. [ 683.824386] env[63515]: DEBUG nova.compute.manager [None req-4fa686a6-dc45-4603-a4ee-7107c810126f tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] [instance: 4087b2c4-6ed1-4b68-8b78-a36e34d935b1] Starting instance... {{(pid=63515) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 683.891044] env[63515]: DEBUG oslo_concurrency.lockutils [None req-d937370c-7531-4295-8d43-07fafc09e90c tempest-ServerDiagnosticsTest-1160321831 tempest-ServerDiagnosticsTest-1160321831-project-member] Acquiring lock "4ad5440f-cd18-44c5-8836-0aa39824cf03" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 683.891044] env[63515]: DEBUG oslo_concurrency.lockutils [None req-d937370c-7531-4295-8d43-07fafc09e90c tempest-ServerDiagnosticsTest-1160321831 tempest-ServerDiagnosticsTest-1160321831-project-member] Lock "4ad5440f-cd18-44c5-8836-0aa39824cf03" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 683.891176] env[63515]: DEBUG oslo_concurrency.lockutils [None req-d937370c-7531-4295-8d43-07fafc09e90c tempest-ServerDiagnosticsTest-1160321831 tempest-ServerDiagnosticsTest-1160321831-project-member] Acquiring lock "4ad5440f-cd18-44c5-8836-0aa39824cf03-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 683.892033] env[63515]: DEBUG oslo_concurrency.lockutils [None req-d937370c-7531-4295-8d43-07fafc09e90c tempest-ServerDiagnosticsTest-1160321831 tempest-ServerDiagnosticsTest-1160321831-project-member] Lock "4ad5440f-cd18-44c5-8836-0aa39824cf03-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 683.892033] env[63515]: DEBUG oslo_concurrency.lockutils [None req-d937370c-7531-4295-8d43-07fafc09e90c tempest-ServerDiagnosticsTest-1160321831 tempest-ServerDiagnosticsTest-1160321831-project-member] Lock "4ad5440f-cd18-44c5-8836-0aa39824cf03-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 683.894100] env[63515]: INFO nova.compute.manager [None req-d937370c-7531-4295-8d43-07fafc09e90c tempest-ServerDiagnosticsTest-1160321831 tempest-ServerDiagnosticsTest-1160321831-project-member] [instance: 4ad5440f-cd18-44c5-8836-0aa39824cf03] Terminating instance [ 683.896066] env[63515]: DEBUG nova.compute.manager [None req-d937370c-7531-4295-8d43-07fafc09e90c tempest-ServerDiagnosticsTest-1160321831 tempest-ServerDiagnosticsTest-1160321831-project-member] [instance: 4ad5440f-cd18-44c5-8836-0aa39824cf03] Start destroying the instance on the hypervisor. {{(pid=63515) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 683.896271] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-d937370c-7531-4295-8d43-07fafc09e90c tempest-ServerDiagnosticsTest-1160321831 tempest-ServerDiagnosticsTest-1160321831-project-member] [instance: 4ad5440f-cd18-44c5-8836-0aa39824cf03] Destroying instance {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 683.897201] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef8caf94-2e98-4bd5-8c53-3c70cd192a1c {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.906451] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-d937370c-7531-4295-8d43-07fafc09e90c tempest-ServerDiagnosticsTest-1160321831 tempest-ServerDiagnosticsTest-1160321831-project-member] [instance: 4ad5440f-cd18-44c5-8836-0aa39824cf03] Powering off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 683.906647] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ff083f17-c30f-47ed-8acb-02d10a496d75 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.913182] env[63515]: DEBUG oslo_vmware.api [None req-d937370c-7531-4295-8d43-07fafc09e90c tempest-ServerDiagnosticsTest-1160321831 tempest-ServerDiagnosticsTest-1160321831-project-member] Waiting for the task: (returnval){ [ 683.913182] env[63515]: value = "task-1110902" [ 683.913182] env[63515]: _type = "Task" [ 683.913182] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 683.922890] env[63515]: DEBUG oslo_vmware.api [None req-d937370c-7531-4295-8d43-07fafc09e90c tempest-ServerDiagnosticsTest-1160321831 tempest-ServerDiagnosticsTest-1160321831-project-member] Task: {'id': task-1110902, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 684.126865] env[63515]: DEBUG oslo_vmware.api [None req-9ecc0501-dc42-4f19-a331-8b337de58eaf tempest-AttachInterfacesUnderV243Test-1943594058 tempest-AttachInterfacesUnderV243Test-1943594058-project-member] Task: {'id': task-1110901, 'name': ReconfigVM_Task, 'duration_secs': 0.281625} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 684.127362] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-9ecc0501-dc42-4f19-a331-8b337de58eaf tempest-AttachInterfacesUnderV243Test-1943594058 tempest-AttachInterfacesUnderV243Test-1943594058-project-member] [instance: b911a5b5-9617-4fb3-9b5e-fb8c492e4931] Reconfigured VM instance instance-0000001f to attach disk [datastore1] b911a5b5-9617-4fb3-9b5e-fb8c492e4931/b911a5b5-9617-4fb3-9b5e-fb8c492e4931.vmdk or device None with type sparse {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 684.128140] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-07541827-6c29-4337-85cc-78694d9d6302 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.135786] env[63515]: DEBUG oslo_vmware.api [None req-9ecc0501-dc42-4f19-a331-8b337de58eaf tempest-AttachInterfacesUnderV243Test-1943594058 tempest-AttachInterfacesUnderV243Test-1943594058-project-member] Waiting for the task: (returnval){ [ 684.135786] env[63515]: value = "task-1110903" [ 684.135786] env[63515]: _type = "Task" [ 684.135786] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 684.146822] env[63515]: DEBUG oslo_vmware.api [None req-9ecc0501-dc42-4f19-a331-8b337de58eaf tempest-AttachInterfacesUnderV243Test-1943594058 tempest-AttachInterfacesUnderV243Test-1943594058-project-member] Task: {'id': task-1110903, 'name': Rename_Task} progress is 5%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 684.155819] env[63515]: DEBUG oslo_concurrency.lockutils [None req-95251dc9-cf22-4a89-a3b0-fce88f9e12db tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.573s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 684.156333] env[63515]: DEBUG nova.compute.manager [None req-95251dc9-cf22-4a89-a3b0-fce88f9e12db tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: 3ad04a9f-62cf-4b1c-aab2-7dc4a219edc6] Start building networks asynchronously for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 684.159075] env[63515]: DEBUG oslo_concurrency.lockutils [None req-b9c961a0-a559-4690-abb6-579359ed447c tempest-ServerPasswordTestJSON-433685770 tempest-ServerPasswordTestJSON-433685770-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.129s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 684.346939] env[63515]: DEBUG oslo_concurrency.lockutils [None req-4fa686a6-dc45-4603-a4ee-7107c810126f tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 684.424969] env[63515]: DEBUG oslo_vmware.api [None req-d937370c-7531-4295-8d43-07fafc09e90c tempest-ServerDiagnosticsTest-1160321831 tempest-ServerDiagnosticsTest-1160321831-project-member] Task: {'id': task-1110902, 'name': PowerOffVM_Task, 'duration_secs': 0.363261} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 684.425484] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-d937370c-7531-4295-8d43-07fafc09e90c tempest-ServerDiagnosticsTest-1160321831 tempest-ServerDiagnosticsTest-1160321831-project-member] [instance: 4ad5440f-cd18-44c5-8836-0aa39824cf03] Powered off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 684.425724] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-d937370c-7531-4295-8d43-07fafc09e90c tempest-ServerDiagnosticsTest-1160321831 tempest-ServerDiagnosticsTest-1160321831-project-member] [instance: 4ad5440f-cd18-44c5-8836-0aa39824cf03] Unregistering the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 684.425990] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-cb73be09-e7fd-4cce-9a53-7ec4fafb6b49 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.448649] env[63515]: DEBUG oslo_concurrency.lockutils [None req-2786351a-303c-45a8-b63b-65e218eb6ef6 tempest-ServersTestJSON-324454119 tempest-ServersTestJSON-324454119-project-member] Acquiring lock "70ed982f-affd-4dd1-bc90-c64e7c6d49d2" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 684.449075] env[63515]: DEBUG oslo_concurrency.lockutils [None req-2786351a-303c-45a8-b63b-65e218eb6ef6 tempest-ServersTestJSON-324454119 tempest-ServersTestJSON-324454119-project-member] Lock "70ed982f-affd-4dd1-bc90-c64e7c6d49d2" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 684.493014] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-d937370c-7531-4295-8d43-07fafc09e90c tempest-ServerDiagnosticsTest-1160321831 tempest-ServerDiagnosticsTest-1160321831-project-member] [instance: 4ad5440f-cd18-44c5-8836-0aa39824cf03] Unregistered the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 684.493014] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-d937370c-7531-4295-8d43-07fafc09e90c tempest-ServerDiagnosticsTest-1160321831 tempest-ServerDiagnosticsTest-1160321831-project-member] [instance: 4ad5440f-cd18-44c5-8836-0aa39824cf03] Deleting contents of the VM from datastore datastore1 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 684.493014] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-d937370c-7531-4295-8d43-07fafc09e90c tempest-ServerDiagnosticsTest-1160321831 tempest-ServerDiagnosticsTest-1160321831-project-member] Deleting the datastore file [datastore1] 4ad5440f-cd18-44c5-8836-0aa39824cf03 {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 684.493014] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-1f344714-02f2-4085-967f-a14a37395de5 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.498692] env[63515]: DEBUG oslo_vmware.api [None req-d937370c-7531-4295-8d43-07fafc09e90c tempest-ServerDiagnosticsTest-1160321831 tempest-ServerDiagnosticsTest-1160321831-project-member] Waiting for the task: (returnval){ [ 684.498692] env[63515]: value = "task-1110905" [ 684.498692] env[63515]: _type = "Task" [ 684.498692] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 684.509206] env[63515]: DEBUG oslo_vmware.api [None req-d937370c-7531-4295-8d43-07fafc09e90c tempest-ServerDiagnosticsTest-1160321831 tempest-ServerDiagnosticsTest-1160321831-project-member] Task: {'id': task-1110905, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 684.646557] env[63515]: DEBUG oslo_vmware.api [None req-9ecc0501-dc42-4f19-a331-8b337de58eaf tempest-AttachInterfacesUnderV243Test-1943594058 tempest-AttachInterfacesUnderV243Test-1943594058-project-member] Task: {'id': task-1110903, 'name': Rename_Task, 'duration_secs': 0.141072} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 684.647016] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-9ecc0501-dc42-4f19-a331-8b337de58eaf tempest-AttachInterfacesUnderV243Test-1943594058 tempest-AttachInterfacesUnderV243Test-1943594058-project-member] [instance: b911a5b5-9617-4fb3-9b5e-fb8c492e4931] Powering on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 684.647346] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1073b4e9-eb17-4f4b-9024-a119982bb143 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.653697] env[63515]: DEBUG oslo_vmware.api [None req-9ecc0501-dc42-4f19-a331-8b337de58eaf tempest-AttachInterfacesUnderV243Test-1943594058 tempest-AttachInterfacesUnderV243Test-1943594058-project-member] Waiting for the task: (returnval){ [ 684.653697] env[63515]: value = "task-1110906" [ 684.653697] env[63515]: _type = "Task" [ 684.653697] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 684.667673] env[63515]: DEBUG nova.compute.utils [None req-95251dc9-cf22-4a89-a3b0-fce88f9e12db tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Using /dev/sd instead of None {{(pid=63515) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 684.672767] env[63515]: DEBUG oslo_vmware.api [None req-9ecc0501-dc42-4f19-a331-8b337de58eaf tempest-AttachInterfacesUnderV243Test-1943594058 tempest-AttachInterfacesUnderV243Test-1943594058-project-member] Task: {'id': task-1110906, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 684.674227] env[63515]: DEBUG nova.compute.manager [None req-95251dc9-cf22-4a89-a3b0-fce88f9e12db tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: 3ad04a9f-62cf-4b1c-aab2-7dc4a219edc6] Allocating IP information in the background. {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 684.674227] env[63515]: DEBUG nova.network.neutron [None req-95251dc9-cf22-4a89-a3b0-fce88f9e12db tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: 3ad04a9f-62cf-4b1c-aab2-7dc4a219edc6] allocate_for_instance() {{(pid=63515) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 684.721533] env[63515]: DEBUG nova.policy [None req-95251dc9-cf22-4a89-a3b0-fce88f9e12db tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f58ac812970845dc825202b842feb5d0', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ea099ce07bfb4a8da014d9303cf552eb', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63515) authorize /opt/stack/nova/nova/policy.py:201}} [ 684.816641] env[63515]: INFO nova.scheduler.client.report [None req-480b7d6d-2f7f-4da3-9eb1-818f0711a940 tempest-ServerRescueTestJSON-1961208277 tempest-ServerRescueTestJSON-1961208277-project-member] Deleted allocations for instance 7f49d79e-3bda-4949-a976-7e3e6513b2c7 [ 685.009190] env[63515]: DEBUG oslo_vmware.api [None req-d937370c-7531-4295-8d43-07fafc09e90c tempest-ServerDiagnosticsTest-1160321831 tempest-ServerDiagnosticsTest-1160321831-project-member] Task: {'id': task-1110905, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.134237} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 685.009461] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-d937370c-7531-4295-8d43-07fafc09e90c tempest-ServerDiagnosticsTest-1160321831 tempest-ServerDiagnosticsTest-1160321831-project-member] Deleted the datastore file {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 685.009652] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-d937370c-7531-4295-8d43-07fafc09e90c tempest-ServerDiagnosticsTest-1160321831 tempest-ServerDiagnosticsTest-1160321831-project-member] [instance: 4ad5440f-cd18-44c5-8836-0aa39824cf03] Deleted contents of the VM from datastore datastore1 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 685.009822] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-d937370c-7531-4295-8d43-07fafc09e90c tempest-ServerDiagnosticsTest-1160321831 tempest-ServerDiagnosticsTest-1160321831-project-member] [instance: 4ad5440f-cd18-44c5-8836-0aa39824cf03] Instance destroyed {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 685.009989] env[63515]: INFO nova.compute.manager [None req-d937370c-7531-4295-8d43-07fafc09e90c tempest-ServerDiagnosticsTest-1160321831 tempest-ServerDiagnosticsTest-1160321831-project-member] [instance: 4ad5440f-cd18-44c5-8836-0aa39824cf03] Took 1.11 seconds to destroy the instance on the hypervisor. [ 685.010247] env[63515]: DEBUG oslo.service.loopingcall [None req-d937370c-7531-4295-8d43-07fafc09e90c tempest-ServerDiagnosticsTest-1160321831 tempest-ServerDiagnosticsTest-1160321831-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 685.010429] env[63515]: DEBUG nova.compute.manager [-] [instance: 4ad5440f-cd18-44c5-8836-0aa39824cf03] Deallocating network for instance {{(pid=63515) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 685.010557] env[63515]: DEBUG nova.network.neutron [-] [instance: 4ad5440f-cd18-44c5-8836-0aa39824cf03] deallocate_for_instance() {{(pid=63515) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 685.063197] env[63515]: DEBUG nova.network.neutron [None req-95251dc9-cf22-4a89-a3b0-fce88f9e12db tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: 3ad04a9f-62cf-4b1c-aab2-7dc4a219edc6] Successfully created port: 3dee3eb6-af97-422d-ac4f-74a994909386 {{(pid=63515) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 685.137256] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b118fb8d-8ec1-48d3-9610-bc759895c3bb {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.145201] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1415b7b-eff0-493d-bd0f-fc292d09ec4a {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.175698] env[63515]: DEBUG nova.compute.manager [None req-95251dc9-cf22-4a89-a3b0-fce88f9e12db tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: 3ad04a9f-62cf-4b1c-aab2-7dc4a219edc6] Start building block device mappings for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 685.181532] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5550fcd0-e9d0-49eb-b547-bcb4aca4fc72 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.197496] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bfc7e90c-6eb8-45b8-8d9d-40d1b59b80ca {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.199011] env[63515]: DEBUG oslo_vmware.api [None req-9ecc0501-dc42-4f19-a331-8b337de58eaf tempest-AttachInterfacesUnderV243Test-1943594058 tempest-AttachInterfacesUnderV243Test-1943594058-project-member] Task: {'id': task-1110906, 'name': PowerOnVM_Task, 'duration_secs': 0.467915} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 685.199011] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-9ecc0501-dc42-4f19-a331-8b337de58eaf tempest-AttachInterfacesUnderV243Test-1943594058 tempest-AttachInterfacesUnderV243Test-1943594058-project-member] [instance: b911a5b5-9617-4fb3-9b5e-fb8c492e4931] Powered on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 685.199174] env[63515]: INFO nova.compute.manager [None req-9ecc0501-dc42-4f19-a331-8b337de58eaf tempest-AttachInterfacesUnderV243Test-1943594058 tempest-AttachInterfacesUnderV243Test-1943594058-project-member] [instance: b911a5b5-9617-4fb3-9b5e-fb8c492e4931] Took 7.75 seconds to spawn the instance on the hypervisor. [ 685.199266] env[63515]: DEBUG nova.compute.manager [None req-9ecc0501-dc42-4f19-a331-8b337de58eaf tempest-AttachInterfacesUnderV243Test-1943594058 tempest-AttachInterfacesUnderV243Test-1943594058-project-member] [instance: b911a5b5-9617-4fb3-9b5e-fb8c492e4931] Checking state {{(pid=63515) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 685.203771] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-227398f2-0bc5-41a3-bfca-3ac4aa4fcdef {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.215342] env[63515]: DEBUG nova.compute.provider_tree [None req-b9c961a0-a559-4690-abb6-579359ed447c tempest-ServerPasswordTestJSON-433685770 tempest-ServerPasswordTestJSON-433685770-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 685.332251] env[63515]: DEBUG oslo_concurrency.lockutils [None req-480b7d6d-2f7f-4da3-9eb1-818f0711a940 tempest-ServerRescueTestJSON-1961208277 tempest-ServerRescueTestJSON-1961208277-project-member] Lock "7f49d79e-3bda-4949-a976-7e3e6513b2c7" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 113.701s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 685.720290] env[63515]: DEBUG nova.scheduler.client.report [None req-b9c961a0-a559-4690-abb6-579359ed447c tempest-ServerPasswordTestJSON-433685770 tempest-ServerPasswordTestJSON-433685770-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 685.737390] env[63515]: INFO nova.compute.manager [None req-9ecc0501-dc42-4f19-a331-8b337de58eaf tempest-AttachInterfacesUnderV243Test-1943594058 tempest-AttachInterfacesUnderV243Test-1943594058-project-member] [instance: b911a5b5-9617-4fb3-9b5e-fb8c492e4931] Took 44.64 seconds to build instance. [ 685.837615] env[63515]: DEBUG nova.compute.manager [None req-44a3080f-f2e2-4f0c-b244-122beee2aea9 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] [instance: 87c468d9-9594-4804-b461-527f01f6118f] Starting instance... {{(pid=63515) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 685.858174] env[63515]: DEBUG nova.compute.manager [req-cb4c0d84-0183-4546-959e-361e151e5571 req-6d6cc9b9-5039-4bb3-8f67-5c5e009ca9e2 service nova] [instance: 4ad5440f-cd18-44c5-8836-0aa39824cf03] Received event network-vif-deleted-c4df35f7-5206-4bec-ac71-0578da364e1a {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 685.858398] env[63515]: INFO nova.compute.manager [req-cb4c0d84-0183-4546-959e-361e151e5571 req-6d6cc9b9-5039-4bb3-8f67-5c5e009ca9e2 service nova] [instance: 4ad5440f-cd18-44c5-8836-0aa39824cf03] Neutron deleted interface c4df35f7-5206-4bec-ac71-0578da364e1a; detaching it from the instance and deleting it from the info cache [ 685.858536] env[63515]: DEBUG nova.network.neutron [req-cb4c0d84-0183-4546-959e-361e151e5571 req-6d6cc9b9-5039-4bb3-8f67-5c5e009ca9e2 service nova] [instance: 4ad5440f-cd18-44c5-8836-0aa39824cf03] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 686.108974] env[63515]: DEBUG nova.network.neutron [-] [instance: 4ad5440f-cd18-44c5-8836-0aa39824cf03] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 686.190880] env[63515]: DEBUG nova.compute.manager [None req-95251dc9-cf22-4a89-a3b0-fce88f9e12db tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: 3ad04a9f-62cf-4b1c-aab2-7dc4a219edc6] Start spawning the instance on the hypervisor. {{(pid=63515) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 686.218910] env[63515]: DEBUG nova.virt.hardware [None req-95251dc9-cf22-4a89-a3b0-fce88f9e12db tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T02:52:14Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T02:51:56Z,direct_url=,disk_format='vmdk',id=8a120570-cb06-4099-b262-554ca0ad15c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b569255cc43e42e7ae3f2b4ad37c6ef4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T02:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 686.219188] env[63515]: DEBUG nova.virt.hardware [None req-95251dc9-cf22-4a89-a3b0-fce88f9e12db tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Flavor limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 686.219349] env[63515]: DEBUG nova.virt.hardware [None req-95251dc9-cf22-4a89-a3b0-fce88f9e12db tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Image limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 686.219541] env[63515]: DEBUG nova.virt.hardware [None req-95251dc9-cf22-4a89-a3b0-fce88f9e12db tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Flavor pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 686.219691] env[63515]: DEBUG nova.virt.hardware [None req-95251dc9-cf22-4a89-a3b0-fce88f9e12db tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Image pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 686.219837] env[63515]: DEBUG nova.virt.hardware [None req-95251dc9-cf22-4a89-a3b0-fce88f9e12db tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 686.220055] env[63515]: DEBUG nova.virt.hardware [None req-95251dc9-cf22-4a89-a3b0-fce88f9e12db tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 686.220249] env[63515]: DEBUG nova.virt.hardware [None req-95251dc9-cf22-4a89-a3b0-fce88f9e12db tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 686.220698] env[63515]: DEBUG nova.virt.hardware [None req-95251dc9-cf22-4a89-a3b0-fce88f9e12db tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Got 1 possible topologies {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 686.220698] env[63515]: DEBUG nova.virt.hardware [None req-95251dc9-cf22-4a89-a3b0-fce88f9e12db tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 686.220956] env[63515]: DEBUG nova.virt.hardware [None req-95251dc9-cf22-4a89-a3b0-fce88f9e12db tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 686.222070] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aba48fac-afe4-43cb-90b0-48551a2c8106 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.230258] env[63515]: DEBUG oslo_concurrency.lockutils [None req-b9c961a0-a559-4690-abb6-579359ed447c tempest-ServerPasswordTestJSON-433685770 tempest-ServerPasswordTestJSON-433685770-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.071s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 686.230842] env[63515]: ERROR nova.compute.manager [None req-b9c961a0-a559-4690-abb6-579359ed447c tempest-ServerPasswordTestJSON-433685770 tempest-ServerPasswordTestJSON-433685770-project-member] [instance: 07499601-62bc-4c31-b295-23f34a6e2e91] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port b497be91-987b-42ce-9f91-0850d502656f, please check neutron logs for more information. [ 686.230842] env[63515]: ERROR nova.compute.manager [instance: 07499601-62bc-4c31-b295-23f34a6e2e91] Traceback (most recent call last): [ 686.230842] env[63515]: ERROR nova.compute.manager [instance: 07499601-62bc-4c31-b295-23f34a6e2e91] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 686.230842] env[63515]: ERROR nova.compute.manager [instance: 07499601-62bc-4c31-b295-23f34a6e2e91] self.driver.spawn(context, instance, image_meta, [ 686.230842] env[63515]: ERROR nova.compute.manager [instance: 07499601-62bc-4c31-b295-23f34a6e2e91] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 686.230842] env[63515]: ERROR nova.compute.manager [instance: 07499601-62bc-4c31-b295-23f34a6e2e91] self._vmops.spawn(context, instance, image_meta, injected_files, [ 686.230842] env[63515]: ERROR nova.compute.manager [instance: 07499601-62bc-4c31-b295-23f34a6e2e91] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 686.230842] env[63515]: ERROR nova.compute.manager [instance: 07499601-62bc-4c31-b295-23f34a6e2e91] vm_ref = self.build_virtual_machine(instance, [ 686.230842] env[63515]: ERROR nova.compute.manager [instance: 07499601-62bc-4c31-b295-23f34a6e2e91] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 686.230842] env[63515]: ERROR nova.compute.manager [instance: 07499601-62bc-4c31-b295-23f34a6e2e91] vif_infos = vmwarevif.get_vif_info(self._session, [ 686.230842] env[63515]: ERROR nova.compute.manager [instance: 07499601-62bc-4c31-b295-23f34a6e2e91] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 686.231151] env[63515]: ERROR nova.compute.manager [instance: 07499601-62bc-4c31-b295-23f34a6e2e91] for vif in network_info: [ 686.231151] env[63515]: ERROR nova.compute.manager [instance: 07499601-62bc-4c31-b295-23f34a6e2e91] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 686.231151] env[63515]: ERROR nova.compute.manager [instance: 07499601-62bc-4c31-b295-23f34a6e2e91] return self._sync_wrapper(fn, *args, **kwargs) [ 686.231151] env[63515]: ERROR nova.compute.manager [instance: 07499601-62bc-4c31-b295-23f34a6e2e91] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 686.231151] env[63515]: ERROR nova.compute.manager [instance: 07499601-62bc-4c31-b295-23f34a6e2e91] self.wait() [ 686.231151] env[63515]: ERROR nova.compute.manager [instance: 07499601-62bc-4c31-b295-23f34a6e2e91] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 686.231151] env[63515]: ERROR nova.compute.manager [instance: 07499601-62bc-4c31-b295-23f34a6e2e91] self[:] = self._gt.wait() [ 686.231151] env[63515]: ERROR nova.compute.manager [instance: 07499601-62bc-4c31-b295-23f34a6e2e91] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 686.231151] env[63515]: ERROR nova.compute.manager [instance: 07499601-62bc-4c31-b295-23f34a6e2e91] return self._exit_event.wait() [ 686.231151] env[63515]: ERROR nova.compute.manager [instance: 07499601-62bc-4c31-b295-23f34a6e2e91] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 686.231151] env[63515]: ERROR nova.compute.manager [instance: 07499601-62bc-4c31-b295-23f34a6e2e91] current.throw(*self._exc) [ 686.231151] env[63515]: ERROR nova.compute.manager [instance: 07499601-62bc-4c31-b295-23f34a6e2e91] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 686.231151] env[63515]: ERROR nova.compute.manager [instance: 07499601-62bc-4c31-b295-23f34a6e2e91] result = function(*args, **kwargs) [ 686.231467] env[63515]: ERROR nova.compute.manager [instance: 07499601-62bc-4c31-b295-23f34a6e2e91] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 686.231467] env[63515]: ERROR nova.compute.manager [instance: 07499601-62bc-4c31-b295-23f34a6e2e91] return func(*args, **kwargs) [ 686.231467] env[63515]: ERROR nova.compute.manager [instance: 07499601-62bc-4c31-b295-23f34a6e2e91] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 686.231467] env[63515]: ERROR nova.compute.manager [instance: 07499601-62bc-4c31-b295-23f34a6e2e91] raise e [ 686.231467] env[63515]: ERROR nova.compute.manager [instance: 07499601-62bc-4c31-b295-23f34a6e2e91] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 686.231467] env[63515]: ERROR nova.compute.manager [instance: 07499601-62bc-4c31-b295-23f34a6e2e91] nwinfo = self.network_api.allocate_for_instance( [ 686.231467] env[63515]: ERROR nova.compute.manager [instance: 07499601-62bc-4c31-b295-23f34a6e2e91] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 686.231467] env[63515]: ERROR nova.compute.manager [instance: 07499601-62bc-4c31-b295-23f34a6e2e91] created_port_ids = self._update_ports_for_instance( [ 686.231467] env[63515]: ERROR nova.compute.manager [instance: 07499601-62bc-4c31-b295-23f34a6e2e91] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 686.231467] env[63515]: ERROR nova.compute.manager [instance: 07499601-62bc-4c31-b295-23f34a6e2e91] with excutils.save_and_reraise_exception(): [ 686.231467] env[63515]: ERROR nova.compute.manager [instance: 07499601-62bc-4c31-b295-23f34a6e2e91] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 686.231467] env[63515]: ERROR nova.compute.manager [instance: 07499601-62bc-4c31-b295-23f34a6e2e91] self.force_reraise() [ 686.231467] env[63515]: ERROR nova.compute.manager [instance: 07499601-62bc-4c31-b295-23f34a6e2e91] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 686.231877] env[63515]: ERROR nova.compute.manager [instance: 07499601-62bc-4c31-b295-23f34a6e2e91] raise self.value [ 686.231877] env[63515]: ERROR nova.compute.manager [instance: 07499601-62bc-4c31-b295-23f34a6e2e91] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 686.231877] env[63515]: ERROR nova.compute.manager [instance: 07499601-62bc-4c31-b295-23f34a6e2e91] updated_port = self._update_port( [ 686.231877] env[63515]: ERROR nova.compute.manager [instance: 07499601-62bc-4c31-b295-23f34a6e2e91] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 686.231877] env[63515]: ERROR nova.compute.manager [instance: 07499601-62bc-4c31-b295-23f34a6e2e91] _ensure_no_port_binding_failure(port) [ 686.231877] env[63515]: ERROR nova.compute.manager [instance: 07499601-62bc-4c31-b295-23f34a6e2e91] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 686.231877] env[63515]: ERROR nova.compute.manager [instance: 07499601-62bc-4c31-b295-23f34a6e2e91] raise exception.PortBindingFailed(port_id=port['id']) [ 686.231877] env[63515]: ERROR nova.compute.manager [instance: 07499601-62bc-4c31-b295-23f34a6e2e91] nova.exception.PortBindingFailed: Binding failed for port b497be91-987b-42ce-9f91-0850d502656f, please check neutron logs for more information. [ 686.231877] env[63515]: ERROR nova.compute.manager [instance: 07499601-62bc-4c31-b295-23f34a6e2e91] [ 686.236200] env[63515]: DEBUG nova.compute.utils [None req-b9c961a0-a559-4690-abb6-579359ed447c tempest-ServerPasswordTestJSON-433685770 tempest-ServerPasswordTestJSON-433685770-project-member] [instance: 07499601-62bc-4c31-b295-23f34a6e2e91] Binding failed for port b497be91-987b-42ce-9f91-0850d502656f, please check neutron logs for more information. {{(pid=63515) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 686.237467] env[63515]: DEBUG oslo_concurrency.lockutils [None req-c7572d27-5b85-4acb-89ec-c56462911d65 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.516s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 686.239012] env[63515]: INFO nova.compute.claims [None req-c7572d27-5b85-4acb-89ec-c56462911d65 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] [instance: 5b42f744-fdd6-45b1-8563-896869648c23] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 686.243154] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a162d67-a5da-4ab8-a8bd-85fd5d076f83 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.247710] env[63515]: DEBUG nova.compute.manager [None req-b9c961a0-a559-4690-abb6-579359ed447c tempest-ServerPasswordTestJSON-433685770 tempest-ServerPasswordTestJSON-433685770-project-member] [instance: 07499601-62bc-4c31-b295-23f34a6e2e91] Build of instance 07499601-62bc-4c31-b295-23f34a6e2e91 was re-scheduled: Binding failed for port b497be91-987b-42ce-9f91-0850d502656f, please check neutron logs for more information. {{(pid=63515) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 686.248180] env[63515]: DEBUG nova.compute.manager [None req-b9c961a0-a559-4690-abb6-579359ed447c tempest-ServerPasswordTestJSON-433685770 tempest-ServerPasswordTestJSON-433685770-project-member] [instance: 07499601-62bc-4c31-b295-23f34a6e2e91] Unplugging VIFs for instance {{(pid=63515) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 686.248397] env[63515]: DEBUG oslo_concurrency.lockutils [None req-b9c961a0-a559-4690-abb6-579359ed447c tempest-ServerPasswordTestJSON-433685770 tempest-ServerPasswordTestJSON-433685770-project-member] Acquiring lock "refresh_cache-07499601-62bc-4c31-b295-23f34a6e2e91" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 686.248561] env[63515]: DEBUG oslo_concurrency.lockutils [None req-b9c961a0-a559-4690-abb6-579359ed447c tempest-ServerPasswordTestJSON-433685770 tempest-ServerPasswordTestJSON-433685770-project-member] Acquired lock "refresh_cache-07499601-62bc-4c31-b295-23f34a6e2e91" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 686.248691] env[63515]: DEBUG nova.network.neutron [None req-b9c961a0-a559-4690-abb6-579359ed447c tempest-ServerPasswordTestJSON-433685770 tempest-ServerPasswordTestJSON-433685770-project-member] [instance: 07499601-62bc-4c31-b295-23f34a6e2e91] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 686.253018] env[63515]: DEBUG oslo_concurrency.lockutils [None req-9ecc0501-dc42-4f19-a331-8b337de58eaf tempest-AttachInterfacesUnderV243Test-1943594058 tempest-AttachInterfacesUnderV243Test-1943594058-project-member] Lock "b911a5b5-9617-4fb3-9b5e-fb8c492e4931" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 103.032s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 686.361082] env[63515]: DEBUG oslo_concurrency.lockutils [None req-44a3080f-f2e2-4f0c-b244-122beee2aea9 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 686.361640] env[63515]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b1838f28-c8da-4c7e-abac-7fcd6bc17439 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.370682] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca809669-880b-4fa9-ad00-472087b0d426 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.394214] env[63515]: DEBUG nova.compute.manager [req-cb4c0d84-0183-4546-959e-361e151e5571 req-6d6cc9b9-5039-4bb3-8f67-5c5e009ca9e2 service nova] [instance: 4ad5440f-cd18-44c5-8836-0aa39824cf03] Detach interface failed, port_id=c4df35f7-5206-4bec-ac71-0578da364e1a, reason: Instance 4ad5440f-cd18-44c5-8836-0aa39824cf03 could not be found. {{(pid=63515) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 686.612040] env[63515]: INFO nova.compute.manager [-] [instance: 4ad5440f-cd18-44c5-8836-0aa39824cf03] Took 1.60 seconds to deallocate network for instance. [ 686.765114] env[63515]: DEBUG nova.compute.manager [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] [instance: 48668736-df27-4f2a-94d9-132f5b49701b] Starting instance... {{(pid=63515) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 686.781920] env[63515]: DEBUG nova.network.neutron [None req-b9c961a0-a559-4690-abb6-579359ed447c tempest-ServerPasswordTestJSON-433685770 tempest-ServerPasswordTestJSON-433685770-project-member] [instance: 07499601-62bc-4c31-b295-23f34a6e2e91] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 686.994957] env[63515]: DEBUG nova.network.neutron [None req-b9c961a0-a559-4690-abb6-579359ed447c tempest-ServerPasswordTestJSON-433685770 tempest-ServerPasswordTestJSON-433685770-project-member] [instance: 07499601-62bc-4c31-b295-23f34a6e2e91] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 687.119139] env[63515]: DEBUG oslo_concurrency.lockutils [None req-d937370c-7531-4295-8d43-07fafc09e90c tempest-ServerDiagnosticsTest-1160321831 tempest-ServerDiagnosticsTest-1160321831-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 687.145480] env[63515]: DEBUG nova.network.neutron [None req-95251dc9-cf22-4a89-a3b0-fce88f9e12db tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: 3ad04a9f-62cf-4b1c-aab2-7dc4a219edc6] Successfully updated port: 3dee3eb6-af97-422d-ac4f-74a994909386 {{(pid=63515) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 687.211106] env[63515]: DEBUG nova.compute.manager [req-aa500bff-46d1-4e89-9df1-372da274825f req-daf91347-393b-447a-b755-d604e67839c5 service nova] [instance: b911a5b5-9617-4fb3-9b5e-fb8c492e4931] Received event network-changed-9bf84cb2-7053-4129-8d8a-dc09e80b384e {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 687.211375] env[63515]: DEBUG nova.compute.manager [req-aa500bff-46d1-4e89-9df1-372da274825f req-daf91347-393b-447a-b755-d604e67839c5 service nova] [instance: b911a5b5-9617-4fb3-9b5e-fb8c492e4931] Refreshing instance network info cache due to event network-changed-9bf84cb2-7053-4129-8d8a-dc09e80b384e. {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 687.213592] env[63515]: DEBUG oslo_concurrency.lockutils [req-aa500bff-46d1-4e89-9df1-372da274825f req-daf91347-393b-447a-b755-d604e67839c5 service nova] Acquiring lock "refresh_cache-b911a5b5-9617-4fb3-9b5e-fb8c492e4931" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 687.213592] env[63515]: DEBUG oslo_concurrency.lockutils [req-aa500bff-46d1-4e89-9df1-372da274825f req-daf91347-393b-447a-b755-d604e67839c5 service nova] Acquired lock "refresh_cache-b911a5b5-9617-4fb3-9b5e-fb8c492e4931" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 687.213592] env[63515]: DEBUG nova.network.neutron [req-aa500bff-46d1-4e89-9df1-372da274825f req-daf91347-393b-447a-b755-d604e67839c5 service nova] [instance: b911a5b5-9617-4fb3-9b5e-fb8c492e4931] Refreshing network info cache for port 9bf84cb2-7053-4129-8d8a-dc09e80b384e {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 687.292834] env[63515]: DEBUG oslo_concurrency.lockutils [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 687.499329] env[63515]: DEBUG oslo_concurrency.lockutils [None req-b9c961a0-a559-4690-abb6-579359ed447c tempest-ServerPasswordTestJSON-433685770 tempest-ServerPasswordTestJSON-433685770-project-member] Releasing lock "refresh_cache-07499601-62bc-4c31-b295-23f34a6e2e91" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 687.499575] env[63515]: DEBUG nova.compute.manager [None req-b9c961a0-a559-4690-abb6-579359ed447c tempest-ServerPasswordTestJSON-433685770 tempest-ServerPasswordTestJSON-433685770-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63515) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 687.499755] env[63515]: DEBUG nova.compute.manager [None req-b9c961a0-a559-4690-abb6-579359ed447c tempest-ServerPasswordTestJSON-433685770 tempest-ServerPasswordTestJSON-433685770-project-member] [instance: 07499601-62bc-4c31-b295-23f34a6e2e91] Deallocating network for instance {{(pid=63515) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 687.499928] env[63515]: DEBUG nova.network.neutron [None req-b9c961a0-a559-4690-abb6-579359ed447c tempest-ServerPasswordTestJSON-433685770 tempest-ServerPasswordTestJSON-433685770-project-member] [instance: 07499601-62bc-4c31-b295-23f34a6e2e91] deallocate_for_instance() {{(pid=63515) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 687.531058] env[63515]: DEBUG nova.network.neutron [None req-b9c961a0-a559-4690-abb6-579359ed447c tempest-ServerPasswordTestJSON-433685770 tempest-ServerPasswordTestJSON-433685770-project-member] [instance: 07499601-62bc-4c31-b295-23f34a6e2e91] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 687.648114] env[63515]: DEBUG oslo_concurrency.lockutils [None req-95251dc9-cf22-4a89-a3b0-fce88f9e12db tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Acquiring lock "refresh_cache-3ad04a9f-62cf-4b1c-aab2-7dc4a219edc6" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 687.648114] env[63515]: DEBUG oslo_concurrency.lockutils [None req-95251dc9-cf22-4a89-a3b0-fce88f9e12db tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Acquired lock "refresh_cache-3ad04a9f-62cf-4b1c-aab2-7dc4a219edc6" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 687.648114] env[63515]: DEBUG nova.network.neutron [None req-95251dc9-cf22-4a89-a3b0-fce88f9e12db tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: 3ad04a9f-62cf-4b1c-aab2-7dc4a219edc6] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 687.693241] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee05efef-d27f-4511-aaf8-ec51767e1a00 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.704606] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37d8ceab-458a-497e-9e27-6ada35ed0758 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.745860] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-621f52cf-0078-4f43-ab3e-cf15aa775cc2 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.754499] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d483ef5-fd21-450d-b2e8-b9f2698dd153 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.769632] env[63515]: DEBUG nova.compute.provider_tree [None req-c7572d27-5b85-4acb-89ec-c56462911d65 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 687.894389] env[63515]: DEBUG nova.compute.manager [req-6ed861fd-ad61-4428-8b3a-483ea20fc17a req-be94b21a-14d1-46ab-9a63-c4a68db99d1d service nova] [instance: 3ad04a9f-62cf-4b1c-aab2-7dc4a219edc6] Received event network-vif-plugged-3dee3eb6-af97-422d-ac4f-74a994909386 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 687.894633] env[63515]: DEBUG oslo_concurrency.lockutils [req-6ed861fd-ad61-4428-8b3a-483ea20fc17a req-be94b21a-14d1-46ab-9a63-c4a68db99d1d service nova] Acquiring lock "3ad04a9f-62cf-4b1c-aab2-7dc4a219edc6-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 687.894832] env[63515]: DEBUG oslo_concurrency.lockutils [req-6ed861fd-ad61-4428-8b3a-483ea20fc17a req-be94b21a-14d1-46ab-9a63-c4a68db99d1d service nova] Lock "3ad04a9f-62cf-4b1c-aab2-7dc4a219edc6-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 687.895197] env[63515]: DEBUG oslo_concurrency.lockutils [req-6ed861fd-ad61-4428-8b3a-483ea20fc17a req-be94b21a-14d1-46ab-9a63-c4a68db99d1d service nova] Lock "3ad04a9f-62cf-4b1c-aab2-7dc4a219edc6-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 687.895428] env[63515]: DEBUG nova.compute.manager [req-6ed861fd-ad61-4428-8b3a-483ea20fc17a req-be94b21a-14d1-46ab-9a63-c4a68db99d1d service nova] [instance: 3ad04a9f-62cf-4b1c-aab2-7dc4a219edc6] No waiting events found dispatching network-vif-plugged-3dee3eb6-af97-422d-ac4f-74a994909386 {{(pid=63515) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 687.895558] env[63515]: WARNING nova.compute.manager [req-6ed861fd-ad61-4428-8b3a-483ea20fc17a req-be94b21a-14d1-46ab-9a63-c4a68db99d1d service nova] [instance: 3ad04a9f-62cf-4b1c-aab2-7dc4a219edc6] Received unexpected event network-vif-plugged-3dee3eb6-af97-422d-ac4f-74a994909386 for instance with vm_state building and task_state spawning. [ 687.895764] env[63515]: DEBUG nova.compute.manager [req-6ed861fd-ad61-4428-8b3a-483ea20fc17a req-be94b21a-14d1-46ab-9a63-c4a68db99d1d service nova] [instance: 3ad04a9f-62cf-4b1c-aab2-7dc4a219edc6] Received event network-changed-3dee3eb6-af97-422d-ac4f-74a994909386 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 687.896024] env[63515]: DEBUG nova.compute.manager [req-6ed861fd-ad61-4428-8b3a-483ea20fc17a req-be94b21a-14d1-46ab-9a63-c4a68db99d1d service nova] [instance: 3ad04a9f-62cf-4b1c-aab2-7dc4a219edc6] Refreshing instance network info cache due to event network-changed-3dee3eb6-af97-422d-ac4f-74a994909386. {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 687.896298] env[63515]: DEBUG oslo_concurrency.lockutils [req-6ed861fd-ad61-4428-8b3a-483ea20fc17a req-be94b21a-14d1-46ab-9a63-c4a68db99d1d service nova] Acquiring lock "refresh_cache-3ad04a9f-62cf-4b1c-aab2-7dc4a219edc6" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 688.026070] env[63515]: DEBUG oslo_concurrency.lockutils [None req-8407c401-88c7-47ad-b91e-e7693f87e0f1 tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] Acquiring lock "192137b0-03e5-4bc4-b911-4b4f1a874f74" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 688.026913] env[63515]: DEBUG oslo_concurrency.lockutils [None req-8407c401-88c7-47ad-b91e-e7693f87e0f1 tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] Lock "192137b0-03e5-4bc4-b911-4b4f1a874f74" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.002s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 688.034179] env[63515]: DEBUG nova.network.neutron [None req-b9c961a0-a559-4690-abb6-579359ed447c tempest-ServerPasswordTestJSON-433685770 tempest-ServerPasswordTestJSON-433685770-project-member] [instance: 07499601-62bc-4c31-b295-23f34a6e2e91] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 688.106684] env[63515]: DEBUG nova.network.neutron [req-aa500bff-46d1-4e89-9df1-372da274825f req-daf91347-393b-447a-b755-d604e67839c5 service nova] [instance: b911a5b5-9617-4fb3-9b5e-fb8c492e4931] Updated VIF entry in instance network info cache for port 9bf84cb2-7053-4129-8d8a-dc09e80b384e. {{(pid=63515) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 688.107591] env[63515]: DEBUG nova.network.neutron [req-aa500bff-46d1-4e89-9df1-372da274825f req-daf91347-393b-447a-b755-d604e67839c5 service nova] [instance: b911a5b5-9617-4fb3-9b5e-fb8c492e4931] Updating instance_info_cache with network_info: [{"id": "9bf84cb2-7053-4129-8d8a-dc09e80b384e", "address": "fa:16:3e:c6:e8:2f", "network": {"id": "22a0cdee-4685-4b2b-9df7-2506b2373a19", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-878688764-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.147", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5086c79c531e4d039896a24a9ea29f73", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "78b49840-c3fc-455c-8491-a253ccd92bb5", "external-id": "nsx-vlan-transportzone-779", "segmentation_id": 779, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9bf84cb2-70", "ovs_interfaceid": "9bf84cb2-7053-4129-8d8a-dc09e80b384e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 688.191353] env[63515]: DEBUG nova.network.neutron [None req-95251dc9-cf22-4a89-a3b0-fce88f9e12db tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: 3ad04a9f-62cf-4b1c-aab2-7dc4a219edc6] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 688.272130] env[63515]: DEBUG nova.scheduler.client.report [None req-c7572d27-5b85-4acb-89ec-c56462911d65 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 688.359716] env[63515]: DEBUG nova.network.neutron [None req-95251dc9-cf22-4a89-a3b0-fce88f9e12db tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: 3ad04a9f-62cf-4b1c-aab2-7dc4a219edc6] Updating instance_info_cache with network_info: [{"id": "3dee3eb6-af97-422d-ac4f-74a994909386", "address": "fa:16:3e:df:b4:4d", "network": {"id": "3e5914c8-0cd4-49ee-81b4-1f3e280440bb", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-641632277-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ea099ce07bfb4a8da014d9303cf552eb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "225b6979-9329-403b-91fa-138bd41f6e83", "external-id": "nsx-vlan-transportzone-38", "segmentation_id": 38, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3dee3eb6-af", "ovs_interfaceid": "3dee3eb6-af97-422d-ac4f-74a994909386", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 688.536703] env[63515]: INFO nova.compute.manager [None req-b9c961a0-a559-4690-abb6-579359ed447c tempest-ServerPasswordTestJSON-433685770 tempest-ServerPasswordTestJSON-433685770-project-member] [instance: 07499601-62bc-4c31-b295-23f34a6e2e91] Took 1.04 seconds to deallocate network for instance. [ 688.609880] env[63515]: DEBUG oslo_concurrency.lockutils [req-aa500bff-46d1-4e89-9df1-372da274825f req-daf91347-393b-447a-b755-d604e67839c5 service nova] Releasing lock "refresh_cache-b911a5b5-9617-4fb3-9b5e-fb8c492e4931" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 688.779581] env[63515]: DEBUG oslo_concurrency.lockutils [None req-c7572d27-5b85-4acb-89ec-c56462911d65 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.542s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 688.780205] env[63515]: DEBUG nova.compute.manager [None req-c7572d27-5b85-4acb-89ec-c56462911d65 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] [instance: 5b42f744-fdd6-45b1-8563-896869648c23] Start building networks asynchronously for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 688.782609] env[63515]: DEBUG oslo_concurrency.lockutils [None req-fa933c74-4a5b-4cc9-a499-9ea16a38d8b7 tempest-ServerActionsTestJSON-1591637270 tempest-ServerActionsTestJSON-1591637270-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.167s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 688.862040] env[63515]: DEBUG oslo_concurrency.lockutils [None req-95251dc9-cf22-4a89-a3b0-fce88f9e12db tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Releasing lock "refresh_cache-3ad04a9f-62cf-4b1c-aab2-7dc4a219edc6" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 688.862381] env[63515]: DEBUG nova.compute.manager [None req-95251dc9-cf22-4a89-a3b0-fce88f9e12db tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: 3ad04a9f-62cf-4b1c-aab2-7dc4a219edc6] Instance network_info: |[{"id": "3dee3eb6-af97-422d-ac4f-74a994909386", "address": "fa:16:3e:df:b4:4d", "network": {"id": "3e5914c8-0cd4-49ee-81b4-1f3e280440bb", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-641632277-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ea099ce07bfb4a8da014d9303cf552eb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "225b6979-9329-403b-91fa-138bd41f6e83", "external-id": "nsx-vlan-transportzone-38", "segmentation_id": 38, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3dee3eb6-af", "ovs_interfaceid": "3dee3eb6-af97-422d-ac4f-74a994909386", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 688.862674] env[63515]: DEBUG oslo_concurrency.lockutils [req-6ed861fd-ad61-4428-8b3a-483ea20fc17a req-be94b21a-14d1-46ab-9a63-c4a68db99d1d service nova] Acquired lock "refresh_cache-3ad04a9f-62cf-4b1c-aab2-7dc4a219edc6" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 688.862845] env[63515]: DEBUG nova.network.neutron [req-6ed861fd-ad61-4428-8b3a-483ea20fc17a req-be94b21a-14d1-46ab-9a63-c4a68db99d1d service nova] [instance: 3ad04a9f-62cf-4b1c-aab2-7dc4a219edc6] Refreshing network info cache for port 3dee3eb6-af97-422d-ac4f-74a994909386 {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 688.864060] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-95251dc9-cf22-4a89-a3b0-fce88f9e12db tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: 3ad04a9f-62cf-4b1c-aab2-7dc4a219edc6] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:df:b4:4d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '225b6979-9329-403b-91fa-138bd41f6e83', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '3dee3eb6-af97-422d-ac4f-74a994909386', 'vif_model': 'vmxnet3'}] {{(pid=63515) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 688.871377] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-95251dc9-cf22-4a89-a3b0-fce88f9e12db tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Creating folder: Project (ea099ce07bfb4a8da014d9303cf552eb). Parent ref: group-v243370. {{(pid=63515) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 688.872328] env[63515]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-32e1a406-6f03-43f3-aad0-ce37235c4bf9 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.883569] env[63515]: INFO nova.virt.vmwareapi.vm_util [None req-95251dc9-cf22-4a89-a3b0-fce88f9e12db tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Created folder: Project (ea099ce07bfb4a8da014d9303cf552eb) in parent group-v243370. [ 688.883754] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-95251dc9-cf22-4a89-a3b0-fce88f9e12db tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Creating folder: Instances. Parent ref: group-v243386. {{(pid=63515) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 688.883978] env[63515]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-8232bd98-9d31-4565-969d-d2a44c9889d1 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.893500] env[63515]: INFO nova.virt.vmwareapi.vm_util [None req-95251dc9-cf22-4a89-a3b0-fce88f9e12db tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Created folder: Instances in parent group-v243386. [ 688.893728] env[63515]: DEBUG oslo.service.loopingcall [None req-95251dc9-cf22-4a89-a3b0-fce88f9e12db tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 688.893945] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3ad04a9f-62cf-4b1c-aab2-7dc4a219edc6] Creating VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 688.894129] env[63515]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-5f8cd56f-df12-4a34-b851-fbd39739fdcb {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.912599] env[63515]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 688.912599] env[63515]: value = "task-1110909" [ 688.912599] env[63515]: _type = "Task" [ 688.912599] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 688.919774] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1110909, 'name': CreateVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 689.288049] env[63515]: DEBUG nova.compute.utils [None req-c7572d27-5b85-4acb-89ec-c56462911d65 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] Using /dev/sd instead of None {{(pid=63515) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 689.292403] env[63515]: DEBUG nova.compute.manager [None req-c7572d27-5b85-4acb-89ec-c56462911d65 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] [instance: 5b42f744-fdd6-45b1-8563-896869648c23] Allocating IP information in the background. {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 689.292591] env[63515]: DEBUG nova.network.neutron [None req-c7572d27-5b85-4acb-89ec-c56462911d65 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] [instance: 5b42f744-fdd6-45b1-8563-896869648c23] allocate_for_instance() {{(pid=63515) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 689.354624] env[63515]: DEBUG nova.policy [None req-c7572d27-5b85-4acb-89ec-c56462911d65 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8b6f6567b7ce4f44a860c1f5cca16964', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a1f130751200408982bffb31a54bbcea', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63515) authorize /opt/stack/nova/nova/policy.py:201}} [ 689.421806] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1110909, 'name': CreateVM_Task, 'duration_secs': 0.385974} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 689.427801] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3ad04a9f-62cf-4b1c-aab2-7dc4a219edc6] Created VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 689.430107] env[63515]: DEBUG oslo_concurrency.lockutils [None req-95251dc9-cf22-4a89-a3b0-fce88f9e12db tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 689.430107] env[63515]: DEBUG oslo_concurrency.lockutils [None req-95251dc9-cf22-4a89-a3b0-fce88f9e12db tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 689.433231] env[63515]: DEBUG oslo_concurrency.lockutils [None req-95251dc9-cf22-4a89-a3b0-fce88f9e12db tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 689.433823] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2c13c0fc-dda1-4077-923e-c596bf37ec33 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.441900] env[63515]: DEBUG oslo_vmware.api [None req-95251dc9-cf22-4a89-a3b0-fce88f9e12db tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Waiting for the task: (returnval){ [ 689.441900] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]5260117a-6b00-5634-98e0-c4e25f15f957" [ 689.441900] env[63515]: _type = "Task" [ 689.441900] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 689.452976] env[63515]: DEBUG oslo_vmware.api [None req-95251dc9-cf22-4a89-a3b0-fce88f9e12db tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]5260117a-6b00-5634-98e0-c4e25f15f957, 'name': SearchDatastore_Task, 'duration_secs': 0.011717} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 689.453263] env[63515]: DEBUG oslo_concurrency.lockutils [None req-95251dc9-cf22-4a89-a3b0-fce88f9e12db tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 689.453488] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-95251dc9-cf22-4a89-a3b0-fce88f9e12db tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: 3ad04a9f-62cf-4b1c-aab2-7dc4a219edc6] Processing image 8a120570-cb06-4099-b262-554ca0ad15c5 {{(pid=63515) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 689.453710] env[63515]: DEBUG oslo_concurrency.lockutils [None req-95251dc9-cf22-4a89-a3b0-fce88f9e12db tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 689.453908] env[63515]: DEBUG oslo_concurrency.lockutils [None req-95251dc9-cf22-4a89-a3b0-fce88f9e12db tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 689.454059] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-95251dc9-cf22-4a89-a3b0-fce88f9e12db tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 689.456492] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-56296684-d421-466f-8168-1bff0d0c1d92 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.464167] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-95251dc9-cf22-4a89-a3b0-fce88f9e12db tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 689.464723] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-95251dc9-cf22-4a89-a3b0-fce88f9e12db tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63515) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 689.465234] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1ffd61ac-01ed-4ff4-bfb7-31a5b65a33de {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.473933] env[63515]: DEBUG oslo_vmware.api [None req-95251dc9-cf22-4a89-a3b0-fce88f9e12db tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Waiting for the task: (returnval){ [ 689.473933] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]52f6c78a-96ee-3925-05b0-df665471f6f4" [ 689.473933] env[63515]: _type = "Task" [ 689.473933] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 689.481209] env[63515]: DEBUG oslo_vmware.api [None req-95251dc9-cf22-4a89-a3b0-fce88f9e12db tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52f6c78a-96ee-3925-05b0-df665471f6f4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 689.581452] env[63515]: INFO nova.scheduler.client.report [None req-b9c961a0-a559-4690-abb6-579359ed447c tempest-ServerPasswordTestJSON-433685770 tempest-ServerPasswordTestJSON-433685770-project-member] Deleted allocations for instance 07499601-62bc-4c31-b295-23f34a6e2e91 [ 689.713239] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b398ba8-e29a-46bf-bf3a-a0e4dbe3cf96 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.726267] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25b55a2a-9de0-44f6-83ca-f8a38a99fc63 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.761599] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6870524-2c75-4fd1-b9e0-e0eb439cabed {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.769496] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aedbe94d-9cca-44be-ae7f-eb8fb6d87df7 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.784706] env[63515]: DEBUG nova.compute.provider_tree [None req-fa933c74-4a5b-4cc9-a499-9ea16a38d8b7 tempest-ServerActionsTestJSON-1591637270 tempest-ServerActionsTestJSON-1591637270-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 689.795133] env[63515]: DEBUG nova.compute.manager [None req-c7572d27-5b85-4acb-89ec-c56462911d65 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] [instance: 5b42f744-fdd6-45b1-8563-896869648c23] Start building block device mappings for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 689.983279] env[63515]: DEBUG oslo_vmware.api [None req-95251dc9-cf22-4a89-a3b0-fce88f9e12db tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52f6c78a-96ee-3925-05b0-df665471f6f4, 'name': SearchDatastore_Task, 'duration_secs': 0.014879} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 689.984080] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-07fe3d87-15b7-4bf5-a6fb-cc6d59e54c97 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.989069] env[63515]: DEBUG oslo_vmware.api [None req-95251dc9-cf22-4a89-a3b0-fce88f9e12db tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Waiting for the task: (returnval){ [ 689.989069] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]52981911-69f9-95ce-cbf6-52ae8a928c2d" [ 689.989069] env[63515]: _type = "Task" [ 689.989069] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 689.993087] env[63515]: DEBUG nova.network.neutron [req-6ed861fd-ad61-4428-8b3a-483ea20fc17a req-be94b21a-14d1-46ab-9a63-c4a68db99d1d service nova] [instance: 3ad04a9f-62cf-4b1c-aab2-7dc4a219edc6] Updated VIF entry in instance network info cache for port 3dee3eb6-af97-422d-ac4f-74a994909386. {{(pid=63515) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 689.993441] env[63515]: DEBUG nova.network.neutron [req-6ed861fd-ad61-4428-8b3a-483ea20fc17a req-be94b21a-14d1-46ab-9a63-c4a68db99d1d service nova] [instance: 3ad04a9f-62cf-4b1c-aab2-7dc4a219edc6] Updating instance_info_cache with network_info: [{"id": "3dee3eb6-af97-422d-ac4f-74a994909386", "address": "fa:16:3e:df:b4:4d", "network": {"id": "3e5914c8-0cd4-49ee-81b4-1f3e280440bb", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-641632277-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ea099ce07bfb4a8da014d9303cf552eb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "225b6979-9329-403b-91fa-138bd41f6e83", "external-id": "nsx-vlan-transportzone-38", "segmentation_id": 38, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3dee3eb6-af", "ovs_interfaceid": "3dee3eb6-af97-422d-ac4f-74a994909386", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 689.998017] env[63515]: DEBUG oslo_vmware.api [None req-95251dc9-cf22-4a89-a3b0-fce88f9e12db tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52981911-69f9-95ce-cbf6-52ae8a928c2d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 690.089866] env[63515]: DEBUG oslo_concurrency.lockutils [None req-b9c961a0-a559-4690-abb6-579359ed447c tempest-ServerPasswordTestJSON-433685770 tempest-ServerPasswordTestJSON-433685770-project-member] Lock "07499601-62bc-4c31-b295-23f34a6e2e91" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 116.925s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 690.173643] env[63515]: DEBUG nova.network.neutron [None req-c7572d27-5b85-4acb-89ec-c56462911d65 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] [instance: 5b42f744-fdd6-45b1-8563-896869648c23] Successfully created port: 3a630639-5423-4e62-9beb-ee1e7054f47d {{(pid=63515) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 690.289641] env[63515]: DEBUG nova.scheduler.client.report [None req-fa933c74-4a5b-4cc9-a499-9ea16a38d8b7 tempest-ServerActionsTestJSON-1591637270 tempest-ServerActionsTestJSON-1591637270-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 690.500491] env[63515]: DEBUG oslo_concurrency.lockutils [req-6ed861fd-ad61-4428-8b3a-483ea20fc17a req-be94b21a-14d1-46ab-9a63-c4a68db99d1d service nova] Releasing lock "refresh_cache-3ad04a9f-62cf-4b1c-aab2-7dc4a219edc6" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 690.500879] env[63515]: DEBUG oslo_vmware.api [None req-95251dc9-cf22-4a89-a3b0-fce88f9e12db tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52981911-69f9-95ce-cbf6-52ae8a928c2d, 'name': SearchDatastore_Task, 'duration_secs': 0.009017} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 690.501123] env[63515]: DEBUG oslo_concurrency.lockutils [None req-95251dc9-cf22-4a89-a3b0-fce88f9e12db tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 690.501369] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-95251dc9-cf22-4a89-a3b0-fce88f9e12db tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk to [datastore1] 3ad04a9f-62cf-4b1c-aab2-7dc4a219edc6/3ad04a9f-62cf-4b1c-aab2-7dc4a219edc6.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 690.501614] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-192facd2-d60a-492b-b2fb-6f792f5f3c40 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.507931] env[63515]: DEBUG oslo_vmware.api [None req-95251dc9-cf22-4a89-a3b0-fce88f9e12db tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Waiting for the task: (returnval){ [ 690.507931] env[63515]: value = "task-1110910" [ 690.507931] env[63515]: _type = "Task" [ 690.507931] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 690.516291] env[63515]: DEBUG oslo_vmware.api [None req-95251dc9-cf22-4a89-a3b0-fce88f9e12db tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Task: {'id': task-1110910, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 690.593028] env[63515]: DEBUG nova.compute.manager [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] [instance: 452f3ca2-6141-43b2-a77a-c9ab5754192d] Starting instance... {{(pid=63515) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 690.797137] env[63515]: DEBUG oslo_concurrency.lockutils [None req-fa933c74-4a5b-4cc9-a499-9ea16a38d8b7 tempest-ServerActionsTestJSON-1591637270 tempest-ServerActionsTestJSON-1591637270-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.011s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 690.797137] env[63515]: ERROR nova.compute.manager [None req-fa933c74-4a5b-4cc9-a499-9ea16a38d8b7 tempest-ServerActionsTestJSON-1591637270 tempest-ServerActionsTestJSON-1591637270-project-member] [instance: 005f45b3-630d-400f-9605-100f6aad88e6] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port e7f7de8b-0bf6-48de-b304-c89a566c1cdc, please check neutron logs for more information. [ 690.797137] env[63515]: ERROR nova.compute.manager [instance: 005f45b3-630d-400f-9605-100f6aad88e6] Traceback (most recent call last): [ 690.797137] env[63515]: ERROR nova.compute.manager [instance: 005f45b3-630d-400f-9605-100f6aad88e6] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 690.797137] env[63515]: ERROR nova.compute.manager [instance: 005f45b3-630d-400f-9605-100f6aad88e6] self.driver.spawn(context, instance, image_meta, [ 690.797137] env[63515]: ERROR nova.compute.manager [instance: 005f45b3-630d-400f-9605-100f6aad88e6] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 690.797137] env[63515]: ERROR nova.compute.manager [instance: 005f45b3-630d-400f-9605-100f6aad88e6] self._vmops.spawn(context, instance, image_meta, injected_files, [ 690.797137] env[63515]: ERROR nova.compute.manager [instance: 005f45b3-630d-400f-9605-100f6aad88e6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 690.797137] env[63515]: ERROR nova.compute.manager [instance: 005f45b3-630d-400f-9605-100f6aad88e6] vm_ref = self.build_virtual_machine(instance, [ 690.797491] env[63515]: ERROR nova.compute.manager [instance: 005f45b3-630d-400f-9605-100f6aad88e6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 690.797491] env[63515]: ERROR nova.compute.manager [instance: 005f45b3-630d-400f-9605-100f6aad88e6] vif_infos = vmwarevif.get_vif_info(self._session, [ 690.797491] env[63515]: ERROR nova.compute.manager [instance: 005f45b3-630d-400f-9605-100f6aad88e6] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 690.797491] env[63515]: ERROR nova.compute.manager [instance: 005f45b3-630d-400f-9605-100f6aad88e6] for vif in network_info: [ 690.797491] env[63515]: ERROR nova.compute.manager [instance: 005f45b3-630d-400f-9605-100f6aad88e6] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 690.797491] env[63515]: ERROR nova.compute.manager [instance: 005f45b3-630d-400f-9605-100f6aad88e6] return self._sync_wrapper(fn, *args, **kwargs) [ 690.797491] env[63515]: ERROR nova.compute.manager [instance: 005f45b3-630d-400f-9605-100f6aad88e6] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 690.797491] env[63515]: ERROR nova.compute.manager [instance: 005f45b3-630d-400f-9605-100f6aad88e6] self.wait() [ 690.797491] env[63515]: ERROR nova.compute.manager [instance: 005f45b3-630d-400f-9605-100f6aad88e6] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 690.797491] env[63515]: ERROR nova.compute.manager [instance: 005f45b3-630d-400f-9605-100f6aad88e6] self[:] = self._gt.wait() [ 690.797491] env[63515]: ERROR nova.compute.manager [instance: 005f45b3-630d-400f-9605-100f6aad88e6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 690.797491] env[63515]: ERROR nova.compute.manager [instance: 005f45b3-630d-400f-9605-100f6aad88e6] return self._exit_event.wait() [ 690.797491] env[63515]: ERROR nova.compute.manager [instance: 005f45b3-630d-400f-9605-100f6aad88e6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 690.797863] env[63515]: ERROR nova.compute.manager [instance: 005f45b3-630d-400f-9605-100f6aad88e6] current.throw(*self._exc) [ 690.797863] env[63515]: ERROR nova.compute.manager [instance: 005f45b3-630d-400f-9605-100f6aad88e6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 690.797863] env[63515]: ERROR nova.compute.manager [instance: 005f45b3-630d-400f-9605-100f6aad88e6] result = function(*args, **kwargs) [ 690.797863] env[63515]: ERROR nova.compute.manager [instance: 005f45b3-630d-400f-9605-100f6aad88e6] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 690.797863] env[63515]: ERROR nova.compute.manager [instance: 005f45b3-630d-400f-9605-100f6aad88e6] return func(*args, **kwargs) [ 690.797863] env[63515]: ERROR nova.compute.manager [instance: 005f45b3-630d-400f-9605-100f6aad88e6] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 690.797863] env[63515]: ERROR nova.compute.manager [instance: 005f45b3-630d-400f-9605-100f6aad88e6] raise e [ 690.797863] env[63515]: ERROR nova.compute.manager [instance: 005f45b3-630d-400f-9605-100f6aad88e6] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 690.797863] env[63515]: ERROR nova.compute.manager [instance: 005f45b3-630d-400f-9605-100f6aad88e6] nwinfo = self.network_api.allocate_for_instance( [ 690.797863] env[63515]: ERROR nova.compute.manager [instance: 005f45b3-630d-400f-9605-100f6aad88e6] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 690.797863] env[63515]: ERROR nova.compute.manager [instance: 005f45b3-630d-400f-9605-100f6aad88e6] created_port_ids = self._update_ports_for_instance( [ 690.797863] env[63515]: ERROR nova.compute.manager [instance: 005f45b3-630d-400f-9605-100f6aad88e6] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 690.797863] env[63515]: ERROR nova.compute.manager [instance: 005f45b3-630d-400f-9605-100f6aad88e6] with excutils.save_and_reraise_exception(): [ 690.798295] env[63515]: ERROR nova.compute.manager [instance: 005f45b3-630d-400f-9605-100f6aad88e6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 690.798295] env[63515]: ERROR nova.compute.manager [instance: 005f45b3-630d-400f-9605-100f6aad88e6] self.force_reraise() [ 690.798295] env[63515]: ERROR nova.compute.manager [instance: 005f45b3-630d-400f-9605-100f6aad88e6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 690.798295] env[63515]: ERROR nova.compute.manager [instance: 005f45b3-630d-400f-9605-100f6aad88e6] raise self.value [ 690.798295] env[63515]: ERROR nova.compute.manager [instance: 005f45b3-630d-400f-9605-100f6aad88e6] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 690.798295] env[63515]: ERROR nova.compute.manager [instance: 005f45b3-630d-400f-9605-100f6aad88e6] updated_port = self._update_port( [ 690.798295] env[63515]: ERROR nova.compute.manager [instance: 005f45b3-630d-400f-9605-100f6aad88e6] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 690.798295] env[63515]: ERROR nova.compute.manager [instance: 005f45b3-630d-400f-9605-100f6aad88e6] _ensure_no_port_binding_failure(port) [ 690.798295] env[63515]: ERROR nova.compute.manager [instance: 005f45b3-630d-400f-9605-100f6aad88e6] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 690.798295] env[63515]: ERROR nova.compute.manager [instance: 005f45b3-630d-400f-9605-100f6aad88e6] raise exception.PortBindingFailed(port_id=port['id']) [ 690.798295] env[63515]: ERROR nova.compute.manager [instance: 005f45b3-630d-400f-9605-100f6aad88e6] nova.exception.PortBindingFailed: Binding failed for port e7f7de8b-0bf6-48de-b304-c89a566c1cdc, please check neutron logs for more information. [ 690.798295] env[63515]: ERROR nova.compute.manager [instance: 005f45b3-630d-400f-9605-100f6aad88e6] [ 690.798673] env[63515]: DEBUG nova.compute.utils [None req-fa933c74-4a5b-4cc9-a499-9ea16a38d8b7 tempest-ServerActionsTestJSON-1591637270 tempest-ServerActionsTestJSON-1591637270-project-member] [instance: 005f45b3-630d-400f-9605-100f6aad88e6] Binding failed for port e7f7de8b-0bf6-48de-b304-c89a566c1cdc, please check neutron logs for more information. {{(pid=63515) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 690.800506] env[63515]: DEBUG oslo_concurrency.lockutils [None req-89884269-25d4-4981-bd3b-a346ece7b681 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.001s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 690.802038] env[63515]: INFO nova.compute.claims [None req-89884269-25d4-4981-bd3b-a346ece7b681 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] [instance: f1d01b75-ac9d-458d-8cc2-ae64cffca4e8] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 690.805149] env[63515]: DEBUG nova.compute.manager [None req-fa933c74-4a5b-4cc9-a499-9ea16a38d8b7 tempest-ServerActionsTestJSON-1591637270 tempest-ServerActionsTestJSON-1591637270-project-member] [instance: 005f45b3-630d-400f-9605-100f6aad88e6] Build of instance 005f45b3-630d-400f-9605-100f6aad88e6 was re-scheduled: Binding failed for port e7f7de8b-0bf6-48de-b304-c89a566c1cdc, please check neutron logs for more information. {{(pid=63515) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 690.805616] env[63515]: DEBUG nova.compute.manager [None req-fa933c74-4a5b-4cc9-a499-9ea16a38d8b7 tempest-ServerActionsTestJSON-1591637270 tempest-ServerActionsTestJSON-1591637270-project-member] [instance: 005f45b3-630d-400f-9605-100f6aad88e6] Unplugging VIFs for instance {{(pid=63515) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 690.805616] env[63515]: DEBUG oslo_concurrency.lockutils [None req-fa933c74-4a5b-4cc9-a499-9ea16a38d8b7 tempest-ServerActionsTestJSON-1591637270 tempest-ServerActionsTestJSON-1591637270-project-member] Acquiring lock "refresh_cache-005f45b3-630d-400f-9605-100f6aad88e6" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 690.805747] env[63515]: DEBUG oslo_concurrency.lockutils [None req-fa933c74-4a5b-4cc9-a499-9ea16a38d8b7 tempest-ServerActionsTestJSON-1591637270 tempest-ServerActionsTestJSON-1591637270-project-member] Acquired lock "refresh_cache-005f45b3-630d-400f-9605-100f6aad88e6" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 690.805898] env[63515]: DEBUG nova.network.neutron [None req-fa933c74-4a5b-4cc9-a499-9ea16a38d8b7 tempest-ServerActionsTestJSON-1591637270 tempest-ServerActionsTestJSON-1591637270-project-member] [instance: 005f45b3-630d-400f-9605-100f6aad88e6] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 690.807635] env[63515]: DEBUG nova.compute.manager [None req-c7572d27-5b85-4acb-89ec-c56462911d65 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] [instance: 5b42f744-fdd6-45b1-8563-896869648c23] Start spawning the instance on the hypervisor. {{(pid=63515) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 690.855577] env[63515]: DEBUG nova.virt.hardware [None req-c7572d27-5b85-4acb-89ec-c56462911d65 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T02:52:14Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T02:51:56Z,direct_url=,disk_format='vmdk',id=8a120570-cb06-4099-b262-554ca0ad15c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b569255cc43e42e7ae3f2b4ad37c6ef4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T02:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 690.855868] env[63515]: DEBUG nova.virt.hardware [None req-c7572d27-5b85-4acb-89ec-c56462911d65 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] Flavor limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 690.856045] env[63515]: DEBUG nova.virt.hardware [None req-c7572d27-5b85-4acb-89ec-c56462911d65 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] Image limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 690.856241] env[63515]: DEBUG nova.virt.hardware [None req-c7572d27-5b85-4acb-89ec-c56462911d65 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] Flavor pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 690.856388] env[63515]: DEBUG nova.virt.hardware [None req-c7572d27-5b85-4acb-89ec-c56462911d65 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] Image pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 690.856536] env[63515]: DEBUG nova.virt.hardware [None req-c7572d27-5b85-4acb-89ec-c56462911d65 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 690.856742] env[63515]: DEBUG nova.virt.hardware [None req-c7572d27-5b85-4acb-89ec-c56462911d65 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 690.856896] env[63515]: DEBUG nova.virt.hardware [None req-c7572d27-5b85-4acb-89ec-c56462911d65 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 690.857126] env[63515]: DEBUG nova.virt.hardware [None req-c7572d27-5b85-4acb-89ec-c56462911d65 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] Got 1 possible topologies {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 690.857230] env[63515]: DEBUG nova.virt.hardware [None req-c7572d27-5b85-4acb-89ec-c56462911d65 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 690.857440] env[63515]: DEBUG nova.virt.hardware [None req-c7572d27-5b85-4acb-89ec-c56462911d65 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 690.858598] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10ed3b09-962a-46c6-b5ae-c4a3815a825a {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.869097] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42a37bd0-c9d3-4a87-adc1-46407f64d558 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.017980] env[63515]: DEBUG oslo_vmware.api [None req-95251dc9-cf22-4a89-a3b0-fce88f9e12db tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Task: {'id': task-1110910, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 691.122051] env[63515]: DEBUG oslo_concurrency.lockutils [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 691.344168] env[63515]: DEBUG nova.network.neutron [None req-fa933c74-4a5b-4cc9-a499-9ea16a38d8b7 tempest-ServerActionsTestJSON-1591637270 tempest-ServerActionsTestJSON-1591637270-project-member] [instance: 005f45b3-630d-400f-9605-100f6aad88e6] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 691.493630] env[63515]: DEBUG nova.network.neutron [None req-fa933c74-4a5b-4cc9-a499-9ea16a38d8b7 tempest-ServerActionsTestJSON-1591637270 tempest-ServerActionsTestJSON-1591637270-project-member] [instance: 005f45b3-630d-400f-9605-100f6aad88e6] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 691.519616] env[63515]: DEBUG oslo_vmware.api [None req-95251dc9-cf22-4a89-a3b0-fce88f9e12db tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Task: {'id': task-1110910, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.530245} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 691.520054] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-95251dc9-cf22-4a89-a3b0-fce88f9e12db tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk to [datastore1] 3ad04a9f-62cf-4b1c-aab2-7dc4a219edc6/3ad04a9f-62cf-4b1c-aab2-7dc4a219edc6.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 691.520187] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-95251dc9-cf22-4a89-a3b0-fce88f9e12db tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: 3ad04a9f-62cf-4b1c-aab2-7dc4a219edc6] Extending root virtual disk to 1048576 {{(pid=63515) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 691.520344] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-08f55dd5-9969-48ac-b12f-d5c9ec02637d {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.527225] env[63515]: DEBUG oslo_vmware.api [None req-95251dc9-cf22-4a89-a3b0-fce88f9e12db tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Waiting for the task: (returnval){ [ 691.527225] env[63515]: value = "task-1110911" [ 691.527225] env[63515]: _type = "Task" [ 691.527225] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 691.537938] env[63515]: DEBUG oslo_vmware.api [None req-95251dc9-cf22-4a89-a3b0-fce88f9e12db tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Task: {'id': task-1110911, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 691.748237] env[63515]: DEBUG nova.compute.manager [req-d8cc81f8-5627-4761-8f32-06fa1580c157 req-cdcb0a18-abe3-4a96-acb3-6c685a83e904 service nova] [instance: 5b42f744-fdd6-45b1-8563-896869648c23] Received event network-vif-plugged-3a630639-5423-4e62-9beb-ee1e7054f47d {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 691.748392] env[63515]: DEBUG oslo_concurrency.lockutils [req-d8cc81f8-5627-4761-8f32-06fa1580c157 req-cdcb0a18-abe3-4a96-acb3-6c685a83e904 service nova] Acquiring lock "5b42f744-fdd6-45b1-8563-896869648c23-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 691.748604] env[63515]: DEBUG oslo_concurrency.lockutils [req-d8cc81f8-5627-4761-8f32-06fa1580c157 req-cdcb0a18-abe3-4a96-acb3-6c685a83e904 service nova] Lock "5b42f744-fdd6-45b1-8563-896869648c23-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 691.748770] env[63515]: DEBUG oslo_concurrency.lockutils [req-d8cc81f8-5627-4761-8f32-06fa1580c157 req-cdcb0a18-abe3-4a96-acb3-6c685a83e904 service nova] Lock "5b42f744-fdd6-45b1-8563-896869648c23-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 691.748930] env[63515]: DEBUG nova.compute.manager [req-d8cc81f8-5627-4761-8f32-06fa1580c157 req-cdcb0a18-abe3-4a96-acb3-6c685a83e904 service nova] [instance: 5b42f744-fdd6-45b1-8563-896869648c23] No waiting events found dispatching network-vif-plugged-3a630639-5423-4e62-9beb-ee1e7054f47d {{(pid=63515) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 691.749273] env[63515]: WARNING nova.compute.manager [req-d8cc81f8-5627-4761-8f32-06fa1580c157 req-cdcb0a18-abe3-4a96-acb3-6c685a83e904 service nova] [instance: 5b42f744-fdd6-45b1-8563-896869648c23] Received unexpected event network-vif-plugged-3a630639-5423-4e62-9beb-ee1e7054f47d for instance with vm_state building and task_state spawning. [ 691.909258] env[63515]: DEBUG nova.network.neutron [None req-c7572d27-5b85-4acb-89ec-c56462911d65 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] [instance: 5b42f744-fdd6-45b1-8563-896869648c23] Successfully updated port: 3a630639-5423-4e62-9beb-ee1e7054f47d {{(pid=63515) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 691.997180] env[63515]: DEBUG oslo_concurrency.lockutils [None req-fa933c74-4a5b-4cc9-a499-9ea16a38d8b7 tempest-ServerActionsTestJSON-1591637270 tempest-ServerActionsTestJSON-1591637270-project-member] Releasing lock "refresh_cache-005f45b3-630d-400f-9605-100f6aad88e6" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 691.997575] env[63515]: DEBUG nova.compute.manager [None req-fa933c74-4a5b-4cc9-a499-9ea16a38d8b7 tempest-ServerActionsTestJSON-1591637270 tempest-ServerActionsTestJSON-1591637270-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=63515) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 691.997771] env[63515]: DEBUG nova.compute.manager [None req-fa933c74-4a5b-4cc9-a499-9ea16a38d8b7 tempest-ServerActionsTestJSON-1591637270 tempest-ServerActionsTestJSON-1591637270-project-member] [instance: 005f45b3-630d-400f-9605-100f6aad88e6] Deallocating network for instance {{(pid=63515) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 691.997933] env[63515]: DEBUG nova.network.neutron [None req-fa933c74-4a5b-4cc9-a499-9ea16a38d8b7 tempest-ServerActionsTestJSON-1591637270 tempest-ServerActionsTestJSON-1591637270-project-member] [instance: 005f45b3-630d-400f-9605-100f6aad88e6] deallocate_for_instance() {{(pid=63515) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 692.021273] env[63515]: DEBUG nova.network.neutron [None req-fa933c74-4a5b-4cc9-a499-9ea16a38d8b7 tempest-ServerActionsTestJSON-1591637270 tempest-ServerActionsTestJSON-1591637270-project-member] [instance: 005f45b3-630d-400f-9605-100f6aad88e6] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 692.045086] env[63515]: DEBUG oslo_vmware.api [None req-95251dc9-cf22-4a89-a3b0-fce88f9e12db tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Task: {'id': task-1110911, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.063955} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 692.045471] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-95251dc9-cf22-4a89-a3b0-fce88f9e12db tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: 3ad04a9f-62cf-4b1c-aab2-7dc4a219edc6] Extended root virtual disk {{(pid=63515) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 692.047395] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e810adba-04fe-4d67-ad99-39af09a3ab12 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.076775] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-95251dc9-cf22-4a89-a3b0-fce88f9e12db tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: 3ad04a9f-62cf-4b1c-aab2-7dc4a219edc6] Reconfiguring VM instance instance-00000020 to attach disk [datastore1] 3ad04a9f-62cf-4b1c-aab2-7dc4a219edc6/3ad04a9f-62cf-4b1c-aab2-7dc4a219edc6.vmdk or device None with type sparse {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 692.079685] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b01cd3d9-d595-42b7-a3ae-56cb42f82ba1 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.102297] env[63515]: DEBUG oslo_vmware.api [None req-95251dc9-cf22-4a89-a3b0-fce88f9e12db tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Waiting for the task: (returnval){ [ 692.102297] env[63515]: value = "task-1110912" [ 692.102297] env[63515]: _type = "Task" [ 692.102297] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 692.109462] env[63515]: DEBUG oslo_vmware.api [None req-95251dc9-cf22-4a89-a3b0-fce88f9e12db tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Task: {'id': task-1110912, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 692.275330] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5f5ee8a-ab89-4575-a3a1-50d7678c8cbc {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.283122] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4a48c0d-2844-4ec3-9098-913772fc2555 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.313584] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f756aa0-3d80-47f1-9875-6b35598f8a49 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.321280] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1153d87c-c075-4a3d-a81a-9ce1ec33a3b0 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.334916] env[63515]: DEBUG nova.compute.provider_tree [None req-89884269-25d4-4981-bd3b-a346ece7b681 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 692.413945] env[63515]: DEBUG oslo_concurrency.lockutils [None req-c7572d27-5b85-4acb-89ec-c56462911d65 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] Acquiring lock "refresh_cache-5b42f744-fdd6-45b1-8563-896869648c23" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 692.414229] env[63515]: DEBUG oslo_concurrency.lockutils [None req-c7572d27-5b85-4acb-89ec-c56462911d65 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] Acquired lock "refresh_cache-5b42f744-fdd6-45b1-8563-896869648c23" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 692.414229] env[63515]: DEBUG nova.network.neutron [None req-c7572d27-5b85-4acb-89ec-c56462911d65 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] [instance: 5b42f744-fdd6-45b1-8563-896869648c23] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 692.524840] env[63515]: DEBUG nova.network.neutron [None req-fa933c74-4a5b-4cc9-a499-9ea16a38d8b7 tempest-ServerActionsTestJSON-1591637270 tempest-ServerActionsTestJSON-1591637270-project-member] [instance: 005f45b3-630d-400f-9605-100f6aad88e6] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 692.612691] env[63515]: DEBUG oslo_vmware.api [None req-95251dc9-cf22-4a89-a3b0-fce88f9e12db tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Task: {'id': task-1110912, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 692.838461] env[63515]: DEBUG nova.scheduler.client.report [None req-89884269-25d4-4981-bd3b-a346ece7b681 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 692.949120] env[63515]: DEBUG nova.network.neutron [None req-c7572d27-5b85-4acb-89ec-c56462911d65 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] [instance: 5b42f744-fdd6-45b1-8563-896869648c23] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 693.027821] env[63515]: INFO nova.compute.manager [None req-fa933c74-4a5b-4cc9-a499-9ea16a38d8b7 tempest-ServerActionsTestJSON-1591637270 tempest-ServerActionsTestJSON-1591637270-project-member] [instance: 005f45b3-630d-400f-9605-100f6aad88e6] Took 1.03 seconds to deallocate network for instance. [ 693.116923] env[63515]: DEBUG oslo_vmware.api [None req-95251dc9-cf22-4a89-a3b0-fce88f9e12db tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Task: {'id': task-1110912, 'name': ReconfigVM_Task, 'duration_secs': 0.737993} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 693.117456] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-95251dc9-cf22-4a89-a3b0-fce88f9e12db tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: 3ad04a9f-62cf-4b1c-aab2-7dc4a219edc6] Reconfigured VM instance instance-00000020 to attach disk [datastore1] 3ad04a9f-62cf-4b1c-aab2-7dc4a219edc6/3ad04a9f-62cf-4b1c-aab2-7dc4a219edc6.vmdk or device None with type sparse {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 693.118602] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b303228c-de2f-41c7-9055-4819f8840de9 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.125761] env[63515]: DEBUG oslo_vmware.api [None req-95251dc9-cf22-4a89-a3b0-fce88f9e12db tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Waiting for the task: (returnval){ [ 693.125761] env[63515]: value = "task-1110913" [ 693.125761] env[63515]: _type = "Task" [ 693.125761] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 693.133508] env[63515]: DEBUG oslo_vmware.api [None req-95251dc9-cf22-4a89-a3b0-fce88f9e12db tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Task: {'id': task-1110913, 'name': Rename_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 693.134498] env[63515]: DEBUG nova.network.neutron [None req-c7572d27-5b85-4acb-89ec-c56462911d65 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] [instance: 5b42f744-fdd6-45b1-8563-896869648c23] Updating instance_info_cache with network_info: [{"id": "3a630639-5423-4e62-9beb-ee1e7054f47d", "address": "fa:16:3e:b6:20:83", "network": {"id": "b4abcbd0-0eac-4bde-ad11-7f93afcedcef", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.28", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "b569255cc43e42e7ae3f2b4ad37c6ef4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0cd5d325-3053-407e-a4ee-f627e82a23f9", "external-id": "nsx-vlan-transportzone-809", "segmentation_id": 809, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3a630639-54", "ovs_interfaceid": "3a630639-5423-4e62-9beb-ee1e7054f47d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 693.343682] env[63515]: DEBUG oslo_concurrency.lockutils [None req-89884269-25d4-4981-bd3b-a346ece7b681 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.543s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 693.344068] env[63515]: DEBUG nova.compute.manager [None req-89884269-25d4-4981-bd3b-a346ece7b681 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] [instance: f1d01b75-ac9d-458d-8cc2-ae64cffca4e8] Start building networks asynchronously for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 693.346846] env[63515]: DEBUG oslo_concurrency.lockutils [None req-c26675e2-4152-417b-99dd-0c18de4dc137 tempest-ServersTestBootFromVolume-2042560423 tempest-ServersTestBootFromVolume-2042560423-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 10.995s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 693.348732] env[63515]: INFO nova.compute.claims [None req-c26675e2-4152-417b-99dd-0c18de4dc137 tempest-ServersTestBootFromVolume-2042560423 tempest-ServersTestBootFromVolume-2042560423-project-member] [instance: 90f4930b-aaa0-4c4b-9ab8-92aed45e200b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 693.637886] env[63515]: DEBUG oslo_concurrency.lockutils [None req-c7572d27-5b85-4acb-89ec-c56462911d65 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] Releasing lock "refresh_cache-5b42f744-fdd6-45b1-8563-896869648c23" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 693.638467] env[63515]: DEBUG nova.compute.manager [None req-c7572d27-5b85-4acb-89ec-c56462911d65 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] [instance: 5b42f744-fdd6-45b1-8563-896869648c23] Instance network_info: |[{"id": "3a630639-5423-4e62-9beb-ee1e7054f47d", "address": "fa:16:3e:b6:20:83", "network": {"id": "b4abcbd0-0eac-4bde-ad11-7f93afcedcef", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.28", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "b569255cc43e42e7ae3f2b4ad37c6ef4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0cd5d325-3053-407e-a4ee-f627e82a23f9", "external-id": "nsx-vlan-transportzone-809", "segmentation_id": 809, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3a630639-54", "ovs_interfaceid": "3a630639-5423-4e62-9beb-ee1e7054f47d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 693.638771] env[63515]: DEBUG oslo_vmware.api [None req-95251dc9-cf22-4a89-a3b0-fce88f9e12db tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Task: {'id': task-1110913, 'name': Rename_Task, 'duration_secs': 0.161854} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 693.639302] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-c7572d27-5b85-4acb-89ec-c56462911d65 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] [instance: 5b42f744-fdd6-45b1-8563-896869648c23] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b6:20:83', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '0cd5d325-3053-407e-a4ee-f627e82a23f9', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '3a630639-5423-4e62-9beb-ee1e7054f47d', 'vif_model': 'vmxnet3'}] {{(pid=63515) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 693.647725] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-c7572d27-5b85-4acb-89ec-c56462911d65 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] Creating folder: Project (a1f130751200408982bffb31a54bbcea). Parent ref: group-v243370. {{(pid=63515) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 693.648024] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-95251dc9-cf22-4a89-a3b0-fce88f9e12db tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: 3ad04a9f-62cf-4b1c-aab2-7dc4a219edc6] Powering on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 693.648744] env[63515]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a77cf0ee-d037-4e09-868f-d4782e1c9b6d {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.650392] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ccfa8ed3-fd88-422f-945a-f9e3f25fe448 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.658082] env[63515]: DEBUG oslo_vmware.api [None req-95251dc9-cf22-4a89-a3b0-fce88f9e12db tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Waiting for the task: (returnval){ [ 693.658082] env[63515]: value = "task-1110915" [ 693.658082] env[63515]: _type = "Task" [ 693.658082] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 693.663116] env[63515]: INFO nova.virt.vmwareapi.vm_util [None req-c7572d27-5b85-4acb-89ec-c56462911d65 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] Created folder: Project (a1f130751200408982bffb31a54bbcea) in parent group-v243370. [ 693.663116] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-c7572d27-5b85-4acb-89ec-c56462911d65 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] Creating folder: Instances. Parent ref: group-v243389. {{(pid=63515) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 693.673227] env[63515]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-1d67fdd8-3220-4ea2-a9cb-da3897ae108d {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.673227] env[63515]: DEBUG oslo_vmware.api [None req-95251dc9-cf22-4a89-a3b0-fce88f9e12db tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Task: {'id': task-1110915, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 693.681524] env[63515]: INFO nova.virt.vmwareapi.vm_util [None req-c7572d27-5b85-4acb-89ec-c56462911d65 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] Created folder: Instances in parent group-v243389. [ 693.681524] env[63515]: DEBUG oslo.service.loopingcall [None req-c7572d27-5b85-4acb-89ec-c56462911d65 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 693.681524] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5b42f744-fdd6-45b1-8563-896869648c23] Creating VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 693.681524] env[63515]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-6b778867-dd33-4084-870f-23dbab867e1c {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.704542] env[63515]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 693.704542] env[63515]: value = "task-1110917" [ 693.704542] env[63515]: _type = "Task" [ 693.704542] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 693.713128] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1110917, 'name': CreateVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 693.776195] env[63515]: DEBUG nova.compute.manager [req-97f6de75-fb96-4601-9988-f56d3386220c req-b581997d-a691-4762-a749-28e3b383e56b service nova] [instance: 5b42f744-fdd6-45b1-8563-896869648c23] Received event network-changed-3a630639-5423-4e62-9beb-ee1e7054f47d {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 693.776286] env[63515]: DEBUG nova.compute.manager [req-97f6de75-fb96-4601-9988-f56d3386220c req-b581997d-a691-4762-a749-28e3b383e56b service nova] [instance: 5b42f744-fdd6-45b1-8563-896869648c23] Refreshing instance network info cache due to event network-changed-3a630639-5423-4e62-9beb-ee1e7054f47d. {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 693.776501] env[63515]: DEBUG oslo_concurrency.lockutils [req-97f6de75-fb96-4601-9988-f56d3386220c req-b581997d-a691-4762-a749-28e3b383e56b service nova] Acquiring lock "refresh_cache-5b42f744-fdd6-45b1-8563-896869648c23" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 693.776659] env[63515]: DEBUG oslo_concurrency.lockutils [req-97f6de75-fb96-4601-9988-f56d3386220c req-b581997d-a691-4762-a749-28e3b383e56b service nova] Acquired lock "refresh_cache-5b42f744-fdd6-45b1-8563-896869648c23" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 693.776743] env[63515]: DEBUG nova.network.neutron [req-97f6de75-fb96-4601-9988-f56d3386220c req-b581997d-a691-4762-a749-28e3b383e56b service nova] [instance: 5b42f744-fdd6-45b1-8563-896869648c23] Refreshing network info cache for port 3a630639-5423-4e62-9beb-ee1e7054f47d {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 693.853650] env[63515]: DEBUG nova.compute.utils [None req-89884269-25d4-4981-bd3b-a346ece7b681 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Using /dev/sd instead of None {{(pid=63515) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 693.857424] env[63515]: DEBUG nova.compute.manager [None req-89884269-25d4-4981-bd3b-a346ece7b681 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] [instance: f1d01b75-ac9d-458d-8cc2-ae64cffca4e8] Allocating IP information in the background. {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 693.858151] env[63515]: DEBUG nova.network.neutron [None req-89884269-25d4-4981-bd3b-a346ece7b681 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] [instance: f1d01b75-ac9d-458d-8cc2-ae64cffca4e8] allocate_for_instance() {{(pid=63515) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 693.913537] env[63515]: DEBUG nova.policy [None req-89884269-25d4-4981-bd3b-a346ece7b681 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '59a67d99419c47e9b4ff92c6d647656f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd3d0957a488b47559794495a451f5437', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63515) authorize /opt/stack/nova/nova/policy.py:201}} [ 694.086157] env[63515]: INFO nova.scheduler.client.report [None req-fa933c74-4a5b-4cc9-a499-9ea16a38d8b7 tempest-ServerActionsTestJSON-1591637270 tempest-ServerActionsTestJSON-1591637270-project-member] Deleted allocations for instance 005f45b3-630d-400f-9605-100f6aad88e6 [ 694.181586] env[63515]: DEBUG oslo_vmware.api [None req-95251dc9-cf22-4a89-a3b0-fce88f9e12db tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Task: {'id': task-1110915, 'name': PowerOnVM_Task} progress is 89%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 694.214371] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1110917, 'name': CreateVM_Task, 'duration_secs': 0.364862} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 694.214537] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5b42f744-fdd6-45b1-8563-896869648c23] Created VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 694.215227] env[63515]: DEBUG oslo_concurrency.lockutils [None req-c7572d27-5b85-4acb-89ec-c56462911d65 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 694.215395] env[63515]: DEBUG oslo_concurrency.lockutils [None req-c7572d27-5b85-4acb-89ec-c56462911d65 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 694.215713] env[63515]: DEBUG oslo_concurrency.lockutils [None req-c7572d27-5b85-4acb-89ec-c56462911d65 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 694.215985] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-91fb836e-6dd7-408e-b4e9-e25b9a80b997 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.220676] env[63515]: DEBUG oslo_vmware.api [None req-c7572d27-5b85-4acb-89ec-c56462911d65 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] Waiting for the task: (returnval){ [ 694.220676] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]52d59374-06e1-2f82-871c-187d9c42a80f" [ 694.220676] env[63515]: _type = "Task" [ 694.220676] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 694.228857] env[63515]: DEBUG oslo_vmware.api [None req-c7572d27-5b85-4acb-89ec-c56462911d65 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52d59374-06e1-2f82-871c-187d9c42a80f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 694.250314] env[63515]: DEBUG nova.network.neutron [None req-89884269-25d4-4981-bd3b-a346ece7b681 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] [instance: f1d01b75-ac9d-458d-8cc2-ae64cffca4e8] Successfully created port: 48f0868f-94db-42f7-8153-d1a27fa02707 {{(pid=63515) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 694.358479] env[63515]: DEBUG nova.compute.manager [None req-89884269-25d4-4981-bd3b-a346ece7b681 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] [instance: f1d01b75-ac9d-458d-8cc2-ae64cffca4e8] Start building block device mappings for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 694.607058] env[63515]: DEBUG oslo_concurrency.lockutils [None req-fa933c74-4a5b-4cc9-a499-9ea16a38d8b7 tempest-ServerActionsTestJSON-1591637270 tempest-ServerActionsTestJSON-1591637270-project-member] Lock "005f45b3-630d-400f-9605-100f6aad88e6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 115.051s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 694.665069] env[63515]: DEBUG nova.network.neutron [req-97f6de75-fb96-4601-9988-f56d3386220c req-b581997d-a691-4762-a749-28e3b383e56b service nova] [instance: 5b42f744-fdd6-45b1-8563-896869648c23] Updated VIF entry in instance network info cache for port 3a630639-5423-4e62-9beb-ee1e7054f47d. {{(pid=63515) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 694.667085] env[63515]: DEBUG nova.network.neutron [req-97f6de75-fb96-4601-9988-f56d3386220c req-b581997d-a691-4762-a749-28e3b383e56b service nova] [instance: 5b42f744-fdd6-45b1-8563-896869648c23] Updating instance_info_cache with network_info: [{"id": "3a630639-5423-4e62-9beb-ee1e7054f47d", "address": "fa:16:3e:b6:20:83", "network": {"id": "b4abcbd0-0eac-4bde-ad11-7f93afcedcef", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.28", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "b569255cc43e42e7ae3f2b4ad37c6ef4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0cd5d325-3053-407e-a4ee-f627e82a23f9", "external-id": "nsx-vlan-transportzone-809", "segmentation_id": 809, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3a630639-54", "ovs_interfaceid": "3a630639-5423-4e62-9beb-ee1e7054f47d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 694.673181] env[63515]: DEBUG oslo_vmware.api [None req-95251dc9-cf22-4a89-a3b0-fce88f9e12db tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Task: {'id': task-1110915, 'name': PowerOnVM_Task, 'duration_secs': 0.666347} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 694.675658] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-95251dc9-cf22-4a89-a3b0-fce88f9e12db tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: 3ad04a9f-62cf-4b1c-aab2-7dc4a219edc6] Powered on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 694.675871] env[63515]: INFO nova.compute.manager [None req-95251dc9-cf22-4a89-a3b0-fce88f9e12db tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: 3ad04a9f-62cf-4b1c-aab2-7dc4a219edc6] Took 8.49 seconds to spawn the instance on the hypervisor. [ 694.676054] env[63515]: DEBUG nova.compute.manager [None req-95251dc9-cf22-4a89-a3b0-fce88f9e12db tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: 3ad04a9f-62cf-4b1c-aab2-7dc4a219edc6] Checking state {{(pid=63515) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 694.677184] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93a95913-262c-4b0f-9d30-e63f4e1780a2 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.735757] env[63515]: DEBUG oslo_vmware.api [None req-c7572d27-5b85-4acb-89ec-c56462911d65 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52d59374-06e1-2f82-871c-187d9c42a80f, 'name': SearchDatastore_Task, 'duration_secs': 0.014891} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 694.738805] env[63515]: DEBUG oslo_concurrency.lockutils [None req-c7572d27-5b85-4acb-89ec-c56462911d65 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 694.739045] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-c7572d27-5b85-4acb-89ec-c56462911d65 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] [instance: 5b42f744-fdd6-45b1-8563-896869648c23] Processing image 8a120570-cb06-4099-b262-554ca0ad15c5 {{(pid=63515) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 694.739580] env[63515]: DEBUG oslo_concurrency.lockutils [None req-c7572d27-5b85-4acb-89ec-c56462911d65 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 694.739727] env[63515]: DEBUG oslo_concurrency.lockutils [None req-c7572d27-5b85-4acb-89ec-c56462911d65 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 694.739909] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-c7572d27-5b85-4acb-89ec-c56462911d65 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 694.741599] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-8af58c6f-ac3a-4ba4-b637-7bd9858f3041 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.754796] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-c7572d27-5b85-4acb-89ec-c56462911d65 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 694.754969] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-c7572d27-5b85-4acb-89ec-c56462911d65 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63515) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 694.755743] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c057e9d2-303d-4f8d-b53e-63732a4ba32e {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.761990] env[63515]: DEBUG oslo_vmware.api [None req-c7572d27-5b85-4acb-89ec-c56462911d65 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] Waiting for the task: (returnval){ [ 694.761990] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]5279e08f-dcfb-7bde-5bdc-997e3d43c49b" [ 694.761990] env[63515]: _type = "Task" [ 694.761990] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 694.767227] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a64b17f-5891-49d4-beca-eb648b49973c {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.774071] env[63515]: DEBUG oslo_vmware.api [None req-c7572d27-5b85-4acb-89ec-c56462911d65 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]5279e08f-dcfb-7bde-5bdc-997e3d43c49b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 694.776847] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8459b968-647a-4855-b2e1-2ba5d1d24dbb {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.809429] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4e18289-903a-4518-81bd-6ef9cb51ccec {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.817199] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7af5b40a-b80d-43d0-ae3b-35acf685c60a {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.830284] env[63515]: DEBUG nova.compute.provider_tree [None req-c26675e2-4152-417b-99dd-0c18de4dc137 tempest-ServersTestBootFromVolume-2042560423 tempest-ServersTestBootFromVolume-2042560423-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 695.110205] env[63515]: DEBUG nova.compute.manager [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] [instance: 37e6c27e-317b-45d2-bd55-2fd78ccf009f] Starting instance... {{(pid=63515) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 695.168821] env[63515]: DEBUG oslo_concurrency.lockutils [req-97f6de75-fb96-4601-9988-f56d3386220c req-b581997d-a691-4762-a749-28e3b383e56b service nova] Releasing lock "refresh_cache-5b42f744-fdd6-45b1-8563-896869648c23" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 695.202767] env[63515]: INFO nova.compute.manager [None req-95251dc9-cf22-4a89-a3b0-fce88f9e12db tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: 3ad04a9f-62cf-4b1c-aab2-7dc4a219edc6] Took 29.11 seconds to build instance. [ 695.277557] env[63515]: DEBUG oslo_vmware.api [None req-c7572d27-5b85-4acb-89ec-c56462911d65 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]5279e08f-dcfb-7bde-5bdc-997e3d43c49b, 'name': SearchDatastore_Task, 'duration_secs': 0.025471} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 695.277557] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6e70a9e8-f98c-4479-a872-42a227feafda {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.281580] env[63515]: DEBUG oslo_vmware.api [None req-c7572d27-5b85-4acb-89ec-c56462911d65 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] Waiting for the task: (returnval){ [ 695.281580] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]5229e385-9ce6-43c1-2c96-a9dc571c4cf4" [ 695.281580] env[63515]: _type = "Task" [ 695.281580] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 695.290766] env[63515]: DEBUG oslo_vmware.api [None req-c7572d27-5b85-4acb-89ec-c56462911d65 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]5229e385-9ce6-43c1-2c96-a9dc571c4cf4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 695.336038] env[63515]: DEBUG nova.scheduler.client.report [None req-c26675e2-4152-417b-99dd-0c18de4dc137 tempest-ServersTestBootFromVolume-2042560423 tempest-ServersTestBootFromVolume-2042560423-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 695.375540] env[63515]: DEBUG nova.compute.manager [None req-89884269-25d4-4981-bd3b-a346ece7b681 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] [instance: f1d01b75-ac9d-458d-8cc2-ae64cffca4e8] Start spawning the instance on the hypervisor. {{(pid=63515) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 695.409310] env[63515]: DEBUG nova.virt.hardware [None req-89884269-25d4-4981-bd3b-a346ece7b681 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T02:54:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='1fae9a09-456d-4e70-9655-fc9d700f1547',id=38,is_public=True,memory_mb=192,name='tempest-test_resize_flavor_-1433437001',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T02:51:56Z,direct_url=,disk_format='vmdk',id=8a120570-cb06-4099-b262-554ca0ad15c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b569255cc43e42e7ae3f2b4ad37c6ef4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T02:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 695.409577] env[63515]: DEBUG nova.virt.hardware [None req-89884269-25d4-4981-bd3b-a346ece7b681 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Flavor limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 695.409734] env[63515]: DEBUG nova.virt.hardware [None req-89884269-25d4-4981-bd3b-a346ece7b681 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Image limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 695.409912] env[63515]: DEBUG nova.virt.hardware [None req-89884269-25d4-4981-bd3b-a346ece7b681 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Flavor pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 695.410072] env[63515]: DEBUG nova.virt.hardware [None req-89884269-25d4-4981-bd3b-a346ece7b681 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Image pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 695.410255] env[63515]: DEBUG nova.virt.hardware [None req-89884269-25d4-4981-bd3b-a346ece7b681 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 695.410553] env[63515]: DEBUG nova.virt.hardware [None req-89884269-25d4-4981-bd3b-a346ece7b681 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 695.410734] env[63515]: DEBUG nova.virt.hardware [None req-89884269-25d4-4981-bd3b-a346ece7b681 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 695.410911] env[63515]: DEBUG nova.virt.hardware [None req-89884269-25d4-4981-bd3b-a346ece7b681 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Got 1 possible topologies {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 695.411084] env[63515]: DEBUG nova.virt.hardware [None req-89884269-25d4-4981-bd3b-a346ece7b681 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 695.411745] env[63515]: DEBUG nova.virt.hardware [None req-89884269-25d4-4981-bd3b-a346ece7b681 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 695.415168] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e462265-e314-45c7-baf5-27b9796ee34d {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.421357] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1edea30e-ee1e-472d-aecf-e596cfc10ff1 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.632696] env[63515]: DEBUG oslo_concurrency.lockutils [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 695.707540] env[63515]: DEBUG oslo_concurrency.lockutils [None req-95251dc9-cf22-4a89-a3b0-fce88f9e12db tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Lock "3ad04a9f-62cf-4b1c-aab2-7dc4a219edc6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 112.002s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 695.794185] env[63515]: DEBUG oslo_vmware.api [None req-c7572d27-5b85-4acb-89ec-c56462911d65 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]5229e385-9ce6-43c1-2c96-a9dc571c4cf4, 'name': SearchDatastore_Task, 'duration_secs': 0.009685} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 695.794632] env[63515]: DEBUG oslo_concurrency.lockutils [None req-c7572d27-5b85-4acb-89ec-c56462911d65 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 695.794881] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-c7572d27-5b85-4acb-89ec-c56462911d65 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk to [datastore1] 5b42f744-fdd6-45b1-8563-896869648c23/5b42f744-fdd6-45b1-8563-896869648c23.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 695.795161] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-58f6ec73-042f-4310-9380-99b6744b3fad {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.804558] env[63515]: DEBUG oslo_vmware.api [None req-c7572d27-5b85-4acb-89ec-c56462911d65 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] Waiting for the task: (returnval){ [ 695.804558] env[63515]: value = "task-1110918" [ 695.804558] env[63515]: _type = "Task" [ 695.804558] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 695.817530] env[63515]: DEBUG oslo_vmware.api [None req-c7572d27-5b85-4acb-89ec-c56462911d65 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] Task: {'id': task-1110918, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 695.844064] env[63515]: DEBUG oslo_concurrency.lockutils [None req-c26675e2-4152-417b-99dd-0c18de4dc137 tempest-ServersTestBootFromVolume-2042560423 tempest-ServersTestBootFromVolume-2042560423-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.497s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 695.844985] env[63515]: DEBUG nova.compute.manager [None req-c26675e2-4152-417b-99dd-0c18de4dc137 tempest-ServersTestBootFromVolume-2042560423 tempest-ServersTestBootFromVolume-2042560423-project-member] [instance: 90f4930b-aaa0-4c4b-9ab8-92aed45e200b] Start building networks asynchronously for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 695.851051] env[63515]: DEBUG oslo_concurrency.lockutils [None req-33f5b097-4c20-4377-b5be-7e558874e5d3 tempest-ServersV294TestFqdnHostnames-2112993702 tempest-ServersV294TestFqdnHostnames-2112993702-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.664s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 695.852662] env[63515]: INFO nova.compute.claims [None req-33f5b097-4c20-4377-b5be-7e558874e5d3 tempest-ServersV294TestFqdnHostnames-2112993702 tempest-ServersV294TestFqdnHostnames-2112993702-project-member] [instance: 396e49dd-48c0-496b-a1ec-190c33a22c5e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 696.216539] env[63515]: DEBUG nova.compute.manager [None req-746f5a16-d8b1-448e-a35c-34720548f33b tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] [instance: b4477e66-ae12-4929-90ed-b7b652e0f207] Starting instance... {{(pid=63515) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 696.317480] env[63515]: DEBUG oslo_vmware.api [None req-c7572d27-5b85-4acb-89ec-c56462911d65 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] Task: {'id': task-1110918, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 696.359323] env[63515]: DEBUG nova.compute.utils [None req-c26675e2-4152-417b-99dd-0c18de4dc137 tempest-ServersTestBootFromVolume-2042560423 tempest-ServersTestBootFromVolume-2042560423-project-member] Using /dev/sd instead of None {{(pid=63515) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 696.366127] env[63515]: DEBUG nova.compute.manager [None req-c26675e2-4152-417b-99dd-0c18de4dc137 tempest-ServersTestBootFromVolume-2042560423 tempest-ServersTestBootFromVolume-2042560423-project-member] [instance: 90f4930b-aaa0-4c4b-9ab8-92aed45e200b] Allocating IP information in the background. {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 696.366127] env[63515]: DEBUG nova.network.neutron [None req-c26675e2-4152-417b-99dd-0c18de4dc137 tempest-ServersTestBootFromVolume-2042560423 tempest-ServersTestBootFromVolume-2042560423-project-member] [instance: 90f4930b-aaa0-4c4b-9ab8-92aed45e200b] allocate_for_instance() {{(pid=63515) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 696.429447] env[63515]: DEBUG nova.network.neutron [None req-89884269-25d4-4981-bd3b-a346ece7b681 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] [instance: f1d01b75-ac9d-458d-8cc2-ae64cffca4e8] Successfully updated port: 48f0868f-94db-42f7-8153-d1a27fa02707 {{(pid=63515) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 696.449254] env[63515]: DEBUG nova.policy [None req-c26675e2-4152-417b-99dd-0c18de4dc137 tempest-ServersTestBootFromVolume-2042560423 tempest-ServersTestBootFromVolume-2042560423-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4581dc538f084fa1a585b0b9ed33deba', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '38b2d712caf34cc08ad9c9694c521b7e', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63515) authorize /opt/stack/nova/nova/policy.py:201}} [ 696.545928] env[63515]: DEBUG oslo_concurrency.lockutils [None req-5690e97d-f63c-4f8e-9c79-d9c523fc3720 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Acquiring lock "3ad04a9f-62cf-4b1c-aab2-7dc4a219edc6" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 696.546198] env[63515]: DEBUG oslo_concurrency.lockutils [None req-5690e97d-f63c-4f8e-9c79-d9c523fc3720 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Lock "3ad04a9f-62cf-4b1c-aab2-7dc4a219edc6" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 696.546399] env[63515]: DEBUG oslo_concurrency.lockutils [None req-5690e97d-f63c-4f8e-9c79-d9c523fc3720 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Acquiring lock "3ad04a9f-62cf-4b1c-aab2-7dc4a219edc6-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 696.547613] env[63515]: DEBUG oslo_concurrency.lockutils [None req-5690e97d-f63c-4f8e-9c79-d9c523fc3720 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Lock "3ad04a9f-62cf-4b1c-aab2-7dc4a219edc6-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 696.548134] env[63515]: DEBUG oslo_concurrency.lockutils [None req-5690e97d-f63c-4f8e-9c79-d9c523fc3720 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Lock "3ad04a9f-62cf-4b1c-aab2-7dc4a219edc6-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.001s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 696.553564] env[63515]: INFO nova.compute.manager [None req-5690e97d-f63c-4f8e-9c79-d9c523fc3720 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: 3ad04a9f-62cf-4b1c-aab2-7dc4a219edc6] Terminating instance [ 696.553564] env[63515]: DEBUG nova.compute.manager [None req-5690e97d-f63c-4f8e-9c79-d9c523fc3720 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: 3ad04a9f-62cf-4b1c-aab2-7dc4a219edc6] Start destroying the instance on the hypervisor. {{(pid=63515) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 696.553564] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-5690e97d-f63c-4f8e-9c79-d9c523fc3720 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: 3ad04a9f-62cf-4b1c-aab2-7dc4a219edc6] Destroying instance {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 696.554255] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e91ad43-f4b7-4811-b446-2682e333e3b9 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.562893] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-5690e97d-f63c-4f8e-9c79-d9c523fc3720 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: 3ad04a9f-62cf-4b1c-aab2-7dc4a219edc6] Powering off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 696.563133] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d3163fe1-84d1-4c20-bb63-3f50c3aa9c98 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.571158] env[63515]: DEBUG oslo_vmware.api [None req-5690e97d-f63c-4f8e-9c79-d9c523fc3720 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Waiting for the task: (returnval){ [ 696.571158] env[63515]: value = "task-1110919" [ 696.571158] env[63515]: _type = "Task" [ 696.571158] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 696.577588] env[63515]: DEBUG oslo_vmware.api [None req-5690e97d-f63c-4f8e-9c79-d9c523fc3720 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Task: {'id': task-1110919, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 696.744497] env[63515]: DEBUG oslo_concurrency.lockutils [None req-746f5a16-d8b1-448e-a35c-34720548f33b tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 696.744961] env[63515]: DEBUG nova.compute.manager [req-17637906-176d-42a2-82a1-8c1260397191 req-3019c6d9-c0e1-4c6c-90a3-0ad69e7596d0 service nova] [instance: f1d01b75-ac9d-458d-8cc2-ae64cffca4e8] Received event network-vif-plugged-48f0868f-94db-42f7-8153-d1a27fa02707 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 696.744997] env[63515]: DEBUG oslo_concurrency.lockutils [req-17637906-176d-42a2-82a1-8c1260397191 req-3019c6d9-c0e1-4c6c-90a3-0ad69e7596d0 service nova] Acquiring lock "f1d01b75-ac9d-458d-8cc2-ae64cffca4e8-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 696.745944] env[63515]: DEBUG oslo_concurrency.lockutils [req-17637906-176d-42a2-82a1-8c1260397191 req-3019c6d9-c0e1-4c6c-90a3-0ad69e7596d0 service nova] Lock "f1d01b75-ac9d-458d-8cc2-ae64cffca4e8-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 696.745944] env[63515]: DEBUG oslo_concurrency.lockutils [req-17637906-176d-42a2-82a1-8c1260397191 req-3019c6d9-c0e1-4c6c-90a3-0ad69e7596d0 service nova] Lock "f1d01b75-ac9d-458d-8cc2-ae64cffca4e8-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 696.745944] env[63515]: DEBUG nova.compute.manager [req-17637906-176d-42a2-82a1-8c1260397191 req-3019c6d9-c0e1-4c6c-90a3-0ad69e7596d0 service nova] [instance: f1d01b75-ac9d-458d-8cc2-ae64cffca4e8] No waiting events found dispatching network-vif-plugged-48f0868f-94db-42f7-8153-d1a27fa02707 {{(pid=63515) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 696.745944] env[63515]: WARNING nova.compute.manager [req-17637906-176d-42a2-82a1-8c1260397191 req-3019c6d9-c0e1-4c6c-90a3-0ad69e7596d0 service nova] [instance: f1d01b75-ac9d-458d-8cc2-ae64cffca4e8] Received unexpected event network-vif-plugged-48f0868f-94db-42f7-8153-d1a27fa02707 for instance with vm_state building and task_state spawning. [ 696.818785] env[63515]: DEBUG oslo_vmware.api [None req-c7572d27-5b85-4acb-89ec-c56462911d65 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] Task: {'id': task-1110918, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.545913} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 696.819498] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-c7572d27-5b85-4acb-89ec-c56462911d65 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk to [datastore1] 5b42f744-fdd6-45b1-8563-896869648c23/5b42f744-fdd6-45b1-8563-896869648c23.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 696.819911] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-c7572d27-5b85-4acb-89ec-c56462911d65 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] [instance: 5b42f744-fdd6-45b1-8563-896869648c23] Extending root virtual disk to 1048576 {{(pid=63515) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 696.820296] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-97225b46-6d06-463c-89f0-a96143b6938c {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.828125] env[63515]: DEBUG oslo_vmware.api [None req-c7572d27-5b85-4acb-89ec-c56462911d65 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] Waiting for the task: (returnval){ [ 696.828125] env[63515]: value = "task-1110920" [ 696.828125] env[63515]: _type = "Task" [ 696.828125] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 696.839432] env[63515]: DEBUG oslo_vmware.api [None req-c7572d27-5b85-4acb-89ec-c56462911d65 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] Task: {'id': task-1110920, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 696.866994] env[63515]: DEBUG nova.compute.manager [None req-c26675e2-4152-417b-99dd-0c18de4dc137 tempest-ServersTestBootFromVolume-2042560423 tempest-ServersTestBootFromVolume-2042560423-project-member] [instance: 90f4930b-aaa0-4c4b-9ab8-92aed45e200b] Start building block device mappings for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 696.872499] env[63515]: DEBUG nova.network.neutron [None req-c26675e2-4152-417b-99dd-0c18de4dc137 tempest-ServersTestBootFromVolume-2042560423 tempest-ServersTestBootFromVolume-2042560423-project-member] [instance: 90f4930b-aaa0-4c4b-9ab8-92aed45e200b] Successfully created port: a35de7c4-c1af-482b-99a5-1059368cefbb {{(pid=63515) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 696.930838] env[63515]: DEBUG oslo_concurrency.lockutils [None req-89884269-25d4-4981-bd3b-a346ece7b681 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Acquiring lock "refresh_cache-f1d01b75-ac9d-458d-8cc2-ae64cffca4e8" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 696.930977] env[63515]: DEBUG oslo_concurrency.lockutils [None req-89884269-25d4-4981-bd3b-a346ece7b681 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Acquired lock "refresh_cache-f1d01b75-ac9d-458d-8cc2-ae64cffca4e8" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 696.931181] env[63515]: DEBUG nova.network.neutron [None req-89884269-25d4-4981-bd3b-a346ece7b681 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] [instance: f1d01b75-ac9d-458d-8cc2-ae64cffca4e8] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 697.097216] env[63515]: DEBUG oslo_concurrency.lockutils [None req-83fbd51b-fc37-49eb-8423-cdaa019d3b0c tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Acquiring lock "5cbce760-0163-4b27-8ae3-e46c926c8916" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 697.097216] env[63515]: DEBUG oslo_concurrency.lockutils [None req-83fbd51b-fc37-49eb-8423-cdaa019d3b0c tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Lock "5cbce760-0163-4b27-8ae3-e46c926c8916" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 697.097216] env[63515]: DEBUG oslo_vmware.api [None req-5690e97d-f63c-4f8e-9c79-d9c523fc3720 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Task: {'id': task-1110919, 'name': PowerOffVM_Task, 'duration_secs': 0.173558} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 697.097216] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-5690e97d-f63c-4f8e-9c79-d9c523fc3720 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: 3ad04a9f-62cf-4b1c-aab2-7dc4a219edc6] Powered off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 697.097689] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-5690e97d-f63c-4f8e-9c79-d9c523fc3720 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: 3ad04a9f-62cf-4b1c-aab2-7dc4a219edc6] Unregistering the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 697.097689] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-359079ef-6b5d-4476-9d86-522039af4254 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.339883] env[63515]: DEBUG oslo_vmware.api [None req-c7572d27-5b85-4acb-89ec-c56462911d65 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] Task: {'id': task-1110920, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.212497} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 697.340240] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-c7572d27-5b85-4acb-89ec-c56462911d65 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] [instance: 5b42f744-fdd6-45b1-8563-896869648c23] Extended root virtual disk {{(pid=63515) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 697.341041] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a63f664a-58e2-48c2-b755-aba5430dc83e {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.366905] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-c7572d27-5b85-4acb-89ec-c56462911d65 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] [instance: 5b42f744-fdd6-45b1-8563-896869648c23] Reconfiguring VM instance instance-00000021 to attach disk [datastore1] 5b42f744-fdd6-45b1-8563-896869648c23/5b42f744-fdd6-45b1-8563-896869648c23.vmdk or device None with type sparse {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 697.369842] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4d9b8bc3-3d54-49ac-b045-acdb347ba9e2 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.385440] env[63515]: INFO nova.virt.block_device [None req-c26675e2-4152-417b-99dd-0c18de4dc137 tempest-ServersTestBootFromVolume-2042560423 tempest-ServersTestBootFromVolume-2042560423-project-member] [instance: 90f4930b-aaa0-4c4b-9ab8-92aed45e200b] Booting with volume 921350b3-bd4a-4fa1-848a-057b53f51f5c at /dev/sda [ 697.403877] env[63515]: DEBUG oslo_vmware.api [None req-c7572d27-5b85-4acb-89ec-c56462911d65 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] Waiting for the task: (returnval){ [ 697.403877] env[63515]: value = "task-1110922" [ 697.403877] env[63515]: _type = "Task" [ 697.403877] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 697.412110] env[63515]: DEBUG oslo_vmware.api [None req-c7572d27-5b85-4acb-89ec-c56462911d65 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] Task: {'id': task-1110922, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 697.417038] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-edd23d87-ff36-4fbe-bf23-8a1f570bbd11 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.423555] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aaa214ee-69b8-4e4c-a65e-8a91ffcc7d4f {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.462075] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cfb092b7-5494-4aa0-9ff7-34124c7f2d90 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.464666] env[63515]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ad82eeb7-bbcc-4bad-9004-50af6e3b2d63 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.476790] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b58160d-39ff-4f8b-add2-76f574737d83 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.490786] env[63515]: DEBUG nova.compute.provider_tree [None req-33f5b097-4c20-4377-b5be-7e558874e5d3 tempest-ServersV294TestFqdnHostnames-2112993702 tempest-ServersV294TestFqdnHostnames-2112993702-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 697.495441] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9cb351b-bbd4-4636-8c71-75f866b441a8 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.513034] env[63515]: DEBUG nova.network.neutron [None req-89884269-25d4-4981-bd3b-a346ece7b681 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] [instance: f1d01b75-ac9d-458d-8cc2-ae64cffca4e8] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 697.521013] env[63515]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ddc8218d-544e-4392-abff-cc8c600c6883 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.528780] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ba9881e-2970-427d-9834-cd6949398689 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.552745] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-351173b1-5886-4d95-97e7-fca4008382ac {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.559685] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bde9bee6-27a8-42ae-b78c-3f48e5ddf47b {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.573948] env[63515]: DEBUG nova.virt.block_device [None req-c26675e2-4152-417b-99dd-0c18de4dc137 tempest-ServersTestBootFromVolume-2042560423 tempest-ServersTestBootFromVolume-2042560423-project-member] [instance: 90f4930b-aaa0-4c4b-9ab8-92aed45e200b] Updating existing volume attachment record: 1387c5da-c79f-4d4d-b1a3-40667ffbd711 {{(pid=63515) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 697.835433] env[63515]: DEBUG nova.network.neutron [None req-89884269-25d4-4981-bd3b-a346ece7b681 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] [instance: f1d01b75-ac9d-458d-8cc2-ae64cffca4e8] Updating instance_info_cache with network_info: [{"id": "48f0868f-94db-42f7-8153-d1a27fa02707", "address": "fa:16:3e:7a:74:da", "network": {"id": "b4abcbd0-0eac-4bde-ad11-7f93afcedcef", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.21", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "b569255cc43e42e7ae3f2b4ad37c6ef4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0cd5d325-3053-407e-a4ee-f627e82a23f9", "external-id": "nsx-vlan-transportzone-809", "segmentation_id": 809, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap48f0868f-94", "ovs_interfaceid": "48f0868f-94db-42f7-8153-d1a27fa02707", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 697.919379] env[63515]: DEBUG oslo_vmware.api [None req-c7572d27-5b85-4acb-89ec-c56462911d65 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] Task: {'id': task-1110922, 'name': ReconfigVM_Task, 'duration_secs': 0.306724} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 697.919593] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-c7572d27-5b85-4acb-89ec-c56462911d65 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] [instance: 5b42f744-fdd6-45b1-8563-896869648c23] Reconfigured VM instance instance-00000021 to attach disk [datastore1] 5b42f744-fdd6-45b1-8563-896869648c23/5b42f744-fdd6-45b1-8563-896869648c23.vmdk or device None with type sparse {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 697.921270] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a2d6190c-35af-4b0d-b405-0fa5e49f54b9 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.928718] env[63515]: DEBUG oslo_vmware.api [None req-c7572d27-5b85-4acb-89ec-c56462911d65 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] Waiting for the task: (returnval){ [ 697.928718] env[63515]: value = "task-1110923" [ 697.928718] env[63515]: _type = "Task" [ 697.928718] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 697.938152] env[63515]: DEBUG oslo_vmware.api [None req-c7572d27-5b85-4acb-89ec-c56462911d65 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] Task: {'id': task-1110923, 'name': Rename_Task} progress is 5%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 697.992770] env[63515]: DEBUG nova.scheduler.client.report [None req-33f5b097-4c20-4377-b5be-7e558874e5d3 tempest-ServersV294TestFqdnHostnames-2112993702 tempest-ServersV294TestFqdnHostnames-2112993702-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 698.130178] env[63515]: DEBUG oslo_concurrency.lockutils [None req-fa0cfe6d-3452-46d4-9987-2cb422956d62 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Acquiring lock "63ce797d-7180-4209-ac2c-81978bf7607a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 698.130863] env[63515]: DEBUG oslo_concurrency.lockutils [None req-fa0cfe6d-3452-46d4-9987-2cb422956d62 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Lock "63ce797d-7180-4209-ac2c-81978bf7607a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 698.341068] env[63515]: DEBUG oslo_concurrency.lockutils [None req-89884269-25d4-4981-bd3b-a346ece7b681 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Releasing lock "refresh_cache-f1d01b75-ac9d-458d-8cc2-ae64cffca4e8" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 698.341440] env[63515]: DEBUG nova.compute.manager [None req-89884269-25d4-4981-bd3b-a346ece7b681 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] [instance: f1d01b75-ac9d-458d-8cc2-ae64cffca4e8] Instance network_info: |[{"id": "48f0868f-94db-42f7-8153-d1a27fa02707", "address": "fa:16:3e:7a:74:da", "network": {"id": "b4abcbd0-0eac-4bde-ad11-7f93afcedcef", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.21", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "b569255cc43e42e7ae3f2b4ad37c6ef4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0cd5d325-3053-407e-a4ee-f627e82a23f9", "external-id": "nsx-vlan-transportzone-809", "segmentation_id": 809, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap48f0868f-94", "ovs_interfaceid": "48f0868f-94db-42f7-8153-d1a27fa02707", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 698.341932] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-89884269-25d4-4981-bd3b-a346ece7b681 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] [instance: f1d01b75-ac9d-458d-8cc2-ae64cffca4e8] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:7a:74:da', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '0cd5d325-3053-407e-a4ee-f627e82a23f9', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '48f0868f-94db-42f7-8153-d1a27fa02707', 'vif_model': 'vmxnet3'}] {{(pid=63515) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 698.356672] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-89884269-25d4-4981-bd3b-a346ece7b681 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Creating folder: Project (d3d0957a488b47559794495a451f5437). Parent ref: group-v243370. {{(pid=63515) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 698.356950] env[63515]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-90366392-57ad-407b-bf0d-91f9fd6f8b9d {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.368246] env[63515]: INFO nova.virt.vmwareapi.vm_util [None req-89884269-25d4-4981-bd3b-a346ece7b681 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Created folder: Project (d3d0957a488b47559794495a451f5437) in parent group-v243370. [ 698.368377] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-89884269-25d4-4981-bd3b-a346ece7b681 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Creating folder: Instances. Parent ref: group-v243392. {{(pid=63515) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 698.368596] env[63515]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a55c8c5e-c80e-46e5-8f75-881ec01b047d {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.382030] env[63515]: INFO nova.virt.vmwareapi.vm_util [None req-89884269-25d4-4981-bd3b-a346ece7b681 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Created folder: Instances in parent group-v243392. [ 698.382030] env[63515]: DEBUG oslo.service.loopingcall [None req-89884269-25d4-4981-bd3b-a346ece7b681 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 698.382154] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f1d01b75-ac9d-458d-8cc2-ae64cffca4e8] Creating VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 698.382370] env[63515]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-5e518252-4615-4117-8840-a3c42c77b926 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.406792] env[63515]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 698.406792] env[63515]: value = "task-1110926" [ 698.406792] env[63515]: _type = "Task" [ 698.406792] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 698.416050] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1110926, 'name': CreateVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 698.444021] env[63515]: DEBUG oslo_vmware.api [None req-c7572d27-5b85-4acb-89ec-c56462911d65 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] Task: {'id': task-1110923, 'name': Rename_Task, 'duration_secs': 0.162399} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 698.444021] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-c7572d27-5b85-4acb-89ec-c56462911d65 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] [instance: 5b42f744-fdd6-45b1-8563-896869648c23] Powering on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 698.444021] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ec6df2ef-460e-4a10-957b-39bb4d3557d6 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.448156] env[63515]: DEBUG oslo_vmware.api [None req-c7572d27-5b85-4acb-89ec-c56462911d65 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] Waiting for the task: (returnval){ [ 698.448156] env[63515]: value = "task-1110927" [ 698.448156] env[63515]: _type = "Task" [ 698.448156] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 698.459107] env[63515]: DEBUG oslo_vmware.api [None req-c7572d27-5b85-4acb-89ec-c56462911d65 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] Task: {'id': task-1110927, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 698.501919] env[63515]: DEBUG oslo_concurrency.lockutils [None req-33f5b097-4c20-4377-b5be-7e558874e5d3 tempest-ServersV294TestFqdnHostnames-2112993702 tempest-ServersV294TestFqdnHostnames-2112993702-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.654s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 698.502757] env[63515]: DEBUG nova.compute.manager [None req-33f5b097-4c20-4377-b5be-7e558874e5d3 tempest-ServersV294TestFqdnHostnames-2112993702 tempest-ServersV294TestFqdnHostnames-2112993702-project-member] [instance: 396e49dd-48c0-496b-a1ec-190c33a22c5e] Start building networks asynchronously for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 698.507366] env[63515]: DEBUG oslo_concurrency.lockutils [None req-4fa686a6-dc45-4603-a4ee-7107c810126f tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.160s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 698.509237] env[63515]: INFO nova.compute.claims [None req-4fa686a6-dc45-4603-a4ee-7107c810126f tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] [instance: 4087b2c4-6ed1-4b68-8b78-a36e34d935b1] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 698.590731] env[63515]: DEBUG nova.network.neutron [None req-c26675e2-4152-417b-99dd-0c18de4dc137 tempest-ServersTestBootFromVolume-2042560423 tempest-ServersTestBootFromVolume-2042560423-project-member] [instance: 90f4930b-aaa0-4c4b-9ab8-92aed45e200b] Successfully updated port: a35de7c4-c1af-482b-99a5-1059368cefbb {{(pid=63515) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 698.631571] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-5690e97d-f63c-4f8e-9c79-d9c523fc3720 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: 3ad04a9f-62cf-4b1c-aab2-7dc4a219edc6] Unregistered the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 698.631852] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-5690e97d-f63c-4f8e-9c79-d9c523fc3720 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: 3ad04a9f-62cf-4b1c-aab2-7dc4a219edc6] Deleting contents of the VM from datastore datastore1 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 698.632063] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-5690e97d-f63c-4f8e-9c79-d9c523fc3720 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Deleting the datastore file [datastore1] 3ad04a9f-62cf-4b1c-aab2-7dc4a219edc6 {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 698.632375] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-36212832-8b8d-4d79-903b-068ce3a063f7 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.640407] env[63515]: DEBUG oslo_vmware.api [None req-5690e97d-f63c-4f8e-9c79-d9c523fc3720 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Waiting for the task: (returnval){ [ 698.640407] env[63515]: value = "task-1110928" [ 698.640407] env[63515]: _type = "Task" [ 698.640407] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 698.648890] env[63515]: DEBUG oslo_vmware.api [None req-5690e97d-f63c-4f8e-9c79-d9c523fc3720 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Task: {'id': task-1110928, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 698.781977] env[63515]: DEBUG nova.compute.manager [req-b0dbd36d-c9a7-48a0-893a-09440004509b req-c532b567-9934-40f2-a098-2090daa2caf0 service nova] [instance: f1d01b75-ac9d-458d-8cc2-ae64cffca4e8] Received event network-changed-48f0868f-94db-42f7-8153-d1a27fa02707 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 698.782435] env[63515]: DEBUG nova.compute.manager [req-b0dbd36d-c9a7-48a0-893a-09440004509b req-c532b567-9934-40f2-a098-2090daa2caf0 service nova] [instance: f1d01b75-ac9d-458d-8cc2-ae64cffca4e8] Refreshing instance network info cache due to event network-changed-48f0868f-94db-42f7-8153-d1a27fa02707. {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 698.783579] env[63515]: DEBUG oslo_concurrency.lockutils [req-b0dbd36d-c9a7-48a0-893a-09440004509b req-c532b567-9934-40f2-a098-2090daa2caf0 service nova] Acquiring lock "refresh_cache-f1d01b75-ac9d-458d-8cc2-ae64cffca4e8" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 698.783579] env[63515]: DEBUG oslo_concurrency.lockutils [req-b0dbd36d-c9a7-48a0-893a-09440004509b req-c532b567-9934-40f2-a098-2090daa2caf0 service nova] Acquired lock "refresh_cache-f1d01b75-ac9d-458d-8cc2-ae64cffca4e8" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 698.783579] env[63515]: DEBUG nova.network.neutron [req-b0dbd36d-c9a7-48a0-893a-09440004509b req-c532b567-9934-40f2-a098-2090daa2caf0 service nova] [instance: f1d01b75-ac9d-458d-8cc2-ae64cffca4e8] Refreshing network info cache for port 48f0868f-94db-42f7-8153-d1a27fa02707 {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 698.918076] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1110926, 'name': CreateVM_Task, 'duration_secs': 0.429169} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 698.918385] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f1d01b75-ac9d-458d-8cc2-ae64cffca4e8] Created VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 698.919406] env[63515]: DEBUG oslo_vmware.service [None req-89884269-25d4-4981-bd3b-a346ece7b681 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d269ad6a-14b3-471f-8817-21ecfafa97da {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.925846] env[63515]: DEBUG oslo_concurrency.lockutils [None req-89884269-25d4-4981-bd3b-a346ece7b681 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 698.926634] env[63515]: DEBUG oslo_concurrency.lockutils [None req-89884269-25d4-4981-bd3b-a346ece7b681 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 698.927215] env[63515]: DEBUG oslo_concurrency.lockutils [None req-89884269-25d4-4981-bd3b-a346ece7b681 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 698.927791] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-63c87bdd-f087-4154-8a44-306c60740c46 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.933075] env[63515]: DEBUG oslo_vmware.api [None req-89884269-25d4-4981-bd3b-a346ece7b681 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Waiting for the task: (returnval){ [ 698.933075] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]5213f88e-91bf-e87b-e401-66e2717408d9" [ 698.933075] env[63515]: _type = "Task" [ 698.933075] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 698.943212] env[63515]: DEBUG oslo_vmware.api [None req-89884269-25d4-4981-bd3b-a346ece7b681 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]5213f88e-91bf-e87b-e401-66e2717408d9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 698.958264] env[63515]: DEBUG oslo_vmware.api [None req-c7572d27-5b85-4acb-89ec-c56462911d65 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] Task: {'id': task-1110927, 'name': PowerOnVM_Task} progress is 100%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 699.015026] env[63515]: DEBUG nova.compute.utils [None req-33f5b097-4c20-4377-b5be-7e558874e5d3 tempest-ServersV294TestFqdnHostnames-2112993702 tempest-ServersV294TestFqdnHostnames-2112993702-project-member] Using /dev/sd instead of None {{(pid=63515) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 699.017313] env[63515]: DEBUG nova.compute.manager [None req-33f5b097-4c20-4377-b5be-7e558874e5d3 tempest-ServersV294TestFqdnHostnames-2112993702 tempest-ServersV294TestFqdnHostnames-2112993702-project-member] [instance: 396e49dd-48c0-496b-a1ec-190c33a22c5e] Allocating IP information in the background. {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 699.017420] env[63515]: DEBUG nova.network.neutron [None req-33f5b097-4c20-4377-b5be-7e558874e5d3 tempest-ServersV294TestFqdnHostnames-2112993702 tempest-ServersV294TestFqdnHostnames-2112993702-project-member] [instance: 396e49dd-48c0-496b-a1ec-190c33a22c5e] allocate_for_instance() {{(pid=63515) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 699.069373] env[63515]: DEBUG nova.policy [None req-33f5b097-4c20-4377-b5be-7e558874e5d3 tempest-ServersV294TestFqdnHostnames-2112993702 tempest-ServersV294TestFqdnHostnames-2112993702-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a750ac57f5f147b292c4bf4d1608b351', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '62ac151dc34e41e7bf134563228d9d63', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63515) authorize /opt/stack/nova/nova/policy.py:201}} [ 699.093667] env[63515]: DEBUG oslo_concurrency.lockutils [None req-c26675e2-4152-417b-99dd-0c18de4dc137 tempest-ServersTestBootFromVolume-2042560423 tempest-ServersTestBootFromVolume-2042560423-project-member] Acquiring lock "refresh_cache-90f4930b-aaa0-4c4b-9ab8-92aed45e200b" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 699.094134] env[63515]: DEBUG oslo_concurrency.lockutils [None req-c26675e2-4152-417b-99dd-0c18de4dc137 tempest-ServersTestBootFromVolume-2042560423 tempest-ServersTestBootFromVolume-2042560423-project-member] Acquired lock "refresh_cache-90f4930b-aaa0-4c4b-9ab8-92aed45e200b" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 699.094134] env[63515]: DEBUG nova.network.neutron [None req-c26675e2-4152-417b-99dd-0c18de4dc137 tempest-ServersTestBootFromVolume-2042560423 tempest-ServersTestBootFromVolume-2042560423-project-member] [instance: 90f4930b-aaa0-4c4b-9ab8-92aed45e200b] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 699.152878] env[63515]: DEBUG oslo_vmware.api [None req-5690e97d-f63c-4f8e-9c79-d9c523fc3720 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Task: {'id': task-1110928, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.17639} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 699.153151] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-5690e97d-f63c-4f8e-9c79-d9c523fc3720 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Deleted the datastore file {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 699.153335] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-5690e97d-f63c-4f8e-9c79-d9c523fc3720 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: 3ad04a9f-62cf-4b1c-aab2-7dc4a219edc6] Deleted contents of the VM from datastore datastore1 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 699.153508] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-5690e97d-f63c-4f8e-9c79-d9c523fc3720 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: 3ad04a9f-62cf-4b1c-aab2-7dc4a219edc6] Instance destroyed {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 699.153674] env[63515]: INFO nova.compute.manager [None req-5690e97d-f63c-4f8e-9c79-d9c523fc3720 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: 3ad04a9f-62cf-4b1c-aab2-7dc4a219edc6] Took 2.60 seconds to destroy the instance on the hypervisor. [ 699.153910] env[63515]: DEBUG oslo.service.loopingcall [None req-5690e97d-f63c-4f8e-9c79-d9c523fc3720 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 699.154145] env[63515]: DEBUG nova.compute.manager [-] [instance: 3ad04a9f-62cf-4b1c-aab2-7dc4a219edc6] Deallocating network for instance {{(pid=63515) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 699.154187] env[63515]: DEBUG nova.network.neutron [-] [instance: 3ad04a9f-62cf-4b1c-aab2-7dc4a219edc6] deallocate_for_instance() {{(pid=63515) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 699.445906] env[63515]: DEBUG oslo_concurrency.lockutils [None req-89884269-25d4-4981-bd3b-a346ece7b681 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 699.447563] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-89884269-25d4-4981-bd3b-a346ece7b681 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] [instance: f1d01b75-ac9d-458d-8cc2-ae64cffca4e8] Processing image 8a120570-cb06-4099-b262-554ca0ad15c5 {{(pid=63515) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 699.447821] env[63515]: DEBUG oslo_concurrency.lockutils [None req-89884269-25d4-4981-bd3b-a346ece7b681 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 699.447972] env[63515]: DEBUG oslo_concurrency.lockutils [None req-89884269-25d4-4981-bd3b-a346ece7b681 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 699.448169] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-89884269-25d4-4981-bd3b-a346ece7b681 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 699.448403] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-34606221-3b0e-4105-b98a-1173af77967e {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.468029] env[63515]: DEBUG oslo_vmware.api [None req-c7572d27-5b85-4acb-89ec-c56462911d65 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] Task: {'id': task-1110927, 'name': PowerOnVM_Task, 'duration_secs': 0.522488} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 699.472019] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-c7572d27-5b85-4acb-89ec-c56462911d65 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] [instance: 5b42f744-fdd6-45b1-8563-896869648c23] Powered on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 699.472019] env[63515]: INFO nova.compute.manager [None req-c7572d27-5b85-4acb-89ec-c56462911d65 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] [instance: 5b42f744-fdd6-45b1-8563-896869648c23] Took 8.66 seconds to spawn the instance on the hypervisor. [ 699.472019] env[63515]: DEBUG nova.compute.manager [None req-c7572d27-5b85-4acb-89ec-c56462911d65 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] [instance: 5b42f744-fdd6-45b1-8563-896869648c23] Checking state {{(pid=63515) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 699.472019] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-89884269-25d4-4981-bd3b-a346ece7b681 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 699.472019] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-89884269-25d4-4981-bd3b-a346ece7b681 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63515) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 699.472263] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3f41960-2836-494f-a93b-46115edab201 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.474710] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-183b5355-d7d9-451f-9827-dc90a18f9028 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.482682] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-592a4ada-6f5a-4600-9ca9-f3dfdd08d47a {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.490597] env[63515]: DEBUG oslo_vmware.api [None req-89884269-25d4-4981-bd3b-a346ece7b681 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Waiting for the task: (returnval){ [ 699.490597] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]52c043ef-d004-9241-f820-e89c442c8a47" [ 699.490597] env[63515]: _type = "Task" [ 699.490597] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 699.499274] env[63515]: DEBUG oslo_vmware.api [None req-89884269-25d4-4981-bd3b-a346ece7b681 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52c043ef-d004-9241-f820-e89c442c8a47, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 699.518835] env[63515]: DEBUG nova.compute.manager [None req-33f5b097-4c20-4377-b5be-7e558874e5d3 tempest-ServersV294TestFqdnHostnames-2112993702 tempest-ServersV294TestFqdnHostnames-2112993702-project-member] [instance: 396e49dd-48c0-496b-a1ec-190c33a22c5e] Start building block device mappings for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 699.669288] env[63515]: DEBUG nova.network.neutron [None req-c26675e2-4152-417b-99dd-0c18de4dc137 tempest-ServersTestBootFromVolume-2042560423 tempest-ServersTestBootFromVolume-2042560423-project-member] [instance: 90f4930b-aaa0-4c4b-9ab8-92aed45e200b] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 699.725227] env[63515]: DEBUG nova.compute.manager [None req-c26675e2-4152-417b-99dd-0c18de4dc137 tempest-ServersTestBootFromVolume-2042560423 tempest-ServersTestBootFromVolume-2042560423-project-member] [instance: 90f4930b-aaa0-4c4b-9ab8-92aed45e200b] Start spawning the instance on the hypervisor. {{(pid=63515) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 699.727142] env[63515]: DEBUG nova.virt.hardware [None req-c26675e2-4152-417b-99dd-0c18de4dc137 tempest-ServersTestBootFromVolume-2042560423 tempest-ServersTestBootFromVolume-2042560423-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T02:52:14Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=0,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=1073741824,status='active',tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 699.727529] env[63515]: DEBUG nova.virt.hardware [None req-c26675e2-4152-417b-99dd-0c18de4dc137 tempest-ServersTestBootFromVolume-2042560423 tempest-ServersTestBootFromVolume-2042560423-project-member] Flavor limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 699.727807] env[63515]: DEBUG nova.virt.hardware [None req-c26675e2-4152-417b-99dd-0c18de4dc137 tempest-ServersTestBootFromVolume-2042560423 tempest-ServersTestBootFromVolume-2042560423-project-member] Image limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 699.728109] env[63515]: DEBUG nova.virt.hardware [None req-c26675e2-4152-417b-99dd-0c18de4dc137 tempest-ServersTestBootFromVolume-2042560423 tempest-ServersTestBootFromVolume-2042560423-project-member] Flavor pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 699.728382] env[63515]: DEBUG nova.virt.hardware [None req-c26675e2-4152-417b-99dd-0c18de4dc137 tempest-ServersTestBootFromVolume-2042560423 tempest-ServersTestBootFromVolume-2042560423-project-member] Image pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 699.729025] env[63515]: DEBUG nova.virt.hardware [None req-c26675e2-4152-417b-99dd-0c18de4dc137 tempest-ServersTestBootFromVolume-2042560423 tempest-ServersTestBootFromVolume-2042560423-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 699.729025] env[63515]: DEBUG nova.virt.hardware [None req-c26675e2-4152-417b-99dd-0c18de4dc137 tempest-ServersTestBootFromVolume-2042560423 tempest-ServersTestBootFromVolume-2042560423-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 699.730021] env[63515]: DEBUG nova.virt.hardware [None req-c26675e2-4152-417b-99dd-0c18de4dc137 tempest-ServersTestBootFromVolume-2042560423 tempest-ServersTestBootFromVolume-2042560423-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 699.730021] env[63515]: DEBUG nova.virt.hardware [None req-c26675e2-4152-417b-99dd-0c18de4dc137 tempest-ServersTestBootFromVolume-2042560423 tempest-ServersTestBootFromVolume-2042560423-project-member] Got 1 possible topologies {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 699.730021] env[63515]: DEBUG nova.virt.hardware [None req-c26675e2-4152-417b-99dd-0c18de4dc137 tempest-ServersTestBootFromVolume-2042560423 tempest-ServersTestBootFromVolume-2042560423-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 699.730021] env[63515]: DEBUG nova.virt.hardware [None req-c26675e2-4152-417b-99dd-0c18de4dc137 tempest-ServersTestBootFromVolume-2042560423 tempest-ServersTestBootFromVolume-2042560423-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 699.730771] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c4b0ea2-2692-4811-aede-5c061f98a041 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.737756] env[63515]: DEBUG nova.network.neutron [None req-33f5b097-4c20-4377-b5be-7e558874e5d3 tempest-ServersV294TestFqdnHostnames-2112993702 tempest-ServersV294TestFqdnHostnames-2112993702-project-member] [instance: 396e49dd-48c0-496b-a1ec-190c33a22c5e] Successfully created port: 77fbf94a-db5a-4569-b4e7-2d9edbe411a1 {{(pid=63515) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 699.746038] env[63515]: DEBUG nova.network.neutron [req-b0dbd36d-c9a7-48a0-893a-09440004509b req-c532b567-9934-40f2-a098-2090daa2caf0 service nova] [instance: f1d01b75-ac9d-458d-8cc2-ae64cffca4e8] Updated VIF entry in instance network info cache for port 48f0868f-94db-42f7-8153-d1a27fa02707. {{(pid=63515) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 699.746385] env[63515]: DEBUG nova.network.neutron [req-b0dbd36d-c9a7-48a0-893a-09440004509b req-c532b567-9934-40f2-a098-2090daa2caf0 service nova] [instance: f1d01b75-ac9d-458d-8cc2-ae64cffca4e8] Updating instance_info_cache with network_info: [{"id": "48f0868f-94db-42f7-8153-d1a27fa02707", "address": "fa:16:3e:7a:74:da", "network": {"id": "b4abcbd0-0eac-4bde-ad11-7f93afcedcef", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.21", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "b569255cc43e42e7ae3f2b4ad37c6ef4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0cd5d325-3053-407e-a4ee-f627e82a23f9", "external-id": "nsx-vlan-transportzone-809", "segmentation_id": 809, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap48f0868f-94", "ovs_interfaceid": "48f0868f-94db-42f7-8153-d1a27fa02707", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 699.751840] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cfd4cc69-a170-4bb7-8206-d3d4b74ca321 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.004695] env[63515]: INFO nova.compute.manager [None req-c7572d27-5b85-4acb-89ec-c56462911d65 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] [instance: 5b42f744-fdd6-45b1-8563-896869648c23] Took 29.30 seconds to build instance. [ 700.015178] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a743eefc-14bc-48ea-8054-1ea41c867ed6 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.018015] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-89884269-25d4-4981-bd3b-a346ece7b681 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] [instance: f1d01b75-ac9d-458d-8cc2-ae64cffca4e8] Preparing fetch location {{(pid=63515) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 700.019758] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-89884269-25d4-4981-bd3b-a346ece7b681 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Creating directory with path [datastore2] vmware_temp/7386036a-4262-4a30-b58b-706f0d52abbb/8a120570-cb06-4099-b262-554ca0ad15c5 {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 700.020044] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-7b1bfdf1-b242-4880-9dee-9187301a698f {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.033738] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fdf74d8d-bcdb-47a8-92ee-7eccf03b3d87 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.037341] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-89884269-25d4-4981-bd3b-a346ece7b681 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Created directory with path [datastore2] vmware_temp/7386036a-4262-4a30-b58b-706f0d52abbb/8a120570-cb06-4099-b262-554ca0ad15c5 {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 700.037542] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-89884269-25d4-4981-bd3b-a346ece7b681 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] [instance: f1d01b75-ac9d-458d-8cc2-ae64cffca4e8] Fetch image to [datastore2] vmware_temp/7386036a-4262-4a30-b58b-706f0d52abbb/8a120570-cb06-4099-b262-554ca0ad15c5/tmp-sparse.vmdk {{(pid=63515) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 700.037966] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-89884269-25d4-4981-bd3b-a346ece7b681 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] [instance: f1d01b75-ac9d-458d-8cc2-ae64cffca4e8] Downloading image file data 8a120570-cb06-4099-b262-554ca0ad15c5 to [datastore2] vmware_temp/7386036a-4262-4a30-b58b-706f0d52abbb/8a120570-cb06-4099-b262-554ca0ad15c5/tmp-sparse.vmdk on the data store datastore2 {{(pid=63515) _fetch_image_as_file /opt/stack/nova/nova/virt/vmwareapi/vmops.py:399}} [ 700.041676] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8093d856-a0ef-4895-8856-17298358c38c {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.078805] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40fb8c9d-bc72-4c91-879a-f2d2a959024a {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.084792] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ac0cbe6-b671-4ccb-92e1-37fa88d1bd24 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.096653] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec40b0af-f728-43d3-a5fb-54e5c6bdf7e9 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.102583] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05042e07-da56-4198-8de8-e431a5be1792 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.107289] env[63515]: DEBUG nova.network.neutron [None req-c26675e2-4152-417b-99dd-0c18de4dc137 tempest-ServersTestBootFromVolume-2042560423 tempest-ServersTestBootFromVolume-2042560423-project-member] [instance: 90f4930b-aaa0-4c4b-9ab8-92aed45e200b] Updating instance_info_cache with network_info: [{"id": "a35de7c4-c1af-482b-99a5-1059368cefbb", "address": "fa:16:3e:df:a6:62", "network": {"id": "d19ced2d-871d-4aa5-9b07-8ca41a46d948", "bridge": "br-int", "label": "tempest-ServersTestBootFromVolume-1210740991-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "38b2d712caf34cc08ad9c9694c521b7e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b2ede0e6-8d7a-4018-bb37-25bf388e9867", "external-id": "nsx-vlan-transportzone-945", "segmentation_id": 945, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa35de7c4-c1", "ovs_interfaceid": "a35de7c4-c1af-482b-99a5-1059368cefbb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 700.122074] env[63515]: DEBUG nova.compute.provider_tree [None req-4fa686a6-dc45-4603-a4ee-7107c810126f tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 700.150043] env[63515]: DEBUG nova.scheduler.client.report [None req-4fa686a6-dc45-4603-a4ee-7107c810126f tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 700.155490] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-084a8dc0-73e4-4344-b64f-3f9a50f6c592 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.165468] env[63515]: DEBUG nova.network.neutron [-] [instance: 3ad04a9f-62cf-4b1c-aab2-7dc4a219edc6] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 700.166524] env[63515]: DEBUG oslo_vmware.service [-] Invoking SessionManager.AcquireGenericServiceTicket with opID=oslo.vmware-6c5a701b-4cc3-4b0a-ab66-601e7b571fc7 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.190312] env[63515]: DEBUG nova.virt.vmwareapi.images [None req-89884269-25d4-4981-bd3b-a346ece7b681 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] [instance: f1d01b75-ac9d-458d-8cc2-ae64cffca4e8] Downloading image file data 8a120570-cb06-4099-b262-554ca0ad15c5 to the data store datastore2 {{(pid=63515) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:245}} [ 700.242229] env[63515]: DEBUG oslo_vmware.rw_handles [None req-89884269-25d4-4981-bd3b-a346ece7b681 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Creating HTTP connection to write to file with size = 21318656 and URL = https://esx7c2n1.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/7386036a-4262-4a30-b58b-706f0d52abbb/8a120570-cb06-4099-b262-554ca0ad15c5/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore2. {{(pid=63515) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 700.299355] env[63515]: DEBUG oslo_concurrency.lockutils [req-b0dbd36d-c9a7-48a0-893a-09440004509b req-c532b567-9934-40f2-a098-2090daa2caf0 service nova] Releasing lock "refresh_cache-f1d01b75-ac9d-458d-8cc2-ae64cffca4e8" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 700.299675] env[63515]: DEBUG nova.compute.manager [req-b0dbd36d-c9a7-48a0-893a-09440004509b req-c532b567-9934-40f2-a098-2090daa2caf0 service nova] [instance: 90f4930b-aaa0-4c4b-9ab8-92aed45e200b] Received event network-vif-plugged-a35de7c4-c1af-482b-99a5-1059368cefbb {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 700.299900] env[63515]: DEBUG oslo_concurrency.lockutils [req-b0dbd36d-c9a7-48a0-893a-09440004509b req-c532b567-9934-40f2-a098-2090daa2caf0 service nova] Acquiring lock "90f4930b-aaa0-4c4b-9ab8-92aed45e200b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 700.300153] env[63515]: DEBUG oslo_concurrency.lockutils [req-b0dbd36d-c9a7-48a0-893a-09440004509b req-c532b567-9934-40f2-a098-2090daa2caf0 service nova] Lock "90f4930b-aaa0-4c4b-9ab8-92aed45e200b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 700.300347] env[63515]: DEBUG oslo_concurrency.lockutils [req-b0dbd36d-c9a7-48a0-893a-09440004509b req-c532b567-9934-40f2-a098-2090daa2caf0 service nova] Lock "90f4930b-aaa0-4c4b-9ab8-92aed45e200b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 700.300520] env[63515]: DEBUG nova.compute.manager [req-b0dbd36d-c9a7-48a0-893a-09440004509b req-c532b567-9934-40f2-a098-2090daa2caf0 service nova] [instance: 90f4930b-aaa0-4c4b-9ab8-92aed45e200b] No waiting events found dispatching network-vif-plugged-a35de7c4-c1af-482b-99a5-1059368cefbb {{(pid=63515) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 700.300692] env[63515]: WARNING nova.compute.manager [req-b0dbd36d-c9a7-48a0-893a-09440004509b req-c532b567-9934-40f2-a098-2090daa2caf0 service nova] [instance: 90f4930b-aaa0-4c4b-9ab8-92aed45e200b] Received unexpected event network-vif-plugged-a35de7c4-c1af-482b-99a5-1059368cefbb for instance with vm_state building and task_state block_device_mapping. [ 700.300856] env[63515]: DEBUG nova.compute.manager [req-b0dbd36d-c9a7-48a0-893a-09440004509b req-c532b567-9934-40f2-a098-2090daa2caf0 service nova] [instance: 90f4930b-aaa0-4c4b-9ab8-92aed45e200b] Received event network-changed-a35de7c4-c1af-482b-99a5-1059368cefbb {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 700.301030] env[63515]: DEBUG nova.compute.manager [req-b0dbd36d-c9a7-48a0-893a-09440004509b req-c532b567-9934-40f2-a098-2090daa2caf0 service nova] [instance: 90f4930b-aaa0-4c4b-9ab8-92aed45e200b] Refreshing instance network info cache due to event network-changed-a35de7c4-c1af-482b-99a5-1059368cefbb. {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 700.301549] env[63515]: DEBUG oslo_concurrency.lockutils [req-b0dbd36d-c9a7-48a0-893a-09440004509b req-c532b567-9934-40f2-a098-2090daa2caf0 service nova] Acquiring lock "refresh_cache-90f4930b-aaa0-4c4b-9ab8-92aed45e200b" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 700.518410] env[63515]: DEBUG oslo_concurrency.lockutils [None req-c7572d27-5b85-4acb-89ec-c56462911d65 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] Lock "5b42f744-fdd6-45b1-8563-896869648c23" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 115.618s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 700.535585] env[63515]: DEBUG nova.compute.manager [None req-33f5b097-4c20-4377-b5be-7e558874e5d3 tempest-ServersV294TestFqdnHostnames-2112993702 tempest-ServersV294TestFqdnHostnames-2112993702-project-member] [instance: 396e49dd-48c0-496b-a1ec-190c33a22c5e] Start spawning the instance on the hypervisor. {{(pid=63515) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 700.558955] env[63515]: DEBUG nova.virt.hardware [None req-33f5b097-4c20-4377-b5be-7e558874e5d3 tempest-ServersV294TestFqdnHostnames-2112993702 tempest-ServersV294TestFqdnHostnames-2112993702-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T02:52:14Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T02:51:56Z,direct_url=,disk_format='vmdk',id=8a120570-cb06-4099-b262-554ca0ad15c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b569255cc43e42e7ae3f2b4ad37c6ef4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T02:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 700.559238] env[63515]: DEBUG nova.virt.hardware [None req-33f5b097-4c20-4377-b5be-7e558874e5d3 tempest-ServersV294TestFqdnHostnames-2112993702 tempest-ServersV294TestFqdnHostnames-2112993702-project-member] Flavor limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 700.559414] env[63515]: DEBUG nova.virt.hardware [None req-33f5b097-4c20-4377-b5be-7e558874e5d3 tempest-ServersV294TestFqdnHostnames-2112993702 tempest-ServersV294TestFqdnHostnames-2112993702-project-member] Image limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 700.559631] env[63515]: DEBUG nova.virt.hardware [None req-33f5b097-4c20-4377-b5be-7e558874e5d3 tempest-ServersV294TestFqdnHostnames-2112993702 tempest-ServersV294TestFqdnHostnames-2112993702-project-member] Flavor pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 700.559783] env[63515]: DEBUG nova.virt.hardware [None req-33f5b097-4c20-4377-b5be-7e558874e5d3 tempest-ServersV294TestFqdnHostnames-2112993702 tempest-ServersV294TestFqdnHostnames-2112993702-project-member] Image pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 700.559929] env[63515]: DEBUG nova.virt.hardware [None req-33f5b097-4c20-4377-b5be-7e558874e5d3 tempest-ServersV294TestFqdnHostnames-2112993702 tempest-ServersV294TestFqdnHostnames-2112993702-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 700.560145] env[63515]: DEBUG nova.virt.hardware [None req-33f5b097-4c20-4377-b5be-7e558874e5d3 tempest-ServersV294TestFqdnHostnames-2112993702 tempest-ServersV294TestFqdnHostnames-2112993702-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 700.560338] env[63515]: DEBUG nova.virt.hardware [None req-33f5b097-4c20-4377-b5be-7e558874e5d3 tempest-ServersV294TestFqdnHostnames-2112993702 tempest-ServersV294TestFqdnHostnames-2112993702-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 700.560539] env[63515]: DEBUG nova.virt.hardware [None req-33f5b097-4c20-4377-b5be-7e558874e5d3 tempest-ServersV294TestFqdnHostnames-2112993702 tempest-ServersV294TestFqdnHostnames-2112993702-project-member] Got 1 possible topologies {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 700.560715] env[63515]: DEBUG nova.virt.hardware [None req-33f5b097-4c20-4377-b5be-7e558874e5d3 tempest-ServersV294TestFqdnHostnames-2112993702 tempest-ServersV294TestFqdnHostnames-2112993702-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 700.560888] env[63515]: DEBUG nova.virt.hardware [None req-33f5b097-4c20-4377-b5be-7e558874e5d3 tempest-ServersV294TestFqdnHostnames-2112993702 tempest-ServersV294TestFqdnHostnames-2112993702-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 700.561908] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81356b3a-e924-43bd-b8f9-8bcc578e8759 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.573695] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7fe04dc2-2cc8-4a83-b476-dd55663d56a4 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.610320] env[63515]: DEBUG oslo_concurrency.lockutils [None req-c26675e2-4152-417b-99dd-0c18de4dc137 tempest-ServersTestBootFromVolume-2042560423 tempest-ServersTestBootFromVolume-2042560423-project-member] Releasing lock "refresh_cache-90f4930b-aaa0-4c4b-9ab8-92aed45e200b" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 700.610735] env[63515]: DEBUG nova.compute.manager [None req-c26675e2-4152-417b-99dd-0c18de4dc137 tempest-ServersTestBootFromVolume-2042560423 tempest-ServersTestBootFromVolume-2042560423-project-member] [instance: 90f4930b-aaa0-4c4b-9ab8-92aed45e200b] Instance network_info: |[{"id": "a35de7c4-c1af-482b-99a5-1059368cefbb", "address": "fa:16:3e:df:a6:62", "network": {"id": "d19ced2d-871d-4aa5-9b07-8ca41a46d948", "bridge": "br-int", "label": "tempest-ServersTestBootFromVolume-1210740991-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "38b2d712caf34cc08ad9c9694c521b7e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b2ede0e6-8d7a-4018-bb37-25bf388e9867", "external-id": "nsx-vlan-transportzone-945", "segmentation_id": 945, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa35de7c4-c1", "ovs_interfaceid": "a35de7c4-c1af-482b-99a5-1059368cefbb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 700.611491] env[63515]: DEBUG oslo_concurrency.lockutils [req-b0dbd36d-c9a7-48a0-893a-09440004509b req-c532b567-9934-40f2-a098-2090daa2caf0 service nova] Acquired lock "refresh_cache-90f4930b-aaa0-4c4b-9ab8-92aed45e200b" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 700.611695] env[63515]: DEBUG nova.network.neutron [req-b0dbd36d-c9a7-48a0-893a-09440004509b req-c532b567-9934-40f2-a098-2090daa2caf0 service nova] [instance: 90f4930b-aaa0-4c4b-9ab8-92aed45e200b] Refreshing network info cache for port a35de7c4-c1af-482b-99a5-1059368cefbb {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 700.612968] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-c26675e2-4152-417b-99dd-0c18de4dc137 tempest-ServersTestBootFromVolume-2042560423 tempest-ServersTestBootFromVolume-2042560423-project-member] [instance: 90f4930b-aaa0-4c4b-9ab8-92aed45e200b] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:df:a6:62', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'b2ede0e6-8d7a-4018-bb37-25bf388e9867', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a35de7c4-c1af-482b-99a5-1059368cefbb', 'vif_model': 'vmxnet3'}] {{(pid=63515) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 700.620253] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-c26675e2-4152-417b-99dd-0c18de4dc137 tempest-ServersTestBootFromVolume-2042560423 tempest-ServersTestBootFromVolume-2042560423-project-member] Creating folder: Project (38b2d712caf34cc08ad9c9694c521b7e). Parent ref: group-v243370. {{(pid=63515) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 700.626010] env[63515]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-ce32416d-dc8b-4c86-a536-60c05528c9e3 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.638116] env[63515]: WARNING suds.client [-] Web service reported a SOAP processing fault using an unexpected HTTP status code 200. Reporting as an internal server error. [ 700.638273] env[63515]: DEBUG oslo_vmware.api [-] Fault list: [DuplicateName] {{(pid=63515) _invoke_api /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:337}} [ 700.638591] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-c26675e2-4152-417b-99dd-0c18de4dc137 tempest-ServersTestBootFromVolume-2042560423 tempest-ServersTestBootFromVolume-2042560423-project-member] Folder already exists: Project (38b2d712caf34cc08ad9c9694c521b7e). Parent ref: group-v243370. {{(pid=63515) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1609}} [ 700.638783] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-c26675e2-4152-417b-99dd-0c18de4dc137 tempest-ServersTestBootFromVolume-2042560423 tempest-ServersTestBootFromVolume-2042560423-project-member] Creating folder: Instances. Parent ref: group-v243374. {{(pid=63515) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 700.639319] env[63515]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b576a925-42bb-4505-9316-eea4d84eece7 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.649658] env[63515]: INFO nova.virt.vmwareapi.vm_util [None req-c26675e2-4152-417b-99dd-0c18de4dc137 tempest-ServersTestBootFromVolume-2042560423 tempest-ServersTestBootFromVolume-2042560423-project-member] Created folder: Instances in parent group-v243374. [ 700.649894] env[63515]: DEBUG oslo.service.loopingcall [None req-c26675e2-4152-417b-99dd-0c18de4dc137 tempest-ServersTestBootFromVolume-2042560423 tempest-ServersTestBootFromVolume-2042560423-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 700.650276] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 90f4930b-aaa0-4c4b-9ab8-92aed45e200b] Creating VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 700.650384] env[63515]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-43af1343-e8b1-4d03-88d4-f582e81f30f4 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.665652] env[63515]: DEBUG oslo_concurrency.lockutils [None req-4fa686a6-dc45-4603-a4ee-7107c810126f tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.158s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 700.666165] env[63515]: DEBUG nova.compute.manager [None req-4fa686a6-dc45-4603-a4ee-7107c810126f tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] [instance: 4087b2c4-6ed1-4b68-8b78-a36e34d935b1] Start building networks asynchronously for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 700.669360] env[63515]: DEBUG oslo_concurrency.lockutils [None req-44a3080f-f2e2-4f0c-b244-122beee2aea9 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.308s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 700.670781] env[63515]: INFO nova.compute.claims [None req-44a3080f-f2e2-4f0c-b244-122beee2aea9 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] [instance: 87c468d9-9594-4804-b461-527f01f6118f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 700.674550] env[63515]: INFO nova.compute.manager [-] [instance: 3ad04a9f-62cf-4b1c-aab2-7dc4a219edc6] Took 1.52 seconds to deallocate network for instance. [ 700.682807] env[63515]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 700.682807] env[63515]: value = "task-1110931" [ 700.682807] env[63515]: _type = "Task" [ 700.682807] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 700.693839] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1110931, 'name': CreateVM_Task} progress is 5%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 700.810926] env[63515]: DEBUG nova.compute.manager [req-bdd3104a-2b49-4b03-9d47-fa4c7927417a req-6bf9564f-783f-4513-bd0e-7553d84083ac service nova] [instance: 3ad04a9f-62cf-4b1c-aab2-7dc4a219edc6] Received event network-vif-deleted-3dee3eb6-af97-422d-ac4f-74a994909386 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 700.892222] env[63515]: DEBUG oslo_vmware.rw_handles [None req-89884269-25d4-4981-bd3b-a346ece7b681 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Completed reading data from the image iterator. {{(pid=63515) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 700.892482] env[63515]: DEBUG oslo_vmware.rw_handles [None req-89884269-25d4-4981-bd3b-a346ece7b681 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Closing write handle for https://esx7c2n1.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/7386036a-4262-4a30-b58b-706f0d52abbb/8a120570-cb06-4099-b262-554ca0ad15c5/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore2. {{(pid=63515) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:281}} [ 700.921063] env[63515]: DEBUG nova.network.neutron [req-b0dbd36d-c9a7-48a0-893a-09440004509b req-c532b567-9934-40f2-a098-2090daa2caf0 service nova] [instance: 90f4930b-aaa0-4c4b-9ab8-92aed45e200b] Updated VIF entry in instance network info cache for port a35de7c4-c1af-482b-99a5-1059368cefbb. {{(pid=63515) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 700.921450] env[63515]: DEBUG nova.network.neutron [req-b0dbd36d-c9a7-48a0-893a-09440004509b req-c532b567-9934-40f2-a098-2090daa2caf0 service nova] [instance: 90f4930b-aaa0-4c4b-9ab8-92aed45e200b] Updating instance_info_cache with network_info: [{"id": "a35de7c4-c1af-482b-99a5-1059368cefbb", "address": "fa:16:3e:df:a6:62", "network": {"id": "d19ced2d-871d-4aa5-9b07-8ca41a46d948", "bridge": "br-int", "label": "tempest-ServersTestBootFromVolume-1210740991-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "38b2d712caf34cc08ad9c9694c521b7e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b2ede0e6-8d7a-4018-bb37-25bf388e9867", "external-id": "nsx-vlan-transportzone-945", "segmentation_id": 945, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa35de7c4-c1", "ovs_interfaceid": "a35de7c4-c1af-482b-99a5-1059368cefbb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 701.021448] env[63515]: DEBUG nova.compute.manager [None req-74f7de08-cdcf-4b98-bfdb-a2dd77c03e6d tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] [instance: b85e9a70-7f5b-4d32-b616-f2a97e3186c8] Starting instance... {{(pid=63515) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 701.040374] env[63515]: DEBUG nova.virt.vmwareapi.images [None req-89884269-25d4-4981-bd3b-a346ece7b681 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] [instance: f1d01b75-ac9d-458d-8cc2-ae64cffca4e8] Downloaded image file data 8a120570-cb06-4099-b262-554ca0ad15c5 to vmware_temp/7386036a-4262-4a30-b58b-706f0d52abbb/8a120570-cb06-4099-b262-554ca0ad15c5/tmp-sparse.vmdk on the data store datastore2 {{(pid=63515) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:258}} [ 701.042766] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-89884269-25d4-4981-bd3b-a346ece7b681 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] [instance: f1d01b75-ac9d-458d-8cc2-ae64cffca4e8] Caching image {{(pid=63515) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 701.043123] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-89884269-25d4-4981-bd3b-a346ece7b681 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Copying Virtual Disk [datastore2] vmware_temp/7386036a-4262-4a30-b58b-706f0d52abbb/8a120570-cb06-4099-b262-554ca0ad15c5/tmp-sparse.vmdk to [datastore2] vmware_temp/7386036a-4262-4a30-b58b-706f0d52abbb/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 701.043458] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7b180c0e-1d8f-4364-8659-44418074cb0f {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.050471] env[63515]: DEBUG oslo_vmware.api [None req-89884269-25d4-4981-bd3b-a346ece7b681 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Waiting for the task: (returnval){ [ 701.050471] env[63515]: value = "task-1110932" [ 701.050471] env[63515]: _type = "Task" [ 701.050471] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 701.059625] env[63515]: DEBUG oslo_vmware.api [None req-89884269-25d4-4981-bd3b-a346ece7b681 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Task: {'id': task-1110932, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 701.179046] env[63515]: DEBUG nova.compute.utils [None req-4fa686a6-dc45-4603-a4ee-7107c810126f tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] Using /dev/sd instead of None {{(pid=63515) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 701.182468] env[63515]: DEBUG nova.compute.manager [None req-4fa686a6-dc45-4603-a4ee-7107c810126f tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] [instance: 4087b2c4-6ed1-4b68-8b78-a36e34d935b1] Allocating IP information in the background. {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 701.182666] env[63515]: DEBUG nova.network.neutron [None req-4fa686a6-dc45-4603-a4ee-7107c810126f tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] [instance: 4087b2c4-6ed1-4b68-8b78-a36e34d935b1] allocate_for_instance() {{(pid=63515) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 701.189807] env[63515]: DEBUG oslo_concurrency.lockutils [None req-5690e97d-f63c-4f8e-9c79-d9c523fc3720 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 701.196538] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1110931, 'name': CreateVM_Task, 'duration_secs': 0.396947} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 701.196749] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 90f4930b-aaa0-4c4b-9ab8-92aed45e200b] Created VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 701.197354] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-c26675e2-4152-417b-99dd-0c18de4dc137 tempest-ServersTestBootFromVolume-2042560423 tempest-ServersTestBootFromVolume-2042560423-project-member] [instance: 90f4930b-aaa0-4c4b-9ab8-92aed45e200b] Block device information present: {'root_device_name': '/dev/sda', 'image': [], 'ephemerals': [], 'block_device_mapping': [{'connection_info': {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-243378', 'volume_id': '921350b3-bd4a-4fa1-848a-057b53f51f5c', 'name': 'volume-921350b3-bd4a-4fa1-848a-057b53f51f5c', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '90f4930b-aaa0-4c4b-9ab8-92aed45e200b', 'attached_at': '', 'detached_at': '', 'volume_id': '921350b3-bd4a-4fa1-848a-057b53f51f5c', 'serial': '921350b3-bd4a-4fa1-848a-057b53f51f5c'}, 'disk_bus': None, 'attachment_id': '1387c5da-c79f-4d4d-b1a3-40667ffbd711', 'guest_format': None, 'mount_device': '/dev/sda', 'delete_on_termination': True, 'device_type': None, 'boot_index': 0, 'volume_type': None}], 'swap': None} {{(pid=63515) spawn /opt/stack/nova/nova/virt/vmwareapi/vmops.py:799}} [ 701.197645] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-c26675e2-4152-417b-99dd-0c18de4dc137 tempest-ServersTestBootFromVolume-2042560423 tempest-ServersTestBootFromVolume-2042560423-project-member] [instance: 90f4930b-aaa0-4c4b-9ab8-92aed45e200b] Root volume attach. Driver type: vmdk {{(pid=63515) attach_root_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:661}} [ 701.198359] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a810947-fb03-4716-94fd-567908ef7df2 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.208010] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f567ebcc-20fd-41de-a348-8b6bd7ce6254 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.213684] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27d407c4-9fd2-4c9e-aff5-9e601fe5b6f8 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.221148] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.RelocateVM_Task with opID=oslo.vmware-a5c6aeef-bfdd-4ec1-98d6-d7426e05df4c {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.227543] env[63515]: DEBUG oslo_vmware.api [None req-c26675e2-4152-417b-99dd-0c18de4dc137 tempest-ServersTestBootFromVolume-2042560423 tempest-ServersTestBootFromVolume-2042560423-project-member] Waiting for the task: (returnval){ [ 701.227543] env[63515]: value = "task-1110933" [ 701.227543] env[63515]: _type = "Task" [ 701.227543] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 701.236013] env[63515]: DEBUG oslo_vmware.api [None req-c26675e2-4152-417b-99dd-0c18de4dc137 tempest-ServersTestBootFromVolume-2042560423 tempest-ServersTestBootFromVolume-2042560423-project-member] Task: {'id': task-1110933, 'name': RelocateVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 701.255374] env[63515]: DEBUG nova.policy [None req-4fa686a6-dc45-4603-a4ee-7107c810126f tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '21c4548852c747528a0b0530f6f42be9', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e0c07bb70a414f1787073d335c9001bf', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63515) authorize /opt/stack/nova/nova/policy.py:201}} [ 701.424328] env[63515]: DEBUG oslo_concurrency.lockutils [req-b0dbd36d-c9a7-48a0-893a-09440004509b req-c532b567-9934-40f2-a098-2090daa2caf0 service nova] Releasing lock "refresh_cache-90f4930b-aaa0-4c4b-9ab8-92aed45e200b" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 701.476451] env[63515]: DEBUG nova.network.neutron [None req-33f5b097-4c20-4377-b5be-7e558874e5d3 tempest-ServersV294TestFqdnHostnames-2112993702 tempest-ServersV294TestFqdnHostnames-2112993702-project-member] [instance: 396e49dd-48c0-496b-a1ec-190c33a22c5e] Successfully updated port: 77fbf94a-db5a-4569-b4e7-2d9edbe411a1 {{(pid=63515) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 701.543115] env[63515]: DEBUG oslo_concurrency.lockutils [None req-74f7de08-cdcf-4b98-bfdb-a2dd77c03e6d tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 701.565584] env[63515]: DEBUG oslo_vmware.api [None req-89884269-25d4-4981-bd3b-a346ece7b681 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Task: {'id': task-1110932, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 701.675598] env[63515]: DEBUG nova.network.neutron [None req-4fa686a6-dc45-4603-a4ee-7107c810126f tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] [instance: 4087b2c4-6ed1-4b68-8b78-a36e34d935b1] Successfully created port: 81eafe68-fff5-401d-812f-73b07de3ad8b {{(pid=63515) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 701.683432] env[63515]: DEBUG nova.compute.manager [None req-4fa686a6-dc45-4603-a4ee-7107c810126f tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] [instance: 4087b2c4-6ed1-4b68-8b78-a36e34d935b1] Start building block device mappings for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 701.744784] env[63515]: DEBUG oslo_vmware.api [None req-c26675e2-4152-417b-99dd-0c18de4dc137 tempest-ServersTestBootFromVolume-2042560423 tempest-ServersTestBootFromVolume-2042560423-project-member] Task: {'id': task-1110933, 'name': RelocateVM_Task} progress is 43%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 701.866076] env[63515]: DEBUG oslo_concurrency.lockutils [None req-4b5baf79-d4c3-467e-88d4-a131ea6620d7 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] Acquiring lock "e666825c-ff4e-4a0e-93c0-43c00f167bbb" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 701.866076] env[63515]: DEBUG oslo_concurrency.lockutils [None req-4b5baf79-d4c3-467e-88d4-a131ea6620d7 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] Lock "e666825c-ff4e-4a0e-93c0-43c00f167bbb" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 701.979593] env[63515]: DEBUG oslo_concurrency.lockutils [None req-33f5b097-4c20-4377-b5be-7e558874e5d3 tempest-ServersV294TestFqdnHostnames-2112993702 tempest-ServersV294TestFqdnHostnames-2112993702-project-member] Acquiring lock "refresh_cache-396e49dd-48c0-496b-a1ec-190c33a22c5e" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 701.979755] env[63515]: DEBUG oslo_concurrency.lockutils [None req-33f5b097-4c20-4377-b5be-7e558874e5d3 tempest-ServersV294TestFqdnHostnames-2112993702 tempest-ServersV294TestFqdnHostnames-2112993702-project-member] Acquired lock "refresh_cache-396e49dd-48c0-496b-a1ec-190c33a22c5e" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 701.979916] env[63515]: DEBUG nova.network.neutron [None req-33f5b097-4c20-4377-b5be-7e558874e5d3 tempest-ServersV294TestFqdnHostnames-2112993702 tempest-ServersV294TestFqdnHostnames-2112993702-project-member] [instance: 396e49dd-48c0-496b-a1ec-190c33a22c5e] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 702.070824] env[63515]: DEBUG oslo_vmware.api [None req-89884269-25d4-4981-bd3b-a346ece7b681 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Task: {'id': task-1110932, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.839893} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 702.071225] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-89884269-25d4-4981-bd3b-a346ece7b681 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Copied Virtual Disk [datastore2] vmware_temp/7386036a-4262-4a30-b58b-706f0d52abbb/8a120570-cb06-4099-b262-554ca0ad15c5/tmp-sparse.vmdk to [datastore2] vmware_temp/7386036a-4262-4a30-b58b-706f0d52abbb/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 702.071782] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-89884269-25d4-4981-bd3b-a346ece7b681 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Deleting the datastore file [datastore2] vmware_temp/7386036a-4262-4a30-b58b-706f0d52abbb/8a120570-cb06-4099-b262-554ca0ad15c5/tmp-sparse.vmdk {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 702.072112] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ef26051d-d888-465b-861a-40e7f4ab91c6 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.085431] env[63515]: DEBUG oslo_vmware.api [None req-89884269-25d4-4981-bd3b-a346ece7b681 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Waiting for the task: (returnval){ [ 702.085431] env[63515]: value = "task-1110934" [ 702.085431] env[63515]: _type = "Task" [ 702.085431] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 702.100510] env[63515]: DEBUG oslo_vmware.api [None req-89884269-25d4-4981-bd3b-a346ece7b681 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Task: {'id': task-1110934, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 702.197747] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee501062-bc55-4b9f-9062-b1d72269ebdc {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.208537] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8fba39f7-834a-48f4-8171-725fdaae931b {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.247861] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a970f91-5188-43d4-9a12-860dcc460e7b {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.259465] env[63515]: DEBUG oslo_vmware.api [None req-c26675e2-4152-417b-99dd-0c18de4dc137 tempest-ServersTestBootFromVolume-2042560423 tempest-ServersTestBootFromVolume-2042560423-project-member] Task: {'id': task-1110933, 'name': RelocateVM_Task} progress is 58%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 702.263175] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec792218-17ee-4015-b57b-fe616641dc94 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.279642] env[63515]: DEBUG nova.compute.provider_tree [None req-44a3080f-f2e2-4f0c-b244-122beee2aea9 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 702.518704] env[63515]: DEBUG nova.network.neutron [None req-33f5b097-4c20-4377-b5be-7e558874e5d3 tempest-ServersV294TestFqdnHostnames-2112993702 tempest-ServersV294TestFqdnHostnames-2112993702-project-member] [instance: 396e49dd-48c0-496b-a1ec-190c33a22c5e] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 702.598747] env[63515]: DEBUG oslo_vmware.api [None req-89884269-25d4-4981-bd3b-a346ece7b681 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Task: {'id': task-1110934, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.039239} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 702.599037] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-89884269-25d4-4981-bd3b-a346ece7b681 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Deleted the datastore file {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 702.599264] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-89884269-25d4-4981-bd3b-a346ece7b681 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Moving file from [datastore2] vmware_temp/7386036a-4262-4a30-b58b-706f0d52abbb/8a120570-cb06-4099-b262-554ca0ad15c5 to [datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5. {{(pid=63515) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:334}} [ 702.599532] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.MoveDatastoreFile_Task with opID=oslo.vmware-7fa97998-8322-4792-aeea-3930dd46ae82 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.610130] env[63515]: DEBUG oslo_vmware.api [None req-89884269-25d4-4981-bd3b-a346ece7b681 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Waiting for the task: (returnval){ [ 702.610130] env[63515]: value = "task-1110935" [ 702.610130] env[63515]: _type = "Task" [ 702.610130] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 702.620951] env[63515]: DEBUG oslo_vmware.api [None req-89884269-25d4-4981-bd3b-a346ece7b681 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Task: {'id': task-1110935, 'name': MoveDatastoreFile_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 702.702455] env[63515]: DEBUG nova.compute.manager [None req-4fa686a6-dc45-4603-a4ee-7107c810126f tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] [instance: 4087b2c4-6ed1-4b68-8b78-a36e34d935b1] Start spawning the instance on the hypervisor. {{(pid=63515) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 702.729830] env[63515]: DEBUG nova.network.neutron [None req-33f5b097-4c20-4377-b5be-7e558874e5d3 tempest-ServersV294TestFqdnHostnames-2112993702 tempest-ServersV294TestFqdnHostnames-2112993702-project-member] [instance: 396e49dd-48c0-496b-a1ec-190c33a22c5e] Updating instance_info_cache with network_info: [{"id": "77fbf94a-db5a-4569-b4e7-2d9edbe411a1", "address": "fa:16:3e:36:07:33", "network": {"id": "1c834c8e-4d0a-4cd9-b2ad-c24a22ed72ff", "bridge": "br-int", "label": "tempest-ServersV294TestFqdnHostnames-1956290090-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "62ac151dc34e41e7bf134563228d9d63", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4adc8ed0-d11a-4510-9be0-b27c0da3a903", "external-id": "nsx-vlan-transportzone-844", "segmentation_id": 844, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap77fbf94a-db", "ovs_interfaceid": "77fbf94a-db5a-4569-b4e7-2d9edbe411a1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 702.733381] env[63515]: DEBUG nova.virt.hardware [None req-4fa686a6-dc45-4603-a4ee-7107c810126f tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T02:52:14Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T02:51:56Z,direct_url=,disk_format='vmdk',id=8a120570-cb06-4099-b262-554ca0ad15c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b569255cc43e42e7ae3f2b4ad37c6ef4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T02:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 702.733637] env[63515]: DEBUG nova.virt.hardware [None req-4fa686a6-dc45-4603-a4ee-7107c810126f tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] Flavor limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 702.733797] env[63515]: DEBUG nova.virt.hardware [None req-4fa686a6-dc45-4603-a4ee-7107c810126f tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] Image limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 702.733980] env[63515]: DEBUG nova.virt.hardware [None req-4fa686a6-dc45-4603-a4ee-7107c810126f tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] Flavor pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 702.734142] env[63515]: DEBUG nova.virt.hardware [None req-4fa686a6-dc45-4603-a4ee-7107c810126f tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] Image pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 702.734287] env[63515]: DEBUG nova.virt.hardware [None req-4fa686a6-dc45-4603-a4ee-7107c810126f tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 702.734516] env[63515]: DEBUG nova.virt.hardware [None req-4fa686a6-dc45-4603-a4ee-7107c810126f tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 702.734702] env[63515]: DEBUG nova.virt.hardware [None req-4fa686a6-dc45-4603-a4ee-7107c810126f tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 702.734880] env[63515]: DEBUG nova.virt.hardware [None req-4fa686a6-dc45-4603-a4ee-7107c810126f tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] Got 1 possible topologies {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 702.735095] env[63515]: DEBUG nova.virt.hardware [None req-4fa686a6-dc45-4603-a4ee-7107c810126f tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 702.735272] env[63515]: DEBUG nova.virt.hardware [None req-4fa686a6-dc45-4603-a4ee-7107c810126f tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 702.736179] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d92aa7a-4a62-479f-bbc3-80646eea40d8 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.752064] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a0db95e-cee3-43cb-8fce-3ce9d9df185e {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.762885] env[63515]: DEBUG oslo_vmware.api [None req-c26675e2-4152-417b-99dd-0c18de4dc137 tempest-ServersTestBootFromVolume-2042560423 tempest-ServersTestBootFromVolume-2042560423-project-member] Task: {'id': task-1110933, 'name': RelocateVM_Task} progress is 71%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 702.783575] env[63515]: DEBUG nova.scheduler.client.report [None req-44a3080f-f2e2-4f0c-b244-122beee2aea9 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 702.854168] env[63515]: DEBUG nova.compute.manager [req-d262db17-cb54-47be-92bb-faab6d869c80 req-35771025-afd2-4c82-9188-07d2396dd185 service nova] [instance: 396e49dd-48c0-496b-a1ec-190c33a22c5e] Received event network-vif-plugged-77fbf94a-db5a-4569-b4e7-2d9edbe411a1 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 702.854314] env[63515]: DEBUG oslo_concurrency.lockutils [req-d262db17-cb54-47be-92bb-faab6d869c80 req-35771025-afd2-4c82-9188-07d2396dd185 service nova] Acquiring lock "396e49dd-48c0-496b-a1ec-190c33a22c5e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 702.854592] env[63515]: DEBUG oslo_concurrency.lockutils [req-d262db17-cb54-47be-92bb-faab6d869c80 req-35771025-afd2-4c82-9188-07d2396dd185 service nova] Lock "396e49dd-48c0-496b-a1ec-190c33a22c5e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 702.854592] env[63515]: DEBUG oslo_concurrency.lockutils [req-d262db17-cb54-47be-92bb-faab6d869c80 req-35771025-afd2-4c82-9188-07d2396dd185 service nova] Lock "396e49dd-48c0-496b-a1ec-190c33a22c5e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 702.854773] env[63515]: DEBUG nova.compute.manager [req-d262db17-cb54-47be-92bb-faab6d869c80 req-35771025-afd2-4c82-9188-07d2396dd185 service nova] [instance: 396e49dd-48c0-496b-a1ec-190c33a22c5e] No waiting events found dispatching network-vif-plugged-77fbf94a-db5a-4569-b4e7-2d9edbe411a1 {{(pid=63515) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 702.854962] env[63515]: WARNING nova.compute.manager [req-d262db17-cb54-47be-92bb-faab6d869c80 req-35771025-afd2-4c82-9188-07d2396dd185 service nova] [instance: 396e49dd-48c0-496b-a1ec-190c33a22c5e] Received unexpected event network-vif-plugged-77fbf94a-db5a-4569-b4e7-2d9edbe411a1 for instance with vm_state building and task_state spawning. [ 702.855148] env[63515]: DEBUG nova.compute.manager [req-d262db17-cb54-47be-92bb-faab6d869c80 req-35771025-afd2-4c82-9188-07d2396dd185 service nova] [instance: 396e49dd-48c0-496b-a1ec-190c33a22c5e] Received event network-changed-77fbf94a-db5a-4569-b4e7-2d9edbe411a1 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 702.855305] env[63515]: DEBUG nova.compute.manager [req-d262db17-cb54-47be-92bb-faab6d869c80 req-35771025-afd2-4c82-9188-07d2396dd185 service nova] [instance: 396e49dd-48c0-496b-a1ec-190c33a22c5e] Refreshing instance network info cache due to event network-changed-77fbf94a-db5a-4569-b4e7-2d9edbe411a1. {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 702.855473] env[63515]: DEBUG oslo_concurrency.lockutils [req-d262db17-cb54-47be-92bb-faab6d869c80 req-35771025-afd2-4c82-9188-07d2396dd185 service nova] Acquiring lock "refresh_cache-396e49dd-48c0-496b-a1ec-190c33a22c5e" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 703.132146] env[63515]: DEBUG oslo_vmware.api [None req-89884269-25d4-4981-bd3b-a346ece7b681 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Task: {'id': task-1110935, 'name': MoveDatastoreFile_Task, 'duration_secs': 0.040806} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 703.132146] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-89884269-25d4-4981-bd3b-a346ece7b681 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] File moved {{(pid=63515) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:346}} [ 703.132146] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-89884269-25d4-4981-bd3b-a346ece7b681 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] [instance: f1d01b75-ac9d-458d-8cc2-ae64cffca4e8] Cleaning up location [datastore2] vmware_temp/7386036a-4262-4a30-b58b-706f0d52abbb {{(pid=63515) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 703.132146] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-89884269-25d4-4981-bd3b-a346ece7b681 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Deleting the datastore file [datastore2] vmware_temp/7386036a-4262-4a30-b58b-706f0d52abbb {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 703.132146] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4b6d4b24-73ef-4a0b-9e28-579b023e148d {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.141445] env[63515]: DEBUG oslo_vmware.api [None req-89884269-25d4-4981-bd3b-a346ece7b681 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Waiting for the task: (returnval){ [ 703.141445] env[63515]: value = "task-1110936" [ 703.141445] env[63515]: _type = "Task" [ 703.141445] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 703.154094] env[63515]: DEBUG oslo_vmware.api [None req-89884269-25d4-4981-bd3b-a346ece7b681 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Task: {'id': task-1110936, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 703.241164] env[63515]: DEBUG oslo_concurrency.lockutils [None req-33f5b097-4c20-4377-b5be-7e558874e5d3 tempest-ServersV294TestFqdnHostnames-2112993702 tempest-ServersV294TestFqdnHostnames-2112993702-project-member] Releasing lock "refresh_cache-396e49dd-48c0-496b-a1ec-190c33a22c5e" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 703.241341] env[63515]: DEBUG nova.compute.manager [None req-33f5b097-4c20-4377-b5be-7e558874e5d3 tempest-ServersV294TestFqdnHostnames-2112993702 tempest-ServersV294TestFqdnHostnames-2112993702-project-member] [instance: 396e49dd-48c0-496b-a1ec-190c33a22c5e] Instance network_info: |[{"id": "77fbf94a-db5a-4569-b4e7-2d9edbe411a1", "address": "fa:16:3e:36:07:33", "network": {"id": "1c834c8e-4d0a-4cd9-b2ad-c24a22ed72ff", "bridge": "br-int", "label": "tempest-ServersV294TestFqdnHostnames-1956290090-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "62ac151dc34e41e7bf134563228d9d63", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4adc8ed0-d11a-4510-9be0-b27c0da3a903", "external-id": "nsx-vlan-transportzone-844", "segmentation_id": 844, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap77fbf94a-db", "ovs_interfaceid": "77fbf94a-db5a-4569-b4e7-2d9edbe411a1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 703.241673] env[63515]: DEBUG oslo_concurrency.lockutils [req-d262db17-cb54-47be-92bb-faab6d869c80 req-35771025-afd2-4c82-9188-07d2396dd185 service nova] Acquired lock "refresh_cache-396e49dd-48c0-496b-a1ec-190c33a22c5e" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 703.241907] env[63515]: DEBUG nova.network.neutron [req-d262db17-cb54-47be-92bb-faab6d869c80 req-35771025-afd2-4c82-9188-07d2396dd185 service nova] [instance: 396e49dd-48c0-496b-a1ec-190c33a22c5e] Refreshing network info cache for port 77fbf94a-db5a-4569-b4e7-2d9edbe411a1 {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 703.243650] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-33f5b097-4c20-4377-b5be-7e558874e5d3 tempest-ServersV294TestFqdnHostnames-2112993702 tempest-ServersV294TestFqdnHostnames-2112993702-project-member] [instance: 396e49dd-48c0-496b-a1ec-190c33a22c5e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:36:07:33', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '4adc8ed0-d11a-4510-9be0-b27c0da3a903', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '77fbf94a-db5a-4569-b4e7-2d9edbe411a1', 'vif_model': 'vmxnet3'}] {{(pid=63515) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 703.251610] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-33f5b097-4c20-4377-b5be-7e558874e5d3 tempest-ServersV294TestFqdnHostnames-2112993702 tempest-ServersV294TestFqdnHostnames-2112993702-project-member] Creating folder: Project (62ac151dc34e41e7bf134563228d9d63). Parent ref: group-v243370. {{(pid=63515) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 703.252938] env[63515]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-75eb608d-a08c-4810-bbd7-f67959037453 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.267661] env[63515]: DEBUG oslo_vmware.api [None req-c26675e2-4152-417b-99dd-0c18de4dc137 tempest-ServersTestBootFromVolume-2042560423 tempest-ServersTestBootFromVolume-2042560423-project-member] Task: {'id': task-1110933, 'name': RelocateVM_Task} progress is 86%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 703.270864] env[63515]: INFO nova.virt.vmwareapi.vm_util [None req-33f5b097-4c20-4377-b5be-7e558874e5d3 tempest-ServersV294TestFqdnHostnames-2112993702 tempest-ServersV294TestFqdnHostnames-2112993702-project-member] Created folder: Project (62ac151dc34e41e7bf134563228d9d63) in parent group-v243370. [ 703.271166] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-33f5b097-4c20-4377-b5be-7e558874e5d3 tempest-ServersV294TestFqdnHostnames-2112993702 tempest-ServersV294TestFqdnHostnames-2112993702-project-member] Creating folder: Instances. Parent ref: group-v243397. {{(pid=63515) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 703.271447] env[63515]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a6802106-96ea-4908-8f2f-21298a7f88b5 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.286378] env[63515]: INFO nova.virt.vmwareapi.vm_util [None req-33f5b097-4c20-4377-b5be-7e558874e5d3 tempest-ServersV294TestFqdnHostnames-2112993702 tempest-ServersV294TestFqdnHostnames-2112993702-project-member] Created folder: Instances in parent group-v243397. [ 703.287373] env[63515]: DEBUG oslo.service.loopingcall [None req-33f5b097-4c20-4377-b5be-7e558874e5d3 tempest-ServersV294TestFqdnHostnames-2112993702 tempest-ServersV294TestFqdnHostnames-2112993702-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 703.287373] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 396e49dd-48c0-496b-a1ec-190c33a22c5e] Creating VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 703.287373] env[63515]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-77d9d1f0-2e06-4974-9f74-fb33d62e6a5c {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.311023] env[63515]: DEBUG oslo_concurrency.lockutils [None req-44a3080f-f2e2-4f0c-b244-122beee2aea9 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.639s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 703.311023] env[63515]: DEBUG nova.compute.manager [None req-44a3080f-f2e2-4f0c-b244-122beee2aea9 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] [instance: 87c468d9-9594-4804-b461-527f01f6118f] Start building networks asynchronously for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 703.317722] env[63515]: DEBUG oslo_concurrency.lockutils [None req-d937370c-7531-4295-8d43-07fafc09e90c tempest-ServerDiagnosticsTest-1160321831 tempest-ServerDiagnosticsTest-1160321831-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 16.198s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 703.318459] env[63515]: DEBUG nova.objects.instance [None req-d937370c-7531-4295-8d43-07fafc09e90c tempest-ServerDiagnosticsTest-1160321831 tempest-ServerDiagnosticsTest-1160321831-project-member] Lazy-loading 'resources' on Instance uuid 4ad5440f-cd18-44c5-8836-0aa39824cf03 {{(pid=63515) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 703.328945] env[63515]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 703.328945] env[63515]: value = "task-1110939" [ 703.328945] env[63515]: _type = "Task" [ 703.328945] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 703.342339] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1110939, 'name': CreateVM_Task} progress is 5%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 703.411894] env[63515]: DEBUG nova.compute.manager [req-700effd8-eead-4c81-82b4-2aeb0312b4ba req-79043345-f097-49ec-a207-5223df412235 service nova] [instance: 4087b2c4-6ed1-4b68-8b78-a36e34d935b1] Received event network-vif-plugged-81eafe68-fff5-401d-812f-73b07de3ad8b {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 703.412157] env[63515]: DEBUG oslo_concurrency.lockutils [req-700effd8-eead-4c81-82b4-2aeb0312b4ba req-79043345-f097-49ec-a207-5223df412235 service nova] Acquiring lock "4087b2c4-6ed1-4b68-8b78-a36e34d935b1-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 703.412396] env[63515]: DEBUG oslo_concurrency.lockutils [req-700effd8-eead-4c81-82b4-2aeb0312b4ba req-79043345-f097-49ec-a207-5223df412235 service nova] Lock "4087b2c4-6ed1-4b68-8b78-a36e34d935b1-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 703.412607] env[63515]: DEBUG oslo_concurrency.lockutils [req-700effd8-eead-4c81-82b4-2aeb0312b4ba req-79043345-f097-49ec-a207-5223df412235 service nova] Lock "4087b2c4-6ed1-4b68-8b78-a36e34d935b1-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 703.412933] env[63515]: DEBUG nova.compute.manager [req-700effd8-eead-4c81-82b4-2aeb0312b4ba req-79043345-f097-49ec-a207-5223df412235 service nova] [instance: 4087b2c4-6ed1-4b68-8b78-a36e34d935b1] No waiting events found dispatching network-vif-plugged-81eafe68-fff5-401d-812f-73b07de3ad8b {{(pid=63515) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 703.412933] env[63515]: WARNING nova.compute.manager [req-700effd8-eead-4c81-82b4-2aeb0312b4ba req-79043345-f097-49ec-a207-5223df412235 service nova] [instance: 4087b2c4-6ed1-4b68-8b78-a36e34d935b1] Received unexpected event network-vif-plugged-81eafe68-fff5-401d-812f-73b07de3ad8b for instance with vm_state building and task_state spawning. [ 703.651199] env[63515]: DEBUG oslo_vmware.api [None req-89884269-25d4-4981-bd3b-a346ece7b681 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Task: {'id': task-1110936, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.050984} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 703.651924] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-89884269-25d4-4981-bd3b-a346ece7b681 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Deleted the datastore file {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 703.652379] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a83a8cd8-c892-43ba-88ef-754fc31ab23a {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.658297] env[63515]: DEBUG oslo_vmware.api [None req-89884269-25d4-4981-bd3b-a346ece7b681 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Waiting for the task: (returnval){ [ 703.658297] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]524002f5-2410-90f5-eaaa-b881e61d174e" [ 703.658297] env[63515]: _type = "Task" [ 703.658297] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 703.666053] env[63515]: DEBUG oslo_vmware.api [None req-89884269-25d4-4981-bd3b-a346ece7b681 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]524002f5-2410-90f5-eaaa-b881e61d174e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 703.763913] env[63515]: DEBUG oslo_vmware.api [None req-c26675e2-4152-417b-99dd-0c18de4dc137 tempest-ServersTestBootFromVolume-2042560423 tempest-ServersTestBootFromVolume-2042560423-project-member] Task: {'id': task-1110933, 'name': RelocateVM_Task} progress is 97%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 703.818744] env[63515]: DEBUG nova.compute.utils [None req-44a3080f-f2e2-4f0c-b244-122beee2aea9 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Using /dev/sd instead of None {{(pid=63515) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 703.821179] env[63515]: DEBUG nova.compute.manager [None req-44a3080f-f2e2-4f0c-b244-122beee2aea9 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] [instance: 87c468d9-9594-4804-b461-527f01f6118f] Allocating IP information in the background. {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 703.821411] env[63515]: DEBUG nova.network.neutron [None req-44a3080f-f2e2-4f0c-b244-122beee2aea9 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] [instance: 87c468d9-9594-4804-b461-527f01f6118f] allocate_for_instance() {{(pid=63515) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 703.839602] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1110939, 'name': CreateVM_Task, 'duration_secs': 0.45005} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 703.839745] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 396e49dd-48c0-496b-a1ec-190c33a22c5e] Created VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 703.840405] env[63515]: DEBUG oslo_concurrency.lockutils [None req-33f5b097-4c20-4377-b5be-7e558874e5d3 tempest-ServersV294TestFqdnHostnames-2112993702 tempest-ServersV294TestFqdnHostnames-2112993702-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 703.840571] env[63515]: DEBUG oslo_concurrency.lockutils [None req-33f5b097-4c20-4377-b5be-7e558874e5d3 tempest-ServersV294TestFqdnHostnames-2112993702 tempest-ServersV294TestFqdnHostnames-2112993702-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 703.840888] env[63515]: DEBUG oslo_concurrency.lockutils [None req-33f5b097-4c20-4377-b5be-7e558874e5d3 tempest-ServersV294TestFqdnHostnames-2112993702 tempest-ServersV294TestFqdnHostnames-2112993702-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 703.841154] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-78bbd9e3-3acb-4ba4-b691-5805b5286ab1 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.845836] env[63515]: DEBUG oslo_vmware.api [None req-33f5b097-4c20-4377-b5be-7e558874e5d3 tempest-ServersV294TestFqdnHostnames-2112993702 tempest-ServersV294TestFqdnHostnames-2112993702-project-member] Waiting for the task: (returnval){ [ 703.845836] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]527345e5-0fee-b226-debe-d97e349c90bd" [ 703.845836] env[63515]: _type = "Task" [ 703.845836] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 703.856588] env[63515]: DEBUG oslo_vmware.api [None req-33f5b097-4c20-4377-b5be-7e558874e5d3 tempest-ServersV294TestFqdnHostnames-2112993702 tempest-ServersV294TestFqdnHostnames-2112993702-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]527345e5-0fee-b226-debe-d97e349c90bd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 703.919634] env[63515]: DEBUG nova.policy [None req-44a3080f-f2e2-4f0c-b244-122beee2aea9 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2bd575c5f972489dabe5c8057db84790', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '19ce3b05eec64475b685faad1fd801f6', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63515) authorize /opt/stack/nova/nova/policy.py:201}} [ 703.921022] env[63515]: DEBUG nova.network.neutron [None req-4fa686a6-dc45-4603-a4ee-7107c810126f tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] [instance: 4087b2c4-6ed1-4b68-8b78-a36e34d935b1] Successfully updated port: 81eafe68-fff5-401d-812f-73b07de3ad8b {{(pid=63515) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 703.944962] env[63515]: DEBUG nova.compute.manager [req-a4d86691-8ed3-49fc-bfde-e840ebf5c24a req-1cbf82b4-41e1-448a-9a7d-48a251761b6c service nova] [instance: 4087b2c4-6ed1-4b68-8b78-a36e34d935b1] Received event network-changed-81eafe68-fff5-401d-812f-73b07de3ad8b {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 703.944962] env[63515]: DEBUG nova.compute.manager [req-a4d86691-8ed3-49fc-bfde-e840ebf5c24a req-1cbf82b4-41e1-448a-9a7d-48a251761b6c service nova] [instance: 4087b2c4-6ed1-4b68-8b78-a36e34d935b1] Refreshing instance network info cache due to event network-changed-81eafe68-fff5-401d-812f-73b07de3ad8b. {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 703.944962] env[63515]: DEBUG oslo_concurrency.lockutils [req-a4d86691-8ed3-49fc-bfde-e840ebf5c24a req-1cbf82b4-41e1-448a-9a7d-48a251761b6c service nova] Acquiring lock "refresh_cache-4087b2c4-6ed1-4b68-8b78-a36e34d935b1" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 703.945288] env[63515]: DEBUG oslo_concurrency.lockutils [req-a4d86691-8ed3-49fc-bfde-e840ebf5c24a req-1cbf82b4-41e1-448a-9a7d-48a251761b6c service nova] Acquired lock "refresh_cache-4087b2c4-6ed1-4b68-8b78-a36e34d935b1" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 703.945288] env[63515]: DEBUG nova.network.neutron [req-a4d86691-8ed3-49fc-bfde-e840ebf5c24a req-1cbf82b4-41e1-448a-9a7d-48a251761b6c service nova] [instance: 4087b2c4-6ed1-4b68-8b78-a36e34d935b1] Refreshing network info cache for port 81eafe68-fff5-401d-812f-73b07de3ad8b {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 704.169979] env[63515]: DEBUG oslo_vmware.api [None req-89884269-25d4-4981-bd3b-a346ece7b681 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]524002f5-2410-90f5-eaaa-b881e61d174e, 'name': SearchDatastore_Task, 'duration_secs': 0.034995} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 704.170313] env[63515]: DEBUG oslo_concurrency.lockutils [None req-89884269-25d4-4981-bd3b-a346ece7b681 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 704.170555] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-89884269-25d4-4981-bd3b-a346ece7b681 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk to [datastore2] f1d01b75-ac9d-458d-8cc2-ae64cffca4e8/f1d01b75-ac9d-458d-8cc2-ae64cffca4e8.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 704.171370] env[63515]: DEBUG nova.network.neutron [req-d262db17-cb54-47be-92bb-faab6d869c80 req-35771025-afd2-4c82-9188-07d2396dd185 service nova] [instance: 396e49dd-48c0-496b-a1ec-190c33a22c5e] Updated VIF entry in instance network info cache for port 77fbf94a-db5a-4569-b4e7-2d9edbe411a1. {{(pid=63515) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 704.171694] env[63515]: DEBUG nova.network.neutron [req-d262db17-cb54-47be-92bb-faab6d869c80 req-35771025-afd2-4c82-9188-07d2396dd185 service nova] [instance: 396e49dd-48c0-496b-a1ec-190c33a22c5e] Updating instance_info_cache with network_info: [{"id": "77fbf94a-db5a-4569-b4e7-2d9edbe411a1", "address": "fa:16:3e:36:07:33", "network": {"id": "1c834c8e-4d0a-4cd9-b2ad-c24a22ed72ff", "bridge": "br-int", "label": "tempest-ServersV294TestFqdnHostnames-1956290090-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "62ac151dc34e41e7bf134563228d9d63", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4adc8ed0-d11a-4510-9be0-b27c0da3a903", "external-id": "nsx-vlan-transportzone-844", "segmentation_id": 844, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap77fbf94a-db", "ovs_interfaceid": "77fbf94a-db5a-4569-b4e7-2d9edbe411a1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 704.173128] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-535e8464-9e23-4163-bf15-87c0d85cf656 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.179874] env[63515]: DEBUG oslo_vmware.api [None req-89884269-25d4-4981-bd3b-a346ece7b681 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Waiting for the task: (returnval){ [ 704.179874] env[63515]: value = "task-1110940" [ 704.179874] env[63515]: _type = "Task" [ 704.179874] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 704.191137] env[63515]: DEBUG oslo_vmware.api [None req-89884269-25d4-4981-bd3b-a346ece7b681 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Task: {'id': task-1110940, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 704.269177] env[63515]: DEBUG oslo_vmware.api [None req-c26675e2-4152-417b-99dd-0c18de4dc137 tempest-ServersTestBootFromVolume-2042560423 tempest-ServersTestBootFromVolume-2042560423-project-member] Task: {'id': task-1110933, 'name': RelocateVM_Task} progress is 98%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 704.322059] env[63515]: DEBUG nova.compute.manager [None req-44a3080f-f2e2-4f0c-b244-122beee2aea9 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] [instance: 87c468d9-9594-4804-b461-527f01f6118f] Start building block device mappings for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 704.328749] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-add1e2f2-eeb5-40ae-ac3b-136e4998dc59 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.336391] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94dc6828-6d0e-4612-9f10-0fa19c56a99b {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.373112] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38d82ad5-77e4-4856-b38a-89d61f309592 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.380254] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a63b5774-0156-4671-81f2-eb9f72a25041 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.385162] env[63515]: DEBUG oslo_vmware.api [None req-33f5b097-4c20-4377-b5be-7e558874e5d3 tempest-ServersV294TestFqdnHostnames-2112993702 tempest-ServersV294TestFqdnHostnames-2112993702-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]527345e5-0fee-b226-debe-d97e349c90bd, 'name': SearchDatastore_Task, 'duration_secs': 0.04197} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 704.385919] env[63515]: DEBUG oslo_concurrency.lockutils [None req-33f5b097-4c20-4377-b5be-7e558874e5d3 tempest-ServersV294TestFqdnHostnames-2112993702 tempest-ServersV294TestFqdnHostnames-2112993702-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 704.386117] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-33f5b097-4c20-4377-b5be-7e558874e5d3 tempest-ServersV294TestFqdnHostnames-2112993702 tempest-ServersV294TestFqdnHostnames-2112993702-project-member] [instance: 396e49dd-48c0-496b-a1ec-190c33a22c5e] Processing image 8a120570-cb06-4099-b262-554ca0ad15c5 {{(pid=63515) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 704.386352] env[63515]: DEBUG oslo_concurrency.lockutils [None req-33f5b097-4c20-4377-b5be-7e558874e5d3 tempest-ServersV294TestFqdnHostnames-2112993702 tempest-ServersV294TestFqdnHostnames-2112993702-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 704.386504] env[63515]: DEBUG oslo_concurrency.lockutils [None req-33f5b097-4c20-4377-b5be-7e558874e5d3 tempest-ServersV294TestFqdnHostnames-2112993702 tempest-ServersV294TestFqdnHostnames-2112993702-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 704.386684] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-33f5b097-4c20-4377-b5be-7e558874e5d3 tempest-ServersV294TestFqdnHostnames-2112993702 tempest-ServersV294TestFqdnHostnames-2112993702-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 704.387312] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-7dcb24aa-1f01-4f20-aa80-47a1daca3f7b {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.397909] env[63515]: DEBUG nova.compute.provider_tree [None req-d937370c-7531-4295-8d43-07fafc09e90c tempest-ServerDiagnosticsTest-1160321831 tempest-ServerDiagnosticsTest-1160321831-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 704.404552] env[63515]: DEBUG nova.network.neutron [None req-44a3080f-f2e2-4f0c-b244-122beee2aea9 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] [instance: 87c468d9-9594-4804-b461-527f01f6118f] Successfully created port: d2098868-0475-4b06-bae3-785c608a861e {{(pid=63515) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 704.408132] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-33f5b097-4c20-4377-b5be-7e558874e5d3 tempest-ServersV294TestFqdnHostnames-2112993702 tempest-ServersV294TestFqdnHostnames-2112993702-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 704.408307] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-33f5b097-4c20-4377-b5be-7e558874e5d3 tempest-ServersV294TestFqdnHostnames-2112993702 tempest-ServersV294TestFqdnHostnames-2112993702-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63515) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 704.409211] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fd8d3348-2f7e-4471-987d-c7ccb29df474 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.413990] env[63515]: DEBUG oslo_vmware.api [None req-33f5b097-4c20-4377-b5be-7e558874e5d3 tempest-ServersV294TestFqdnHostnames-2112993702 tempest-ServersV294TestFqdnHostnames-2112993702-project-member] Waiting for the task: (returnval){ [ 704.413990] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]52b2d0d4-79e1-37e7-b0d9-dcc3f55b2540" [ 704.413990] env[63515]: _type = "Task" [ 704.413990] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 704.421557] env[63515]: DEBUG oslo_vmware.api [None req-33f5b097-4c20-4377-b5be-7e558874e5d3 tempest-ServersV294TestFqdnHostnames-2112993702 tempest-ServersV294TestFqdnHostnames-2112993702-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52b2d0d4-79e1-37e7-b0d9-dcc3f55b2540, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 704.425169] env[63515]: DEBUG oslo_concurrency.lockutils [None req-4fa686a6-dc45-4603-a4ee-7107c810126f tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] Acquiring lock "refresh_cache-4087b2c4-6ed1-4b68-8b78-a36e34d935b1" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 704.488885] env[63515]: DEBUG nova.network.neutron [req-a4d86691-8ed3-49fc-bfde-e840ebf5c24a req-1cbf82b4-41e1-448a-9a7d-48a251761b6c service nova] [instance: 4087b2c4-6ed1-4b68-8b78-a36e34d935b1] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 704.643838] env[63515]: DEBUG nova.network.neutron [req-a4d86691-8ed3-49fc-bfde-e840ebf5c24a req-1cbf82b4-41e1-448a-9a7d-48a251761b6c service nova] [instance: 4087b2c4-6ed1-4b68-8b78-a36e34d935b1] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 704.676120] env[63515]: DEBUG oslo_concurrency.lockutils [req-d262db17-cb54-47be-92bb-faab6d869c80 req-35771025-afd2-4c82-9188-07d2396dd185 service nova] Releasing lock "refresh_cache-396e49dd-48c0-496b-a1ec-190c33a22c5e" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 704.690185] env[63515]: DEBUG oslo_vmware.api [None req-89884269-25d4-4981-bd3b-a346ece7b681 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Task: {'id': task-1110940, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 704.768534] env[63515]: DEBUG oslo_vmware.api [None req-c26675e2-4152-417b-99dd-0c18de4dc137 tempest-ServersTestBootFromVolume-2042560423 tempest-ServersTestBootFromVolume-2042560423-project-member] Task: {'id': task-1110933, 'name': RelocateVM_Task, 'duration_secs': 3.233628} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 704.769038] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-c26675e2-4152-417b-99dd-0c18de4dc137 tempest-ServersTestBootFromVolume-2042560423 tempest-ServersTestBootFromVolume-2042560423-project-member] [instance: 90f4930b-aaa0-4c4b-9ab8-92aed45e200b] Volume attach. Driver type: vmdk {{(pid=63515) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 704.769365] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-c26675e2-4152-417b-99dd-0c18de4dc137 tempest-ServersTestBootFromVolume-2042560423 tempest-ServersTestBootFromVolume-2042560423-project-member] [instance: 90f4930b-aaa0-4c4b-9ab8-92aed45e200b] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-243378', 'volume_id': '921350b3-bd4a-4fa1-848a-057b53f51f5c', 'name': 'volume-921350b3-bd4a-4fa1-848a-057b53f51f5c', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '90f4930b-aaa0-4c4b-9ab8-92aed45e200b', 'attached_at': '', 'detached_at': '', 'volume_id': '921350b3-bd4a-4fa1-848a-057b53f51f5c', 'serial': '921350b3-bd4a-4fa1-848a-057b53f51f5c'} {{(pid=63515) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 704.770280] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-086ac192-b966-4fcf-a6c6-a0ad3a25e67a {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.786725] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ab68e66-5834-4875-9932-3c1f0e0f58b9 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.808936] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-c26675e2-4152-417b-99dd-0c18de4dc137 tempest-ServersTestBootFromVolume-2042560423 tempest-ServersTestBootFromVolume-2042560423-project-member] [instance: 90f4930b-aaa0-4c4b-9ab8-92aed45e200b] Reconfiguring VM instance instance-00000023 to attach disk [datastore2] volume-921350b3-bd4a-4fa1-848a-057b53f51f5c/volume-921350b3-bd4a-4fa1-848a-057b53f51f5c.vmdk or device None with type thin {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 704.809359] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2e1b0ad5-179f-4f1e-909a-02cbcde9bfd6 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.834427] env[63515]: DEBUG oslo_vmware.api [None req-c26675e2-4152-417b-99dd-0c18de4dc137 tempest-ServersTestBootFromVolume-2042560423 tempest-ServersTestBootFromVolume-2042560423-project-member] Waiting for the task: (returnval){ [ 704.834427] env[63515]: value = "task-1110941" [ 704.834427] env[63515]: _type = "Task" [ 704.834427] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 704.842458] env[63515]: DEBUG oslo_vmware.api [None req-c26675e2-4152-417b-99dd-0c18de4dc137 tempest-ServersTestBootFromVolume-2042560423 tempest-ServersTestBootFromVolume-2042560423-project-member] Task: {'id': task-1110941, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 704.901292] env[63515]: DEBUG nova.scheduler.client.report [None req-d937370c-7531-4295-8d43-07fafc09e90c tempest-ServerDiagnosticsTest-1160321831 tempest-ServerDiagnosticsTest-1160321831-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 704.925935] env[63515]: DEBUG oslo_vmware.api [None req-33f5b097-4c20-4377-b5be-7e558874e5d3 tempest-ServersV294TestFqdnHostnames-2112993702 tempest-ServersV294TestFqdnHostnames-2112993702-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52b2d0d4-79e1-37e7-b0d9-dcc3f55b2540, 'name': SearchDatastore_Task, 'duration_secs': 0.027059} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 704.926711] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b31b17c8-2105-43d4-986c-3dd699e52aa1 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.931783] env[63515]: DEBUG oslo_vmware.api [None req-33f5b097-4c20-4377-b5be-7e558874e5d3 tempest-ServersV294TestFqdnHostnames-2112993702 tempest-ServersV294TestFqdnHostnames-2112993702-project-member] Waiting for the task: (returnval){ [ 704.931783] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]52562aee-d00b-a567-5326-33d35d14b7aa" [ 704.931783] env[63515]: _type = "Task" [ 704.931783] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 704.939925] env[63515]: DEBUG oslo_vmware.api [None req-33f5b097-4c20-4377-b5be-7e558874e5d3 tempest-ServersV294TestFqdnHostnames-2112993702 tempest-ServersV294TestFqdnHostnames-2112993702-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52562aee-d00b-a567-5326-33d35d14b7aa, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 705.148651] env[63515]: DEBUG oslo_concurrency.lockutils [req-a4d86691-8ed3-49fc-bfde-e840ebf5c24a req-1cbf82b4-41e1-448a-9a7d-48a251761b6c service nova] Releasing lock "refresh_cache-4087b2c4-6ed1-4b68-8b78-a36e34d935b1" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 705.149379] env[63515]: DEBUG oslo_concurrency.lockutils [None req-4fa686a6-dc45-4603-a4ee-7107c810126f tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] Acquired lock "refresh_cache-4087b2c4-6ed1-4b68-8b78-a36e34d935b1" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 705.149825] env[63515]: DEBUG nova.network.neutron [None req-4fa686a6-dc45-4603-a4ee-7107c810126f tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] [instance: 4087b2c4-6ed1-4b68-8b78-a36e34d935b1] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 705.191253] env[63515]: DEBUG oslo_vmware.api [None req-89884269-25d4-4981-bd3b-a346ece7b681 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Task: {'id': task-1110940, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 705.334490] env[63515]: DEBUG nova.compute.manager [None req-44a3080f-f2e2-4f0c-b244-122beee2aea9 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] [instance: 87c468d9-9594-4804-b461-527f01f6118f] Start spawning the instance on the hypervisor. {{(pid=63515) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 705.348496] env[63515]: DEBUG oslo_vmware.api [None req-c26675e2-4152-417b-99dd-0c18de4dc137 tempest-ServersTestBootFromVolume-2042560423 tempest-ServersTestBootFromVolume-2042560423-project-member] Task: {'id': task-1110941, 'name': ReconfigVM_Task, 'duration_secs': 0.295861} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 705.353104] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-c26675e2-4152-417b-99dd-0c18de4dc137 tempest-ServersTestBootFromVolume-2042560423 tempest-ServersTestBootFromVolume-2042560423-project-member] [instance: 90f4930b-aaa0-4c4b-9ab8-92aed45e200b] Reconfigured VM instance instance-00000023 to attach disk [datastore2] volume-921350b3-bd4a-4fa1-848a-057b53f51f5c/volume-921350b3-bd4a-4fa1-848a-057b53f51f5c.vmdk or device None with type thin {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 705.356373] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6d95d74b-6d03-41c0-9edf-bec575b4c94b {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.373431] env[63515]: DEBUG nova.virt.hardware [None req-44a3080f-f2e2-4f0c-b244-122beee2aea9 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T02:52:14Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T02:51:56Z,direct_url=,disk_format='vmdk',id=8a120570-cb06-4099-b262-554ca0ad15c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b569255cc43e42e7ae3f2b4ad37c6ef4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T02:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 705.373729] env[63515]: DEBUG nova.virt.hardware [None req-44a3080f-f2e2-4f0c-b244-122beee2aea9 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Flavor limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 705.373815] env[63515]: DEBUG nova.virt.hardware [None req-44a3080f-f2e2-4f0c-b244-122beee2aea9 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Image limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 705.373993] env[63515]: DEBUG nova.virt.hardware [None req-44a3080f-f2e2-4f0c-b244-122beee2aea9 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Flavor pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 705.374152] env[63515]: DEBUG nova.virt.hardware [None req-44a3080f-f2e2-4f0c-b244-122beee2aea9 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Image pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 705.374299] env[63515]: DEBUG nova.virt.hardware [None req-44a3080f-f2e2-4f0c-b244-122beee2aea9 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 705.374526] env[63515]: DEBUG nova.virt.hardware [None req-44a3080f-f2e2-4f0c-b244-122beee2aea9 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 705.374704] env[63515]: DEBUG nova.virt.hardware [None req-44a3080f-f2e2-4f0c-b244-122beee2aea9 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 705.374896] env[63515]: DEBUG nova.virt.hardware [None req-44a3080f-f2e2-4f0c-b244-122beee2aea9 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Got 1 possible topologies {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 705.375035] env[63515]: DEBUG nova.virt.hardware [None req-44a3080f-f2e2-4f0c-b244-122beee2aea9 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 705.375209] env[63515]: DEBUG nova.virt.hardware [None req-44a3080f-f2e2-4f0c-b244-122beee2aea9 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 705.376055] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06403031-eb32-407a-8717-f82103f92314 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.379682] env[63515]: DEBUG oslo_vmware.api [None req-c26675e2-4152-417b-99dd-0c18de4dc137 tempest-ServersTestBootFromVolume-2042560423 tempest-ServersTestBootFromVolume-2042560423-project-member] Waiting for the task: (returnval){ [ 705.379682] env[63515]: value = "task-1110942" [ 705.379682] env[63515]: _type = "Task" [ 705.379682] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 705.388047] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ccc0f295-7b48-4d27-aa7c-e6891f7b7399 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.395095] env[63515]: DEBUG oslo_vmware.api [None req-c26675e2-4152-417b-99dd-0c18de4dc137 tempest-ServersTestBootFromVolume-2042560423 tempest-ServersTestBootFromVolume-2042560423-project-member] Task: {'id': task-1110942, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 705.406422] env[63515]: DEBUG oslo_concurrency.lockutils [None req-d937370c-7531-4295-8d43-07fafc09e90c tempest-ServerDiagnosticsTest-1160321831 tempest-ServerDiagnosticsTest-1160321831-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.089s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 705.408690] env[63515]: DEBUG oslo_concurrency.lockutils [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.116s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 705.410231] env[63515]: INFO nova.compute.claims [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] [instance: 48668736-df27-4f2a-94d9-132f5b49701b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 705.437650] env[63515]: INFO nova.scheduler.client.report [None req-d937370c-7531-4295-8d43-07fafc09e90c tempest-ServerDiagnosticsTest-1160321831 tempest-ServerDiagnosticsTest-1160321831-project-member] Deleted allocations for instance 4ad5440f-cd18-44c5-8836-0aa39824cf03 [ 705.445413] env[63515]: DEBUG oslo_vmware.api [None req-33f5b097-4c20-4377-b5be-7e558874e5d3 tempest-ServersV294TestFqdnHostnames-2112993702 tempest-ServersV294TestFqdnHostnames-2112993702-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52562aee-d00b-a567-5326-33d35d14b7aa, 'name': SearchDatastore_Task, 'duration_secs': 0.064293} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 705.446444] env[63515]: DEBUG oslo_concurrency.lockutils [None req-33f5b097-4c20-4377-b5be-7e558874e5d3 tempest-ServersV294TestFqdnHostnames-2112993702 tempest-ServersV294TestFqdnHostnames-2112993702-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 705.446738] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-33f5b097-4c20-4377-b5be-7e558874e5d3 tempest-ServersV294TestFqdnHostnames-2112993702 tempest-ServersV294TestFqdnHostnames-2112993702-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk to [datastore2] 396e49dd-48c0-496b-a1ec-190c33a22c5e/396e49dd-48c0-496b-a1ec-190c33a22c5e.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 705.447046] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-62e4f49a-bd80-4f27-8385-ced51e6f9984 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.453777] env[63515]: DEBUG oslo_vmware.api [None req-33f5b097-4c20-4377-b5be-7e558874e5d3 tempest-ServersV294TestFqdnHostnames-2112993702 tempest-ServersV294TestFqdnHostnames-2112993702-project-member] Waiting for the task: (returnval){ [ 705.453777] env[63515]: value = "task-1110943" [ 705.453777] env[63515]: _type = "Task" [ 705.453777] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 705.463596] env[63515]: DEBUG oslo_vmware.api [None req-33f5b097-4c20-4377-b5be-7e558874e5d3 tempest-ServersV294TestFqdnHostnames-2112993702 tempest-ServersV294TestFqdnHostnames-2112993702-project-member] Task: {'id': task-1110943, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 705.687708] env[63515]: DEBUG nova.network.neutron [None req-4fa686a6-dc45-4603-a4ee-7107c810126f tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] [instance: 4087b2c4-6ed1-4b68-8b78-a36e34d935b1] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 705.694760] env[63515]: DEBUG oslo_vmware.api [None req-89884269-25d4-4981-bd3b-a346ece7b681 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Task: {'id': task-1110940, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.062745} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 705.694897] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-89884269-25d4-4981-bd3b-a346ece7b681 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk to [datastore2] f1d01b75-ac9d-458d-8cc2-ae64cffca4e8/f1d01b75-ac9d-458d-8cc2-ae64cffca4e8.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 705.695137] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-89884269-25d4-4981-bd3b-a346ece7b681 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] [instance: f1d01b75-ac9d-458d-8cc2-ae64cffca4e8] Extending root virtual disk to 1048576 {{(pid=63515) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 705.695493] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-cf0e8298-8d9f-478b-abf7-2c1dcbd2656c {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.702648] env[63515]: DEBUG oslo_vmware.api [None req-89884269-25d4-4981-bd3b-a346ece7b681 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Waiting for the task: (returnval){ [ 705.702648] env[63515]: value = "task-1110944" [ 705.702648] env[63515]: _type = "Task" [ 705.702648] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 705.711050] env[63515]: DEBUG oslo_vmware.api [None req-89884269-25d4-4981-bd3b-a346ece7b681 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Task: {'id': task-1110944, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 705.891483] env[63515]: DEBUG oslo_vmware.api [None req-c26675e2-4152-417b-99dd-0c18de4dc137 tempest-ServersTestBootFromVolume-2042560423 tempest-ServersTestBootFromVolume-2042560423-project-member] Task: {'id': task-1110942, 'name': ReconfigVM_Task, 'duration_secs': 0.127469} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 705.891919] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-c26675e2-4152-417b-99dd-0c18de4dc137 tempest-ServersTestBootFromVolume-2042560423 tempest-ServersTestBootFromVolume-2042560423-project-member] [instance: 90f4930b-aaa0-4c4b-9ab8-92aed45e200b] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-243378', 'volume_id': '921350b3-bd4a-4fa1-848a-057b53f51f5c', 'name': 'volume-921350b3-bd4a-4fa1-848a-057b53f51f5c', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '90f4930b-aaa0-4c4b-9ab8-92aed45e200b', 'attached_at': '', 'detached_at': '', 'volume_id': '921350b3-bd4a-4fa1-848a-057b53f51f5c', 'serial': '921350b3-bd4a-4fa1-848a-057b53f51f5c'} {{(pid=63515) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 705.892687] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-7865746e-4791-4f27-8537-8bab13c61386 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.902186] env[63515]: DEBUG oslo_vmware.api [None req-c26675e2-4152-417b-99dd-0c18de4dc137 tempest-ServersTestBootFromVolume-2042560423 tempest-ServersTestBootFromVolume-2042560423-project-member] Waiting for the task: (returnval){ [ 705.902186] env[63515]: value = "task-1110945" [ 705.902186] env[63515]: _type = "Task" [ 705.902186] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 705.912082] env[63515]: DEBUG oslo_vmware.api [None req-c26675e2-4152-417b-99dd-0c18de4dc137 tempest-ServersTestBootFromVolume-2042560423 tempest-ServersTestBootFromVolume-2042560423-project-member] Task: {'id': task-1110945, 'name': Rename_Task} progress is 5%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 705.949735] env[63515]: DEBUG oslo_concurrency.lockutils [None req-d937370c-7531-4295-8d43-07fafc09e90c tempest-ServerDiagnosticsTest-1160321831 tempest-ServerDiagnosticsTest-1160321831-project-member] Lock "4ad5440f-cd18-44c5-8836-0aa39824cf03" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 22.059s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 705.951966] env[63515]: DEBUG nova.network.neutron [None req-4fa686a6-dc45-4603-a4ee-7107c810126f tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] [instance: 4087b2c4-6ed1-4b68-8b78-a36e34d935b1] Updating instance_info_cache with network_info: [{"id": "81eafe68-fff5-401d-812f-73b07de3ad8b", "address": "fa:16:3e:83:c6:b6", "network": {"id": "76b5c20b-de28-4b1e-8acb-9032c4c3860f", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-1031942723-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e0c07bb70a414f1787073d335c9001bf", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "171aeae0-6a27-44fc-bc3d-a2d5581fc702", "external-id": "nsx-vlan-transportzone-410", "segmentation_id": 410, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap81eafe68-ff", "ovs_interfaceid": "81eafe68-fff5-401d-812f-73b07de3ad8b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 705.966740] env[63515]: DEBUG oslo_vmware.api [None req-33f5b097-4c20-4377-b5be-7e558874e5d3 tempest-ServersV294TestFqdnHostnames-2112993702 tempest-ServersV294TestFqdnHostnames-2112993702-project-member] Task: {'id': task-1110943, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 706.218022] env[63515]: DEBUG oslo_vmware.api [None req-89884269-25d4-4981-bd3b-a346ece7b681 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Task: {'id': task-1110944, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.082489} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 706.218022] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-89884269-25d4-4981-bd3b-a346ece7b681 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] [instance: f1d01b75-ac9d-458d-8cc2-ae64cffca4e8] Extended root virtual disk {{(pid=63515) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 706.218022] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03d0f048-b8e3-456c-bb50-f10174e42e57 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.241411] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-89884269-25d4-4981-bd3b-a346ece7b681 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] [instance: f1d01b75-ac9d-458d-8cc2-ae64cffca4e8] Reconfiguring VM instance instance-00000022 to attach disk [datastore2] f1d01b75-ac9d-458d-8cc2-ae64cffca4e8/f1d01b75-ac9d-458d-8cc2-ae64cffca4e8.vmdk or device None with type sparse {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 706.241713] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ef16beee-ecf2-4889-ac58-aa535ec0b105 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.262484] env[63515]: DEBUG oslo_vmware.api [None req-89884269-25d4-4981-bd3b-a346ece7b681 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Waiting for the task: (returnval){ [ 706.262484] env[63515]: value = "task-1110946" [ 706.262484] env[63515]: _type = "Task" [ 706.262484] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 706.270600] env[63515]: DEBUG oslo_vmware.api [None req-89884269-25d4-4981-bd3b-a346ece7b681 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Task: {'id': task-1110946, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 706.276997] env[63515]: DEBUG nova.network.neutron [None req-44a3080f-f2e2-4f0c-b244-122beee2aea9 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] [instance: 87c468d9-9594-4804-b461-527f01f6118f] Successfully updated port: d2098868-0475-4b06-bae3-785c608a861e {{(pid=63515) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 706.288285] env[63515]: DEBUG nova.compute.manager [req-5b9baba3-ca4f-45f7-8a24-6bfd7031b320 req-c2991e05-d98d-49a3-bcf4-5eb11c8a1a54 service nova] [instance: 87c468d9-9594-4804-b461-527f01f6118f] Received event network-vif-plugged-d2098868-0475-4b06-bae3-785c608a861e {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 706.288444] env[63515]: DEBUG oslo_concurrency.lockutils [req-5b9baba3-ca4f-45f7-8a24-6bfd7031b320 req-c2991e05-d98d-49a3-bcf4-5eb11c8a1a54 service nova] Acquiring lock "87c468d9-9594-4804-b461-527f01f6118f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 706.288655] env[63515]: DEBUG oslo_concurrency.lockutils [req-5b9baba3-ca4f-45f7-8a24-6bfd7031b320 req-c2991e05-d98d-49a3-bcf4-5eb11c8a1a54 service nova] Lock "87c468d9-9594-4804-b461-527f01f6118f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 706.288806] env[63515]: DEBUG oslo_concurrency.lockutils [req-5b9baba3-ca4f-45f7-8a24-6bfd7031b320 req-c2991e05-d98d-49a3-bcf4-5eb11c8a1a54 service nova] Lock "87c468d9-9594-4804-b461-527f01f6118f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 706.289059] env[63515]: DEBUG nova.compute.manager [req-5b9baba3-ca4f-45f7-8a24-6bfd7031b320 req-c2991e05-d98d-49a3-bcf4-5eb11c8a1a54 service nova] [instance: 87c468d9-9594-4804-b461-527f01f6118f] No waiting events found dispatching network-vif-plugged-d2098868-0475-4b06-bae3-785c608a861e {{(pid=63515) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 706.289186] env[63515]: WARNING nova.compute.manager [req-5b9baba3-ca4f-45f7-8a24-6bfd7031b320 req-c2991e05-d98d-49a3-bcf4-5eb11c8a1a54 service nova] [instance: 87c468d9-9594-4804-b461-527f01f6118f] Received unexpected event network-vif-plugged-d2098868-0475-4b06-bae3-785c608a861e for instance with vm_state building and task_state spawning. [ 706.411907] env[63515]: DEBUG oslo_vmware.api [None req-c26675e2-4152-417b-99dd-0c18de4dc137 tempest-ServersTestBootFromVolume-2042560423 tempest-ServersTestBootFromVolume-2042560423-project-member] Task: {'id': task-1110945, 'name': Rename_Task, 'duration_secs': 0.348282} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 706.412201] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-c26675e2-4152-417b-99dd-0c18de4dc137 tempest-ServersTestBootFromVolume-2042560423 tempest-ServersTestBootFromVolume-2042560423-project-member] [instance: 90f4930b-aaa0-4c4b-9ab8-92aed45e200b] Powering on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 706.412474] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ce5a338e-8881-4ca0-b275-5b66bbce2a27 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.420517] env[63515]: DEBUG oslo_vmware.api [None req-c26675e2-4152-417b-99dd-0c18de4dc137 tempest-ServersTestBootFromVolume-2042560423 tempest-ServersTestBootFromVolume-2042560423-project-member] Waiting for the task: (returnval){ [ 706.420517] env[63515]: value = "task-1110947" [ 706.420517] env[63515]: _type = "Task" [ 706.420517] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 706.429068] env[63515]: DEBUG oslo_vmware.api [None req-c26675e2-4152-417b-99dd-0c18de4dc137 tempest-ServersTestBootFromVolume-2042560423 tempest-ServersTestBootFromVolume-2042560423-project-member] Task: {'id': task-1110947, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 706.456352] env[63515]: DEBUG oslo_concurrency.lockutils [None req-4fa686a6-dc45-4603-a4ee-7107c810126f tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] Releasing lock "refresh_cache-4087b2c4-6ed1-4b68-8b78-a36e34d935b1" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 706.456706] env[63515]: DEBUG nova.compute.manager [None req-4fa686a6-dc45-4603-a4ee-7107c810126f tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] [instance: 4087b2c4-6ed1-4b68-8b78-a36e34d935b1] Instance network_info: |[{"id": "81eafe68-fff5-401d-812f-73b07de3ad8b", "address": "fa:16:3e:83:c6:b6", "network": {"id": "76b5c20b-de28-4b1e-8acb-9032c4c3860f", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-1031942723-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e0c07bb70a414f1787073d335c9001bf", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "171aeae0-6a27-44fc-bc3d-a2d5581fc702", "external-id": "nsx-vlan-transportzone-410", "segmentation_id": 410, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap81eafe68-ff", "ovs_interfaceid": "81eafe68-fff5-401d-812f-73b07de3ad8b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 706.457490] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-4fa686a6-dc45-4603-a4ee-7107c810126f tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] [instance: 4087b2c4-6ed1-4b68-8b78-a36e34d935b1] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:83:c6:b6', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '171aeae0-6a27-44fc-bc3d-a2d5581fc702', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '81eafe68-fff5-401d-812f-73b07de3ad8b', 'vif_model': 'vmxnet3'}] {{(pid=63515) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 706.465528] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-4fa686a6-dc45-4603-a4ee-7107c810126f tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] Creating folder: Project (e0c07bb70a414f1787073d335c9001bf). Parent ref: group-v243370. {{(pid=63515) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 706.471207] env[63515]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-db43cbcb-8176-4a6f-a4f5-005b46a78b67 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.478334] env[63515]: DEBUG oslo_vmware.api [None req-33f5b097-4c20-4377-b5be-7e558874e5d3 tempest-ServersV294TestFqdnHostnames-2112993702 tempest-ServersV294TestFqdnHostnames-2112993702-project-member] Task: {'id': task-1110943, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.549604} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 706.478590] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-33f5b097-4c20-4377-b5be-7e558874e5d3 tempest-ServersV294TestFqdnHostnames-2112993702 tempest-ServersV294TestFqdnHostnames-2112993702-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk to [datastore2] 396e49dd-48c0-496b-a1ec-190c33a22c5e/396e49dd-48c0-496b-a1ec-190c33a22c5e.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 706.478799] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-33f5b097-4c20-4377-b5be-7e558874e5d3 tempest-ServersV294TestFqdnHostnames-2112993702 tempest-ServersV294TestFqdnHostnames-2112993702-project-member] [instance: 396e49dd-48c0-496b-a1ec-190c33a22c5e] Extending root virtual disk to 1048576 {{(pid=63515) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 706.479058] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c64e25b2-9d53-4305-973a-eeea2f1eb970 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.481805] env[63515]: INFO nova.virt.vmwareapi.vm_util [None req-4fa686a6-dc45-4603-a4ee-7107c810126f tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] Created folder: Project (e0c07bb70a414f1787073d335c9001bf) in parent group-v243370. [ 706.482449] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-4fa686a6-dc45-4603-a4ee-7107c810126f tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] Creating folder: Instances. Parent ref: group-v243400. {{(pid=63515) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 706.484527] env[63515]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-dec3f3b9-887f-4873-9071-fba578a60efb {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.487685] env[63515]: DEBUG oslo_vmware.api [None req-33f5b097-4c20-4377-b5be-7e558874e5d3 tempest-ServersV294TestFqdnHostnames-2112993702 tempest-ServersV294TestFqdnHostnames-2112993702-project-member] Waiting for the task: (returnval){ [ 706.487685] env[63515]: value = "task-1110949" [ 706.487685] env[63515]: _type = "Task" [ 706.487685] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 706.495824] env[63515]: DEBUG oslo_vmware.api [None req-33f5b097-4c20-4377-b5be-7e558874e5d3 tempest-ServersV294TestFqdnHostnames-2112993702 tempest-ServersV294TestFqdnHostnames-2112993702-project-member] Task: {'id': task-1110949, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 706.498261] env[63515]: INFO nova.virt.vmwareapi.vm_util [None req-4fa686a6-dc45-4603-a4ee-7107c810126f tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] Created folder: Instances in parent group-v243400. [ 706.498261] env[63515]: DEBUG oslo.service.loopingcall [None req-4fa686a6-dc45-4603-a4ee-7107c810126f tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 706.499300] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4087b2c4-6ed1-4b68-8b78-a36e34d935b1] Creating VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 706.499697] env[63515]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-13518259-bfe9-4653-8fa0-61f37e361ffd {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.520774] env[63515]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 706.520774] env[63515]: value = "task-1110951" [ 706.520774] env[63515]: _type = "Task" [ 706.520774] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 706.528435] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1110951, 'name': CreateVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 706.775573] env[63515]: DEBUG oslo_vmware.api [None req-89884269-25d4-4981-bd3b-a346ece7b681 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Task: {'id': task-1110946, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 706.782328] env[63515]: DEBUG oslo_concurrency.lockutils [None req-44a3080f-f2e2-4f0c-b244-122beee2aea9 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Acquiring lock "refresh_cache-87c468d9-9594-4804-b461-527f01f6118f" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 706.782495] env[63515]: DEBUG oslo_concurrency.lockutils [None req-44a3080f-f2e2-4f0c-b244-122beee2aea9 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Acquired lock "refresh_cache-87c468d9-9594-4804-b461-527f01f6118f" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 706.782974] env[63515]: DEBUG nova.network.neutron [None req-44a3080f-f2e2-4f0c-b244-122beee2aea9 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] [instance: 87c468d9-9594-4804-b461-527f01f6118f] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 706.828786] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e255fef5-0d44-4582-bde9-a1dbe5b958f4 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.836055] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b2ef2d8-3dfd-4673-b6ae-d0367c49817b {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.868980] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dafab514-cf6b-4180-9ae3-516b0131525a {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.877393] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b819d530-0c9e-4ac0-afac-793146477aef {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.893184] env[63515]: DEBUG nova.compute.provider_tree [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 706.932720] env[63515]: DEBUG oslo_vmware.api [None req-c26675e2-4152-417b-99dd-0c18de4dc137 tempest-ServersTestBootFromVolume-2042560423 tempest-ServersTestBootFromVolume-2042560423-project-member] Task: {'id': task-1110947, 'name': PowerOnVM_Task, 'duration_secs': 0.484582} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 706.933109] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-c26675e2-4152-417b-99dd-0c18de4dc137 tempest-ServersTestBootFromVolume-2042560423 tempest-ServersTestBootFromVolume-2042560423-project-member] [instance: 90f4930b-aaa0-4c4b-9ab8-92aed45e200b] Powered on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 706.933375] env[63515]: INFO nova.compute.manager [None req-c26675e2-4152-417b-99dd-0c18de4dc137 tempest-ServersTestBootFromVolume-2042560423 tempest-ServersTestBootFromVolume-2042560423-project-member] [instance: 90f4930b-aaa0-4c4b-9ab8-92aed45e200b] Took 7.21 seconds to spawn the instance on the hypervisor. [ 706.933612] env[63515]: DEBUG nova.compute.manager [None req-c26675e2-4152-417b-99dd-0c18de4dc137 tempest-ServersTestBootFromVolume-2042560423 tempest-ServersTestBootFromVolume-2042560423-project-member] [instance: 90f4930b-aaa0-4c4b-9ab8-92aed45e200b] Checking state {{(pid=63515) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 706.934578] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a77e32c-eeae-44ba-90a9-67683edc1c63 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.001701] env[63515]: DEBUG oslo_vmware.api [None req-33f5b097-4c20-4377-b5be-7e558874e5d3 tempest-ServersV294TestFqdnHostnames-2112993702 tempest-ServersV294TestFqdnHostnames-2112993702-project-member] Task: {'id': task-1110949, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.067772} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 707.002800] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-33f5b097-4c20-4377-b5be-7e558874e5d3 tempest-ServersV294TestFqdnHostnames-2112993702 tempest-ServersV294TestFqdnHostnames-2112993702-project-member] [instance: 396e49dd-48c0-496b-a1ec-190c33a22c5e] Extended root virtual disk {{(pid=63515) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 707.003928] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3509cc27-4d20-4948-b500-babfcfb85755 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.028136] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-33f5b097-4c20-4377-b5be-7e558874e5d3 tempest-ServersV294TestFqdnHostnames-2112993702 tempest-ServersV294TestFqdnHostnames-2112993702-project-member] [instance: 396e49dd-48c0-496b-a1ec-190c33a22c5e] Reconfiguring VM instance instance-00000024 to attach disk [datastore2] 396e49dd-48c0-496b-a1ec-190c33a22c5e/396e49dd-48c0-496b-a1ec-190c33a22c5e.vmdk or device None with type sparse {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 707.028876] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b0afe234-1544-4cd8-8f4a-295ae1234aff {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.051959] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1110951, 'name': CreateVM_Task} progress is 99%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 707.053431] env[63515]: DEBUG oslo_vmware.api [None req-33f5b097-4c20-4377-b5be-7e558874e5d3 tempest-ServersV294TestFqdnHostnames-2112993702 tempest-ServersV294TestFqdnHostnames-2112993702-project-member] Waiting for the task: (returnval){ [ 707.053431] env[63515]: value = "task-1110952" [ 707.053431] env[63515]: _type = "Task" [ 707.053431] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 707.273859] env[63515]: DEBUG oslo_vmware.api [None req-89884269-25d4-4981-bd3b-a346ece7b681 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Task: {'id': task-1110946, 'name': ReconfigVM_Task, 'duration_secs': 0.817836} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 707.273859] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-89884269-25d4-4981-bd3b-a346ece7b681 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] [instance: f1d01b75-ac9d-458d-8cc2-ae64cffca4e8] Reconfigured VM instance instance-00000022 to attach disk [datastore2] f1d01b75-ac9d-458d-8cc2-ae64cffca4e8/f1d01b75-ac9d-458d-8cc2-ae64cffca4e8.vmdk or device None with type sparse {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 707.274538] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-23d09d76-4793-4967-8029-d9d07f098d37 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.281433] env[63515]: DEBUG oslo_vmware.api [None req-89884269-25d4-4981-bd3b-a346ece7b681 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Waiting for the task: (returnval){ [ 707.281433] env[63515]: value = "task-1110953" [ 707.281433] env[63515]: _type = "Task" [ 707.281433] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 707.292595] env[63515]: DEBUG oslo_vmware.api [None req-89884269-25d4-4981-bd3b-a346ece7b681 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Task: {'id': task-1110953, 'name': Rename_Task} progress is 5%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 707.342028] env[63515]: DEBUG nova.network.neutron [None req-44a3080f-f2e2-4f0c-b244-122beee2aea9 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] [instance: 87c468d9-9594-4804-b461-527f01f6118f] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 707.398644] env[63515]: DEBUG nova.scheduler.client.report [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 707.458700] env[63515]: INFO nova.compute.manager [None req-c26675e2-4152-417b-99dd-0c18de4dc137 tempest-ServersTestBootFromVolume-2042560423 tempest-ServersTestBootFromVolume-2042560423-project-member] [instance: 90f4930b-aaa0-4c4b-9ab8-92aed45e200b] Took 25.12 seconds to build instance. [ 707.538023] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1110951, 'name': CreateVM_Task, 'duration_secs': 0.576399} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 707.538199] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4087b2c4-6ed1-4b68-8b78-a36e34d935b1] Created VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 707.538861] env[63515]: DEBUG oslo_concurrency.lockutils [None req-4fa686a6-dc45-4603-a4ee-7107c810126f tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 707.539030] env[63515]: DEBUG oslo_concurrency.lockutils [None req-4fa686a6-dc45-4603-a4ee-7107c810126f tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 707.539333] env[63515]: DEBUG oslo_concurrency.lockutils [None req-4fa686a6-dc45-4603-a4ee-7107c810126f tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 707.539575] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3e5b3e3a-7ec8-40d3-bfb8-dbf629c4cd5a {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.546375] env[63515]: DEBUG oslo_vmware.api [None req-4fa686a6-dc45-4603-a4ee-7107c810126f tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] Waiting for the task: (returnval){ [ 707.546375] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]52eefea6-74d1-0b49-30d5-43e2b702344f" [ 707.546375] env[63515]: _type = "Task" [ 707.546375] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 707.556069] env[63515]: DEBUG oslo_vmware.api [None req-4fa686a6-dc45-4603-a4ee-7107c810126f tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52eefea6-74d1-0b49-30d5-43e2b702344f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 707.566487] env[63515]: DEBUG oslo_vmware.api [None req-33f5b097-4c20-4377-b5be-7e558874e5d3 tempest-ServersV294TestFqdnHostnames-2112993702 tempest-ServersV294TestFqdnHostnames-2112993702-project-member] Task: {'id': task-1110952, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 707.574052] env[63515]: DEBUG nova.network.neutron [None req-44a3080f-f2e2-4f0c-b244-122beee2aea9 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] [instance: 87c468d9-9594-4804-b461-527f01f6118f] Updating instance_info_cache with network_info: [{"id": "d2098868-0475-4b06-bae3-785c608a861e", "address": "fa:16:3e:af:f9:f0", "network": {"id": "09fb20ca-2b17-4d46-b915-6aff85b862d8", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1815529427-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "19ce3b05eec64475b685faad1fd801f6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ded18042-834c-4792-b3e8-b1c377446432", "external-id": "nsx-vlan-transportzone-293", "segmentation_id": 293, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd2098868-04", "ovs_interfaceid": "d2098868-0475-4b06-bae3-785c608a861e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 707.793350] env[63515]: DEBUG oslo_vmware.api [None req-89884269-25d4-4981-bd3b-a346ece7b681 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Task: {'id': task-1110953, 'name': Rename_Task, 'duration_secs': 0.130041} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 707.794134] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-89884269-25d4-4981-bd3b-a346ece7b681 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] [instance: f1d01b75-ac9d-458d-8cc2-ae64cffca4e8] Powering on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 707.794134] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8703ef37-1b9b-4536-81e5-22735aa87749 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.800558] env[63515]: DEBUG oslo_vmware.api [None req-89884269-25d4-4981-bd3b-a346ece7b681 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Waiting for the task: (returnval){ [ 707.800558] env[63515]: value = "task-1110954" [ 707.800558] env[63515]: _type = "Task" [ 707.800558] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 707.808070] env[63515]: DEBUG oslo_vmware.api [None req-89884269-25d4-4981-bd3b-a346ece7b681 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Task: {'id': task-1110954, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 707.904813] env[63515]: DEBUG oslo_concurrency.lockutils [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.496s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 707.905502] env[63515]: DEBUG nova.compute.manager [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] [instance: 48668736-df27-4f2a-94d9-132f5b49701b] Start building networks asynchronously for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 707.908257] env[63515]: DEBUG oslo_concurrency.lockutils [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.786s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 707.909696] env[63515]: INFO nova.compute.claims [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] [instance: 452f3ca2-6141-43b2-a77a-c9ab5754192d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 707.961218] env[63515]: DEBUG oslo_concurrency.lockutils [None req-c26675e2-4152-417b-99dd-0c18de4dc137 tempest-ServersTestBootFromVolume-2042560423 tempest-ServersTestBootFromVolume-2042560423-project-member] Lock "90f4930b-aaa0-4c4b-9ab8-92aed45e200b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 116.061s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 708.058052] env[63515]: DEBUG oslo_vmware.api [None req-4fa686a6-dc45-4603-a4ee-7107c810126f tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52eefea6-74d1-0b49-30d5-43e2b702344f, 'name': SearchDatastore_Task, 'duration_secs': 0.009835} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 708.061737] env[63515]: DEBUG oslo_concurrency.lockutils [None req-4fa686a6-dc45-4603-a4ee-7107c810126f tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 708.062070] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-4fa686a6-dc45-4603-a4ee-7107c810126f tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] [instance: 4087b2c4-6ed1-4b68-8b78-a36e34d935b1] Processing image 8a120570-cb06-4099-b262-554ca0ad15c5 {{(pid=63515) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 708.062670] env[63515]: DEBUG oslo_concurrency.lockutils [None req-4fa686a6-dc45-4603-a4ee-7107c810126f tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 708.062932] env[63515]: DEBUG oslo_concurrency.lockutils [None req-4fa686a6-dc45-4603-a4ee-7107c810126f tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 708.063127] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-4fa686a6-dc45-4603-a4ee-7107c810126f tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 708.063412] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-01547a3a-902a-4048-b9ed-bd963769b0ce {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.071418] env[63515]: DEBUG oslo_vmware.api [None req-33f5b097-4c20-4377-b5be-7e558874e5d3 tempest-ServersV294TestFqdnHostnames-2112993702 tempest-ServersV294TestFqdnHostnames-2112993702-project-member] Task: {'id': task-1110952, 'name': ReconfigVM_Task, 'duration_secs': 0.578368} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 708.072854] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-33f5b097-4c20-4377-b5be-7e558874e5d3 tempest-ServersV294TestFqdnHostnames-2112993702 tempest-ServersV294TestFqdnHostnames-2112993702-project-member] [instance: 396e49dd-48c0-496b-a1ec-190c33a22c5e] Reconfigured VM instance instance-00000024 to attach disk [datastore2] 396e49dd-48c0-496b-a1ec-190c33a22c5e/396e49dd-48c0-496b-a1ec-190c33a22c5e.vmdk or device None with type sparse {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 708.074334] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-4fa686a6-dc45-4603-a4ee-7107c810126f tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 708.074334] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-4fa686a6-dc45-4603-a4ee-7107c810126f tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63515) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 708.075020] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-76f738f9-8b8b-4e87-9e3f-5dd4934c8743 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.076615] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c1895fd6-c167-42e2-9e3a-dff99acdcbc4 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.078871] env[63515]: DEBUG oslo_concurrency.lockutils [None req-44a3080f-f2e2-4f0c-b244-122beee2aea9 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Releasing lock "refresh_cache-87c468d9-9594-4804-b461-527f01f6118f" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 708.079168] env[63515]: DEBUG nova.compute.manager [None req-44a3080f-f2e2-4f0c-b244-122beee2aea9 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] [instance: 87c468d9-9594-4804-b461-527f01f6118f] Instance network_info: |[{"id": "d2098868-0475-4b06-bae3-785c608a861e", "address": "fa:16:3e:af:f9:f0", "network": {"id": "09fb20ca-2b17-4d46-b915-6aff85b862d8", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1815529427-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "19ce3b05eec64475b685faad1fd801f6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ded18042-834c-4792-b3e8-b1c377446432", "external-id": "nsx-vlan-transportzone-293", "segmentation_id": 293, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd2098868-04", "ovs_interfaceid": "d2098868-0475-4b06-bae3-785c608a861e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 708.079529] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-44a3080f-f2e2-4f0c-b244-122beee2aea9 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] [instance: 87c468d9-9594-4804-b461-527f01f6118f] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:af:f9:f0', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ded18042-834c-4792-b3e8-b1c377446432', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'd2098868-0475-4b06-bae3-785c608a861e', 'vif_model': 'vmxnet3'}] {{(pid=63515) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 708.089857] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-44a3080f-f2e2-4f0c-b244-122beee2aea9 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Creating folder: Project (19ce3b05eec64475b685faad1fd801f6). Parent ref: group-v243370. {{(pid=63515) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 708.091268] env[63515]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-7ed930ab-9fe9-4680-9517-9d2aa0087ed2 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.095109] env[63515]: DEBUG oslo_vmware.api [None req-4fa686a6-dc45-4603-a4ee-7107c810126f tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] Waiting for the task: (returnval){ [ 708.095109] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]5262cc9d-ba7e-fb51-8da5-681220b3eb0e" [ 708.095109] env[63515]: _type = "Task" [ 708.095109] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 708.096864] env[63515]: DEBUG oslo_vmware.api [None req-33f5b097-4c20-4377-b5be-7e558874e5d3 tempest-ServersV294TestFqdnHostnames-2112993702 tempest-ServersV294TestFqdnHostnames-2112993702-project-member] Waiting for the task: (returnval){ [ 708.096864] env[63515]: value = "task-1110955" [ 708.096864] env[63515]: _type = "Task" [ 708.096864] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 708.105227] env[63515]: INFO nova.virt.vmwareapi.vm_util [None req-44a3080f-f2e2-4f0c-b244-122beee2aea9 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Created folder: Project (19ce3b05eec64475b685faad1fd801f6) in parent group-v243370. [ 708.105428] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-44a3080f-f2e2-4f0c-b244-122beee2aea9 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Creating folder: Instances. Parent ref: group-v243403. {{(pid=63515) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 708.108976] env[63515]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-14df9114-ab35-4c2f-851b-36518aba4265 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.110635] env[63515]: DEBUG oslo_vmware.api [None req-4fa686a6-dc45-4603-a4ee-7107c810126f tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]5262cc9d-ba7e-fb51-8da5-681220b3eb0e, 'name': SearchDatastore_Task, 'duration_secs': 0.011495} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 708.114579] env[63515]: DEBUG oslo_vmware.api [None req-33f5b097-4c20-4377-b5be-7e558874e5d3 tempest-ServersV294TestFqdnHostnames-2112993702 tempest-ServersV294TestFqdnHostnames-2112993702-project-member] Task: {'id': task-1110955, 'name': Rename_Task} progress is 6%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 708.115184] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-28278f42-ca2c-4ec9-b62b-152ba3c601da {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.122684] env[63515]: DEBUG oslo_vmware.api [None req-4fa686a6-dc45-4603-a4ee-7107c810126f tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] Waiting for the task: (returnval){ [ 708.122684] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]525a6c1d-4287-48aa-6299-552f60b06c69" [ 708.122684] env[63515]: _type = "Task" [ 708.122684] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 708.124511] env[63515]: INFO nova.virt.vmwareapi.vm_util [None req-44a3080f-f2e2-4f0c-b244-122beee2aea9 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Created folder: Instances in parent group-v243403. [ 708.124806] env[63515]: DEBUG oslo.service.loopingcall [None req-44a3080f-f2e2-4f0c-b244-122beee2aea9 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 708.130255] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 87c468d9-9594-4804-b461-527f01f6118f] Creating VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 708.130547] env[63515]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-de001edc-5921-4c45-bfa1-8fec43b8f133 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.152103] env[63515]: DEBUG oslo_vmware.api [None req-4fa686a6-dc45-4603-a4ee-7107c810126f tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]525a6c1d-4287-48aa-6299-552f60b06c69, 'name': SearchDatastore_Task, 'duration_secs': 0.009403} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 708.154261] env[63515]: DEBUG oslo_concurrency.lockutils [None req-4fa686a6-dc45-4603-a4ee-7107c810126f tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 708.154361] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-4fa686a6-dc45-4603-a4ee-7107c810126f tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk to [datastore1] 4087b2c4-6ed1-4b68-8b78-a36e34d935b1/4087b2c4-6ed1-4b68-8b78-a36e34d935b1.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 708.154625] env[63515]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 708.154625] env[63515]: value = "task-1110958" [ 708.154625] env[63515]: _type = "Task" [ 708.154625] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 708.154829] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-4ea1ea5b-02b6-44d6-aadc-032db5360323 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.164799] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1110958, 'name': CreateVM_Task} progress is 5%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 708.166227] env[63515]: DEBUG oslo_vmware.api [None req-4fa686a6-dc45-4603-a4ee-7107c810126f tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] Waiting for the task: (returnval){ [ 708.166227] env[63515]: value = "task-1110959" [ 708.166227] env[63515]: _type = "Task" [ 708.166227] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 708.174036] env[63515]: DEBUG oslo_vmware.api [None req-4fa686a6-dc45-4603-a4ee-7107c810126f tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] Task: {'id': task-1110959, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 708.311143] env[63515]: DEBUG oslo_vmware.api [None req-89884269-25d4-4981-bd3b-a346ece7b681 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Task: {'id': task-1110954, 'name': PowerOnVM_Task, 'duration_secs': 0.483916} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 708.311485] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-89884269-25d4-4981-bd3b-a346ece7b681 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] [instance: f1d01b75-ac9d-458d-8cc2-ae64cffca4e8] Powered on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 708.311485] env[63515]: INFO nova.compute.manager [None req-89884269-25d4-4981-bd3b-a346ece7b681 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] [instance: f1d01b75-ac9d-458d-8cc2-ae64cffca4e8] Took 12.94 seconds to spawn the instance on the hypervisor. [ 708.311687] env[63515]: DEBUG nova.compute.manager [None req-89884269-25d4-4981-bd3b-a346ece7b681 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] [instance: f1d01b75-ac9d-458d-8cc2-ae64cffca4e8] Checking state {{(pid=63515) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 708.312496] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9e99fe1-4848-48d8-8b6c-bc80bcfe7c2a {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.337490] env[63515]: DEBUG nova.compute.manager [req-17f173b5-8bad-45cb-aaaa-77b2f1a31d98 req-42ebae56-614d-4c8e-a907-06755d33bbde service nova] [instance: 87c468d9-9594-4804-b461-527f01f6118f] Received event network-changed-d2098868-0475-4b06-bae3-785c608a861e {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 708.337490] env[63515]: DEBUG nova.compute.manager [req-17f173b5-8bad-45cb-aaaa-77b2f1a31d98 req-42ebae56-614d-4c8e-a907-06755d33bbde service nova] [instance: 87c468d9-9594-4804-b461-527f01f6118f] Refreshing instance network info cache due to event network-changed-d2098868-0475-4b06-bae3-785c608a861e. {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 708.337490] env[63515]: DEBUG oslo_concurrency.lockutils [req-17f173b5-8bad-45cb-aaaa-77b2f1a31d98 req-42ebae56-614d-4c8e-a907-06755d33bbde service nova] Acquiring lock "refresh_cache-87c468d9-9594-4804-b461-527f01f6118f" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 708.337490] env[63515]: DEBUG oslo_concurrency.lockutils [req-17f173b5-8bad-45cb-aaaa-77b2f1a31d98 req-42ebae56-614d-4c8e-a907-06755d33bbde service nova] Acquired lock "refresh_cache-87c468d9-9594-4804-b461-527f01f6118f" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 708.337490] env[63515]: DEBUG nova.network.neutron [req-17f173b5-8bad-45cb-aaaa-77b2f1a31d98 req-42ebae56-614d-4c8e-a907-06755d33bbde service nova] [instance: 87c468d9-9594-4804-b461-527f01f6118f] Refreshing network info cache for port d2098868-0475-4b06-bae3-785c608a861e {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 708.416030] env[63515]: DEBUG nova.compute.utils [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Using /dev/sd instead of None {{(pid=63515) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 708.420382] env[63515]: DEBUG nova.compute.manager [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] [instance: 48668736-df27-4f2a-94d9-132f5b49701b] Allocating IP information in the background. {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 708.420964] env[63515]: DEBUG nova.network.neutron [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] [instance: 48668736-df27-4f2a-94d9-132f5b49701b] allocate_for_instance() {{(pid=63515) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 708.466951] env[63515]: DEBUG nova.compute.manager [None req-e71315b0-9587-4929-82a2-71bebc6fcbfc tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] [instance: d7e8dcdb-41cb-46fb-8b61-d251e7c2d372] Starting instance... {{(pid=63515) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 708.476636] env[63515]: DEBUG nova.policy [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '479be60850604cf19669871a78f4b710', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3374114af6314f86be3559223d70dfdb', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63515) authorize /opt/stack/nova/nova/policy.py:201}} [ 708.612681] env[63515]: DEBUG oslo_vmware.api [None req-33f5b097-4c20-4377-b5be-7e558874e5d3 tempest-ServersV294TestFqdnHostnames-2112993702 tempest-ServersV294TestFqdnHostnames-2112993702-project-member] Task: {'id': task-1110955, 'name': Rename_Task, 'duration_secs': 0.194644} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 708.612975] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-33f5b097-4c20-4377-b5be-7e558874e5d3 tempest-ServersV294TestFqdnHostnames-2112993702 tempest-ServersV294TestFqdnHostnames-2112993702-project-member] [instance: 396e49dd-48c0-496b-a1ec-190c33a22c5e] Powering on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 708.613263] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2149d3b0-fbf8-4d9c-86e4-c9d6f2ef88dc {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.624392] env[63515]: DEBUG oslo_vmware.api [None req-33f5b097-4c20-4377-b5be-7e558874e5d3 tempest-ServersV294TestFqdnHostnames-2112993702 tempest-ServersV294TestFqdnHostnames-2112993702-project-member] Waiting for the task: (returnval){ [ 708.624392] env[63515]: value = "task-1110960" [ 708.624392] env[63515]: _type = "Task" [ 708.624392] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 708.630859] env[63515]: DEBUG oslo_vmware.api [None req-33f5b097-4c20-4377-b5be-7e558874e5d3 tempest-ServersV294TestFqdnHostnames-2112993702 tempest-ServersV294TestFqdnHostnames-2112993702-project-member] Task: {'id': task-1110960, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 708.670736] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1110958, 'name': CreateVM_Task, 'duration_secs': 0.384323} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 708.676720] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 87c468d9-9594-4804-b461-527f01f6118f] Created VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 708.677935] env[63515]: DEBUG oslo_concurrency.lockutils [None req-44a3080f-f2e2-4f0c-b244-122beee2aea9 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 708.678260] env[63515]: DEBUG oslo_concurrency.lockutils [None req-44a3080f-f2e2-4f0c-b244-122beee2aea9 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 708.680216] env[63515]: DEBUG oslo_concurrency.lockutils [None req-44a3080f-f2e2-4f0c-b244-122beee2aea9 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 708.680216] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-db0f605b-50ef-4d75-b9a5-df33c33703a5 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.691334] env[63515]: DEBUG oslo_vmware.api [None req-4fa686a6-dc45-4603-a4ee-7107c810126f tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] Task: {'id': task-1110959, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.507354} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 708.691334] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-4fa686a6-dc45-4603-a4ee-7107c810126f tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk to [datastore1] 4087b2c4-6ed1-4b68-8b78-a36e34d935b1/4087b2c4-6ed1-4b68-8b78-a36e34d935b1.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 708.692818] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-4fa686a6-dc45-4603-a4ee-7107c810126f tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] [instance: 4087b2c4-6ed1-4b68-8b78-a36e34d935b1] Extending root virtual disk to 1048576 {{(pid=63515) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 708.692818] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-39154fff-d346-4a6f-b1d6-8f76f630ea91 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.699058] env[63515]: DEBUG oslo_vmware.api [None req-44a3080f-f2e2-4f0c-b244-122beee2aea9 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Waiting for the task: (returnval){ [ 708.699058] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]520cfdbf-5f8c-89cc-f439-5625d894d524" [ 708.699058] env[63515]: _type = "Task" [ 708.699058] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 708.708205] env[63515]: DEBUG oslo_vmware.api [None req-4fa686a6-dc45-4603-a4ee-7107c810126f tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] Waiting for the task: (returnval){ [ 708.708205] env[63515]: value = "task-1110961" [ 708.708205] env[63515]: _type = "Task" [ 708.708205] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 708.722285] env[63515]: DEBUG oslo_vmware.api [None req-44a3080f-f2e2-4f0c-b244-122beee2aea9 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]520cfdbf-5f8c-89cc-f439-5625d894d524, 'name': SearchDatastore_Task, 'duration_secs': 0.014053} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 708.723396] env[63515]: DEBUG oslo_concurrency.lockutils [None req-44a3080f-f2e2-4f0c-b244-122beee2aea9 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 708.723794] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-44a3080f-f2e2-4f0c-b244-122beee2aea9 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] [instance: 87c468d9-9594-4804-b461-527f01f6118f] Processing image 8a120570-cb06-4099-b262-554ca0ad15c5 {{(pid=63515) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 708.725233] env[63515]: DEBUG oslo_concurrency.lockutils [None req-44a3080f-f2e2-4f0c-b244-122beee2aea9 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 708.725233] env[63515]: DEBUG oslo_concurrency.lockutils [None req-44a3080f-f2e2-4f0c-b244-122beee2aea9 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 708.725233] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-44a3080f-f2e2-4f0c-b244-122beee2aea9 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 708.729684] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1be25866-d7b0-4104-a6cc-80b3bbc36326 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.732640] env[63515]: DEBUG oslo_vmware.api [None req-4fa686a6-dc45-4603-a4ee-7107c810126f tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] Task: {'id': task-1110961, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 708.739840] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-44a3080f-f2e2-4f0c-b244-122beee2aea9 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 708.740166] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-44a3080f-f2e2-4f0c-b244-122beee2aea9 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63515) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 708.740963] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3c146ad2-a38e-47c7-867d-eed40d2dc155 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.746701] env[63515]: DEBUG oslo_vmware.api [None req-44a3080f-f2e2-4f0c-b244-122beee2aea9 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Waiting for the task: (returnval){ [ 708.746701] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]52252239-cd57-6f4a-d3e3-4101a552e3db" [ 708.746701] env[63515]: _type = "Task" [ 708.746701] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 708.757136] env[63515]: DEBUG oslo_vmware.api [None req-44a3080f-f2e2-4f0c-b244-122beee2aea9 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52252239-cd57-6f4a-d3e3-4101a552e3db, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 708.832371] env[63515]: INFO nova.compute.manager [None req-89884269-25d4-4981-bd3b-a346ece7b681 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] [instance: f1d01b75-ac9d-458d-8cc2-ae64cffca4e8] Took 31.06 seconds to build instance. [ 708.920609] env[63515]: DEBUG nova.compute.manager [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] [instance: 48668736-df27-4f2a-94d9-132f5b49701b] Start building block device mappings for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 708.987450] env[63515]: DEBUG oslo_concurrency.lockutils [None req-e71315b0-9587-4929-82a2-71bebc6fcbfc tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 709.018040] env[63515]: DEBUG nova.network.neutron [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] [instance: 48668736-df27-4f2a-94d9-132f5b49701b] Successfully created port: 9379e602-f7be-4c75-a6e6-c42a8702f6da {{(pid=63515) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 709.134697] env[63515]: DEBUG oslo_vmware.api [None req-33f5b097-4c20-4377-b5be-7e558874e5d3 tempest-ServersV294TestFqdnHostnames-2112993702 tempest-ServersV294TestFqdnHostnames-2112993702-project-member] Task: {'id': task-1110960, 'name': PowerOnVM_Task} progress is 78%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 709.219187] env[63515]: DEBUG oslo_vmware.api [None req-4fa686a6-dc45-4603-a4ee-7107c810126f tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] Task: {'id': task-1110961, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.082175} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 709.219455] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-4fa686a6-dc45-4603-a4ee-7107c810126f tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] [instance: 4087b2c4-6ed1-4b68-8b78-a36e34d935b1] Extended root virtual disk {{(pid=63515) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 709.220304] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba493a5e-03da-4b9f-89b3-f77b55d6bda2 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.249032] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-4fa686a6-dc45-4603-a4ee-7107c810126f tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] [instance: 4087b2c4-6ed1-4b68-8b78-a36e34d935b1] Reconfiguring VM instance instance-00000025 to attach disk [datastore1] 4087b2c4-6ed1-4b68-8b78-a36e34d935b1/4087b2c4-6ed1-4b68-8b78-a36e34d935b1.vmdk or device None with type sparse {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 709.257074] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a983955e-885d-4c43-a7a6-281a03aa0772 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.281600] env[63515]: DEBUG oslo_vmware.api [None req-44a3080f-f2e2-4f0c-b244-122beee2aea9 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52252239-cd57-6f4a-d3e3-4101a552e3db, 'name': SearchDatastore_Task, 'duration_secs': 0.00897} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 709.285825] env[63515]: DEBUG oslo_vmware.api [None req-4fa686a6-dc45-4603-a4ee-7107c810126f tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] Waiting for the task: (returnval){ [ 709.285825] env[63515]: value = "task-1110962" [ 709.285825] env[63515]: _type = "Task" [ 709.285825] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 709.286228] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3220b6dc-6bcf-4eea-9248-449f5209fdcc {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.296482] env[63515]: DEBUG oslo_vmware.api [None req-4fa686a6-dc45-4603-a4ee-7107c810126f tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] Task: {'id': task-1110962, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 709.297439] env[63515]: DEBUG oslo_vmware.api [None req-44a3080f-f2e2-4f0c-b244-122beee2aea9 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Waiting for the task: (returnval){ [ 709.297439] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]52e75c3d-246f-828b-22c4-c7c7ddbe5837" [ 709.297439] env[63515]: _type = "Task" [ 709.297439] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 709.310025] env[63515]: DEBUG oslo_vmware.api [None req-44a3080f-f2e2-4f0c-b244-122beee2aea9 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52e75c3d-246f-828b-22c4-c7c7ddbe5837, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 709.328738] env[63515]: DEBUG nova.compute.manager [req-47f4d0c6-4515-4a08-9241-f900ccf0de3f req-a7eecd75-c61a-42ee-ad17-cf663bf943f8 service nova] [instance: 90f4930b-aaa0-4c4b-9ab8-92aed45e200b] Received event network-changed-a35de7c4-c1af-482b-99a5-1059368cefbb {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 709.329011] env[63515]: DEBUG nova.compute.manager [req-47f4d0c6-4515-4a08-9241-f900ccf0de3f req-a7eecd75-c61a-42ee-ad17-cf663bf943f8 service nova] [instance: 90f4930b-aaa0-4c4b-9ab8-92aed45e200b] Refreshing instance network info cache due to event network-changed-a35de7c4-c1af-482b-99a5-1059368cefbb. {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 709.330086] env[63515]: DEBUG oslo_concurrency.lockutils [req-47f4d0c6-4515-4a08-9241-f900ccf0de3f req-a7eecd75-c61a-42ee-ad17-cf663bf943f8 service nova] Acquiring lock "refresh_cache-90f4930b-aaa0-4c4b-9ab8-92aed45e200b" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 709.330086] env[63515]: DEBUG oslo_concurrency.lockutils [req-47f4d0c6-4515-4a08-9241-f900ccf0de3f req-a7eecd75-c61a-42ee-ad17-cf663bf943f8 service nova] Acquired lock "refresh_cache-90f4930b-aaa0-4c4b-9ab8-92aed45e200b" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 709.330086] env[63515]: DEBUG nova.network.neutron [req-47f4d0c6-4515-4a08-9241-f900ccf0de3f req-a7eecd75-c61a-42ee-ad17-cf663bf943f8 service nova] [instance: 90f4930b-aaa0-4c4b-9ab8-92aed45e200b] Refreshing network info cache for port a35de7c4-c1af-482b-99a5-1059368cefbb {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 709.333740] env[63515]: DEBUG oslo_concurrency.lockutils [None req-89884269-25d4-4981-bd3b-a346ece7b681 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Lock "f1d01b75-ac9d-458d-8cc2-ae64cffca4e8" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 121.062s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 709.423043] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-824e409b-a074-4b58-af3d-f58171b64096 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.436621] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1bc95a44-22f8-47d5-bb9a-1d5e255c1010 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.470767] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a52184a-4d09-4167-97d2-82a63413c2e4 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.479726] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58d08d68-bcc9-4d29-8a17-33be764188bf {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.495902] env[63515]: DEBUG nova.compute.provider_tree [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 709.517957] env[63515]: DEBUG nova.network.neutron [req-17f173b5-8bad-45cb-aaaa-77b2f1a31d98 req-42ebae56-614d-4c8e-a907-06755d33bbde service nova] [instance: 87c468d9-9594-4804-b461-527f01f6118f] Updated VIF entry in instance network info cache for port d2098868-0475-4b06-bae3-785c608a861e. {{(pid=63515) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 709.518393] env[63515]: DEBUG nova.network.neutron [req-17f173b5-8bad-45cb-aaaa-77b2f1a31d98 req-42ebae56-614d-4c8e-a907-06755d33bbde service nova] [instance: 87c468d9-9594-4804-b461-527f01f6118f] Updating instance_info_cache with network_info: [{"id": "d2098868-0475-4b06-bae3-785c608a861e", "address": "fa:16:3e:af:f9:f0", "network": {"id": "09fb20ca-2b17-4d46-b915-6aff85b862d8", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1815529427-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "19ce3b05eec64475b685faad1fd801f6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ded18042-834c-4792-b3e8-b1c377446432", "external-id": "nsx-vlan-transportzone-293", "segmentation_id": 293, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd2098868-04", "ovs_interfaceid": "d2098868-0475-4b06-bae3-785c608a861e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 709.632125] env[63515]: DEBUG oslo_vmware.api [None req-33f5b097-4c20-4377-b5be-7e558874e5d3 tempest-ServersV294TestFqdnHostnames-2112993702 tempest-ServersV294TestFqdnHostnames-2112993702-project-member] Task: {'id': task-1110960, 'name': PowerOnVM_Task, 'duration_secs': 0.947582} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 709.632418] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-33f5b097-4c20-4377-b5be-7e558874e5d3 tempest-ServersV294TestFqdnHostnames-2112993702 tempest-ServersV294TestFqdnHostnames-2112993702-project-member] [instance: 396e49dd-48c0-496b-a1ec-190c33a22c5e] Powered on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 709.632739] env[63515]: INFO nova.compute.manager [None req-33f5b097-4c20-4377-b5be-7e558874e5d3 tempest-ServersV294TestFqdnHostnames-2112993702 tempest-ServersV294TestFqdnHostnames-2112993702-project-member] [instance: 396e49dd-48c0-496b-a1ec-190c33a22c5e] Took 9.10 seconds to spawn the instance on the hypervisor. [ 709.633024] env[63515]: DEBUG nova.compute.manager [None req-33f5b097-4c20-4377-b5be-7e558874e5d3 tempest-ServersV294TestFqdnHostnames-2112993702 tempest-ServersV294TestFqdnHostnames-2112993702-project-member] [instance: 396e49dd-48c0-496b-a1ec-190c33a22c5e] Checking state {{(pid=63515) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 709.633905] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74462efc-5d70-40fe-9c02-b23f1f1983d6 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.799362] env[63515]: DEBUG oslo_vmware.api [None req-4fa686a6-dc45-4603-a4ee-7107c810126f tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] Task: {'id': task-1110962, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 709.808748] env[63515]: DEBUG oslo_vmware.api [None req-44a3080f-f2e2-4f0c-b244-122beee2aea9 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52e75c3d-246f-828b-22c4-c7c7ddbe5837, 'name': SearchDatastore_Task, 'duration_secs': 0.020382} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 709.809024] env[63515]: DEBUG oslo_concurrency.lockutils [None req-44a3080f-f2e2-4f0c-b244-122beee2aea9 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 709.809514] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-44a3080f-f2e2-4f0c-b244-122beee2aea9 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk to [datastore1] 87c468d9-9594-4804-b461-527f01f6118f/87c468d9-9594-4804-b461-527f01f6118f.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 709.809974] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e910814a-d067-4ff7-907c-a92d230685ad {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.818384] env[63515]: DEBUG oslo_vmware.api [None req-44a3080f-f2e2-4f0c-b244-122beee2aea9 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Waiting for the task: (returnval){ [ 709.818384] env[63515]: value = "task-1110963" [ 709.818384] env[63515]: _type = "Task" [ 709.818384] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 709.827157] env[63515]: DEBUG oslo_vmware.api [None req-44a3080f-f2e2-4f0c-b244-122beee2aea9 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Task: {'id': task-1110963, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 709.838952] env[63515]: DEBUG nova.compute.manager [None req-dcda8d78-de0d-44e9-b61f-e84b4a49533f tempest-ServersTestManualDisk-1114050206 tempest-ServersTestManualDisk-1114050206-project-member] [instance: e91aa479-1540-4950-851b-b2409e5f89f1] Starting instance... {{(pid=63515) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 709.933496] env[63515]: DEBUG nova.compute.manager [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] [instance: 48668736-df27-4f2a-94d9-132f5b49701b] Start spawning the instance on the hypervisor. {{(pid=63515) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 709.985288] env[63515]: DEBUG nova.virt.hardware [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T02:52:14Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T02:51:56Z,direct_url=,disk_format='vmdk',id=8a120570-cb06-4099-b262-554ca0ad15c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b569255cc43e42e7ae3f2b4ad37c6ef4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T02:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 709.985288] env[63515]: DEBUG nova.virt.hardware [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Flavor limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 709.985288] env[63515]: DEBUG nova.virt.hardware [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Image limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 709.985503] env[63515]: DEBUG nova.virt.hardware [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Flavor pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 709.985503] env[63515]: DEBUG nova.virt.hardware [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Image pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 709.985503] env[63515]: DEBUG nova.virt.hardware [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 709.985503] env[63515]: DEBUG nova.virt.hardware [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 709.985503] env[63515]: DEBUG nova.virt.hardware [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 709.985699] env[63515]: DEBUG nova.virt.hardware [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Got 1 possible topologies {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 709.985699] env[63515]: DEBUG nova.virt.hardware [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 709.985699] env[63515]: DEBUG nova.virt.hardware [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 709.985996] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f0c8cab-9238-49c9-8b0a-f0af2d1e0663 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.995044] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5edf50ce-7ae3-49ef-9ff5-8a52bd1fb239 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.001940] env[63515]: DEBUG nova.scheduler.client.report [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 710.024237] env[63515]: DEBUG oslo_concurrency.lockutils [req-17f173b5-8bad-45cb-aaaa-77b2f1a31d98 req-42ebae56-614d-4c8e-a907-06755d33bbde service nova] Releasing lock "refresh_cache-87c468d9-9594-4804-b461-527f01f6118f" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 710.158063] env[63515]: INFO nova.compute.manager [None req-33f5b097-4c20-4377-b5be-7e558874e5d3 tempest-ServersV294TestFqdnHostnames-2112993702 tempest-ServersV294TestFqdnHostnames-2112993702-project-member] [instance: 396e49dd-48c0-496b-a1ec-190c33a22c5e] Took 26.99 seconds to build instance. [ 710.300841] env[63515]: DEBUG oslo_vmware.api [None req-4fa686a6-dc45-4603-a4ee-7107c810126f tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] Task: {'id': task-1110962, 'name': ReconfigVM_Task, 'duration_secs': 0.732458} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 710.301268] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-4fa686a6-dc45-4603-a4ee-7107c810126f tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] [instance: 4087b2c4-6ed1-4b68-8b78-a36e34d935b1] Reconfigured VM instance instance-00000025 to attach disk [datastore1] 4087b2c4-6ed1-4b68-8b78-a36e34d935b1/4087b2c4-6ed1-4b68-8b78-a36e34d935b1.vmdk or device None with type sparse {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 710.302170] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-49d98850-e68e-405d-a792-929a1ad337b6 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.309721] env[63515]: DEBUG oslo_vmware.api [None req-4fa686a6-dc45-4603-a4ee-7107c810126f tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] Waiting for the task: (returnval){ [ 710.309721] env[63515]: value = "task-1110964" [ 710.309721] env[63515]: _type = "Task" [ 710.309721] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 710.322084] env[63515]: DEBUG oslo_vmware.api [None req-4fa686a6-dc45-4603-a4ee-7107c810126f tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] Task: {'id': task-1110964, 'name': Rename_Task} progress is 5%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 710.335027] env[63515]: DEBUG oslo_vmware.api [None req-44a3080f-f2e2-4f0c-b244-122beee2aea9 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Task: {'id': task-1110963, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 710.369743] env[63515]: DEBUG oslo_concurrency.lockutils [None req-dcda8d78-de0d-44e9-b61f-e84b4a49533f tempest-ServersTestManualDisk-1114050206 tempest-ServersTestManualDisk-1114050206-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 710.467997] env[63515]: DEBUG nova.network.neutron [req-47f4d0c6-4515-4a08-9241-f900ccf0de3f req-a7eecd75-c61a-42ee-ad17-cf663bf943f8 service nova] [instance: 90f4930b-aaa0-4c4b-9ab8-92aed45e200b] Updated VIF entry in instance network info cache for port a35de7c4-c1af-482b-99a5-1059368cefbb. {{(pid=63515) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 710.468403] env[63515]: DEBUG nova.network.neutron [req-47f4d0c6-4515-4a08-9241-f900ccf0de3f req-a7eecd75-c61a-42ee-ad17-cf663bf943f8 service nova] [instance: 90f4930b-aaa0-4c4b-9ab8-92aed45e200b] Updating instance_info_cache with network_info: [{"id": "a35de7c4-c1af-482b-99a5-1059368cefbb", "address": "fa:16:3e:df:a6:62", "network": {"id": "d19ced2d-871d-4aa5-9b07-8ca41a46d948", "bridge": "br-int", "label": "tempest-ServersTestBootFromVolume-1210740991-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.241", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "38b2d712caf34cc08ad9c9694c521b7e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b2ede0e6-8d7a-4018-bb37-25bf388e9867", "external-id": "nsx-vlan-transportzone-945", "segmentation_id": 945, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa35de7c4-c1", "ovs_interfaceid": "a35de7c4-c1af-482b-99a5-1059368cefbb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 710.506700] env[63515]: DEBUG oslo_concurrency.lockutils [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.596s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 710.506700] env[63515]: DEBUG nova.compute.manager [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] [instance: 452f3ca2-6141-43b2-a77a-c9ab5754192d] Start building networks asynchronously for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 710.508886] env[63515]: DEBUG oslo_concurrency.lockutils [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.877s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 710.512249] env[63515]: INFO nova.compute.claims [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] [instance: 37e6c27e-317b-45d2-bd55-2fd78ccf009f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 710.663289] env[63515]: DEBUG oslo_concurrency.lockutils [None req-33f5b097-4c20-4377-b5be-7e558874e5d3 tempest-ServersV294TestFqdnHostnames-2112993702 tempest-ServersV294TestFqdnHostnames-2112993702-project-member] Lock "396e49dd-48c0-496b-a1ec-190c33a22c5e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 116.125s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 710.825806] env[63515]: DEBUG oslo_vmware.api [None req-4fa686a6-dc45-4603-a4ee-7107c810126f tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] Task: {'id': task-1110964, 'name': Rename_Task, 'duration_secs': 0.316547} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 710.834210] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-4fa686a6-dc45-4603-a4ee-7107c810126f tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] [instance: 4087b2c4-6ed1-4b68-8b78-a36e34d935b1] Powering on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 710.834664] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1a0307cd-ffd1-4d25-9dad-ec7874da2f0d {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.844232] env[63515]: DEBUG oslo_vmware.api [None req-44a3080f-f2e2-4f0c-b244-122beee2aea9 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Task: {'id': task-1110963, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.586976} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 710.845486] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-44a3080f-f2e2-4f0c-b244-122beee2aea9 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk to [datastore1] 87c468d9-9594-4804-b461-527f01f6118f/87c468d9-9594-4804-b461-527f01f6118f.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 710.846145] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-44a3080f-f2e2-4f0c-b244-122beee2aea9 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] [instance: 87c468d9-9594-4804-b461-527f01f6118f] Extending root virtual disk to 1048576 {{(pid=63515) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 710.846145] env[63515]: DEBUG oslo_vmware.api [None req-4fa686a6-dc45-4603-a4ee-7107c810126f tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] Waiting for the task: (returnval){ [ 710.846145] env[63515]: value = "task-1110965" [ 710.846145] env[63515]: _type = "Task" [ 710.846145] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 710.846407] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d708f346-6947-44ea-b185-266843ddb321 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.858835] env[63515]: DEBUG oslo_vmware.api [None req-4fa686a6-dc45-4603-a4ee-7107c810126f tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] Task: {'id': task-1110965, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 710.860225] env[63515]: DEBUG oslo_vmware.api [None req-44a3080f-f2e2-4f0c-b244-122beee2aea9 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Waiting for the task: (returnval){ [ 710.860225] env[63515]: value = "task-1110966" [ 710.860225] env[63515]: _type = "Task" [ 710.860225] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 710.972427] env[63515]: DEBUG oslo_concurrency.lockutils [req-47f4d0c6-4515-4a08-9241-f900ccf0de3f req-a7eecd75-c61a-42ee-ad17-cf663bf943f8 service nova] Releasing lock "refresh_cache-90f4930b-aaa0-4c4b-9ab8-92aed45e200b" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 711.014898] env[63515]: DEBUG nova.compute.utils [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Using /dev/sd instead of None {{(pid=63515) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 711.019970] env[63515]: DEBUG nova.compute.manager [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] [instance: 452f3ca2-6141-43b2-a77a-c9ab5754192d] Allocating IP information in the background. {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 711.019970] env[63515]: DEBUG nova.network.neutron [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] [instance: 452f3ca2-6141-43b2-a77a-c9ab5754192d] allocate_for_instance() {{(pid=63515) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 711.146939] env[63515]: DEBUG nova.policy [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '479be60850604cf19669871a78f4b710', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3374114af6314f86be3559223d70dfdb', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63515) authorize /opt/stack/nova/nova/policy.py:201}} [ 711.171019] env[63515]: DEBUG nova.compute.manager [None req-4bc5b4e9-701c-413d-abd5-d300a7447cb8 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: ed550b10-d58f-45b8-b766-198f431c3788] Starting instance... {{(pid=63515) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 711.360926] env[63515]: DEBUG oslo_vmware.api [None req-4fa686a6-dc45-4603-a4ee-7107c810126f tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] Task: {'id': task-1110965, 'name': PowerOnVM_Task} progress is 37%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 711.371757] env[63515]: DEBUG oslo_vmware.api [None req-44a3080f-f2e2-4f0c-b244-122beee2aea9 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Task: {'id': task-1110966, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.090241} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 711.372590] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-44a3080f-f2e2-4f0c-b244-122beee2aea9 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] [instance: 87c468d9-9594-4804-b461-527f01f6118f] Extended root virtual disk {{(pid=63515) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 711.373695] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57976d16-1b54-4fae-a052-a3e7bb27fc31 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.405430] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-44a3080f-f2e2-4f0c-b244-122beee2aea9 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] [instance: 87c468d9-9594-4804-b461-527f01f6118f] Reconfiguring VM instance instance-00000026 to attach disk [datastore1] 87c468d9-9594-4804-b461-527f01f6118f/87c468d9-9594-4804-b461-527f01f6118f.vmdk or device None with type sparse {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 711.408448] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e8d4a136-dcb0-4cce-a2bc-6763a16431e7 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.429208] env[63515]: DEBUG nova.compute.manager [None req-99f0df88-8ade-4c67-afd9-dbdc6b385015 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] [instance: f1d01b75-ac9d-458d-8cc2-ae64cffca4e8] Stashing vm_state: active {{(pid=63515) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 711.439019] env[63515]: DEBUG oslo_vmware.api [None req-44a3080f-f2e2-4f0c-b244-122beee2aea9 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Waiting for the task: (returnval){ [ 711.439019] env[63515]: value = "task-1110967" [ 711.439019] env[63515]: _type = "Task" [ 711.439019] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 711.446647] env[63515]: DEBUG oslo_vmware.api [None req-44a3080f-f2e2-4f0c-b244-122beee2aea9 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Task: {'id': task-1110967, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 711.467771] env[63515]: DEBUG nova.compute.manager [req-124fd163-00d8-43a1-8d29-98e59d8bd2f1 req-7a9d2a81-0bfb-44c9-bedf-36f799e69b53 service nova] [instance: 396e49dd-48c0-496b-a1ec-190c33a22c5e] Received event network-changed-77fbf94a-db5a-4569-b4e7-2d9edbe411a1 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 711.467965] env[63515]: DEBUG nova.compute.manager [req-124fd163-00d8-43a1-8d29-98e59d8bd2f1 req-7a9d2a81-0bfb-44c9-bedf-36f799e69b53 service nova] [instance: 396e49dd-48c0-496b-a1ec-190c33a22c5e] Refreshing instance network info cache due to event network-changed-77fbf94a-db5a-4569-b4e7-2d9edbe411a1. {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 711.469685] env[63515]: DEBUG oslo_concurrency.lockutils [req-124fd163-00d8-43a1-8d29-98e59d8bd2f1 req-7a9d2a81-0bfb-44c9-bedf-36f799e69b53 service nova] Acquiring lock "refresh_cache-396e49dd-48c0-496b-a1ec-190c33a22c5e" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 711.469685] env[63515]: DEBUG oslo_concurrency.lockutils [req-124fd163-00d8-43a1-8d29-98e59d8bd2f1 req-7a9d2a81-0bfb-44c9-bedf-36f799e69b53 service nova] Acquired lock "refresh_cache-396e49dd-48c0-496b-a1ec-190c33a22c5e" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 711.469685] env[63515]: DEBUG nova.network.neutron [req-124fd163-00d8-43a1-8d29-98e59d8bd2f1 req-7a9d2a81-0bfb-44c9-bedf-36f799e69b53 service nova] [instance: 396e49dd-48c0-496b-a1ec-190c33a22c5e] Refreshing network info cache for port 77fbf94a-db5a-4569-b4e7-2d9edbe411a1 {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 711.489085] env[63515]: DEBUG nova.compute.manager [req-ee1870c6-aba9-42fd-a6cd-fda41acc6721 req-051efa79-68da-4a1b-9c77-21d9b5dc0bf4 service nova] [instance: 48668736-df27-4f2a-94d9-132f5b49701b] Received event network-vif-plugged-9379e602-f7be-4c75-a6e6-c42a8702f6da {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 711.489085] env[63515]: DEBUG oslo_concurrency.lockutils [req-ee1870c6-aba9-42fd-a6cd-fda41acc6721 req-051efa79-68da-4a1b-9c77-21d9b5dc0bf4 service nova] Acquiring lock "48668736-df27-4f2a-94d9-132f5b49701b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 711.489085] env[63515]: DEBUG oslo_concurrency.lockutils [req-ee1870c6-aba9-42fd-a6cd-fda41acc6721 req-051efa79-68da-4a1b-9c77-21d9b5dc0bf4 service nova] Lock "48668736-df27-4f2a-94d9-132f5b49701b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 711.489085] env[63515]: DEBUG oslo_concurrency.lockutils [req-ee1870c6-aba9-42fd-a6cd-fda41acc6721 req-051efa79-68da-4a1b-9c77-21d9b5dc0bf4 service nova] Lock "48668736-df27-4f2a-94d9-132f5b49701b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 711.489085] env[63515]: DEBUG nova.compute.manager [req-ee1870c6-aba9-42fd-a6cd-fda41acc6721 req-051efa79-68da-4a1b-9c77-21d9b5dc0bf4 service nova] [instance: 48668736-df27-4f2a-94d9-132f5b49701b] No waiting events found dispatching network-vif-plugged-9379e602-f7be-4c75-a6e6-c42a8702f6da {{(pid=63515) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 711.489317] env[63515]: WARNING nova.compute.manager [req-ee1870c6-aba9-42fd-a6cd-fda41acc6721 req-051efa79-68da-4a1b-9c77-21d9b5dc0bf4 service nova] [instance: 48668736-df27-4f2a-94d9-132f5b49701b] Received unexpected event network-vif-plugged-9379e602-f7be-4c75-a6e6-c42a8702f6da for instance with vm_state building and task_state spawning. [ 711.520322] env[63515]: DEBUG nova.compute.manager [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] [instance: 452f3ca2-6141-43b2-a77a-c9ab5754192d] Start building block device mappings for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 711.700684] env[63515]: DEBUG oslo_concurrency.lockutils [None req-4bc5b4e9-701c-413d-abd5-d300a7447cb8 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 711.753275] env[63515]: DEBUG nova.network.neutron [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] [instance: 48668736-df27-4f2a-94d9-132f5b49701b] Successfully updated port: 9379e602-f7be-4c75-a6e6-c42a8702f6da {{(pid=63515) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 711.858968] env[63515]: DEBUG oslo_vmware.api [None req-4fa686a6-dc45-4603-a4ee-7107c810126f tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] Task: {'id': task-1110965, 'name': PowerOnVM_Task} progress is 91%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 711.893455] env[63515]: DEBUG nova.network.neutron [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] [instance: 452f3ca2-6141-43b2-a77a-c9ab5754192d] Successfully created port: 26b2403b-166a-432e-9692-27511a1a4ae6 {{(pid=63515) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 711.947470] env[63515]: DEBUG oslo_vmware.api [None req-44a3080f-f2e2-4f0c-b244-122beee2aea9 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Task: {'id': task-1110967, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 711.953236] env[63515]: DEBUG oslo_concurrency.lockutils [None req-99f0df88-8ade-4c67-afd9-dbdc6b385015 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 712.075365] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f290e2b-d757-4f45-86c2-ec431bf5b125 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.087451] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1309c7b7-b540-4f57-9c2c-cdb4cdd0f86b {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.122119] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c044b3f-26c2-4d4a-9c00-cd57936ca659 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.130286] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-450ed569-dbb6-48e6-9363-adc42a94bec3 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.148278] env[63515]: DEBUG nova.compute.provider_tree [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 712.258586] env[63515]: DEBUG oslo_concurrency.lockutils [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Acquiring lock "refresh_cache-48668736-df27-4f2a-94d9-132f5b49701b" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 712.258586] env[63515]: DEBUG oslo_concurrency.lockutils [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Acquired lock "refresh_cache-48668736-df27-4f2a-94d9-132f5b49701b" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 712.258586] env[63515]: DEBUG nova.network.neutron [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] [instance: 48668736-df27-4f2a-94d9-132f5b49701b] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 712.362397] env[63515]: DEBUG oslo_vmware.api [None req-4fa686a6-dc45-4603-a4ee-7107c810126f tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] Task: {'id': task-1110965, 'name': PowerOnVM_Task, 'duration_secs': 1.032331} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 712.362725] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-4fa686a6-dc45-4603-a4ee-7107c810126f tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] [instance: 4087b2c4-6ed1-4b68-8b78-a36e34d935b1] Powered on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 712.362881] env[63515]: INFO nova.compute.manager [None req-4fa686a6-dc45-4603-a4ee-7107c810126f tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] [instance: 4087b2c4-6ed1-4b68-8b78-a36e34d935b1] Took 9.66 seconds to spawn the instance on the hypervisor. [ 712.363065] env[63515]: DEBUG nova.compute.manager [None req-4fa686a6-dc45-4603-a4ee-7107c810126f tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] [instance: 4087b2c4-6ed1-4b68-8b78-a36e34d935b1] Checking state {{(pid=63515) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 712.363841] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04d32ce7-3a5c-4e76-ba82-9e07ff5848ee {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.450984] env[63515]: DEBUG oslo_vmware.api [None req-44a3080f-f2e2-4f0c-b244-122beee2aea9 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Task: {'id': task-1110967, 'name': ReconfigVM_Task, 'duration_secs': 0.769514} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 712.451302] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-44a3080f-f2e2-4f0c-b244-122beee2aea9 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] [instance: 87c468d9-9594-4804-b461-527f01f6118f] Reconfigured VM instance instance-00000026 to attach disk [datastore1] 87c468d9-9594-4804-b461-527f01f6118f/87c468d9-9594-4804-b461-527f01f6118f.vmdk or device None with type sparse {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 712.453217] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-bfb85f66-15e0-4455-80e0-1725000f836f {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.461198] env[63515]: DEBUG oslo_vmware.api [None req-44a3080f-f2e2-4f0c-b244-122beee2aea9 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Waiting for the task: (returnval){ [ 712.461198] env[63515]: value = "task-1110968" [ 712.461198] env[63515]: _type = "Task" [ 712.461198] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 712.467660] env[63515]: DEBUG oslo_vmware.api [None req-44a3080f-f2e2-4f0c-b244-122beee2aea9 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Task: {'id': task-1110968, 'name': Rename_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 712.520841] env[63515]: DEBUG nova.network.neutron [req-124fd163-00d8-43a1-8d29-98e59d8bd2f1 req-7a9d2a81-0bfb-44c9-bedf-36f799e69b53 service nova] [instance: 396e49dd-48c0-496b-a1ec-190c33a22c5e] Updated VIF entry in instance network info cache for port 77fbf94a-db5a-4569-b4e7-2d9edbe411a1. {{(pid=63515) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 712.521241] env[63515]: DEBUG nova.network.neutron [req-124fd163-00d8-43a1-8d29-98e59d8bd2f1 req-7a9d2a81-0bfb-44c9-bedf-36f799e69b53 service nova] [instance: 396e49dd-48c0-496b-a1ec-190c33a22c5e] Updating instance_info_cache with network_info: [{"id": "77fbf94a-db5a-4569-b4e7-2d9edbe411a1", "address": "fa:16:3e:36:07:33", "network": {"id": "1c834c8e-4d0a-4cd9-b2ad-c24a22ed72ff", "bridge": "br-int", "label": "tempest-ServersV294TestFqdnHostnames-1956290090-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.140", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "62ac151dc34e41e7bf134563228d9d63", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4adc8ed0-d11a-4510-9be0-b27c0da3a903", "external-id": "nsx-vlan-transportzone-844", "segmentation_id": 844, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap77fbf94a-db", "ovs_interfaceid": "77fbf94a-db5a-4569-b4e7-2d9edbe411a1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 712.536379] env[63515]: DEBUG nova.compute.manager [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] [instance: 452f3ca2-6141-43b2-a77a-c9ab5754192d] Start spawning the instance on the hypervisor. {{(pid=63515) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 712.558852] env[63515]: DEBUG nova.virt.hardware [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T02:52:14Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T02:51:56Z,direct_url=,disk_format='vmdk',id=8a120570-cb06-4099-b262-554ca0ad15c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b569255cc43e42e7ae3f2b4ad37c6ef4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T02:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 712.559057] env[63515]: DEBUG nova.virt.hardware [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Flavor limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 712.559221] env[63515]: DEBUG nova.virt.hardware [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Image limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 712.559397] env[63515]: DEBUG nova.virt.hardware [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Flavor pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 712.559540] env[63515]: DEBUG nova.virt.hardware [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Image pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 712.559684] env[63515]: DEBUG nova.virt.hardware [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 712.559882] env[63515]: DEBUG nova.virt.hardware [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 712.560047] env[63515]: DEBUG nova.virt.hardware [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 712.560218] env[63515]: DEBUG nova.virt.hardware [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Got 1 possible topologies {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 712.560384] env[63515]: DEBUG nova.virt.hardware [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 712.560573] env[63515]: DEBUG nova.virt.hardware [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 712.561807] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48a3a8b2-268c-4b0d-8489-b21721fa3005 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.570158] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4203cfcf-078c-486e-9b37-def4ef6e45ca {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.652416] env[63515]: DEBUG nova.scheduler.client.report [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 712.796486] env[63515]: DEBUG nova.network.neutron [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] [instance: 48668736-df27-4f2a-94d9-132f5b49701b] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 712.884465] env[63515]: INFO nova.compute.manager [None req-4fa686a6-dc45-4603-a4ee-7107c810126f tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] [instance: 4087b2c4-6ed1-4b68-8b78-a36e34d935b1] Took 28.55 seconds to build instance. [ 712.969910] env[63515]: DEBUG oslo_vmware.api [None req-44a3080f-f2e2-4f0c-b244-122beee2aea9 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Task: {'id': task-1110968, 'name': Rename_Task, 'duration_secs': 0.384645} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 712.970217] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-44a3080f-f2e2-4f0c-b244-122beee2aea9 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] [instance: 87c468d9-9594-4804-b461-527f01f6118f] Powering on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 712.970464] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-fb0892a4-b06e-431a-a9d3-a211b44f37d3 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.978698] env[63515]: DEBUG oslo_vmware.api [None req-44a3080f-f2e2-4f0c-b244-122beee2aea9 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Waiting for the task: (returnval){ [ 712.978698] env[63515]: value = "task-1110969" [ 712.978698] env[63515]: _type = "Task" [ 712.978698] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 712.988640] env[63515]: DEBUG oslo_vmware.api [None req-44a3080f-f2e2-4f0c-b244-122beee2aea9 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Task: {'id': task-1110969, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 713.009310] env[63515]: DEBUG nova.network.neutron [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] [instance: 48668736-df27-4f2a-94d9-132f5b49701b] Updating instance_info_cache with network_info: [{"id": "9379e602-f7be-4c75-a6e6-c42a8702f6da", "address": "fa:16:3e:f6:8f:c5", "network": {"id": "a2fdda6f-f99f-420f-b3c7-61c9ebcbeaa8", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-1981716599-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3374114af6314f86be3559223d70dfdb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "13d625c9-77ec-4edb-a56b-9f37a314cc39", "external-id": "nsx-vlan-transportzone-358", "segmentation_id": 358, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9379e602-f7", "ovs_interfaceid": "9379e602-f7be-4c75-a6e6-c42a8702f6da", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 713.024729] env[63515]: DEBUG oslo_concurrency.lockutils [req-124fd163-00d8-43a1-8d29-98e59d8bd2f1 req-7a9d2a81-0bfb-44c9-bedf-36f799e69b53 service nova] Releasing lock "refresh_cache-396e49dd-48c0-496b-a1ec-190c33a22c5e" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 713.155954] env[63515]: DEBUG oslo_concurrency.lockutils [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.647s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 713.156475] env[63515]: DEBUG nova.compute.manager [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] [instance: 37e6c27e-317b-45d2-bd55-2fd78ccf009f] Start building networks asynchronously for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 713.159081] env[63515]: DEBUG oslo_concurrency.lockutils [None req-746f5a16-d8b1-448e-a35c-34720548f33b tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.418s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 713.160933] env[63515]: INFO nova.compute.claims [None req-746f5a16-d8b1-448e-a35c-34720548f33b tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] [instance: b4477e66-ae12-4929-90ed-b7b652e0f207] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 713.384939] env[63515]: DEBUG oslo_concurrency.lockutils [None req-4fa686a6-dc45-4603-a4ee-7107c810126f tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] Lock "4087b2c4-6ed1-4b68-8b78-a36e34d935b1" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 118.516s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 713.447158] env[63515]: DEBUG oslo_concurrency.lockutils [None req-66c3e202-5a3e-433e-b543-4bfb02c8ad30 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775-project-member] Acquiring lock "4e453127-1f3e-40ea-819f-6678479826c8" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 713.447401] env[63515]: DEBUG oslo_concurrency.lockutils [None req-66c3e202-5a3e-433e-b543-4bfb02c8ad30 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775-project-member] Lock "4e453127-1f3e-40ea-819f-6678479826c8" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 713.491185] env[63515]: DEBUG oslo_vmware.api [None req-44a3080f-f2e2-4f0c-b244-122beee2aea9 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Task: {'id': task-1110969, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 713.515020] env[63515]: DEBUG oslo_concurrency.lockutils [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Releasing lock "refresh_cache-48668736-df27-4f2a-94d9-132f5b49701b" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 713.515020] env[63515]: DEBUG nova.compute.manager [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] [instance: 48668736-df27-4f2a-94d9-132f5b49701b] Instance network_info: |[{"id": "9379e602-f7be-4c75-a6e6-c42a8702f6da", "address": "fa:16:3e:f6:8f:c5", "network": {"id": "a2fdda6f-f99f-420f-b3c7-61c9ebcbeaa8", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-1981716599-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3374114af6314f86be3559223d70dfdb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "13d625c9-77ec-4edb-a56b-9f37a314cc39", "external-id": "nsx-vlan-transportzone-358", "segmentation_id": 358, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9379e602-f7", "ovs_interfaceid": "9379e602-f7be-4c75-a6e6-c42a8702f6da", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 713.515297] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] [instance: 48668736-df27-4f2a-94d9-132f5b49701b] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:f6:8f:c5', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '13d625c9-77ec-4edb-a56b-9f37a314cc39', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '9379e602-f7be-4c75-a6e6-c42a8702f6da', 'vif_model': 'vmxnet3'}] {{(pid=63515) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 713.522769] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Creating folder: Project (3374114af6314f86be3559223d70dfdb). Parent ref: group-v243370. {{(pid=63515) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 713.524314] env[63515]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d93bc929-8a7e-4f73-b52d-74ef7fbb8efd {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.528148] env[63515]: DEBUG nova.compute.manager [req-947ce066-8113-4264-9828-e1bf93fc826c req-0def0ef6-eccf-415f-a1bd-d537faa8f79e service nova] [instance: 48668736-df27-4f2a-94d9-132f5b49701b] Received event network-changed-9379e602-f7be-4c75-a6e6-c42a8702f6da {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 713.528148] env[63515]: DEBUG nova.compute.manager [req-947ce066-8113-4264-9828-e1bf93fc826c req-0def0ef6-eccf-415f-a1bd-d537faa8f79e service nova] [instance: 48668736-df27-4f2a-94d9-132f5b49701b] Refreshing instance network info cache due to event network-changed-9379e602-f7be-4c75-a6e6-c42a8702f6da. {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 713.528493] env[63515]: DEBUG oslo_concurrency.lockutils [req-947ce066-8113-4264-9828-e1bf93fc826c req-0def0ef6-eccf-415f-a1bd-d537faa8f79e service nova] Acquiring lock "refresh_cache-48668736-df27-4f2a-94d9-132f5b49701b" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 713.528757] env[63515]: DEBUG oslo_concurrency.lockutils [req-947ce066-8113-4264-9828-e1bf93fc826c req-0def0ef6-eccf-415f-a1bd-d537faa8f79e service nova] Acquired lock "refresh_cache-48668736-df27-4f2a-94d9-132f5b49701b" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 713.529037] env[63515]: DEBUG nova.network.neutron [req-947ce066-8113-4264-9828-e1bf93fc826c req-0def0ef6-eccf-415f-a1bd-d537faa8f79e service nova] [instance: 48668736-df27-4f2a-94d9-132f5b49701b] Refreshing network info cache for port 9379e602-f7be-4c75-a6e6-c42a8702f6da {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 713.539542] env[63515]: INFO nova.virt.vmwareapi.vm_util [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Created folder: Project (3374114af6314f86be3559223d70dfdb) in parent group-v243370. [ 713.542021] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Creating folder: Instances. Parent ref: group-v243406. {{(pid=63515) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 713.542021] env[63515]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-45b8e052-3a99-480d-9fc8-331e0587a3da {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.552751] env[63515]: INFO nova.virt.vmwareapi.vm_util [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Created folder: Instances in parent group-v243406. [ 713.552751] env[63515]: DEBUG oslo.service.loopingcall [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 713.552751] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 48668736-df27-4f2a-94d9-132f5b49701b] Creating VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 713.552751] env[63515]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-db4f2ad6-94b7-42a7-ac6b-bb999d954383 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.572071] env[63515]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 713.572071] env[63515]: value = "task-1110972" [ 713.572071] env[63515]: _type = "Task" [ 713.572071] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 713.580324] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1110972, 'name': CreateVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 713.665932] env[63515]: DEBUG nova.compute.utils [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Using /dev/sd instead of None {{(pid=63515) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 713.670306] env[63515]: DEBUG nova.compute.manager [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] [instance: 37e6c27e-317b-45d2-bd55-2fd78ccf009f] Allocating IP information in the background. {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 713.670513] env[63515]: DEBUG nova.network.neutron [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] [instance: 37e6c27e-317b-45d2-bd55-2fd78ccf009f] allocate_for_instance() {{(pid=63515) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 713.736382] env[63515]: DEBUG nova.policy [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '479be60850604cf19669871a78f4b710', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3374114af6314f86be3559223d70dfdb', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63515) authorize /opt/stack/nova/nova/policy.py:201}} [ 713.891999] env[63515]: DEBUG nova.compute.manager [None req-3b658824-6e25-4b06-878c-99b3be1f1d31 tempest-ServerTagsTestJSON-1170819896 tempest-ServerTagsTestJSON-1170819896-project-member] [instance: de210780-5c0f-4fba-883c-13707566a2e1] Starting instance... {{(pid=63515) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 713.992980] env[63515]: DEBUG oslo_vmware.api [None req-44a3080f-f2e2-4f0c-b244-122beee2aea9 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Task: {'id': task-1110969, 'name': PowerOnVM_Task, 'duration_secs': 0.843923} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 713.993296] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-44a3080f-f2e2-4f0c-b244-122beee2aea9 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] [instance: 87c468d9-9594-4804-b461-527f01f6118f] Powered on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 713.993503] env[63515]: INFO nova.compute.manager [None req-44a3080f-f2e2-4f0c-b244-122beee2aea9 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] [instance: 87c468d9-9594-4804-b461-527f01f6118f] Took 8.66 seconds to spawn the instance on the hypervisor. [ 713.993684] env[63515]: DEBUG nova.compute.manager [None req-44a3080f-f2e2-4f0c-b244-122beee2aea9 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] [instance: 87c468d9-9594-4804-b461-527f01f6118f] Checking state {{(pid=63515) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 713.994497] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-752b1b82-8bd7-423e-ac8e-df32b5fe8fda {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.083932] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1110972, 'name': CreateVM_Task, 'duration_secs': 0.452686} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 714.084383] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 48668736-df27-4f2a-94d9-132f5b49701b] Created VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 714.085252] env[63515]: DEBUG oslo_concurrency.lockutils [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 714.085671] env[63515]: DEBUG oslo_concurrency.lockutils [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 714.086176] env[63515]: DEBUG oslo_concurrency.lockutils [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 714.086638] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8ea1823d-a3b5-41bb-9a00-b3e6c69c719c {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.091653] env[63515]: DEBUG oslo_vmware.api [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Waiting for the task: (returnval){ [ 714.091653] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]524e4052-e36b-5a77-f8a1-1886a6cade57" [ 714.091653] env[63515]: _type = "Task" [ 714.091653] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 714.100299] env[63515]: DEBUG oslo_vmware.api [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]524e4052-e36b-5a77-f8a1-1886a6cade57, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 714.150674] env[63515]: DEBUG nova.network.neutron [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] [instance: 452f3ca2-6141-43b2-a77a-c9ab5754192d] Successfully updated port: 26b2403b-166a-432e-9692-27511a1a4ae6 {{(pid=63515) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 714.173428] env[63515]: DEBUG nova.compute.manager [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] [instance: 37e6c27e-317b-45d2-bd55-2fd78ccf009f] Start building block device mappings for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 714.325852] env[63515]: DEBUG nova.network.neutron [req-947ce066-8113-4264-9828-e1bf93fc826c req-0def0ef6-eccf-415f-a1bd-d537faa8f79e service nova] [instance: 48668736-df27-4f2a-94d9-132f5b49701b] Updated VIF entry in instance network info cache for port 9379e602-f7be-4c75-a6e6-c42a8702f6da. {{(pid=63515) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 714.326175] env[63515]: DEBUG nova.network.neutron [req-947ce066-8113-4264-9828-e1bf93fc826c req-0def0ef6-eccf-415f-a1bd-d537faa8f79e service nova] [instance: 48668736-df27-4f2a-94d9-132f5b49701b] Updating instance_info_cache with network_info: [{"id": "9379e602-f7be-4c75-a6e6-c42a8702f6da", "address": "fa:16:3e:f6:8f:c5", "network": {"id": "a2fdda6f-f99f-420f-b3c7-61c9ebcbeaa8", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-1981716599-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3374114af6314f86be3559223d70dfdb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "13d625c9-77ec-4edb-a56b-9f37a314cc39", "external-id": "nsx-vlan-transportzone-358", "segmentation_id": 358, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9379e602-f7", "ovs_interfaceid": "9379e602-f7be-4c75-a6e6-c42a8702f6da", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 714.373237] env[63515]: DEBUG nova.network.neutron [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] [instance: 37e6c27e-317b-45d2-bd55-2fd78ccf009f] Successfully created port: 21c8f7a2-4403-41cd-93f8-3091de7cdc40 {{(pid=63515) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 714.418711] env[63515]: DEBUG oslo_concurrency.lockutils [None req-3b658824-6e25-4b06-878c-99b3be1f1d31 tempest-ServerTagsTestJSON-1170819896 tempest-ServerTagsTestJSON-1170819896-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 714.440285] env[63515]: DEBUG oslo_service.periodic_task [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=63515) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 714.440530] env[63515]: DEBUG oslo_service.periodic_task [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=63515) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 714.440677] env[63515]: DEBUG nova.compute.manager [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Starting heal instance info cache {{(pid=63515) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9926}} [ 714.440793] env[63515]: DEBUG nova.compute.manager [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Rebuilding the list of instances to heal {{(pid=63515) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9930}} [ 714.518427] env[63515]: INFO nova.compute.manager [None req-44a3080f-f2e2-4f0c-b244-122beee2aea9 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] [instance: 87c468d9-9594-4804-b461-527f01f6118f] Took 28.18 seconds to build instance. [ 714.603101] env[63515]: DEBUG oslo_vmware.api [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]524e4052-e36b-5a77-f8a1-1886a6cade57, 'name': SearchDatastore_Task, 'duration_secs': 0.012932} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 714.606025] env[63515]: DEBUG oslo_concurrency.lockutils [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 714.606280] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] [instance: 48668736-df27-4f2a-94d9-132f5b49701b] Processing image 8a120570-cb06-4099-b262-554ca0ad15c5 {{(pid=63515) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 714.606515] env[63515]: DEBUG oslo_concurrency.lockutils [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 714.606663] env[63515]: DEBUG oslo_concurrency.lockutils [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 714.607404] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 714.607548] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-199de4fe-1d0e-4419-93d9-abc3c4d9120c {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.616029] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 714.616029] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63515) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 714.616849] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b76b7340-35ed-4976-bf19-ea815edb5396 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.624877] env[63515]: DEBUG oslo_vmware.api [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Waiting for the task: (returnval){ [ 714.624877] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]52f091c7-7921-3c2c-5b51-47fb34946c29" [ 714.624877] env[63515]: _type = "Task" [ 714.624877] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 714.632693] env[63515]: DEBUG oslo_vmware.api [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52f091c7-7921-3c2c-5b51-47fb34946c29, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 714.653478] env[63515]: DEBUG oslo_concurrency.lockutils [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Acquiring lock "refresh_cache-452f3ca2-6141-43b2-a77a-c9ab5754192d" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 714.653605] env[63515]: DEBUG oslo_concurrency.lockutils [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Acquired lock "refresh_cache-452f3ca2-6141-43b2-a77a-c9ab5754192d" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 714.653770] env[63515]: DEBUG nova.network.neutron [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] [instance: 452f3ca2-6141-43b2-a77a-c9ab5754192d] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 714.712113] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f4a8d8b-51b9-48cc-91d5-3bdb31d51915 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.719744] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7128a27e-ec81-40b1-a637-a51f835cf133 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.752124] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf0cee6d-f0f9-406d-9329-a69cac18d172 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.763493] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26e5f649-e46d-4677-b822-cf5d6fb87f7c {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.779077] env[63515]: DEBUG nova.compute.provider_tree [None req-746f5a16-d8b1-448e-a35c-34720548f33b tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 714.829081] env[63515]: DEBUG oslo_concurrency.lockutils [req-947ce066-8113-4264-9828-e1bf93fc826c req-0def0ef6-eccf-415f-a1bd-d537faa8f79e service nova] Releasing lock "refresh_cache-48668736-df27-4f2a-94d9-132f5b49701b" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 714.948019] env[63515]: DEBUG nova.compute.manager [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] [instance: 48668736-df27-4f2a-94d9-132f5b49701b] Skipping network cache update for instance because it is Building. {{(pid=63515) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 714.948277] env[63515]: DEBUG nova.compute.manager [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] [instance: 452f3ca2-6141-43b2-a77a-c9ab5754192d] Skipping network cache update for instance because it is Building. {{(pid=63515) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 714.948277] env[63515]: DEBUG nova.compute.manager [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] [instance: 37e6c27e-317b-45d2-bd55-2fd78ccf009f] Skipping network cache update for instance because it is Building. {{(pid=63515) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 714.948449] env[63515]: DEBUG nova.compute.manager [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] [instance: b4477e66-ae12-4929-90ed-b7b652e0f207] Skipping network cache update for instance because it is Building. {{(pid=63515) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 714.978208] env[63515]: DEBUG oslo_concurrency.lockutils [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Acquiring lock "refresh_cache-b911a5b5-9617-4fb3-9b5e-fb8c492e4931" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 714.978362] env[63515]: DEBUG oslo_concurrency.lockutils [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Acquired lock "refresh_cache-b911a5b5-9617-4fb3-9b5e-fb8c492e4931" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 714.978509] env[63515]: DEBUG nova.network.neutron [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] [instance: b911a5b5-9617-4fb3-9b5e-fb8c492e4931] Forcefully refreshing network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2004}} [ 714.978668] env[63515]: DEBUG nova.objects.instance [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Lazy-loading 'info_cache' on Instance uuid b911a5b5-9617-4fb3-9b5e-fb8c492e4931 {{(pid=63515) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 715.020863] env[63515]: DEBUG oslo_concurrency.lockutils [None req-44a3080f-f2e2-4f0c-b244-122beee2aea9 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Lock "87c468d9-9594-4804-b461-527f01f6118f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 119.620s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 715.135091] env[63515]: DEBUG oslo_vmware.api [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52f091c7-7921-3c2c-5b51-47fb34946c29, 'name': SearchDatastore_Task, 'duration_secs': 0.009446} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 715.135882] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-102b6b9c-635e-4ec5-ac3d-ba4c6402c46b {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.140950] env[63515]: DEBUG oslo_vmware.api [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Waiting for the task: (returnval){ [ 715.140950] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]5245ea47-d578-c012-0ba3-3d6424cb578b" [ 715.140950] env[63515]: _type = "Task" [ 715.140950] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 715.149257] env[63515]: DEBUG oslo_vmware.api [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]5245ea47-d578-c012-0ba3-3d6424cb578b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 715.187163] env[63515]: DEBUG nova.compute.manager [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] [instance: 37e6c27e-317b-45d2-bd55-2fd78ccf009f] Start spawning the instance on the hypervisor. {{(pid=63515) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 715.213653] env[63515]: DEBUG nova.virt.hardware [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T02:52:14Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T02:51:56Z,direct_url=,disk_format='vmdk',id=8a120570-cb06-4099-b262-554ca0ad15c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b569255cc43e42e7ae3f2b4ad37c6ef4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T02:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 715.213896] env[63515]: DEBUG nova.virt.hardware [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Flavor limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 715.214061] env[63515]: DEBUG nova.virt.hardware [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Image limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 715.214244] env[63515]: DEBUG nova.virt.hardware [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Flavor pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 715.214434] env[63515]: DEBUG nova.virt.hardware [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Image pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 715.214565] env[63515]: DEBUG nova.virt.hardware [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 715.214862] env[63515]: DEBUG nova.virt.hardware [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 715.215053] env[63515]: DEBUG nova.virt.hardware [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 715.215227] env[63515]: DEBUG nova.virt.hardware [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Got 1 possible topologies {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 715.215501] env[63515]: DEBUG nova.virt.hardware [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 715.215564] env[63515]: DEBUG nova.virt.hardware [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 715.216488] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52283b75-3438-4c27-a075-efc7ae5aaab6 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.219596] env[63515]: DEBUG nova.network.neutron [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] [instance: 452f3ca2-6141-43b2-a77a-c9ab5754192d] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 715.226871] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-363ed9a7-0b1b-4749-8d43-325a399b28b2 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.283452] env[63515]: DEBUG nova.scheduler.client.report [None req-746f5a16-d8b1-448e-a35c-34720548f33b tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 715.431967] env[63515]: DEBUG nova.network.neutron [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] [instance: 452f3ca2-6141-43b2-a77a-c9ab5754192d] Updating instance_info_cache with network_info: [{"id": "26b2403b-166a-432e-9692-27511a1a4ae6", "address": "fa:16:3e:50:05:b2", "network": {"id": "a2fdda6f-f99f-420f-b3c7-61c9ebcbeaa8", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-1981716599-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3374114af6314f86be3559223d70dfdb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "13d625c9-77ec-4edb-a56b-9f37a314cc39", "external-id": "nsx-vlan-transportzone-358", "segmentation_id": 358, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap26b2403b-16", "ovs_interfaceid": "26b2403b-166a-432e-9692-27511a1a4ae6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 715.523180] env[63515]: DEBUG nova.compute.manager [None req-fd4a0cad-edc5-4c70-94f6-406aa5060c1e tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: 001a48e1-b237-41f8-a6f4-2fe98d173512] Starting instance... {{(pid=63515) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 715.654625] env[63515]: DEBUG oslo_vmware.api [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]5245ea47-d578-c012-0ba3-3d6424cb578b, 'name': SearchDatastore_Task, 'duration_secs': 0.012915} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 715.654625] env[63515]: DEBUG oslo_concurrency.lockutils [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 715.654625] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk to [datastore1] 48668736-df27-4f2a-94d9-132f5b49701b/48668736-df27-4f2a-94d9-132f5b49701b.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 715.654625] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b5ecc0c9-54e2-4ea8-ae6e-09ac27802f22 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.659117] env[63515]: DEBUG nova.compute.manager [req-ed83a135-13af-44da-9eae-1fb8831f71ef req-8a46ef23-4e8b-489d-ab0e-86ec285b6f32 service nova] [instance: 452f3ca2-6141-43b2-a77a-c9ab5754192d] Received event network-vif-plugged-26b2403b-166a-432e-9692-27511a1a4ae6 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 715.659452] env[63515]: DEBUG oslo_concurrency.lockutils [req-ed83a135-13af-44da-9eae-1fb8831f71ef req-8a46ef23-4e8b-489d-ab0e-86ec285b6f32 service nova] Acquiring lock "452f3ca2-6141-43b2-a77a-c9ab5754192d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 715.659508] env[63515]: DEBUG oslo_concurrency.lockutils [req-ed83a135-13af-44da-9eae-1fb8831f71ef req-8a46ef23-4e8b-489d-ab0e-86ec285b6f32 service nova] Lock "452f3ca2-6141-43b2-a77a-c9ab5754192d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 715.661024] env[63515]: DEBUG oslo_concurrency.lockutils [req-ed83a135-13af-44da-9eae-1fb8831f71ef req-8a46ef23-4e8b-489d-ab0e-86ec285b6f32 service nova] Lock "452f3ca2-6141-43b2-a77a-c9ab5754192d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 715.661024] env[63515]: DEBUG nova.compute.manager [req-ed83a135-13af-44da-9eae-1fb8831f71ef req-8a46ef23-4e8b-489d-ab0e-86ec285b6f32 service nova] [instance: 452f3ca2-6141-43b2-a77a-c9ab5754192d] No waiting events found dispatching network-vif-plugged-26b2403b-166a-432e-9692-27511a1a4ae6 {{(pid=63515) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 715.661024] env[63515]: WARNING nova.compute.manager [req-ed83a135-13af-44da-9eae-1fb8831f71ef req-8a46ef23-4e8b-489d-ab0e-86ec285b6f32 service nova] [instance: 452f3ca2-6141-43b2-a77a-c9ab5754192d] Received unexpected event network-vif-plugged-26b2403b-166a-432e-9692-27511a1a4ae6 for instance with vm_state building and task_state spawning. [ 715.661024] env[63515]: DEBUG nova.compute.manager [req-ed83a135-13af-44da-9eae-1fb8831f71ef req-8a46ef23-4e8b-489d-ab0e-86ec285b6f32 service nova] [instance: 452f3ca2-6141-43b2-a77a-c9ab5754192d] Received event network-changed-26b2403b-166a-432e-9692-27511a1a4ae6 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 715.661024] env[63515]: DEBUG nova.compute.manager [req-ed83a135-13af-44da-9eae-1fb8831f71ef req-8a46ef23-4e8b-489d-ab0e-86ec285b6f32 service nova] [instance: 452f3ca2-6141-43b2-a77a-c9ab5754192d] Refreshing instance network info cache due to event network-changed-26b2403b-166a-432e-9692-27511a1a4ae6. {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 715.661360] env[63515]: DEBUG oslo_concurrency.lockutils [req-ed83a135-13af-44da-9eae-1fb8831f71ef req-8a46ef23-4e8b-489d-ab0e-86ec285b6f32 service nova] Acquiring lock "refresh_cache-452f3ca2-6141-43b2-a77a-c9ab5754192d" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 715.662259] env[63515]: DEBUG oslo_vmware.api [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Waiting for the task: (returnval){ [ 715.662259] env[63515]: value = "task-1110973" [ 715.662259] env[63515]: _type = "Task" [ 715.662259] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 715.670764] env[63515]: DEBUG oslo_vmware.api [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Task: {'id': task-1110973, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 715.789634] env[63515]: DEBUG oslo_concurrency.lockutils [None req-746f5a16-d8b1-448e-a35c-34720548f33b tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.630s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 715.790441] env[63515]: DEBUG nova.compute.manager [None req-746f5a16-d8b1-448e-a35c-34720548f33b tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] [instance: b4477e66-ae12-4929-90ed-b7b652e0f207] Start building networks asynchronously for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 715.793873] env[63515]: DEBUG oslo_concurrency.lockutils [None req-5690e97d-f63c-4f8e-9c79-d9c523fc3720 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 14.604s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 715.793873] env[63515]: DEBUG nova.objects.instance [None req-5690e97d-f63c-4f8e-9c79-d9c523fc3720 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Lazy-loading 'resources' on Instance uuid 3ad04a9f-62cf-4b1c-aab2-7dc4a219edc6 {{(pid=63515) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 715.935367] env[63515]: DEBUG oslo_concurrency.lockutils [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Releasing lock "refresh_cache-452f3ca2-6141-43b2-a77a-c9ab5754192d" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 715.935741] env[63515]: DEBUG nova.compute.manager [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] [instance: 452f3ca2-6141-43b2-a77a-c9ab5754192d] Instance network_info: |[{"id": "26b2403b-166a-432e-9692-27511a1a4ae6", "address": "fa:16:3e:50:05:b2", "network": {"id": "a2fdda6f-f99f-420f-b3c7-61c9ebcbeaa8", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-1981716599-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3374114af6314f86be3559223d70dfdb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "13d625c9-77ec-4edb-a56b-9f37a314cc39", "external-id": "nsx-vlan-transportzone-358", "segmentation_id": 358, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap26b2403b-16", "ovs_interfaceid": "26b2403b-166a-432e-9692-27511a1a4ae6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 715.936078] env[63515]: DEBUG oslo_concurrency.lockutils [req-ed83a135-13af-44da-9eae-1fb8831f71ef req-8a46ef23-4e8b-489d-ab0e-86ec285b6f32 service nova] Acquired lock "refresh_cache-452f3ca2-6141-43b2-a77a-c9ab5754192d" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 715.936274] env[63515]: DEBUG nova.network.neutron [req-ed83a135-13af-44da-9eae-1fb8831f71ef req-8a46ef23-4e8b-489d-ab0e-86ec285b6f32 service nova] [instance: 452f3ca2-6141-43b2-a77a-c9ab5754192d] Refreshing network info cache for port 26b2403b-166a-432e-9692-27511a1a4ae6 {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 715.937511] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] [instance: 452f3ca2-6141-43b2-a77a-c9ab5754192d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:50:05:b2', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '13d625c9-77ec-4edb-a56b-9f37a314cc39', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '26b2403b-166a-432e-9692-27511a1a4ae6', 'vif_model': 'vmxnet3'}] {{(pid=63515) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 715.952209] env[63515]: DEBUG oslo.service.loopingcall [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 715.955336] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 452f3ca2-6141-43b2-a77a-c9ab5754192d] Creating VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 715.955877] env[63515]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-1b041a47-e8d6-41d2-b49d-44612e1db8da {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.979314] env[63515]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 715.979314] env[63515]: value = "task-1110974" [ 715.979314] env[63515]: _type = "Task" [ 715.979314] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 715.989830] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1110974, 'name': CreateVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 716.063390] env[63515]: DEBUG oslo_concurrency.lockutils [None req-fd4a0cad-edc5-4c70-94f6-406aa5060c1e tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 716.100411] env[63515]: DEBUG nova.network.neutron [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] [instance: 37e6c27e-317b-45d2-bd55-2fd78ccf009f] Successfully updated port: 21c8f7a2-4403-41cd-93f8-3091de7cdc40 {{(pid=63515) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 716.176360] env[63515]: DEBUG oslo_vmware.api [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Task: {'id': task-1110973, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 716.298279] env[63515]: DEBUG nova.compute.utils [None req-746f5a16-d8b1-448e-a35c-34720548f33b tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Using /dev/sd instead of None {{(pid=63515) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 716.305765] env[63515]: DEBUG nova.compute.manager [None req-746f5a16-d8b1-448e-a35c-34720548f33b tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] [instance: b4477e66-ae12-4929-90ed-b7b652e0f207] Not allocating networking since 'none' was specified. {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 716.442333] env[63515]: DEBUG nova.network.neutron [req-ed83a135-13af-44da-9eae-1fb8831f71ef req-8a46ef23-4e8b-489d-ab0e-86ec285b6f32 service nova] [instance: 452f3ca2-6141-43b2-a77a-c9ab5754192d] Updated VIF entry in instance network info cache for port 26b2403b-166a-432e-9692-27511a1a4ae6. {{(pid=63515) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 716.442970] env[63515]: DEBUG nova.network.neutron [req-ed83a135-13af-44da-9eae-1fb8831f71ef req-8a46ef23-4e8b-489d-ab0e-86ec285b6f32 service nova] [instance: 452f3ca2-6141-43b2-a77a-c9ab5754192d] Updating instance_info_cache with network_info: [{"id": "26b2403b-166a-432e-9692-27511a1a4ae6", "address": "fa:16:3e:50:05:b2", "network": {"id": "a2fdda6f-f99f-420f-b3c7-61c9ebcbeaa8", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-1981716599-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3374114af6314f86be3559223d70dfdb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "13d625c9-77ec-4edb-a56b-9f37a314cc39", "external-id": "nsx-vlan-transportzone-358", "segmentation_id": 358, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap26b2403b-16", "ovs_interfaceid": "26b2403b-166a-432e-9692-27511a1a4ae6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 716.486124] env[63515]: DEBUG nova.compute.manager [req-c82cebe1-45cc-4648-be8e-b55d45230645 req-165a31d8-d6a2-47a8-843a-644d3a0952c3 service nova] [instance: 87c468d9-9594-4804-b461-527f01f6118f] Received event network-changed-d2098868-0475-4b06-bae3-785c608a861e {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 716.486323] env[63515]: DEBUG nova.compute.manager [req-c82cebe1-45cc-4648-be8e-b55d45230645 req-165a31d8-d6a2-47a8-843a-644d3a0952c3 service nova] [instance: 87c468d9-9594-4804-b461-527f01f6118f] Refreshing instance network info cache due to event network-changed-d2098868-0475-4b06-bae3-785c608a861e. {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 716.486531] env[63515]: DEBUG oslo_concurrency.lockutils [req-c82cebe1-45cc-4648-be8e-b55d45230645 req-165a31d8-d6a2-47a8-843a-644d3a0952c3 service nova] Acquiring lock "refresh_cache-87c468d9-9594-4804-b461-527f01f6118f" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 716.486677] env[63515]: DEBUG oslo_concurrency.lockutils [req-c82cebe1-45cc-4648-be8e-b55d45230645 req-165a31d8-d6a2-47a8-843a-644d3a0952c3 service nova] Acquired lock "refresh_cache-87c468d9-9594-4804-b461-527f01f6118f" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 716.486835] env[63515]: DEBUG nova.network.neutron [req-c82cebe1-45cc-4648-be8e-b55d45230645 req-165a31d8-d6a2-47a8-843a-644d3a0952c3 service nova] [instance: 87c468d9-9594-4804-b461-527f01f6118f] Refreshing network info cache for port d2098868-0475-4b06-bae3-785c608a861e {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 716.498573] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1110974, 'name': CreateVM_Task, 'duration_secs': 0.393482} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 716.501886] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 452f3ca2-6141-43b2-a77a-c9ab5754192d] Created VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 716.504090] env[63515]: DEBUG oslo_concurrency.lockutils [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 716.504280] env[63515]: DEBUG oslo_concurrency.lockutils [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 716.504746] env[63515]: DEBUG oslo_concurrency.lockutils [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 716.505120] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e48bfcbe-1f3c-4ac8-9c6e-ff1b58a59fc0 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.512140] env[63515]: DEBUG oslo_vmware.api [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Waiting for the task: (returnval){ [ 716.512140] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]521170a3-b208-f8d4-6544-5ad545970db7" [ 716.512140] env[63515]: _type = "Task" [ 716.512140] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 716.524461] env[63515]: DEBUG oslo_vmware.api [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]521170a3-b208-f8d4-6544-5ad545970db7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 716.528172] env[63515]: DEBUG oslo_concurrency.lockutils [None req-bb0fc58f-3c69-4616-a8cc-91b6b5f340c3 tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] Acquiring lock "4087b2c4-6ed1-4b68-8b78-a36e34d935b1" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 716.528364] env[63515]: DEBUG oslo_concurrency.lockutils [None req-bb0fc58f-3c69-4616-a8cc-91b6b5f340c3 tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] Lock "4087b2c4-6ed1-4b68-8b78-a36e34d935b1" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 716.528563] env[63515]: DEBUG oslo_concurrency.lockutils [None req-bb0fc58f-3c69-4616-a8cc-91b6b5f340c3 tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] Acquiring lock "4087b2c4-6ed1-4b68-8b78-a36e34d935b1-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 716.528741] env[63515]: DEBUG oslo_concurrency.lockutils [None req-bb0fc58f-3c69-4616-a8cc-91b6b5f340c3 tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] Lock "4087b2c4-6ed1-4b68-8b78-a36e34d935b1-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 716.528962] env[63515]: DEBUG oslo_concurrency.lockutils [None req-bb0fc58f-3c69-4616-a8cc-91b6b5f340c3 tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] Lock "4087b2c4-6ed1-4b68-8b78-a36e34d935b1-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 716.533430] env[63515]: INFO nova.compute.manager [None req-bb0fc58f-3c69-4616-a8cc-91b6b5f340c3 tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] [instance: 4087b2c4-6ed1-4b68-8b78-a36e34d935b1] Terminating instance [ 716.535569] env[63515]: DEBUG nova.compute.manager [None req-bb0fc58f-3c69-4616-a8cc-91b6b5f340c3 tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] [instance: 4087b2c4-6ed1-4b68-8b78-a36e34d935b1] Start destroying the instance on the hypervisor. {{(pid=63515) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 716.538434] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-bb0fc58f-3c69-4616-a8cc-91b6b5f340c3 tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] [instance: 4087b2c4-6ed1-4b68-8b78-a36e34d935b1] Destroying instance {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 716.538434] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8dbfaa9a-1498-4c6b-93ec-6b80672a1e26 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.546208] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-bb0fc58f-3c69-4616-a8cc-91b6b5f340c3 tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] [instance: 4087b2c4-6ed1-4b68-8b78-a36e34d935b1] Powering off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 716.548670] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9fd57774-1b66-49f0-9001-cdb6de598c8d {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.556175] env[63515]: DEBUG oslo_vmware.api [None req-bb0fc58f-3c69-4616-a8cc-91b6b5f340c3 tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] Waiting for the task: (returnval){ [ 716.556175] env[63515]: value = "task-1110975" [ 716.556175] env[63515]: _type = "Task" [ 716.556175] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 716.565620] env[63515]: DEBUG oslo_vmware.api [None req-bb0fc58f-3c69-4616-a8cc-91b6b5f340c3 tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] Task: {'id': task-1110975, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 716.607904] env[63515]: DEBUG oslo_concurrency.lockutils [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Acquiring lock "refresh_cache-37e6c27e-317b-45d2-bd55-2fd78ccf009f" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 716.608112] env[63515]: DEBUG oslo_concurrency.lockutils [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Acquired lock "refresh_cache-37e6c27e-317b-45d2-bd55-2fd78ccf009f" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 716.608359] env[63515]: DEBUG nova.network.neutron [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] [instance: 37e6c27e-317b-45d2-bd55-2fd78ccf009f] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 716.673584] env[63515]: DEBUG oslo_vmware.api [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Task: {'id': task-1110973, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.830057} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 716.674298] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk to [datastore1] 48668736-df27-4f2a-94d9-132f5b49701b/48668736-df27-4f2a-94d9-132f5b49701b.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 716.674298] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] [instance: 48668736-df27-4f2a-94d9-132f5b49701b] Extending root virtual disk to 1048576 {{(pid=63515) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 716.674298] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-3531d6f7-2369-45ad-a499-c6c2076bc7d2 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.683758] env[63515]: DEBUG oslo_vmware.api [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Waiting for the task: (returnval){ [ 716.683758] env[63515]: value = "task-1110976" [ 716.683758] env[63515]: _type = "Task" [ 716.683758] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 716.692370] env[63515]: DEBUG oslo_vmware.api [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Task: {'id': task-1110976, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 716.807456] env[63515]: DEBUG nova.compute.manager [None req-746f5a16-d8b1-448e-a35c-34720548f33b tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] [instance: b4477e66-ae12-4929-90ed-b7b652e0f207] Start building block device mappings for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 716.841808] env[63515]: DEBUG nova.network.neutron [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] [instance: b911a5b5-9617-4fb3-9b5e-fb8c492e4931] Updating instance_info_cache with network_info: [{"id": "9bf84cb2-7053-4129-8d8a-dc09e80b384e", "address": "fa:16:3e:c6:e8:2f", "network": {"id": "22a0cdee-4685-4b2b-9df7-2506b2373a19", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-878688764-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.147", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5086c79c531e4d039896a24a9ea29f73", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "78b49840-c3fc-455c-8491-a253ccd92bb5", "external-id": "nsx-vlan-transportzone-779", "segmentation_id": 779, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9bf84cb2-70", "ovs_interfaceid": "9bf84cb2-7053-4129-8d8a-dc09e80b384e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 716.888425] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c16a1dd4-1a2f-4c9e-b42a-6cf9ab2e8a02 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.899195] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9375cbec-9439-421a-8f0d-2621bed3716a {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.934286] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-766a0faf-c5ba-4410-a252-3aa3b5cfcaaa {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.945021] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47736e9c-d666-4737-aace-dc2820af4092 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.946717] env[63515]: DEBUG oslo_concurrency.lockutils [req-ed83a135-13af-44da-9eae-1fb8831f71ef req-8a46ef23-4e8b-489d-ab0e-86ec285b6f32 service nova] Releasing lock "refresh_cache-452f3ca2-6141-43b2-a77a-c9ab5754192d" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 716.957267] env[63515]: DEBUG nova.compute.provider_tree [None req-5690e97d-f63c-4f8e-9c79-d9c523fc3720 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 717.026905] env[63515]: DEBUG oslo_vmware.api [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]521170a3-b208-f8d4-6544-5ad545970db7, 'name': SearchDatastore_Task, 'duration_secs': 0.012833} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 717.026905] env[63515]: DEBUG oslo_concurrency.lockutils [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 717.026905] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] [instance: 452f3ca2-6141-43b2-a77a-c9ab5754192d] Processing image 8a120570-cb06-4099-b262-554ca0ad15c5 {{(pid=63515) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 717.026905] env[63515]: DEBUG oslo_concurrency.lockutils [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 717.027145] env[63515]: DEBUG oslo_concurrency.lockutils [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 717.027145] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 717.027145] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-97ea4b4e-5722-4310-b3da-fb613a69097d {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.037698] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 717.037938] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63515) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 717.038860] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-28f7e490-25af-4f29-b8d4-85363c5c8fe5 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.046901] env[63515]: DEBUG oslo_vmware.api [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Waiting for the task: (returnval){ [ 717.046901] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]5273e9ff-5c2c-b9f0-e059-f2455b5c30ba" [ 717.046901] env[63515]: _type = "Task" [ 717.046901] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 717.056032] env[63515]: DEBUG oslo_vmware.api [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]5273e9ff-5c2c-b9f0-e059-f2455b5c30ba, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 717.065893] env[63515]: DEBUG oslo_vmware.api [None req-bb0fc58f-3c69-4616-a8cc-91b6b5f340c3 tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] Task: {'id': task-1110975, 'name': PowerOffVM_Task, 'duration_secs': 0.196907} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 717.065893] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-bb0fc58f-3c69-4616-a8cc-91b6b5f340c3 tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] [instance: 4087b2c4-6ed1-4b68-8b78-a36e34d935b1] Powered off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 717.065893] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-bb0fc58f-3c69-4616-a8cc-91b6b5f340c3 tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] [instance: 4087b2c4-6ed1-4b68-8b78-a36e34d935b1] Unregistering the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 717.065893] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-691cace7-83e8-42b0-9dce-ab6496aa06ee {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.144929] env[63515]: DEBUG nova.network.neutron [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] [instance: 37e6c27e-317b-45d2-bd55-2fd78ccf009f] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 717.194667] env[63515]: DEBUG oslo_vmware.api [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Task: {'id': task-1110976, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.072252} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 717.194959] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] [instance: 48668736-df27-4f2a-94d9-132f5b49701b] Extended root virtual disk {{(pid=63515) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 717.195773] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-afea5a12-d67f-46fc-808c-bfe79e2f8fae {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.223904] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] [instance: 48668736-df27-4f2a-94d9-132f5b49701b] Reconfiguring VM instance instance-00000027 to attach disk [datastore1] 48668736-df27-4f2a-94d9-132f5b49701b/48668736-df27-4f2a-94d9-132f5b49701b.vmdk or device None with type sparse {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 717.226869] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-42a3b14a-b6fb-4ca4-85b8-8d420e6da79a {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.249985] env[63515]: DEBUG oslo_vmware.api [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Waiting for the task: (returnval){ [ 717.249985] env[63515]: value = "task-1110978" [ 717.249985] env[63515]: _type = "Task" [ 717.249985] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 717.258024] env[63515]: DEBUG oslo_vmware.api [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Task: {'id': task-1110978, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 717.286455] env[63515]: DEBUG nova.network.neutron [req-c82cebe1-45cc-4648-be8e-b55d45230645 req-165a31d8-d6a2-47a8-843a-644d3a0952c3 service nova] [instance: 87c468d9-9594-4804-b461-527f01f6118f] Updated VIF entry in instance network info cache for port d2098868-0475-4b06-bae3-785c608a861e. {{(pid=63515) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 717.286828] env[63515]: DEBUG nova.network.neutron [req-c82cebe1-45cc-4648-be8e-b55d45230645 req-165a31d8-d6a2-47a8-843a-644d3a0952c3 service nova] [instance: 87c468d9-9594-4804-b461-527f01f6118f] Updating instance_info_cache with network_info: [{"id": "d2098868-0475-4b06-bae3-785c608a861e", "address": "fa:16:3e:af:f9:f0", "network": {"id": "09fb20ca-2b17-4d46-b915-6aff85b862d8", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1815529427-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.212", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "19ce3b05eec64475b685faad1fd801f6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ded18042-834c-4792-b3e8-b1c377446432", "external-id": "nsx-vlan-transportzone-293", "segmentation_id": 293, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd2098868-04", "ovs_interfaceid": "d2098868-0475-4b06-bae3-785c608a861e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 717.322952] env[63515]: DEBUG nova.network.neutron [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] [instance: 37e6c27e-317b-45d2-bd55-2fd78ccf009f] Updating instance_info_cache with network_info: [{"id": "21c8f7a2-4403-41cd-93f8-3091de7cdc40", "address": "fa:16:3e:44:aa:08", "network": {"id": "a2fdda6f-f99f-420f-b3c7-61c9ebcbeaa8", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-1981716599-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3374114af6314f86be3559223d70dfdb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "13d625c9-77ec-4edb-a56b-9f37a314cc39", "external-id": "nsx-vlan-transportzone-358", "segmentation_id": 358, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap21c8f7a2-44", "ovs_interfaceid": "21c8f7a2-4403-41cd-93f8-3091de7cdc40", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 717.343976] env[63515]: DEBUG oslo_concurrency.lockutils [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Releasing lock "refresh_cache-b911a5b5-9617-4fb3-9b5e-fb8c492e4931" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 717.344215] env[63515]: DEBUG nova.compute.manager [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] [instance: b911a5b5-9617-4fb3-9b5e-fb8c492e4931] Updated the network info_cache for instance {{(pid=63515) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9997}} [ 717.344407] env[63515]: DEBUG oslo_service.periodic_task [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=63515) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 717.344650] env[63515]: DEBUG oslo_service.periodic_task [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=63515) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 717.345296] env[63515]: DEBUG oslo_service.periodic_task [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=63515) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 717.345480] env[63515]: DEBUG oslo_service.periodic_task [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=63515) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 717.345632] env[63515]: DEBUG oslo_service.periodic_task [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=63515) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 717.345863] env[63515]: DEBUG oslo_service.periodic_task [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=63515) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 717.346112] env[63515]: DEBUG nova.compute.manager [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=63515) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10545}} [ 717.346282] env[63515]: DEBUG oslo_service.periodic_task [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Running periodic task ComputeManager.update_available_resource {{(pid=63515) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 717.460357] env[63515]: DEBUG nova.scheduler.client.report [None req-5690e97d-f63c-4f8e-9c79-d9c523fc3720 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 717.559128] env[63515]: DEBUG oslo_vmware.api [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]5273e9ff-5c2c-b9f0-e059-f2455b5c30ba, 'name': SearchDatastore_Task, 'duration_secs': 0.009497} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 717.559128] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-62eca018-1cf2-4738-a037-9c308c3d0a43 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.566314] env[63515]: DEBUG oslo_vmware.api [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Waiting for the task: (returnval){ [ 717.566314] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]52bf3cb6-23ec-f6b8-b48f-d2dcd0f1f83f" [ 717.566314] env[63515]: _type = "Task" [ 717.566314] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 717.574869] env[63515]: DEBUG oslo_vmware.api [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52bf3cb6-23ec-f6b8-b48f-d2dcd0f1f83f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 717.715274] env[63515]: DEBUG nova.compute.manager [req-5594c4b2-d1be-4e4a-967e-532bd44d8334 req-14857b49-469e-4363-8a07-535afcbdfa73 service nova] [instance: 37e6c27e-317b-45d2-bd55-2fd78ccf009f] Received event network-vif-plugged-21c8f7a2-4403-41cd-93f8-3091de7cdc40 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 717.715482] env[63515]: DEBUG oslo_concurrency.lockutils [req-5594c4b2-d1be-4e4a-967e-532bd44d8334 req-14857b49-469e-4363-8a07-535afcbdfa73 service nova] Acquiring lock "37e6c27e-317b-45d2-bd55-2fd78ccf009f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 717.715696] env[63515]: DEBUG oslo_concurrency.lockutils [req-5594c4b2-d1be-4e4a-967e-532bd44d8334 req-14857b49-469e-4363-8a07-535afcbdfa73 service nova] Lock "37e6c27e-317b-45d2-bd55-2fd78ccf009f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 717.715863] env[63515]: DEBUG oslo_concurrency.lockutils [req-5594c4b2-d1be-4e4a-967e-532bd44d8334 req-14857b49-469e-4363-8a07-535afcbdfa73 service nova] Lock "37e6c27e-317b-45d2-bd55-2fd78ccf009f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 717.716123] env[63515]: DEBUG nova.compute.manager [req-5594c4b2-d1be-4e4a-967e-532bd44d8334 req-14857b49-469e-4363-8a07-535afcbdfa73 service nova] [instance: 37e6c27e-317b-45d2-bd55-2fd78ccf009f] No waiting events found dispatching network-vif-plugged-21c8f7a2-4403-41cd-93f8-3091de7cdc40 {{(pid=63515) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 717.716346] env[63515]: WARNING nova.compute.manager [req-5594c4b2-d1be-4e4a-967e-532bd44d8334 req-14857b49-469e-4363-8a07-535afcbdfa73 service nova] [instance: 37e6c27e-317b-45d2-bd55-2fd78ccf009f] Received unexpected event network-vif-plugged-21c8f7a2-4403-41cd-93f8-3091de7cdc40 for instance with vm_state building and task_state spawning. [ 717.716526] env[63515]: DEBUG nova.compute.manager [req-5594c4b2-d1be-4e4a-967e-532bd44d8334 req-14857b49-469e-4363-8a07-535afcbdfa73 service nova] [instance: 37e6c27e-317b-45d2-bd55-2fd78ccf009f] Received event network-changed-21c8f7a2-4403-41cd-93f8-3091de7cdc40 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 717.716601] env[63515]: DEBUG nova.compute.manager [req-5594c4b2-d1be-4e4a-967e-532bd44d8334 req-14857b49-469e-4363-8a07-535afcbdfa73 service nova] [instance: 37e6c27e-317b-45d2-bd55-2fd78ccf009f] Refreshing instance network info cache due to event network-changed-21c8f7a2-4403-41cd-93f8-3091de7cdc40. {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 717.716772] env[63515]: DEBUG oslo_concurrency.lockutils [req-5594c4b2-d1be-4e4a-967e-532bd44d8334 req-14857b49-469e-4363-8a07-535afcbdfa73 service nova] Acquiring lock "refresh_cache-37e6c27e-317b-45d2-bd55-2fd78ccf009f" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 717.759464] env[63515]: DEBUG oslo_vmware.api [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Task: {'id': task-1110978, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 717.792103] env[63515]: DEBUG oslo_concurrency.lockutils [req-c82cebe1-45cc-4648-be8e-b55d45230645 req-165a31d8-d6a2-47a8-843a-644d3a0952c3 service nova] Releasing lock "refresh_cache-87c468d9-9594-4804-b461-527f01f6118f" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 717.816698] env[63515]: DEBUG nova.compute.manager [None req-746f5a16-d8b1-448e-a35c-34720548f33b tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] [instance: b4477e66-ae12-4929-90ed-b7b652e0f207] Start spawning the instance on the hypervisor. {{(pid=63515) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 717.825511] env[63515]: DEBUG oslo_concurrency.lockutils [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Releasing lock "refresh_cache-37e6c27e-317b-45d2-bd55-2fd78ccf009f" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 717.825674] env[63515]: DEBUG nova.compute.manager [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] [instance: 37e6c27e-317b-45d2-bd55-2fd78ccf009f] Instance network_info: |[{"id": "21c8f7a2-4403-41cd-93f8-3091de7cdc40", "address": "fa:16:3e:44:aa:08", "network": {"id": "a2fdda6f-f99f-420f-b3c7-61c9ebcbeaa8", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-1981716599-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3374114af6314f86be3559223d70dfdb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "13d625c9-77ec-4edb-a56b-9f37a314cc39", "external-id": "nsx-vlan-transportzone-358", "segmentation_id": 358, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap21c8f7a2-44", "ovs_interfaceid": "21c8f7a2-4403-41cd-93f8-3091de7cdc40", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 717.826040] env[63515]: DEBUG oslo_concurrency.lockutils [req-5594c4b2-d1be-4e4a-967e-532bd44d8334 req-14857b49-469e-4363-8a07-535afcbdfa73 service nova] Acquired lock "refresh_cache-37e6c27e-317b-45d2-bd55-2fd78ccf009f" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 717.826224] env[63515]: DEBUG nova.network.neutron [req-5594c4b2-d1be-4e4a-967e-532bd44d8334 req-14857b49-469e-4363-8a07-535afcbdfa73 service nova] [instance: 37e6c27e-317b-45d2-bd55-2fd78ccf009f] Refreshing network info cache for port 21c8f7a2-4403-41cd-93f8-3091de7cdc40 {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 717.827439] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] [instance: 37e6c27e-317b-45d2-bd55-2fd78ccf009f] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:44:aa:08', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '13d625c9-77ec-4edb-a56b-9f37a314cc39', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '21c8f7a2-4403-41cd-93f8-3091de7cdc40', 'vif_model': 'vmxnet3'}] {{(pid=63515) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 717.835121] env[63515]: DEBUG oslo.service.loopingcall [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 717.837487] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 37e6c27e-317b-45d2-bd55-2fd78ccf009f] Creating VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 717.838376] env[63515]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-1e7f017b-795a-430a-80ae-7badbe823363 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.854799] env[63515]: DEBUG nova.virt.hardware [None req-746f5a16-d8b1-448e-a35c-34720548f33b tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T02:52:14Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T02:51:56Z,direct_url=,disk_format='vmdk',id=8a120570-cb06-4099-b262-554ca0ad15c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b569255cc43e42e7ae3f2b4ad37c6ef4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T02:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 717.855116] env[63515]: DEBUG nova.virt.hardware [None req-746f5a16-d8b1-448e-a35c-34720548f33b tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Flavor limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 717.855335] env[63515]: DEBUG nova.virt.hardware [None req-746f5a16-d8b1-448e-a35c-34720548f33b tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Image limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 717.855591] env[63515]: DEBUG nova.virt.hardware [None req-746f5a16-d8b1-448e-a35c-34720548f33b tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Flavor pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 717.855705] env[63515]: DEBUG nova.virt.hardware [None req-746f5a16-d8b1-448e-a35c-34720548f33b tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Image pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 717.855830] env[63515]: DEBUG nova.virt.hardware [None req-746f5a16-d8b1-448e-a35c-34720548f33b tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 717.856029] env[63515]: DEBUG nova.virt.hardware [None req-746f5a16-d8b1-448e-a35c-34720548f33b tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 717.856188] env[63515]: DEBUG nova.virt.hardware [None req-746f5a16-d8b1-448e-a35c-34720548f33b tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 717.856349] env[63515]: DEBUG nova.virt.hardware [None req-746f5a16-d8b1-448e-a35c-34720548f33b tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Got 1 possible topologies {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 717.856510] env[63515]: DEBUG nova.virt.hardware [None req-746f5a16-d8b1-448e-a35c-34720548f33b tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 717.856676] env[63515]: DEBUG nova.virt.hardware [None req-746f5a16-d8b1-448e-a35c-34720548f33b tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 717.858070] env[63515]: DEBUG oslo_concurrency.lockutils [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 717.858939] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-302db116-6be6-47c1-99dd-fe8f5a82e46e {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.867691] env[63515]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 717.867691] env[63515]: value = "task-1110979" [ 717.867691] env[63515]: _type = "Task" [ 717.867691] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 717.873872] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ac08b84-84ac-4398-ac61-5aeac1e5e94c {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.879228] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-bb0fc58f-3c69-4616-a8cc-91b6b5f340c3 tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] [instance: 4087b2c4-6ed1-4b68-8b78-a36e34d935b1] Unregistered the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 717.879431] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-bb0fc58f-3c69-4616-a8cc-91b6b5f340c3 tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] [instance: 4087b2c4-6ed1-4b68-8b78-a36e34d935b1] Deleting contents of the VM from datastore datastore1 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 717.879604] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-bb0fc58f-3c69-4616-a8cc-91b6b5f340c3 tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] Deleting the datastore file [datastore1] 4087b2c4-6ed1-4b68-8b78-a36e34d935b1 {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 717.880414] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-83cbcbd1-63f6-4c9f-a011-b95bab24a4f6 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.893256] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-746f5a16-d8b1-448e-a35c-34720548f33b tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] [instance: b4477e66-ae12-4929-90ed-b7b652e0f207] Instance VIF info [] {{(pid=63515) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 717.898991] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-746f5a16-d8b1-448e-a35c-34720548f33b tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Creating folder: Project (7c42466885f541368ceb00603fe7a8d0). Parent ref: group-v243370. {{(pid=63515) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 717.899272] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1110979, 'name': CreateVM_Task} progress is 6%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 717.900516] env[63515]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-61fb6d91-8084-4088-8a3a-0b4aff9e317b {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.903522] env[63515]: DEBUG oslo_vmware.api [None req-bb0fc58f-3c69-4616-a8cc-91b6b5f340c3 tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] Waiting for the task: (returnval){ [ 717.903522] env[63515]: value = "task-1110980" [ 717.903522] env[63515]: _type = "Task" [ 717.903522] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 717.908065] env[63515]: INFO nova.virt.vmwareapi.vm_util [None req-746f5a16-d8b1-448e-a35c-34720548f33b tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Created folder: Project (7c42466885f541368ceb00603fe7a8d0) in parent group-v243370. [ 717.908266] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-746f5a16-d8b1-448e-a35c-34720548f33b tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Creating folder: Instances. Parent ref: group-v243411. {{(pid=63515) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 717.908454] env[63515]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-c33e3fc7-4b53-4c7d-8fd8-807dfc9de185 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.912869] env[63515]: DEBUG oslo_vmware.api [None req-bb0fc58f-3c69-4616-a8cc-91b6b5f340c3 tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] Task: {'id': task-1110980, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 717.919743] env[63515]: INFO nova.virt.vmwareapi.vm_util [None req-746f5a16-d8b1-448e-a35c-34720548f33b tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Created folder: Instances in parent group-v243411. [ 717.919990] env[63515]: DEBUG oslo.service.loopingcall [None req-746f5a16-d8b1-448e-a35c-34720548f33b tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 717.920214] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b4477e66-ae12-4929-90ed-b7b652e0f207] Creating VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 717.920425] env[63515]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-37caac0a-2885-484b-8935-077e41227a3b {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.937396] env[63515]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 717.937396] env[63515]: value = "task-1110983" [ 717.937396] env[63515]: _type = "Task" [ 717.937396] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 717.948806] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1110983, 'name': CreateVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 717.968914] env[63515]: DEBUG oslo_concurrency.lockutils [None req-5690e97d-f63c-4f8e-9c79-d9c523fc3720 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.175s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 717.971242] env[63515]: DEBUG oslo_concurrency.lockutils [None req-74f7de08-cdcf-4b98-bfdb-a2dd77c03e6d tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.428s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 717.972804] env[63515]: INFO nova.compute.claims [None req-74f7de08-cdcf-4b98-bfdb-a2dd77c03e6d tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] [instance: b85e9a70-7f5b-4d32-b616-f2a97e3186c8] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 717.993452] env[63515]: INFO nova.scheduler.client.report [None req-5690e97d-f63c-4f8e-9c79-d9c523fc3720 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Deleted allocations for instance 3ad04a9f-62cf-4b1c-aab2-7dc4a219edc6 [ 718.078874] env[63515]: DEBUG oslo_vmware.api [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52bf3cb6-23ec-f6b8-b48f-d2dcd0f1f83f, 'name': SearchDatastore_Task, 'duration_secs': 0.01448} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 718.079257] env[63515]: DEBUG oslo_concurrency.lockutils [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 718.079589] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk to [datastore1] 452f3ca2-6141-43b2-a77a-c9ab5754192d/452f3ca2-6141-43b2-a77a-c9ab5754192d.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 718.079883] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a72af59b-bdc8-4bb5-bf79-d2170e43709f {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.087910] env[63515]: DEBUG oslo_vmware.api [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Waiting for the task: (returnval){ [ 718.087910] env[63515]: value = "task-1110984" [ 718.087910] env[63515]: _type = "Task" [ 718.087910] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 718.097607] env[63515]: DEBUG oslo_vmware.api [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Task: {'id': task-1110984, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 718.260681] env[63515]: DEBUG oslo_vmware.api [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Task: {'id': task-1110978, 'name': ReconfigVM_Task, 'duration_secs': 0.898685} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 718.261062] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] [instance: 48668736-df27-4f2a-94d9-132f5b49701b] Reconfigured VM instance instance-00000027 to attach disk [datastore1] 48668736-df27-4f2a-94d9-132f5b49701b/48668736-df27-4f2a-94d9-132f5b49701b.vmdk or device None with type sparse {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 718.261731] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-2ba7f359-242e-41ee-a0d8-6feb049f7dea {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.267660] env[63515]: DEBUG oslo_vmware.api [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Waiting for the task: (returnval){ [ 718.267660] env[63515]: value = "task-1110985" [ 718.267660] env[63515]: _type = "Task" [ 718.267660] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 718.275392] env[63515]: DEBUG oslo_vmware.api [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Task: {'id': task-1110985, 'name': Rename_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 718.383906] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1110979, 'name': CreateVM_Task, 'duration_secs': 0.350467} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 718.383906] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 37e6c27e-317b-45d2-bd55-2fd78ccf009f] Created VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 718.383906] env[63515]: DEBUG oslo_concurrency.lockutils [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 718.384190] env[63515]: DEBUG oslo_concurrency.lockutils [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 718.384308] env[63515]: DEBUG oslo_concurrency.lockutils [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 718.384642] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a0b28921-1dfc-4538-9717-d3d632b47941 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.389591] env[63515]: DEBUG oslo_vmware.api [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Waiting for the task: (returnval){ [ 718.389591] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]52ab72dd-13b5-b7ed-0c23-c536694736f2" [ 718.389591] env[63515]: _type = "Task" [ 718.389591] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 718.399735] env[63515]: DEBUG oslo_vmware.api [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52ab72dd-13b5-b7ed-0c23-c536694736f2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 718.412984] env[63515]: DEBUG oslo_vmware.api [None req-bb0fc58f-3c69-4616-a8cc-91b6b5f340c3 tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] Task: {'id': task-1110980, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.209101} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 718.412984] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-bb0fc58f-3c69-4616-a8cc-91b6b5f340c3 tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] Deleted the datastore file {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 718.412984] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-bb0fc58f-3c69-4616-a8cc-91b6b5f340c3 tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] [instance: 4087b2c4-6ed1-4b68-8b78-a36e34d935b1] Deleted contents of the VM from datastore datastore1 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 718.412984] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-bb0fc58f-3c69-4616-a8cc-91b6b5f340c3 tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] [instance: 4087b2c4-6ed1-4b68-8b78-a36e34d935b1] Instance destroyed {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 718.412984] env[63515]: INFO nova.compute.manager [None req-bb0fc58f-3c69-4616-a8cc-91b6b5f340c3 tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] [instance: 4087b2c4-6ed1-4b68-8b78-a36e34d935b1] Took 1.88 seconds to destroy the instance on the hypervisor. [ 718.413396] env[63515]: DEBUG oslo.service.loopingcall [None req-bb0fc58f-3c69-4616-a8cc-91b6b5f340c3 tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 718.413444] env[63515]: DEBUG nova.compute.manager [-] [instance: 4087b2c4-6ed1-4b68-8b78-a36e34d935b1] Deallocating network for instance {{(pid=63515) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 718.414067] env[63515]: DEBUG nova.network.neutron [-] [instance: 4087b2c4-6ed1-4b68-8b78-a36e34d935b1] deallocate_for_instance() {{(pid=63515) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 718.448292] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1110983, 'name': CreateVM_Task, 'duration_secs': 0.278392} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 718.452857] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b4477e66-ae12-4929-90ed-b7b652e0f207] Created VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 718.453524] env[63515]: DEBUG oslo_concurrency.lockutils [None req-746f5a16-d8b1-448e-a35c-34720548f33b tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 718.453688] env[63515]: DEBUG oslo_concurrency.lockutils [None req-746f5a16-d8b1-448e-a35c-34720548f33b tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 718.454090] env[63515]: DEBUG oslo_concurrency.lockutils [None req-746f5a16-d8b1-448e-a35c-34720548f33b tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 718.455010] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0fda5ad8-b5cc-4e2d-bc34-40a8bb4a73c9 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.460484] env[63515]: DEBUG oslo_vmware.api [None req-746f5a16-d8b1-448e-a35c-34720548f33b tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Waiting for the task: (returnval){ [ 718.460484] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]52fc99e5-e7ee-05be-0273-065ecd1fe207" [ 718.460484] env[63515]: _type = "Task" [ 718.460484] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 718.471040] env[63515]: DEBUG oslo_vmware.api [None req-746f5a16-d8b1-448e-a35c-34720548f33b tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52fc99e5-e7ee-05be-0273-065ecd1fe207, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 718.502480] env[63515]: DEBUG oslo_concurrency.lockutils [None req-5690e97d-f63c-4f8e-9c79-d9c523fc3720 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Lock "3ad04a9f-62cf-4b1c-aab2-7dc4a219edc6" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 21.956s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 718.600146] env[63515]: DEBUG oslo_vmware.api [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Task: {'id': task-1110984, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.485019} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 718.600447] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk to [datastore1] 452f3ca2-6141-43b2-a77a-c9ab5754192d/452f3ca2-6141-43b2-a77a-c9ab5754192d.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 718.600795] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] [instance: 452f3ca2-6141-43b2-a77a-c9ab5754192d] Extending root virtual disk to 1048576 {{(pid=63515) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 718.601106] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-dd85241d-b267-4e1a-b3a9-cbe1fec7f1a1 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.609134] env[63515]: DEBUG oslo_vmware.api [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Waiting for the task: (returnval){ [ 718.609134] env[63515]: value = "task-1110986" [ 718.609134] env[63515]: _type = "Task" [ 718.609134] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 718.612540] env[63515]: DEBUG nova.network.neutron [req-5594c4b2-d1be-4e4a-967e-532bd44d8334 req-14857b49-469e-4363-8a07-535afcbdfa73 service nova] [instance: 37e6c27e-317b-45d2-bd55-2fd78ccf009f] Updated VIF entry in instance network info cache for port 21c8f7a2-4403-41cd-93f8-3091de7cdc40. {{(pid=63515) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 718.612976] env[63515]: DEBUG nova.network.neutron [req-5594c4b2-d1be-4e4a-967e-532bd44d8334 req-14857b49-469e-4363-8a07-535afcbdfa73 service nova] [instance: 37e6c27e-317b-45d2-bd55-2fd78ccf009f] Updating instance_info_cache with network_info: [{"id": "21c8f7a2-4403-41cd-93f8-3091de7cdc40", "address": "fa:16:3e:44:aa:08", "network": {"id": "a2fdda6f-f99f-420f-b3c7-61c9ebcbeaa8", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-1981716599-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3374114af6314f86be3559223d70dfdb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "13d625c9-77ec-4edb-a56b-9f37a314cc39", "external-id": "nsx-vlan-transportzone-358", "segmentation_id": 358, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap21c8f7a2-44", "ovs_interfaceid": "21c8f7a2-4403-41cd-93f8-3091de7cdc40", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 718.617143] env[63515]: DEBUG oslo_vmware.api [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Task: {'id': task-1110986, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 718.777602] env[63515]: DEBUG oslo_vmware.api [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Task: {'id': task-1110985, 'name': Rename_Task, 'duration_secs': 0.181061} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 718.777872] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] [instance: 48668736-df27-4f2a-94d9-132f5b49701b] Powering on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 718.778123] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-493a2905-f12e-40cf-bec9-64e5d2fa5181 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.783549] env[63515]: DEBUG oslo_vmware.api [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Waiting for the task: (returnval){ [ 718.783549] env[63515]: value = "task-1110987" [ 718.783549] env[63515]: _type = "Task" [ 718.783549] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 718.792513] env[63515]: DEBUG oslo_vmware.api [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Task: {'id': task-1110987, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 718.902905] env[63515]: DEBUG oslo_vmware.api [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52ab72dd-13b5-b7ed-0c23-c536694736f2, 'name': SearchDatastore_Task, 'duration_secs': 0.05458} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 718.902905] env[63515]: DEBUG oslo_concurrency.lockutils [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 718.902905] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] [instance: 37e6c27e-317b-45d2-bd55-2fd78ccf009f] Processing image 8a120570-cb06-4099-b262-554ca0ad15c5 {{(pid=63515) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 718.902905] env[63515]: DEBUG oslo_concurrency.lockutils [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 718.903166] env[63515]: DEBUG oslo_concurrency.lockutils [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 718.903166] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 718.903166] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e0937587-b9ac-405e-bea2-edb26351f89a {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.911432] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 718.911432] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63515) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 718.911843] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2c7c59e6-6b09-4eda-8efc-411bb38866df {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.917383] env[63515]: DEBUG oslo_vmware.api [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Waiting for the task: (returnval){ [ 718.917383] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]5286557d-1701-5153-2582-c575dc852c24" [ 718.917383] env[63515]: _type = "Task" [ 718.917383] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 718.925926] env[63515]: DEBUG oslo_vmware.api [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]5286557d-1701-5153-2582-c575dc852c24, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 718.952013] env[63515]: DEBUG nova.compute.manager [req-938d804f-a41c-491c-bf8c-f6c12ec385be req-b4e78f48-ee28-4219-9c56-b6cc8eafaebb service nova] [instance: 4087b2c4-6ed1-4b68-8b78-a36e34d935b1] Received event network-vif-deleted-81eafe68-fff5-401d-812f-73b07de3ad8b {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 718.952232] env[63515]: INFO nova.compute.manager [req-938d804f-a41c-491c-bf8c-f6c12ec385be req-b4e78f48-ee28-4219-9c56-b6cc8eafaebb service nova] [instance: 4087b2c4-6ed1-4b68-8b78-a36e34d935b1] Neutron deleted interface 81eafe68-fff5-401d-812f-73b07de3ad8b; detaching it from the instance and deleting it from the info cache [ 718.952402] env[63515]: DEBUG nova.network.neutron [req-938d804f-a41c-491c-bf8c-f6c12ec385be req-b4e78f48-ee28-4219-9c56-b6cc8eafaebb service nova] [instance: 4087b2c4-6ed1-4b68-8b78-a36e34d935b1] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 718.970273] env[63515]: DEBUG oslo_vmware.api [None req-746f5a16-d8b1-448e-a35c-34720548f33b tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52fc99e5-e7ee-05be-0273-065ecd1fe207, 'name': SearchDatastore_Task, 'duration_secs': 0.023376} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 718.970694] env[63515]: DEBUG oslo_concurrency.lockutils [None req-746f5a16-d8b1-448e-a35c-34720548f33b tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 718.970811] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-746f5a16-d8b1-448e-a35c-34720548f33b tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] [instance: b4477e66-ae12-4929-90ed-b7b652e0f207] Processing image 8a120570-cb06-4099-b262-554ca0ad15c5 {{(pid=63515) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 718.971032] env[63515]: DEBUG oslo_concurrency.lockutils [None req-746f5a16-d8b1-448e-a35c-34720548f33b tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 718.971182] env[63515]: DEBUG oslo_concurrency.lockutils [None req-746f5a16-d8b1-448e-a35c-34720548f33b tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 718.971356] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-746f5a16-d8b1-448e-a35c-34720548f33b tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 718.971609] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-aba4591f-2947-4a2c-8e5a-dcf489a0bc57 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.979905] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-746f5a16-d8b1-448e-a35c-34720548f33b tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 718.980087] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-746f5a16-d8b1-448e-a35c-34720548f33b tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63515) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 718.980707] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8eccc102-3ade-4502-a544-cfc5ee921da6 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.988556] env[63515]: DEBUG oslo_vmware.api [None req-746f5a16-d8b1-448e-a35c-34720548f33b tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Waiting for the task: (returnval){ [ 718.988556] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]52910003-4552-8e9b-f201-b8daaa746b59" [ 718.988556] env[63515]: _type = "Task" [ 718.988556] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 718.997505] env[63515]: DEBUG oslo_vmware.api [None req-746f5a16-d8b1-448e-a35c-34720548f33b tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52910003-4552-8e9b-f201-b8daaa746b59, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 719.117942] env[63515]: DEBUG oslo_vmware.api [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Task: {'id': task-1110986, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.063703} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 719.118067] env[63515]: DEBUG oslo_concurrency.lockutils [req-5594c4b2-d1be-4e4a-967e-532bd44d8334 req-14857b49-469e-4363-8a07-535afcbdfa73 service nova] Releasing lock "refresh_cache-37e6c27e-317b-45d2-bd55-2fd78ccf009f" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 719.118386] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] [instance: 452f3ca2-6141-43b2-a77a-c9ab5754192d] Extended root virtual disk {{(pid=63515) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 719.119171] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-400f7a6f-a468-4a98-bfd1-430eae3a66d6 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.143606] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] [instance: 452f3ca2-6141-43b2-a77a-c9ab5754192d] Reconfiguring VM instance instance-00000028 to attach disk [datastore1] 452f3ca2-6141-43b2-a77a-c9ab5754192d/452f3ca2-6141-43b2-a77a-c9ab5754192d.vmdk or device None with type sparse {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 719.146518] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e0191143-5cd3-49eb-adf8-fee91fb2dbc1 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.166230] env[63515]: DEBUG oslo_vmware.api [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Waiting for the task: (returnval){ [ 719.166230] env[63515]: value = "task-1110988" [ 719.166230] env[63515]: _type = "Task" [ 719.166230] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 719.177785] env[63515]: DEBUG oslo_vmware.api [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Task: {'id': task-1110988, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 719.240042] env[63515]: DEBUG nova.network.neutron [-] [instance: 4087b2c4-6ed1-4b68-8b78-a36e34d935b1] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 719.295913] env[63515]: DEBUG oslo_vmware.api [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Task: {'id': task-1110987, 'name': PowerOnVM_Task} progress is 88%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 719.391917] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-457462e7-fd5f-4f86-b90a-994348fb71cd {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.399110] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eeb8f0f1-cf42-4f7c-a090-43659a123852 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.431209] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65fdd13f-3aa3-4f06-b45e-fb6a2b156027 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.438680] env[63515]: DEBUG oslo_vmware.api [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]5286557d-1701-5153-2582-c575dc852c24, 'name': SearchDatastore_Task, 'duration_secs': 0.008974} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 719.441117] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e81d84d4-d904-424e-9d42-a6ddecf50538 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.444432] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94a07431-daae-4e61-86b3-eba146e11a42 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.453854] env[63515]: DEBUG oslo_vmware.api [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Waiting for the task: (returnval){ [ 719.453854] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]525aef20-b20e-3403-1d71-a75aac993ef7" [ 719.453854] env[63515]: _type = "Task" [ 719.453854] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 719.462985] env[63515]: DEBUG nova.compute.provider_tree [None req-74f7de08-cdcf-4b98-bfdb-a2dd77c03e6d tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 719.467196] env[63515]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-5fcb29be-57e9-43de-a6b0-4dd278c0c09b {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.475094] env[63515]: DEBUG oslo_vmware.api [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]525aef20-b20e-3403-1d71-a75aac993ef7, 'name': SearchDatastore_Task, 'duration_secs': 0.009842} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 719.476842] env[63515]: DEBUG oslo_concurrency.lockutils [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 719.477173] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk to [datastore1] 37e6c27e-317b-45d2-bd55-2fd78ccf009f/37e6c27e-317b-45d2-bd55-2fd78ccf009f.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 719.477699] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-eb8477d0-a1fc-4656-a8f0-0f72aa5dc8fb {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.482380] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b42e57e-67c4-457a-b9be-bb8d98086ebd {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.499993] env[63515]: DEBUG oslo_vmware.api [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Waiting for the task: (returnval){ [ 719.499993] env[63515]: value = "task-1110989" [ 719.499993] env[63515]: _type = "Task" [ 719.499993] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 719.507916] env[63515]: DEBUG oslo_vmware.api [None req-746f5a16-d8b1-448e-a35c-34720548f33b tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52910003-4552-8e9b-f201-b8daaa746b59, 'name': SearchDatastore_Task, 'duration_secs': 0.008045} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 719.519015] env[63515]: DEBUG nova.compute.manager [req-938d804f-a41c-491c-bf8c-f6c12ec385be req-b4e78f48-ee28-4219-9c56-b6cc8eafaebb service nova] [instance: 4087b2c4-6ed1-4b68-8b78-a36e34d935b1] Detach interface failed, port_id=81eafe68-fff5-401d-812f-73b07de3ad8b, reason: Instance 4087b2c4-6ed1-4b68-8b78-a36e34d935b1 could not be found. {{(pid=63515) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 719.519566] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ee0256ec-c3ac-41bf-946d-7785b2142de7 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.526060] env[63515]: DEBUG oslo_vmware.api [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Task: {'id': task-1110989, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 719.529154] env[63515]: DEBUG oslo_vmware.api [None req-746f5a16-d8b1-448e-a35c-34720548f33b tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Waiting for the task: (returnval){ [ 719.529154] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]52077f7f-527a-fa68-22c1-37a39789fc39" [ 719.529154] env[63515]: _type = "Task" [ 719.529154] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 719.538062] env[63515]: DEBUG oslo_vmware.api [None req-746f5a16-d8b1-448e-a35c-34720548f33b tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52077f7f-527a-fa68-22c1-37a39789fc39, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 719.568230] env[63515]: DEBUG oslo_concurrency.lockutils [None req-95950a02-20c9-4a3f-9ea8-1f7f19725cc1 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Acquiring lock "4ef8066a-b68c-457b-9964-b1c34bab0fc3" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 719.568481] env[63515]: DEBUG oslo_concurrency.lockutils [None req-95950a02-20c9-4a3f-9ea8-1f7f19725cc1 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Lock "4ef8066a-b68c-457b-9964-b1c34bab0fc3" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 719.677849] env[63515]: DEBUG oslo_vmware.api [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Task: {'id': task-1110988, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 719.742434] env[63515]: INFO nova.compute.manager [-] [instance: 4087b2c4-6ed1-4b68-8b78-a36e34d935b1] Took 1.33 seconds to deallocate network for instance. [ 719.795866] env[63515]: DEBUG oslo_vmware.api [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Task: {'id': task-1110987, 'name': PowerOnVM_Task, 'duration_secs': 0.899168} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 719.796204] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] [instance: 48668736-df27-4f2a-94d9-132f5b49701b] Powered on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 719.796413] env[63515]: INFO nova.compute.manager [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] [instance: 48668736-df27-4f2a-94d9-132f5b49701b] Took 9.86 seconds to spawn the instance on the hypervisor. [ 719.796590] env[63515]: DEBUG nova.compute.manager [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] [instance: 48668736-df27-4f2a-94d9-132f5b49701b] Checking state {{(pid=63515) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 719.797610] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10dbb05d-c8be-4b1d-b056-e1a3c5c6f650 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.970394] env[63515]: DEBUG nova.scheduler.client.report [None req-74f7de08-cdcf-4b98-bfdb-a2dd77c03e6d tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 720.010322] env[63515]: DEBUG oslo_vmware.api [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Task: {'id': task-1110989, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.475205} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 720.010582] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk to [datastore1] 37e6c27e-317b-45d2-bd55-2fd78ccf009f/37e6c27e-317b-45d2-bd55-2fd78ccf009f.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 720.010798] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] [instance: 37e6c27e-317b-45d2-bd55-2fd78ccf009f] Extending root virtual disk to 1048576 {{(pid=63515) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 720.011052] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b0f30c80-bab5-4095-b2b3-c2eb49d0cce2 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.017754] env[63515]: DEBUG oslo_vmware.api [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Waiting for the task: (returnval){ [ 720.017754] env[63515]: value = "task-1110990" [ 720.017754] env[63515]: _type = "Task" [ 720.017754] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 720.025159] env[63515]: DEBUG oslo_vmware.api [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Task: {'id': task-1110990, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 720.039175] env[63515]: DEBUG oslo_vmware.api [None req-746f5a16-d8b1-448e-a35c-34720548f33b tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52077f7f-527a-fa68-22c1-37a39789fc39, 'name': SearchDatastore_Task, 'duration_secs': 0.008509} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 720.039473] env[63515]: DEBUG oslo_concurrency.lockutils [None req-746f5a16-d8b1-448e-a35c-34720548f33b tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 720.039795] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-746f5a16-d8b1-448e-a35c-34720548f33b tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk to [datastore2] b4477e66-ae12-4929-90ed-b7b652e0f207/b4477e66-ae12-4929-90ed-b7b652e0f207.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 720.040079] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e2914800-ff1f-41f3-9c4d-69c092f24f72 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.047168] env[63515]: DEBUG oslo_vmware.api [None req-746f5a16-d8b1-448e-a35c-34720548f33b tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Waiting for the task: (returnval){ [ 720.047168] env[63515]: value = "task-1110991" [ 720.047168] env[63515]: _type = "Task" [ 720.047168] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 720.054690] env[63515]: DEBUG oslo_vmware.api [None req-746f5a16-d8b1-448e-a35c-34720548f33b tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Task: {'id': task-1110991, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 720.176013] env[63515]: DEBUG oslo_vmware.api [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Task: {'id': task-1110988, 'name': ReconfigVM_Task, 'duration_secs': 0.610774} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 720.176307] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] [instance: 452f3ca2-6141-43b2-a77a-c9ab5754192d] Reconfigured VM instance instance-00000028 to attach disk [datastore1] 452f3ca2-6141-43b2-a77a-c9ab5754192d/452f3ca2-6141-43b2-a77a-c9ab5754192d.vmdk or device None with type sparse {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 720.176930] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e9e6e8dc-741a-4255-8618-477b61f96ffd {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.183189] env[63515]: DEBUG oslo_vmware.api [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Waiting for the task: (returnval){ [ 720.183189] env[63515]: value = "task-1110992" [ 720.183189] env[63515]: _type = "Task" [ 720.183189] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 720.191067] env[63515]: DEBUG oslo_vmware.api [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Task: {'id': task-1110992, 'name': Rename_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 720.249919] env[63515]: DEBUG oslo_concurrency.lockutils [None req-bb0fc58f-3c69-4616-a8cc-91b6b5f340c3 tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 720.316673] env[63515]: INFO nova.compute.manager [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] [instance: 48668736-df27-4f2a-94d9-132f5b49701b] Took 33.04 seconds to build instance. [ 720.475680] env[63515]: DEBUG oslo_concurrency.lockutils [None req-74f7de08-cdcf-4b98-bfdb-a2dd77c03e6d tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.504s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 720.476236] env[63515]: DEBUG nova.compute.manager [None req-74f7de08-cdcf-4b98-bfdb-a2dd77c03e6d tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] [instance: b85e9a70-7f5b-4d32-b616-f2a97e3186c8] Start building networks asynchronously for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 720.479564] env[63515]: DEBUG oslo_concurrency.lockutils [None req-e71315b0-9587-4929-82a2-71bebc6fcbfc tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 11.492s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 720.481608] env[63515]: INFO nova.compute.claims [None req-e71315b0-9587-4929-82a2-71bebc6fcbfc tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] [instance: d7e8dcdb-41cb-46fb-8b61-d251e7c2d372] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 720.527172] env[63515]: DEBUG oslo_vmware.api [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Task: {'id': task-1110990, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.062148} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 720.527429] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] [instance: 37e6c27e-317b-45d2-bd55-2fd78ccf009f] Extended root virtual disk {{(pid=63515) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 720.528197] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e8edac4-18fa-42eb-91b6-580e0bb2943f {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.549809] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] [instance: 37e6c27e-317b-45d2-bd55-2fd78ccf009f] Reconfiguring VM instance instance-00000029 to attach disk [datastore1] 37e6c27e-317b-45d2-bd55-2fd78ccf009f/37e6c27e-317b-45d2-bd55-2fd78ccf009f.vmdk or device None with type sparse {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 720.550334] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-99cfeff7-e474-45b3-a15d-0bf4d667f480 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.572820] env[63515]: DEBUG oslo_vmware.api [None req-746f5a16-d8b1-448e-a35c-34720548f33b tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Task: {'id': task-1110991, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 720.573418] env[63515]: DEBUG oslo_vmware.api [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Waiting for the task: (returnval){ [ 720.573418] env[63515]: value = "task-1110993" [ 720.573418] env[63515]: _type = "Task" [ 720.573418] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 720.580970] env[63515]: DEBUG oslo_vmware.api [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Task: {'id': task-1110993, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 720.694655] env[63515]: DEBUG oslo_vmware.api [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Task: {'id': task-1110992, 'name': Rename_Task, 'duration_secs': 0.203956} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 720.694931] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] [instance: 452f3ca2-6141-43b2-a77a-c9ab5754192d] Powering on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 720.695186] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-df406395-79b0-49ad-9bfb-bebbd4ae2e73 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.701019] env[63515]: DEBUG oslo_vmware.api [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Waiting for the task: (returnval){ [ 720.701019] env[63515]: value = "task-1110994" [ 720.701019] env[63515]: _type = "Task" [ 720.701019] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 720.708473] env[63515]: DEBUG oslo_vmware.api [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Task: {'id': task-1110994, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 720.816758] env[63515]: DEBUG oslo_concurrency.lockutils [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Lock "48668736-df27-4f2a-94d9-132f5b49701b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 117.272s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 720.986702] env[63515]: DEBUG nova.compute.utils [None req-74f7de08-cdcf-4b98-bfdb-a2dd77c03e6d tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Using /dev/sd instead of None {{(pid=63515) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 720.990180] env[63515]: DEBUG nova.compute.manager [None req-74f7de08-cdcf-4b98-bfdb-a2dd77c03e6d tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] [instance: b85e9a70-7f5b-4d32-b616-f2a97e3186c8] Not allocating networking since 'none' was specified. {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 721.060744] env[63515]: DEBUG oslo_vmware.api [None req-746f5a16-d8b1-448e-a35c-34720548f33b tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Task: {'id': task-1110991, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.539994} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 721.060847] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-746f5a16-d8b1-448e-a35c-34720548f33b tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk to [datastore2] b4477e66-ae12-4929-90ed-b7b652e0f207/b4477e66-ae12-4929-90ed-b7b652e0f207.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 721.061066] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-746f5a16-d8b1-448e-a35c-34720548f33b tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] [instance: b4477e66-ae12-4929-90ed-b7b652e0f207] Extending root virtual disk to 1048576 {{(pid=63515) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 721.061322] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-50d5a298-65ed-44e8-927f-b0727f821c78 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.068095] env[63515]: DEBUG oslo_vmware.api [None req-746f5a16-d8b1-448e-a35c-34720548f33b tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Waiting for the task: (returnval){ [ 721.068095] env[63515]: value = "task-1110995" [ 721.068095] env[63515]: _type = "Task" [ 721.068095] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 721.076596] env[63515]: DEBUG oslo_vmware.api [None req-746f5a16-d8b1-448e-a35c-34720548f33b tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Task: {'id': task-1110995, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 721.085228] env[63515]: DEBUG oslo_vmware.api [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Task: {'id': task-1110993, 'name': ReconfigVM_Task, 'duration_secs': 0.261723} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 721.085499] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] [instance: 37e6c27e-317b-45d2-bd55-2fd78ccf009f] Reconfigured VM instance instance-00000029 to attach disk [datastore1] 37e6c27e-317b-45d2-bd55-2fd78ccf009f/37e6c27e-317b-45d2-bd55-2fd78ccf009f.vmdk or device None with type sparse {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 721.086226] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-976fe502-60d9-449f-97b2-f0000ded7683 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.091489] env[63515]: DEBUG oslo_vmware.api [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Waiting for the task: (returnval){ [ 721.091489] env[63515]: value = "task-1110996" [ 721.091489] env[63515]: _type = "Task" [ 721.091489] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 721.099259] env[63515]: DEBUG oslo_vmware.api [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Task: {'id': task-1110996, 'name': Rename_Task} progress is 5%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 721.210771] env[63515]: DEBUG oslo_vmware.api [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Task: {'id': task-1110994, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 721.320572] env[63515]: DEBUG nova.compute.manager [None req-99ae4d48-1fae-4ace-913c-d249ca72bbeb tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] [instance: d2a4f66a-f1a2-4f0d-826d-0b7c4f04bd23] Starting instance... {{(pid=63515) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 721.491739] env[63515]: DEBUG nova.compute.manager [None req-74f7de08-cdcf-4b98-bfdb-a2dd77c03e6d tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] [instance: b85e9a70-7f5b-4d32-b616-f2a97e3186c8] Start building block device mappings for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 721.580416] env[63515]: DEBUG oslo_vmware.api [None req-746f5a16-d8b1-448e-a35c-34720548f33b tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Task: {'id': task-1110995, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.065321} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 721.582922] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-746f5a16-d8b1-448e-a35c-34720548f33b tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] [instance: b4477e66-ae12-4929-90ed-b7b652e0f207] Extended root virtual disk {{(pid=63515) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 721.584248] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-390c658d-ca9a-4edc-b0d2-dcbfe14069a1 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.603673] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-746f5a16-d8b1-448e-a35c-34720548f33b tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] [instance: b4477e66-ae12-4929-90ed-b7b652e0f207] Reconfiguring VM instance instance-0000002a to attach disk [datastore2] b4477e66-ae12-4929-90ed-b7b652e0f207/b4477e66-ae12-4929-90ed-b7b652e0f207.vmdk or device None with type sparse {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 721.609865] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-57e5dc7c-5355-4616-a9bf-d1028d3cb209 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.628937] env[63515]: DEBUG oslo_vmware.api [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Task: {'id': task-1110996, 'name': Rename_Task, 'duration_secs': 0.141675} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 721.630352] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] [instance: 37e6c27e-317b-45d2-bd55-2fd78ccf009f] Powering on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 721.630640] env[63515]: DEBUG oslo_vmware.api [None req-746f5a16-d8b1-448e-a35c-34720548f33b tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Waiting for the task: (returnval){ [ 721.630640] env[63515]: value = "task-1110997" [ 721.630640] env[63515]: _type = "Task" [ 721.630640] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 721.630821] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1300cd20-3314-45b0-8fdd-1bdc87f4357e {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.642753] env[63515]: DEBUG oslo_vmware.api [None req-746f5a16-d8b1-448e-a35c-34720548f33b tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Task: {'id': task-1110997, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 721.643904] env[63515]: DEBUG oslo_vmware.api [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Waiting for the task: (returnval){ [ 721.643904] env[63515]: value = "task-1110998" [ 721.643904] env[63515]: _type = "Task" [ 721.643904] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 721.652074] env[63515]: DEBUG oslo_vmware.api [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Task: {'id': task-1110998, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 721.714319] env[63515]: DEBUG oslo_vmware.api [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Task: {'id': task-1110994, 'name': PowerOnVM_Task} progress is 88%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 721.842057] env[63515]: DEBUG oslo_concurrency.lockutils [None req-99ae4d48-1fae-4ace-913c-d249ca72bbeb tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 721.949254] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80667de4-3061-4214-b168-52e414c2152b {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.957377] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4cc2d79f-37fc-497e-8b64-642eb04ad0e1 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.987598] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d24a3f3e-084c-4035-b1ef-32f88ac9f916 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.996121] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07ad2a53-7fac-4c14-9969-a7330dc17363 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.013778] env[63515]: DEBUG nova.compute.provider_tree [None req-e71315b0-9587-4929-82a2-71bebc6fcbfc tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 722.142255] env[63515]: DEBUG oslo_vmware.api [None req-746f5a16-d8b1-448e-a35c-34720548f33b tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Task: {'id': task-1110997, 'name': ReconfigVM_Task, 'duration_secs': 0.307821} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 722.142517] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-746f5a16-d8b1-448e-a35c-34720548f33b tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] [instance: b4477e66-ae12-4929-90ed-b7b652e0f207] Reconfigured VM instance instance-0000002a to attach disk [datastore2] b4477e66-ae12-4929-90ed-b7b652e0f207/b4477e66-ae12-4929-90ed-b7b652e0f207.vmdk or device None with type sparse {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 722.143137] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6996455a-746b-4f72-b71f-f50208e6d7fa {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.153371] env[63515]: DEBUG oslo_vmware.api [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Task: {'id': task-1110998, 'name': PowerOnVM_Task, 'duration_secs': 0.458122} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 722.154356] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] [instance: 37e6c27e-317b-45d2-bd55-2fd78ccf009f] Powered on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 722.154548] env[63515]: INFO nova.compute.manager [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] [instance: 37e6c27e-317b-45d2-bd55-2fd78ccf009f] Took 6.97 seconds to spawn the instance on the hypervisor. [ 722.154790] env[63515]: DEBUG nova.compute.manager [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] [instance: 37e6c27e-317b-45d2-bd55-2fd78ccf009f] Checking state {{(pid=63515) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 722.155965] env[63515]: DEBUG oslo_vmware.api [None req-746f5a16-d8b1-448e-a35c-34720548f33b tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Waiting for the task: (returnval){ [ 722.155965] env[63515]: value = "task-1110999" [ 722.155965] env[63515]: _type = "Task" [ 722.155965] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 722.155965] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c8bf32f-ac5c-4d19-ba1d-208087490002 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.165758] env[63515]: DEBUG oslo_vmware.api [None req-746f5a16-d8b1-448e-a35c-34720548f33b tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Task: {'id': task-1110999, 'name': Rename_Task} progress is 6%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 722.214296] env[63515]: DEBUG oslo_vmware.api [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Task: {'id': task-1110994, 'name': PowerOnVM_Task, 'duration_secs': 1.182994} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 722.214557] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] [instance: 452f3ca2-6141-43b2-a77a-c9ab5754192d] Powered on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 722.214813] env[63515]: INFO nova.compute.manager [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] [instance: 452f3ca2-6141-43b2-a77a-c9ab5754192d] Took 9.68 seconds to spawn the instance on the hypervisor. [ 722.215355] env[63515]: DEBUG nova.compute.manager [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] [instance: 452f3ca2-6141-43b2-a77a-c9ab5754192d] Checking state {{(pid=63515) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 722.215776] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9922b89c-9138-4c84-9a18-56d4fc565a26 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.506181] env[63515]: DEBUG nova.compute.manager [None req-74f7de08-cdcf-4b98-bfdb-a2dd77c03e6d tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] [instance: b85e9a70-7f5b-4d32-b616-f2a97e3186c8] Start spawning the instance on the hypervisor. {{(pid=63515) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 722.516246] env[63515]: DEBUG nova.scheduler.client.report [None req-e71315b0-9587-4929-82a2-71bebc6fcbfc tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 722.530596] env[63515]: DEBUG nova.virt.hardware [None req-74f7de08-cdcf-4b98-bfdb-a2dd77c03e6d tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T02:52:14Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T02:51:56Z,direct_url=,disk_format='vmdk',id=8a120570-cb06-4099-b262-554ca0ad15c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b569255cc43e42e7ae3f2b4ad37c6ef4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T02:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 722.530842] env[63515]: DEBUG nova.virt.hardware [None req-74f7de08-cdcf-4b98-bfdb-a2dd77c03e6d tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Flavor limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 722.530998] env[63515]: DEBUG nova.virt.hardware [None req-74f7de08-cdcf-4b98-bfdb-a2dd77c03e6d tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Image limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 722.531199] env[63515]: DEBUG nova.virt.hardware [None req-74f7de08-cdcf-4b98-bfdb-a2dd77c03e6d tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Flavor pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 722.531343] env[63515]: DEBUG nova.virt.hardware [None req-74f7de08-cdcf-4b98-bfdb-a2dd77c03e6d tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Image pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 722.531487] env[63515]: DEBUG nova.virt.hardware [None req-74f7de08-cdcf-4b98-bfdb-a2dd77c03e6d tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 722.531690] env[63515]: DEBUG nova.virt.hardware [None req-74f7de08-cdcf-4b98-bfdb-a2dd77c03e6d tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 722.531847] env[63515]: DEBUG nova.virt.hardware [None req-74f7de08-cdcf-4b98-bfdb-a2dd77c03e6d tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 722.532028] env[63515]: DEBUG nova.virt.hardware [None req-74f7de08-cdcf-4b98-bfdb-a2dd77c03e6d tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Got 1 possible topologies {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 722.532215] env[63515]: DEBUG nova.virt.hardware [None req-74f7de08-cdcf-4b98-bfdb-a2dd77c03e6d tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 722.532385] env[63515]: DEBUG nova.virt.hardware [None req-74f7de08-cdcf-4b98-bfdb-a2dd77c03e6d tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 722.533266] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dfe5c2f0-71cb-4328-9b31-80c73a02efb7 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.541721] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-544d1052-9a4a-42be-8be6-66e789a8164e {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.555037] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-74f7de08-cdcf-4b98-bfdb-a2dd77c03e6d tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] [instance: b85e9a70-7f5b-4d32-b616-f2a97e3186c8] Instance VIF info [] {{(pid=63515) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 722.560567] env[63515]: DEBUG oslo.service.loopingcall [None req-74f7de08-cdcf-4b98-bfdb-a2dd77c03e6d tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 722.560795] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b85e9a70-7f5b-4d32-b616-f2a97e3186c8] Creating VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 722.560992] env[63515]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-acca966e-360c-4684-a43d-3769285e2a3d {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.577157] env[63515]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 722.577157] env[63515]: value = "task-1111000" [ 722.577157] env[63515]: _type = "Task" [ 722.577157] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 722.584480] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1111000, 'name': CreateVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 722.668030] env[63515]: DEBUG oslo_vmware.api [None req-746f5a16-d8b1-448e-a35c-34720548f33b tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Task: {'id': task-1110999, 'name': Rename_Task, 'duration_secs': 0.130835} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 722.668384] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-746f5a16-d8b1-448e-a35c-34720548f33b tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] [instance: b4477e66-ae12-4929-90ed-b7b652e0f207] Powering on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 722.668653] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-30853f00-e21d-46ba-8f29-8b48d3bce654 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.676631] env[63515]: INFO nova.compute.manager [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] [instance: 37e6c27e-317b-45d2-bd55-2fd78ccf009f] Took 27.06 seconds to build instance. [ 722.679428] env[63515]: DEBUG oslo_vmware.api [None req-746f5a16-d8b1-448e-a35c-34720548f33b tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Waiting for the task: (returnval){ [ 722.679428] env[63515]: value = "task-1111001" [ 722.679428] env[63515]: _type = "Task" [ 722.679428] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 722.687486] env[63515]: DEBUG oslo_vmware.api [None req-746f5a16-d8b1-448e-a35c-34720548f33b tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Task: {'id': task-1111001, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 722.734916] env[63515]: INFO nova.compute.manager [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] [instance: 452f3ca2-6141-43b2-a77a-c9ab5754192d] Took 31.64 seconds to build instance. [ 723.021502] env[63515]: DEBUG oslo_concurrency.lockutils [None req-e71315b0-9587-4929-82a2-71bebc6fcbfc tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.542s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 723.022076] env[63515]: DEBUG nova.compute.manager [None req-e71315b0-9587-4929-82a2-71bebc6fcbfc tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] [instance: d7e8dcdb-41cb-46fb-8b61-d251e7c2d372] Start building networks asynchronously for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 723.024758] env[63515]: DEBUG oslo_concurrency.lockutils [None req-dcda8d78-de0d-44e9-b61f-e84b4a49533f tempest-ServersTestManualDisk-1114050206 tempest-ServersTestManualDisk-1114050206-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.655s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 723.026285] env[63515]: INFO nova.compute.claims [None req-dcda8d78-de0d-44e9-b61f-e84b4a49533f tempest-ServersTestManualDisk-1114050206 tempest-ServersTestManualDisk-1114050206-project-member] [instance: e91aa479-1540-4950-851b-b2409e5f89f1] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 723.089243] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1111000, 'name': CreateVM_Task, 'duration_secs': 0.429496} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 723.089409] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b85e9a70-7f5b-4d32-b616-f2a97e3186c8] Created VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 723.089827] env[63515]: DEBUG oslo_concurrency.lockutils [None req-74f7de08-cdcf-4b98-bfdb-a2dd77c03e6d tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 723.089993] env[63515]: DEBUG oslo_concurrency.lockutils [None req-74f7de08-cdcf-4b98-bfdb-a2dd77c03e6d tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 723.090331] env[63515]: DEBUG oslo_concurrency.lockutils [None req-74f7de08-cdcf-4b98-bfdb-a2dd77c03e6d tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 723.090577] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c772abad-e782-4ecb-8a5f-db0192bc519e {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.095640] env[63515]: DEBUG oslo_vmware.api [None req-74f7de08-cdcf-4b98-bfdb-a2dd77c03e6d tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Waiting for the task: (returnval){ [ 723.095640] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]529ace97-594f-3d27-0783-72012c7af9ed" [ 723.095640] env[63515]: _type = "Task" [ 723.095640] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 723.107817] env[63515]: DEBUG oslo_vmware.api [None req-74f7de08-cdcf-4b98-bfdb-a2dd77c03e6d tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]529ace97-594f-3d27-0783-72012c7af9ed, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 723.181801] env[63515]: DEBUG oslo_concurrency.lockutils [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Lock "37e6c27e-317b-45d2-bd55-2fd78ccf009f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 119.579s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 723.195769] env[63515]: DEBUG oslo_vmware.api [None req-746f5a16-d8b1-448e-a35c-34720548f33b tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Task: {'id': task-1111001, 'name': PowerOnVM_Task, 'duration_secs': 0.470086} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 723.196433] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-746f5a16-d8b1-448e-a35c-34720548f33b tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] [instance: b4477e66-ae12-4929-90ed-b7b652e0f207] Powered on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 723.196686] env[63515]: INFO nova.compute.manager [None req-746f5a16-d8b1-448e-a35c-34720548f33b tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] [instance: b4477e66-ae12-4929-90ed-b7b652e0f207] Took 5.38 seconds to spawn the instance on the hypervisor. [ 723.196873] env[63515]: DEBUG nova.compute.manager [None req-746f5a16-d8b1-448e-a35c-34720548f33b tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] [instance: b4477e66-ae12-4929-90ed-b7b652e0f207] Checking state {{(pid=63515) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 723.197872] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3212637-9c87-45d7-a0c5-d00b8ad8f6a8 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.237274] env[63515]: DEBUG oslo_concurrency.lockutils [None req-ae72c23a-be38-4811-80ab-db4e77765072 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Lock "452f3ca2-6141-43b2-a77a-c9ab5754192d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 119.662s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 723.531647] env[63515]: DEBUG nova.compute.utils [None req-e71315b0-9587-4929-82a2-71bebc6fcbfc tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Using /dev/sd instead of None {{(pid=63515) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 723.535297] env[63515]: DEBUG nova.compute.manager [None req-e71315b0-9587-4929-82a2-71bebc6fcbfc tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] [instance: d7e8dcdb-41cb-46fb-8b61-d251e7c2d372] Allocating IP information in the background. {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 723.536384] env[63515]: DEBUG nova.network.neutron [None req-e71315b0-9587-4929-82a2-71bebc6fcbfc tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] [instance: d7e8dcdb-41cb-46fb-8b61-d251e7c2d372] allocate_for_instance() {{(pid=63515) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 723.578512] env[63515]: DEBUG nova.policy [None req-e71315b0-9587-4929-82a2-71bebc6fcbfc tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c7dc08223aa943e1968a9c62da3a117d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '12a90929c98e40d9a1263929e2b97532', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63515) authorize /opt/stack/nova/nova/policy.py:201}} [ 723.606609] env[63515]: DEBUG oslo_vmware.api [None req-74f7de08-cdcf-4b98-bfdb-a2dd77c03e6d tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]529ace97-594f-3d27-0783-72012c7af9ed, 'name': SearchDatastore_Task, 'duration_secs': 0.011536} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 723.606978] env[63515]: DEBUG oslo_concurrency.lockutils [None req-74f7de08-cdcf-4b98-bfdb-a2dd77c03e6d tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 723.607265] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-74f7de08-cdcf-4b98-bfdb-a2dd77c03e6d tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] [instance: b85e9a70-7f5b-4d32-b616-f2a97e3186c8] Processing image 8a120570-cb06-4099-b262-554ca0ad15c5 {{(pid=63515) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 723.607547] env[63515]: DEBUG oslo_concurrency.lockutils [None req-74f7de08-cdcf-4b98-bfdb-a2dd77c03e6d tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 723.607726] env[63515]: DEBUG oslo_concurrency.lockutils [None req-74f7de08-cdcf-4b98-bfdb-a2dd77c03e6d tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 723.607918] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-74f7de08-cdcf-4b98-bfdb-a2dd77c03e6d tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 723.608179] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d1c3389a-f21a-4bdb-a904-3f1f84d389b4 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.618026] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-74f7de08-cdcf-4b98-bfdb-a2dd77c03e6d tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 723.618193] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-74f7de08-cdcf-4b98-bfdb-a2dd77c03e6d tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63515) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 723.619099] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-96b63ee6-528b-406c-a418-361a1c9d38cc {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.624055] env[63515]: DEBUG oslo_vmware.api [None req-74f7de08-cdcf-4b98-bfdb-a2dd77c03e6d tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Waiting for the task: (returnval){ [ 723.624055] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]52fb1522-133a-ed36-b07e-076687c9d575" [ 723.624055] env[63515]: _type = "Task" [ 723.624055] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 723.633258] env[63515]: DEBUG oslo_vmware.api [None req-74f7de08-cdcf-4b98-bfdb-a2dd77c03e6d tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52fb1522-133a-ed36-b07e-076687c9d575, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 723.671507] env[63515]: DEBUG oslo_concurrency.lockutils [None req-9695633a-0f62-476f-8466-deb3365d2b47 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Acquiring lock "48668736-df27-4f2a-94d9-132f5b49701b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 723.671760] env[63515]: DEBUG oslo_concurrency.lockutils [None req-9695633a-0f62-476f-8466-deb3365d2b47 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Lock "48668736-df27-4f2a-94d9-132f5b49701b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 723.671965] env[63515]: DEBUG oslo_concurrency.lockutils [None req-9695633a-0f62-476f-8466-deb3365d2b47 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Acquiring lock "48668736-df27-4f2a-94d9-132f5b49701b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 723.672156] env[63515]: DEBUG oslo_concurrency.lockutils [None req-9695633a-0f62-476f-8466-deb3365d2b47 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Lock "48668736-df27-4f2a-94d9-132f5b49701b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 723.672324] env[63515]: DEBUG oslo_concurrency.lockutils [None req-9695633a-0f62-476f-8466-deb3365d2b47 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Lock "48668736-df27-4f2a-94d9-132f5b49701b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 723.675352] env[63515]: INFO nova.compute.manager [None req-9695633a-0f62-476f-8466-deb3365d2b47 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] [instance: 48668736-df27-4f2a-94d9-132f5b49701b] Terminating instance [ 723.678056] env[63515]: DEBUG nova.compute.manager [None req-9695633a-0f62-476f-8466-deb3365d2b47 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] [instance: 48668736-df27-4f2a-94d9-132f5b49701b] Start destroying the instance on the hypervisor. {{(pid=63515) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 723.678056] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-9695633a-0f62-476f-8466-deb3365d2b47 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] [instance: 48668736-df27-4f2a-94d9-132f5b49701b] Destroying instance {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 723.678848] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed94fedf-0522-4120-82e0-7548fd72efdc {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.689149] env[63515]: DEBUG nova.compute.manager [None req-5bd41396-218d-4ce7-9765-053dcc695301 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] [instance: 6688cdaa-29ba-413a-8131-4f834cdb70e4] Starting instance... {{(pid=63515) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 723.689149] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-9695633a-0f62-476f-8466-deb3365d2b47 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] [instance: 48668736-df27-4f2a-94d9-132f5b49701b] Powering off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 723.689273] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7fef9926-cef4-4323-b5d8-393f3ab2f5f4 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.696350] env[63515]: DEBUG oslo_vmware.api [None req-9695633a-0f62-476f-8466-deb3365d2b47 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Waiting for the task: (returnval){ [ 723.696350] env[63515]: value = "task-1111002" [ 723.696350] env[63515]: _type = "Task" [ 723.696350] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 723.704166] env[63515]: DEBUG oslo_vmware.api [None req-9695633a-0f62-476f-8466-deb3365d2b47 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Task: {'id': task-1111002, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 723.713491] env[63515]: INFO nova.compute.manager [None req-746f5a16-d8b1-448e-a35c-34720548f33b tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] [instance: b4477e66-ae12-4929-90ed-b7b652e0f207] Took 26.99 seconds to build instance. [ 723.740065] env[63515]: DEBUG nova.compute.manager [None req-dfa187f2-adf0-4e24-9cf4-7df6d0016e12 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] [instance: 38d6d030-06b0-4185-904d-44a038b3a752] Starting instance... {{(pid=63515) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 723.862778] env[63515]: DEBUG nova.network.neutron [None req-e71315b0-9587-4929-82a2-71bebc6fcbfc tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] [instance: d7e8dcdb-41cb-46fb-8b61-d251e7c2d372] Successfully created port: 9378f59d-f9bc-4c2e-8336-2c5710f987cf {{(pid=63515) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 724.036814] env[63515]: DEBUG nova.compute.manager [None req-e71315b0-9587-4929-82a2-71bebc6fcbfc tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] [instance: d7e8dcdb-41cb-46fb-8b61-d251e7c2d372] Start building block device mappings for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 724.136289] env[63515]: DEBUG oslo_vmware.api [None req-74f7de08-cdcf-4b98-bfdb-a2dd77c03e6d tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52fb1522-133a-ed36-b07e-076687c9d575, 'name': SearchDatastore_Task, 'duration_secs': 0.039127} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 724.139998] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4e259b65-ee9f-4d08-996f-5eb3ca1cf742 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.145722] env[63515]: DEBUG oslo_vmware.api [None req-74f7de08-cdcf-4b98-bfdb-a2dd77c03e6d tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Waiting for the task: (returnval){ [ 724.145722] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]5220549c-5987-ac7a-dc6c-f6d7292c47ed" [ 724.145722] env[63515]: _type = "Task" [ 724.145722] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 724.160199] env[63515]: DEBUG oslo_vmware.api [None req-74f7de08-cdcf-4b98-bfdb-a2dd77c03e6d tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]5220549c-5987-ac7a-dc6c-f6d7292c47ed, 'name': SearchDatastore_Task, 'duration_secs': 0.008666} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 724.160199] env[63515]: DEBUG oslo_concurrency.lockutils [None req-74f7de08-cdcf-4b98-bfdb-a2dd77c03e6d tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 724.160199] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-74f7de08-cdcf-4b98-bfdb-a2dd77c03e6d tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk to [datastore2] b85e9a70-7f5b-4d32-b616-f2a97e3186c8/b85e9a70-7f5b-4d32-b616-f2a97e3186c8.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 724.160199] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6ca8d412-7d05-4f18-9cff-84ed682d91e4 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.166064] env[63515]: DEBUG oslo_vmware.api [None req-74f7de08-cdcf-4b98-bfdb-a2dd77c03e6d tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Waiting for the task: (returnval){ [ 724.166064] env[63515]: value = "task-1111003" [ 724.166064] env[63515]: _type = "Task" [ 724.166064] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 724.176828] env[63515]: DEBUG oslo_vmware.api [None req-74f7de08-cdcf-4b98-bfdb-a2dd77c03e6d tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Task: {'id': task-1111003, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 724.207605] env[63515]: DEBUG oslo_concurrency.lockutils [None req-5bd41396-218d-4ce7-9765-053dcc695301 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 724.209522] env[63515]: DEBUG oslo_vmware.api [None req-9695633a-0f62-476f-8466-deb3365d2b47 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Task: {'id': task-1111002, 'name': PowerOffVM_Task, 'duration_secs': 0.423903} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 724.209800] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-9695633a-0f62-476f-8466-deb3365d2b47 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] [instance: 48668736-df27-4f2a-94d9-132f5b49701b] Powered off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 724.209998] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-9695633a-0f62-476f-8466-deb3365d2b47 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] [instance: 48668736-df27-4f2a-94d9-132f5b49701b] Unregistering the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 724.210279] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d843a903-4f73-4fc1-b211-75cb36de6a9f {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.215073] env[63515]: DEBUG oslo_concurrency.lockutils [None req-746f5a16-d8b1-448e-a35c-34720548f33b tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Lock "b4477e66-ae12-4929-90ed-b7b652e0f207" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 107.644s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 724.258604] env[63515]: DEBUG oslo_concurrency.lockutils [None req-dfa187f2-adf0-4e24-9cf4-7df6d0016e12 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 724.271644] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-9695633a-0f62-476f-8466-deb3365d2b47 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] [instance: 48668736-df27-4f2a-94d9-132f5b49701b] Unregistered the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 724.271902] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-9695633a-0f62-476f-8466-deb3365d2b47 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] [instance: 48668736-df27-4f2a-94d9-132f5b49701b] Deleting contents of the VM from datastore datastore1 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 724.272094] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-9695633a-0f62-476f-8466-deb3365d2b47 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Deleting the datastore file [datastore1] 48668736-df27-4f2a-94d9-132f5b49701b {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 724.272348] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-9fb53c05-c2c7-4936-80ac-ae9a5ecbdaa1 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.281409] env[63515]: DEBUG oslo_vmware.api [None req-9695633a-0f62-476f-8466-deb3365d2b47 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Waiting for the task: (returnval){ [ 724.281409] env[63515]: value = "task-1111005" [ 724.281409] env[63515]: _type = "Task" [ 724.281409] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 724.289496] env[63515]: DEBUG oslo_vmware.api [None req-9695633a-0f62-476f-8466-deb3365d2b47 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Task: {'id': task-1111005, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 724.535322] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-badacd5d-a402-47d1-b259-21ac4a3c3ee4 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.548964] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b830fcde-6ea0-49a9-987e-4f577de669cf {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.582586] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7bbb7b8e-77dd-4355-b907-8b8bb3d9a727 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.593168] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d708772-982a-434f-bcd8-026af4517ba0 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.610223] env[63515]: DEBUG nova.compute.provider_tree [None req-dcda8d78-de0d-44e9-b61f-e84b4a49533f tempest-ServersTestManualDisk-1114050206 tempest-ServersTestManualDisk-1114050206-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 724.677560] env[63515]: DEBUG oslo_vmware.api [None req-74f7de08-cdcf-4b98-bfdb-a2dd77c03e6d tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Task: {'id': task-1111003, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.45217} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 724.677852] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-74f7de08-cdcf-4b98-bfdb-a2dd77c03e6d tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk to [datastore2] b85e9a70-7f5b-4d32-b616-f2a97e3186c8/b85e9a70-7f5b-4d32-b616-f2a97e3186c8.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 724.678141] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-74f7de08-cdcf-4b98-bfdb-a2dd77c03e6d tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] [instance: b85e9a70-7f5b-4d32-b616-f2a97e3186c8] Extending root virtual disk to 1048576 {{(pid=63515) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 724.678423] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-6570f616-63c3-4405-bea8-aca99992f13e {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.685305] env[63515]: DEBUG oslo_vmware.api [None req-74f7de08-cdcf-4b98-bfdb-a2dd77c03e6d tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Waiting for the task: (returnval){ [ 724.685305] env[63515]: value = "task-1111006" [ 724.685305] env[63515]: _type = "Task" [ 724.685305] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 724.694221] env[63515]: DEBUG oslo_vmware.api [None req-74f7de08-cdcf-4b98-bfdb-a2dd77c03e6d tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Task: {'id': task-1111006, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 724.717990] env[63515]: DEBUG nova.compute.manager [None req-750d4038-2a95-4b14-b9d7-c9c6803c64b6 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] [instance: 8eb17506-ff93-4d25-b9af-ec5886569e65] Starting instance... {{(pid=63515) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 724.791562] env[63515]: DEBUG oslo_vmware.api [None req-9695633a-0f62-476f-8466-deb3365d2b47 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Task: {'id': task-1111005, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.331447} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 724.792159] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-9695633a-0f62-476f-8466-deb3365d2b47 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Deleted the datastore file {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 724.792388] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-9695633a-0f62-476f-8466-deb3365d2b47 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] [instance: 48668736-df27-4f2a-94d9-132f5b49701b] Deleted contents of the VM from datastore datastore1 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 724.792597] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-9695633a-0f62-476f-8466-deb3365d2b47 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] [instance: 48668736-df27-4f2a-94d9-132f5b49701b] Instance destroyed {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 724.792769] env[63515]: INFO nova.compute.manager [None req-9695633a-0f62-476f-8466-deb3365d2b47 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] [instance: 48668736-df27-4f2a-94d9-132f5b49701b] Took 1.12 seconds to destroy the instance on the hypervisor. [ 724.793023] env[63515]: DEBUG oslo.service.loopingcall [None req-9695633a-0f62-476f-8466-deb3365d2b47 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 724.793219] env[63515]: DEBUG nova.compute.manager [-] [instance: 48668736-df27-4f2a-94d9-132f5b49701b] Deallocating network for instance {{(pid=63515) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 724.793317] env[63515]: DEBUG nova.network.neutron [-] [instance: 48668736-df27-4f2a-94d9-132f5b49701b] deallocate_for_instance() {{(pid=63515) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 725.026497] env[63515]: DEBUG nova.compute.manager [req-8eb1027a-3f32-4469-87fb-5381b69191ac req-a8f70487-98b7-4d77-9275-ea5785a7d05b service nova] [instance: 48668736-df27-4f2a-94d9-132f5b49701b] Received event network-vif-deleted-9379e602-f7be-4c75-a6e6-c42a8702f6da {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 725.026671] env[63515]: INFO nova.compute.manager [req-8eb1027a-3f32-4469-87fb-5381b69191ac req-a8f70487-98b7-4d77-9275-ea5785a7d05b service nova] [instance: 48668736-df27-4f2a-94d9-132f5b49701b] Neutron deleted interface 9379e602-f7be-4c75-a6e6-c42a8702f6da; detaching it from the instance and deleting it from the info cache [ 725.026842] env[63515]: DEBUG nova.network.neutron [req-8eb1027a-3f32-4469-87fb-5381b69191ac req-a8f70487-98b7-4d77-9275-ea5785a7d05b service nova] [instance: 48668736-df27-4f2a-94d9-132f5b49701b] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 725.055291] env[63515]: DEBUG nova.compute.manager [None req-e71315b0-9587-4929-82a2-71bebc6fcbfc tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] [instance: d7e8dcdb-41cb-46fb-8b61-d251e7c2d372] Start spawning the instance on the hypervisor. {{(pid=63515) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 725.083202] env[63515]: DEBUG nova.virt.hardware [None req-e71315b0-9587-4929-82a2-71bebc6fcbfc tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T02:52:14Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T02:51:56Z,direct_url=,disk_format='vmdk',id=8a120570-cb06-4099-b262-554ca0ad15c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b569255cc43e42e7ae3f2b4ad37c6ef4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T02:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 725.083475] env[63515]: DEBUG nova.virt.hardware [None req-e71315b0-9587-4929-82a2-71bebc6fcbfc tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Flavor limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 725.083622] env[63515]: DEBUG nova.virt.hardware [None req-e71315b0-9587-4929-82a2-71bebc6fcbfc tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Image limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 725.083821] env[63515]: DEBUG nova.virt.hardware [None req-e71315b0-9587-4929-82a2-71bebc6fcbfc tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Flavor pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 725.083973] env[63515]: DEBUG nova.virt.hardware [None req-e71315b0-9587-4929-82a2-71bebc6fcbfc tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Image pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 725.084137] env[63515]: DEBUG nova.virt.hardware [None req-e71315b0-9587-4929-82a2-71bebc6fcbfc tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 725.084354] env[63515]: DEBUG nova.virt.hardware [None req-e71315b0-9587-4929-82a2-71bebc6fcbfc tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 725.084507] env[63515]: DEBUG nova.virt.hardware [None req-e71315b0-9587-4929-82a2-71bebc6fcbfc tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 725.084762] env[63515]: DEBUG nova.virt.hardware [None req-e71315b0-9587-4929-82a2-71bebc6fcbfc tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Got 1 possible topologies {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 725.084973] env[63515]: DEBUG nova.virt.hardware [None req-e71315b0-9587-4929-82a2-71bebc6fcbfc tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 725.085181] env[63515]: DEBUG nova.virt.hardware [None req-e71315b0-9587-4929-82a2-71bebc6fcbfc tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 725.086080] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30464d13-0038-4e1e-ac6a-37f3fd3b2aa4 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.094382] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d319189-3246-428e-9f04-2a57fe790a2f {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.113224] env[63515]: DEBUG nova.scheduler.client.report [None req-dcda8d78-de0d-44e9-b61f-e84b4a49533f tempest-ServersTestManualDisk-1114050206 tempest-ServersTestManualDisk-1114050206-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 725.199310] env[63515]: DEBUG oslo_vmware.api [None req-74f7de08-cdcf-4b98-bfdb-a2dd77c03e6d tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Task: {'id': task-1111006, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.062787} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 725.199990] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-74f7de08-cdcf-4b98-bfdb-a2dd77c03e6d tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] [instance: b85e9a70-7f5b-4d32-b616-f2a97e3186c8] Extended root virtual disk {{(pid=63515) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 725.201637] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-729b812a-aa40-448e-abc0-22eb3b7fd39a {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.224108] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-74f7de08-cdcf-4b98-bfdb-a2dd77c03e6d tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] [instance: b85e9a70-7f5b-4d32-b616-f2a97e3186c8] Reconfiguring VM instance instance-0000002b to attach disk [datastore2] b85e9a70-7f5b-4d32-b616-f2a97e3186c8/b85e9a70-7f5b-4d32-b616-f2a97e3186c8.vmdk or device None with type sparse {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 725.228028] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-14c55c1b-4e7a-41e8-9db7-6d0815829c1f {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.247797] env[63515]: DEBUG oslo_vmware.api [None req-74f7de08-cdcf-4b98-bfdb-a2dd77c03e6d tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Waiting for the task: (returnval){ [ 725.247797] env[63515]: value = "task-1111007" [ 725.247797] env[63515]: _type = "Task" [ 725.247797] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 725.256392] env[63515]: DEBUG oslo_vmware.api [None req-74f7de08-cdcf-4b98-bfdb-a2dd77c03e6d tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Task: {'id': task-1111007, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 725.260569] env[63515]: DEBUG oslo_concurrency.lockutils [None req-750d4038-2a95-4b14-b9d7-c9c6803c64b6 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 725.307294] env[63515]: DEBUG nova.compute.manager [req-a496ed4b-aaaf-4a9c-a1ff-94a31a5c23be req-8ea8637d-1aaa-4fe1-9ad1-de5d9c03b201 service nova] [instance: d7e8dcdb-41cb-46fb-8b61-d251e7c2d372] Received event network-vif-plugged-9378f59d-f9bc-4c2e-8336-2c5710f987cf {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 725.307504] env[63515]: DEBUG oslo_concurrency.lockutils [req-a496ed4b-aaaf-4a9c-a1ff-94a31a5c23be req-8ea8637d-1aaa-4fe1-9ad1-de5d9c03b201 service nova] Acquiring lock "d7e8dcdb-41cb-46fb-8b61-d251e7c2d372-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 725.307707] env[63515]: DEBUG oslo_concurrency.lockutils [req-a496ed4b-aaaf-4a9c-a1ff-94a31a5c23be req-8ea8637d-1aaa-4fe1-9ad1-de5d9c03b201 service nova] Lock "d7e8dcdb-41cb-46fb-8b61-d251e7c2d372-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 725.307872] env[63515]: DEBUG oslo_concurrency.lockutils [req-a496ed4b-aaaf-4a9c-a1ff-94a31a5c23be req-8ea8637d-1aaa-4fe1-9ad1-de5d9c03b201 service nova] Lock "d7e8dcdb-41cb-46fb-8b61-d251e7c2d372-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 725.308044] env[63515]: DEBUG nova.compute.manager [req-a496ed4b-aaaf-4a9c-a1ff-94a31a5c23be req-8ea8637d-1aaa-4fe1-9ad1-de5d9c03b201 service nova] [instance: d7e8dcdb-41cb-46fb-8b61-d251e7c2d372] No waiting events found dispatching network-vif-plugged-9378f59d-f9bc-4c2e-8336-2c5710f987cf {{(pid=63515) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 725.308211] env[63515]: WARNING nova.compute.manager [req-a496ed4b-aaaf-4a9c-a1ff-94a31a5c23be req-8ea8637d-1aaa-4fe1-9ad1-de5d9c03b201 service nova] [instance: d7e8dcdb-41cb-46fb-8b61-d251e7c2d372] Received unexpected event network-vif-plugged-9378f59d-f9bc-4c2e-8336-2c5710f987cf for instance with vm_state building and task_state spawning. [ 725.401236] env[63515]: DEBUG nova.network.neutron [None req-e71315b0-9587-4929-82a2-71bebc6fcbfc tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] [instance: d7e8dcdb-41cb-46fb-8b61-d251e7c2d372] Successfully updated port: 9378f59d-f9bc-4c2e-8336-2c5710f987cf {{(pid=63515) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 725.501519] env[63515]: DEBUG nova.network.neutron [-] [instance: 48668736-df27-4f2a-94d9-132f5b49701b] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 725.529512] env[63515]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d8111cb6-8df8-4b58-b9ce-cbc6b636e861 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.539162] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac1c9f40-6fb5-4b8e-bcf4-41466557aa26 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.568971] env[63515]: DEBUG nova.compute.manager [req-8eb1027a-3f32-4469-87fb-5381b69191ac req-a8f70487-98b7-4d77-9275-ea5785a7d05b service nova] [instance: 48668736-df27-4f2a-94d9-132f5b49701b] Detach interface failed, port_id=9379e602-f7be-4c75-a6e6-c42a8702f6da, reason: Instance 48668736-df27-4f2a-94d9-132f5b49701b could not be found. {{(pid=63515) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 725.619023] env[63515]: DEBUG oslo_concurrency.lockutils [None req-dcda8d78-de0d-44e9-b61f-e84b4a49533f tempest-ServersTestManualDisk-1114050206 tempest-ServersTestManualDisk-1114050206-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.594s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 725.619023] env[63515]: DEBUG nova.compute.manager [None req-dcda8d78-de0d-44e9-b61f-e84b4a49533f tempest-ServersTestManualDisk-1114050206 tempest-ServersTestManualDisk-1114050206-project-member] [instance: e91aa479-1540-4950-851b-b2409e5f89f1] Start building networks asynchronously for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 725.622200] env[63515]: DEBUG oslo_concurrency.lockutils [None req-4bc5b4e9-701c-413d-abd5-d300a7447cb8 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.922s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 725.624025] env[63515]: INFO nova.compute.claims [None req-4bc5b4e9-701c-413d-abd5-d300a7447cb8 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: ed550b10-d58f-45b8-b766-198f431c3788] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 725.759194] env[63515]: DEBUG oslo_vmware.api [None req-74f7de08-cdcf-4b98-bfdb-a2dd77c03e6d tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Task: {'id': task-1111007, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 725.903741] env[63515]: DEBUG oslo_concurrency.lockutils [None req-e71315b0-9587-4929-82a2-71bebc6fcbfc tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Acquiring lock "refresh_cache-d7e8dcdb-41cb-46fb-8b61-d251e7c2d372" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 725.903903] env[63515]: DEBUG oslo_concurrency.lockutils [None req-e71315b0-9587-4929-82a2-71bebc6fcbfc tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Acquired lock "refresh_cache-d7e8dcdb-41cb-46fb-8b61-d251e7c2d372" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 725.904069] env[63515]: DEBUG nova.network.neutron [None req-e71315b0-9587-4929-82a2-71bebc6fcbfc tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] [instance: d7e8dcdb-41cb-46fb-8b61-d251e7c2d372] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 726.004066] env[63515]: INFO nova.compute.manager [-] [instance: 48668736-df27-4f2a-94d9-132f5b49701b] Took 1.21 seconds to deallocate network for instance. [ 726.128629] env[63515]: DEBUG nova.compute.utils [None req-dcda8d78-de0d-44e9-b61f-e84b4a49533f tempest-ServersTestManualDisk-1114050206 tempest-ServersTestManualDisk-1114050206-project-member] Using /dev/sd instead of None {{(pid=63515) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 726.131837] env[63515]: DEBUG nova.compute.manager [None req-dcda8d78-de0d-44e9-b61f-e84b4a49533f tempest-ServersTestManualDisk-1114050206 tempest-ServersTestManualDisk-1114050206-project-member] [instance: e91aa479-1540-4950-851b-b2409e5f89f1] Allocating IP information in the background. {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 726.132016] env[63515]: DEBUG nova.network.neutron [None req-dcda8d78-de0d-44e9-b61f-e84b4a49533f tempest-ServersTestManualDisk-1114050206 tempest-ServersTestManualDisk-1114050206-project-member] [instance: e91aa479-1540-4950-851b-b2409e5f89f1] allocate_for_instance() {{(pid=63515) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 726.170607] env[63515]: DEBUG nova.policy [None req-dcda8d78-de0d-44e9-b61f-e84b4a49533f tempest-ServersTestManualDisk-1114050206 tempest-ServersTestManualDisk-1114050206-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8bfdd6ed0fb3493bb0caf2425fd189d0', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '957bed1311824087944c5d19cfd0879a', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63515) authorize /opt/stack/nova/nova/policy.py:201}} [ 726.258133] env[63515]: DEBUG oslo_vmware.api [None req-74f7de08-cdcf-4b98-bfdb-a2dd77c03e6d tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Task: {'id': task-1111007, 'name': ReconfigVM_Task, 'duration_secs': 0.864721} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 726.258418] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-74f7de08-cdcf-4b98-bfdb-a2dd77c03e6d tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] [instance: b85e9a70-7f5b-4d32-b616-f2a97e3186c8] Reconfigured VM instance instance-0000002b to attach disk [datastore2] b85e9a70-7f5b-4d32-b616-f2a97e3186c8/b85e9a70-7f5b-4d32-b616-f2a97e3186c8.vmdk or device None with type sparse {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 726.259049] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-92011f17-5fd5-4424-b3e8-d179f92b0f09 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.265230] env[63515]: DEBUG oslo_vmware.api [None req-74f7de08-cdcf-4b98-bfdb-a2dd77c03e6d tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Waiting for the task: (returnval){ [ 726.265230] env[63515]: value = "task-1111008" [ 726.265230] env[63515]: _type = "Task" [ 726.265230] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 726.273081] env[63515]: DEBUG oslo_vmware.api [None req-74f7de08-cdcf-4b98-bfdb-a2dd77c03e6d tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Task: {'id': task-1111008, 'name': Rename_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 726.448100] env[63515]: DEBUG nova.network.neutron [None req-e71315b0-9587-4929-82a2-71bebc6fcbfc tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] [instance: d7e8dcdb-41cb-46fb-8b61-d251e7c2d372] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 726.510842] env[63515]: DEBUG oslo_concurrency.lockutils [None req-9695633a-0f62-476f-8466-deb3365d2b47 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 726.622389] env[63515]: DEBUG nova.network.neutron [None req-e71315b0-9587-4929-82a2-71bebc6fcbfc tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] [instance: d7e8dcdb-41cb-46fb-8b61-d251e7c2d372] Updating instance_info_cache with network_info: [{"id": "9378f59d-f9bc-4c2e-8336-2c5710f987cf", "address": "fa:16:3e:7b:d9:e1", "network": {"id": "57de0b07-78e5-453d-83ba-b12b7dcaf6ce", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-272799484-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "12a90929c98e40d9a1263929e2b97532", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a1895250-76cc-41f7-b7f8-2e5679494607", "external-id": "nsx-vlan-transportzone-785", "segmentation_id": 785, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9378f59d-f9", "ovs_interfaceid": "9378f59d-f9bc-4c2e-8336-2c5710f987cf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 726.632675] env[63515]: DEBUG nova.compute.manager [None req-dcda8d78-de0d-44e9-b61f-e84b4a49533f tempest-ServersTestManualDisk-1114050206 tempest-ServersTestManualDisk-1114050206-project-member] [instance: e91aa479-1540-4950-851b-b2409e5f89f1] Start building block device mappings for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 726.745697] env[63515]: DEBUG nova.network.neutron [None req-dcda8d78-de0d-44e9-b61f-e84b4a49533f tempest-ServersTestManualDisk-1114050206 tempest-ServersTestManualDisk-1114050206-project-member] [instance: e91aa479-1540-4950-851b-b2409e5f89f1] Successfully created port: fcab3239-8739-4ac7-9f35-4a121630aa6f {{(pid=63515) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 726.778786] env[63515]: DEBUG oslo_vmware.api [None req-74f7de08-cdcf-4b98-bfdb-a2dd77c03e6d tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Task: {'id': task-1111008, 'name': Rename_Task, 'duration_secs': 0.436265} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 726.778839] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-74f7de08-cdcf-4b98-bfdb-a2dd77c03e6d tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] [instance: b85e9a70-7f5b-4d32-b616-f2a97e3186c8] Powering on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 726.779109] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-cd8b2368-a8e1-4e7e-beb3-2323c3774f8d {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.790886] env[63515]: DEBUG oslo_vmware.api [None req-74f7de08-cdcf-4b98-bfdb-a2dd77c03e6d tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Waiting for the task: (returnval){ [ 726.790886] env[63515]: value = "task-1111009" [ 726.790886] env[63515]: _type = "Task" [ 726.790886] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 726.802111] env[63515]: DEBUG oslo_vmware.api [None req-74f7de08-cdcf-4b98-bfdb-a2dd77c03e6d tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Task: {'id': task-1111009, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 727.125039] env[63515]: DEBUG oslo_concurrency.lockutils [None req-e71315b0-9587-4929-82a2-71bebc6fcbfc tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Releasing lock "refresh_cache-d7e8dcdb-41cb-46fb-8b61-d251e7c2d372" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 727.125594] env[63515]: DEBUG nova.compute.manager [None req-e71315b0-9587-4929-82a2-71bebc6fcbfc tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] [instance: d7e8dcdb-41cb-46fb-8b61-d251e7c2d372] Instance network_info: |[{"id": "9378f59d-f9bc-4c2e-8336-2c5710f987cf", "address": "fa:16:3e:7b:d9:e1", "network": {"id": "57de0b07-78e5-453d-83ba-b12b7dcaf6ce", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-272799484-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "12a90929c98e40d9a1263929e2b97532", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a1895250-76cc-41f7-b7f8-2e5679494607", "external-id": "nsx-vlan-transportzone-785", "segmentation_id": 785, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9378f59d-f9", "ovs_interfaceid": "9378f59d-f9bc-4c2e-8336-2c5710f987cf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 727.125808] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-e71315b0-9587-4929-82a2-71bebc6fcbfc tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] [instance: d7e8dcdb-41cb-46fb-8b61-d251e7c2d372] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:7b:d9:e1', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a1895250-76cc-41f7-b7f8-2e5679494607', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '9378f59d-f9bc-4c2e-8336-2c5710f987cf', 'vif_model': 'vmxnet3'}] {{(pid=63515) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 727.133887] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-e71315b0-9587-4929-82a2-71bebc6fcbfc tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Creating folder: Project (12a90929c98e40d9a1263929e2b97532). Parent ref: group-v243370. {{(pid=63515) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 727.137082] env[63515]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-7510fa2c-29d8-4eaf-95ca-216c42d4effc {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.152508] env[63515]: INFO nova.virt.vmwareapi.vm_util [None req-e71315b0-9587-4929-82a2-71bebc6fcbfc tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Created folder: Project (12a90929c98e40d9a1263929e2b97532) in parent group-v243370. [ 727.152741] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-e71315b0-9587-4929-82a2-71bebc6fcbfc tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Creating folder: Instances. Parent ref: group-v243415. {{(pid=63515) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 727.152976] env[63515]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-06c2c890-daff-43e3-84d3-aa09ceae058a {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.165098] env[63515]: INFO nova.virt.vmwareapi.vm_util [None req-e71315b0-9587-4929-82a2-71bebc6fcbfc tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Created folder: Instances in parent group-v243415. [ 727.165197] env[63515]: DEBUG oslo.service.loopingcall [None req-e71315b0-9587-4929-82a2-71bebc6fcbfc tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 727.165956] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d7e8dcdb-41cb-46fb-8b61-d251e7c2d372] Creating VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 727.165956] env[63515]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-83862ae6-b6d8-4f4e-bb22-cc82a129e731 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.183125] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5fa7831d-ab91-4ad8-8ce5-c130b26fe260 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.192283] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0deeb8cc-8284-418e-8c49-d0f49a1b507c {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.195194] env[63515]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 727.195194] env[63515]: value = "task-1111012" [ 727.195194] env[63515]: _type = "Task" [ 727.195194] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 727.225933] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9fc8a81d-d952-42b4-8f1d-75aa34e66258 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.228225] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1111012, 'name': CreateVM_Task} progress is 5%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 727.232940] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba8df564-16f8-4bb1-af9b-f1ddf9393e2e {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.245819] env[63515]: DEBUG nova.compute.provider_tree [None req-4bc5b4e9-701c-413d-abd5-d300a7447cb8 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 727.302901] env[63515]: DEBUG oslo_vmware.api [None req-74f7de08-cdcf-4b98-bfdb-a2dd77c03e6d tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Task: {'id': task-1111009, 'name': PowerOnVM_Task} progress is 74%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 727.336994] env[63515]: DEBUG nova.compute.manager [req-820b2075-360c-48e0-b81c-05d6451f624a req-1aa697f6-5146-4f3c-b9f7-9c5f4ba41a05 service nova] [instance: d7e8dcdb-41cb-46fb-8b61-d251e7c2d372] Received event network-changed-9378f59d-f9bc-4c2e-8336-2c5710f987cf {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 727.337297] env[63515]: DEBUG nova.compute.manager [req-820b2075-360c-48e0-b81c-05d6451f624a req-1aa697f6-5146-4f3c-b9f7-9c5f4ba41a05 service nova] [instance: d7e8dcdb-41cb-46fb-8b61-d251e7c2d372] Refreshing instance network info cache due to event network-changed-9378f59d-f9bc-4c2e-8336-2c5710f987cf. {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 727.337541] env[63515]: DEBUG oslo_concurrency.lockutils [req-820b2075-360c-48e0-b81c-05d6451f624a req-1aa697f6-5146-4f3c-b9f7-9c5f4ba41a05 service nova] Acquiring lock "refresh_cache-d7e8dcdb-41cb-46fb-8b61-d251e7c2d372" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 727.337717] env[63515]: DEBUG oslo_concurrency.lockutils [req-820b2075-360c-48e0-b81c-05d6451f624a req-1aa697f6-5146-4f3c-b9f7-9c5f4ba41a05 service nova] Acquired lock "refresh_cache-d7e8dcdb-41cb-46fb-8b61-d251e7c2d372" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 727.337876] env[63515]: DEBUG nova.network.neutron [req-820b2075-360c-48e0-b81c-05d6451f624a req-1aa697f6-5146-4f3c-b9f7-9c5f4ba41a05 service nova] [instance: d7e8dcdb-41cb-46fb-8b61-d251e7c2d372] Refreshing network info cache for port 9378f59d-f9bc-4c2e-8336-2c5710f987cf {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 727.647636] env[63515]: DEBUG nova.compute.manager [None req-dcda8d78-de0d-44e9-b61f-e84b4a49533f tempest-ServersTestManualDisk-1114050206 tempest-ServersTestManualDisk-1114050206-project-member] [instance: e91aa479-1540-4950-851b-b2409e5f89f1] Start spawning the instance on the hypervisor. {{(pid=63515) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 727.671401] env[63515]: DEBUG nova.virt.hardware [None req-dcda8d78-de0d-44e9-b61f-e84b4a49533f tempest-ServersTestManualDisk-1114050206 tempest-ServersTestManualDisk-1114050206-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T02:52:14Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T02:51:56Z,direct_url=,disk_format='vmdk',id=8a120570-cb06-4099-b262-554ca0ad15c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b569255cc43e42e7ae3f2b4ad37c6ef4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T02:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 727.671657] env[63515]: DEBUG nova.virt.hardware [None req-dcda8d78-de0d-44e9-b61f-e84b4a49533f tempest-ServersTestManualDisk-1114050206 tempest-ServersTestManualDisk-1114050206-project-member] Flavor limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 727.671813] env[63515]: DEBUG nova.virt.hardware [None req-dcda8d78-de0d-44e9-b61f-e84b4a49533f tempest-ServersTestManualDisk-1114050206 tempest-ServersTestManualDisk-1114050206-project-member] Image limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 727.671989] env[63515]: DEBUG nova.virt.hardware [None req-dcda8d78-de0d-44e9-b61f-e84b4a49533f tempest-ServersTestManualDisk-1114050206 tempest-ServersTestManualDisk-1114050206-project-member] Flavor pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 727.672145] env[63515]: DEBUG nova.virt.hardware [None req-dcda8d78-de0d-44e9-b61f-e84b4a49533f tempest-ServersTestManualDisk-1114050206 tempest-ServersTestManualDisk-1114050206-project-member] Image pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 727.672289] env[63515]: DEBUG nova.virt.hardware [None req-dcda8d78-de0d-44e9-b61f-e84b4a49533f tempest-ServersTestManualDisk-1114050206 tempest-ServersTestManualDisk-1114050206-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 727.672490] env[63515]: DEBUG nova.virt.hardware [None req-dcda8d78-de0d-44e9-b61f-e84b4a49533f tempest-ServersTestManualDisk-1114050206 tempest-ServersTestManualDisk-1114050206-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 727.672881] env[63515]: DEBUG nova.virt.hardware [None req-dcda8d78-de0d-44e9-b61f-e84b4a49533f tempest-ServersTestManualDisk-1114050206 tempest-ServersTestManualDisk-1114050206-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 727.672881] env[63515]: DEBUG nova.virt.hardware [None req-dcda8d78-de0d-44e9-b61f-e84b4a49533f tempest-ServersTestManualDisk-1114050206 tempest-ServersTestManualDisk-1114050206-project-member] Got 1 possible topologies {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 727.673019] env[63515]: DEBUG nova.virt.hardware [None req-dcda8d78-de0d-44e9-b61f-e84b4a49533f tempest-ServersTestManualDisk-1114050206 tempest-ServersTestManualDisk-1114050206-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 727.673145] env[63515]: DEBUG nova.virt.hardware [None req-dcda8d78-de0d-44e9-b61f-e84b4a49533f tempest-ServersTestManualDisk-1114050206 tempest-ServersTestManualDisk-1114050206-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 727.674011] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ffd5a73-c7f0-40ae-9d14-e93b9b814bcb {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.681621] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4d7587e-0909-43fe-808b-bbb54aca458f {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.704693] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1111012, 'name': CreateVM_Task, 'duration_secs': 0.502408} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 727.704693] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d7e8dcdb-41cb-46fb-8b61-d251e7c2d372] Created VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 727.704935] env[63515]: DEBUG oslo_concurrency.lockutils [None req-e71315b0-9587-4929-82a2-71bebc6fcbfc tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 727.704935] env[63515]: DEBUG oslo_concurrency.lockutils [None req-e71315b0-9587-4929-82a2-71bebc6fcbfc tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 727.705233] env[63515]: DEBUG oslo_concurrency.lockutils [None req-e71315b0-9587-4929-82a2-71bebc6fcbfc tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 727.705464] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e6d9347f-0d82-493b-b21d-4c556cf57eef {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.709627] env[63515]: DEBUG oslo_vmware.api [None req-e71315b0-9587-4929-82a2-71bebc6fcbfc tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Waiting for the task: (returnval){ [ 727.709627] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]52d3edb9-d084-2886-aba0-e881734510d1" [ 727.709627] env[63515]: _type = "Task" [ 727.709627] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 727.716834] env[63515]: DEBUG oslo_vmware.api [None req-e71315b0-9587-4929-82a2-71bebc6fcbfc tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52d3edb9-d084-2886-aba0-e881734510d1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 727.749176] env[63515]: DEBUG nova.scheduler.client.report [None req-4bc5b4e9-701c-413d-abd5-d300a7447cb8 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 727.801858] env[63515]: DEBUG oslo_vmware.api [None req-74f7de08-cdcf-4b98-bfdb-a2dd77c03e6d tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Task: {'id': task-1111009, 'name': PowerOnVM_Task, 'duration_secs': 0.731239} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 727.802147] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-74f7de08-cdcf-4b98-bfdb-a2dd77c03e6d tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] [instance: b85e9a70-7f5b-4d32-b616-f2a97e3186c8] Powered on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 727.802348] env[63515]: INFO nova.compute.manager [None req-74f7de08-cdcf-4b98-bfdb-a2dd77c03e6d tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] [instance: b85e9a70-7f5b-4d32-b616-f2a97e3186c8] Took 5.30 seconds to spawn the instance on the hypervisor. [ 727.802570] env[63515]: DEBUG nova.compute.manager [None req-74f7de08-cdcf-4b98-bfdb-a2dd77c03e6d tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] [instance: b85e9a70-7f5b-4d32-b616-f2a97e3186c8] Checking state {{(pid=63515) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 727.803433] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88fab364-1780-48d7-9d63-5d86ac7653ff {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.115703] env[63515]: DEBUG nova.network.neutron [req-820b2075-360c-48e0-b81c-05d6451f624a req-1aa697f6-5146-4f3c-b9f7-9c5f4ba41a05 service nova] [instance: d7e8dcdb-41cb-46fb-8b61-d251e7c2d372] Updated VIF entry in instance network info cache for port 9378f59d-f9bc-4c2e-8336-2c5710f987cf. {{(pid=63515) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 728.115703] env[63515]: DEBUG nova.network.neutron [req-820b2075-360c-48e0-b81c-05d6451f624a req-1aa697f6-5146-4f3c-b9f7-9c5f4ba41a05 service nova] [instance: d7e8dcdb-41cb-46fb-8b61-d251e7c2d372] Updating instance_info_cache with network_info: [{"id": "9378f59d-f9bc-4c2e-8336-2c5710f987cf", "address": "fa:16:3e:7b:d9:e1", "network": {"id": "57de0b07-78e5-453d-83ba-b12b7dcaf6ce", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-272799484-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "12a90929c98e40d9a1263929e2b97532", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a1895250-76cc-41f7-b7f8-2e5679494607", "external-id": "nsx-vlan-transportzone-785", "segmentation_id": 785, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9378f59d-f9", "ovs_interfaceid": "9378f59d-f9bc-4c2e-8336-2c5710f987cf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 728.220623] env[63515]: DEBUG oslo_vmware.api [None req-e71315b0-9587-4929-82a2-71bebc6fcbfc tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52d3edb9-d084-2886-aba0-e881734510d1, 'name': SearchDatastore_Task, 'duration_secs': 0.047512} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 728.220921] env[63515]: DEBUG oslo_concurrency.lockutils [None req-e71315b0-9587-4929-82a2-71bebc6fcbfc tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 728.221170] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-e71315b0-9587-4929-82a2-71bebc6fcbfc tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] [instance: d7e8dcdb-41cb-46fb-8b61-d251e7c2d372] Processing image 8a120570-cb06-4099-b262-554ca0ad15c5 {{(pid=63515) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 728.221399] env[63515]: DEBUG oslo_concurrency.lockutils [None req-e71315b0-9587-4929-82a2-71bebc6fcbfc tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 728.221540] env[63515]: DEBUG oslo_concurrency.lockutils [None req-e71315b0-9587-4929-82a2-71bebc6fcbfc tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 728.221713] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-e71315b0-9587-4929-82a2-71bebc6fcbfc tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 728.221966] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-479e2714-b2bb-4aaf-8e23-9caaf8e17a92 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.230732] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-e71315b0-9587-4929-82a2-71bebc6fcbfc tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 728.230923] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-e71315b0-9587-4929-82a2-71bebc6fcbfc tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63515) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 728.231670] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dbd47f41-5d49-419c-b06e-0c97f6b7e7a6 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.237251] env[63515]: DEBUG oslo_vmware.api [None req-e71315b0-9587-4929-82a2-71bebc6fcbfc tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Waiting for the task: (returnval){ [ 728.237251] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]5205367d-f13d-4051-9d8b-c78463d2cda8" [ 728.237251] env[63515]: _type = "Task" [ 728.237251] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 728.244995] env[63515]: DEBUG oslo_vmware.api [None req-e71315b0-9587-4929-82a2-71bebc6fcbfc tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]5205367d-f13d-4051-9d8b-c78463d2cda8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 728.253855] env[63515]: DEBUG oslo_concurrency.lockutils [None req-4bc5b4e9-701c-413d-abd5-d300a7447cb8 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.632s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 728.254383] env[63515]: DEBUG nova.compute.manager [None req-4bc5b4e9-701c-413d-abd5-d300a7447cb8 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: ed550b10-d58f-45b8-b766-198f431c3788] Start building networks asynchronously for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 728.257159] env[63515]: DEBUG oslo_concurrency.lockutils [None req-99f0df88-8ade-4c67-afd9-dbdc6b385015 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 16.304s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 728.321934] env[63515]: INFO nova.compute.manager [None req-74f7de08-cdcf-4b98-bfdb-a2dd77c03e6d tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] [instance: b85e9a70-7f5b-4d32-b616-f2a97e3186c8] Took 26.80 seconds to build instance. [ 728.476709] env[63515]: DEBUG nova.network.neutron [None req-dcda8d78-de0d-44e9-b61f-e84b4a49533f tempest-ServersTestManualDisk-1114050206 tempest-ServersTestManualDisk-1114050206-project-member] [instance: e91aa479-1540-4950-851b-b2409e5f89f1] Successfully updated port: fcab3239-8739-4ac7-9f35-4a121630aa6f {{(pid=63515) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 728.617961] env[63515]: DEBUG oslo_concurrency.lockutils [req-820b2075-360c-48e0-b81c-05d6451f624a req-1aa697f6-5146-4f3c-b9f7-9c5f4ba41a05 service nova] Releasing lock "refresh_cache-d7e8dcdb-41cb-46fb-8b61-d251e7c2d372" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 728.747455] env[63515]: DEBUG oslo_vmware.api [None req-e71315b0-9587-4929-82a2-71bebc6fcbfc tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]5205367d-f13d-4051-9d8b-c78463d2cda8, 'name': SearchDatastore_Task, 'duration_secs': 0.009841} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 728.748289] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-73321f0e-5c17-4d88-8ed5-b8099b6b9d37 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.753957] env[63515]: DEBUG oslo_vmware.api [None req-e71315b0-9587-4929-82a2-71bebc6fcbfc tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Waiting for the task: (returnval){ [ 728.753957] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]52ebb736-2d54-dcdb-c4ff-c05f74c4dbdc" [ 728.753957] env[63515]: _type = "Task" [ 728.753957] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 728.764317] env[63515]: INFO nova.compute.claims [None req-99f0df88-8ade-4c67-afd9-dbdc6b385015 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] [instance: f1d01b75-ac9d-458d-8cc2-ae64cffca4e8] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 728.768435] env[63515]: DEBUG nova.compute.utils [None req-4bc5b4e9-701c-413d-abd5-d300a7447cb8 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Using /dev/sd instead of None {{(pid=63515) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 728.769563] env[63515]: DEBUG oslo_vmware.api [None req-e71315b0-9587-4929-82a2-71bebc6fcbfc tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52ebb736-2d54-dcdb-c4ff-c05f74c4dbdc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 728.770054] env[63515]: DEBUG nova.compute.manager [None req-4bc5b4e9-701c-413d-abd5-d300a7447cb8 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: ed550b10-d58f-45b8-b766-198f431c3788] Allocating IP information in the background. {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 728.770526] env[63515]: DEBUG nova.network.neutron [None req-4bc5b4e9-701c-413d-abd5-d300a7447cb8 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: ed550b10-d58f-45b8-b766-198f431c3788] allocate_for_instance() {{(pid=63515) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 728.809681] env[63515]: DEBUG nova.policy [None req-4bc5b4e9-701c-413d-abd5-d300a7447cb8 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '16880abf1c5b4341800c94ada2c756c7', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7d4ada9ec35f42b19c6480a9101d21a4', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63515) authorize /opt/stack/nova/nova/policy.py:201}} [ 728.824457] env[63515]: DEBUG oslo_concurrency.lockutils [None req-74f7de08-cdcf-4b98-bfdb-a2dd77c03e6d tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Lock "b85e9a70-7f5b-4d32-b616-f2a97e3186c8" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 111.966s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 728.902287] env[63515]: INFO nova.compute.manager [None req-996c1e42-eed8-414b-a716-3f08d65ec493 tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] [instance: b85e9a70-7f5b-4d32-b616-f2a97e3186c8] Rebuilding instance [ 728.942681] env[63515]: DEBUG nova.compute.manager [None req-996c1e42-eed8-414b-a716-3f08d65ec493 tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] [instance: b85e9a70-7f5b-4d32-b616-f2a97e3186c8] Checking state {{(pid=63515) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 728.943563] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0fbe95f3-907b-4f88-a6f7-71f83beb2524 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.979968] env[63515]: DEBUG oslo_concurrency.lockutils [None req-dcda8d78-de0d-44e9-b61f-e84b4a49533f tempest-ServersTestManualDisk-1114050206 tempest-ServersTestManualDisk-1114050206-project-member] Acquiring lock "refresh_cache-e91aa479-1540-4950-851b-b2409e5f89f1" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 728.980070] env[63515]: DEBUG oslo_concurrency.lockutils [None req-dcda8d78-de0d-44e9-b61f-e84b4a49533f tempest-ServersTestManualDisk-1114050206 tempest-ServersTestManualDisk-1114050206-project-member] Acquired lock "refresh_cache-e91aa479-1540-4950-851b-b2409e5f89f1" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 728.980156] env[63515]: DEBUG nova.network.neutron [None req-dcda8d78-de0d-44e9-b61f-e84b4a49533f tempest-ServersTestManualDisk-1114050206 tempest-ServersTestManualDisk-1114050206-project-member] [instance: e91aa479-1540-4950-851b-b2409e5f89f1] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 729.192474] env[63515]: DEBUG nova.network.neutron [None req-4bc5b4e9-701c-413d-abd5-d300a7447cb8 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: ed550b10-d58f-45b8-b766-198f431c3788] Successfully created port: ad9dda29-9326-472b-8c82-294a33a7faf0 {{(pid=63515) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 729.265331] env[63515]: DEBUG oslo_vmware.api [None req-e71315b0-9587-4929-82a2-71bebc6fcbfc tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52ebb736-2d54-dcdb-c4ff-c05f74c4dbdc, 'name': SearchDatastore_Task, 'duration_secs': 0.012454} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 729.265876] env[63515]: DEBUG oslo_concurrency.lockutils [None req-e71315b0-9587-4929-82a2-71bebc6fcbfc tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 729.266297] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-e71315b0-9587-4929-82a2-71bebc6fcbfc tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk to [datastore1] d7e8dcdb-41cb-46fb-8b61-d251e7c2d372/d7e8dcdb-41cb-46fb-8b61-d251e7c2d372.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 729.268941] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f6bb273b-ada5-4269-a5b2-f252d6795f89 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.271637] env[63515]: INFO nova.compute.resource_tracker [None req-99f0df88-8ade-4c67-afd9-dbdc6b385015 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] [instance: f1d01b75-ac9d-458d-8cc2-ae64cffca4e8] Updating resource usage from migration 802930d6-330b-4516-9996-d196743b8ab8 [ 729.277179] env[63515]: DEBUG nova.compute.manager [None req-4bc5b4e9-701c-413d-abd5-d300a7447cb8 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: ed550b10-d58f-45b8-b766-198f431c3788] Start building block device mappings for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 729.281670] env[63515]: DEBUG oslo_vmware.api [None req-e71315b0-9587-4929-82a2-71bebc6fcbfc tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Waiting for the task: (returnval){ [ 729.281670] env[63515]: value = "task-1111013" [ 729.281670] env[63515]: _type = "Task" [ 729.281670] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 729.292157] env[63515]: DEBUG oslo_vmware.api [None req-e71315b0-9587-4929-82a2-71bebc6fcbfc tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Task: {'id': task-1111013, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 729.329404] env[63515]: DEBUG nova.compute.manager [None req-b8b2a7f6-808d-40d4-8de9-c8863e5b55c5 tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] [instance: 30b8b872-40ca-4297-b98e-a64c83a75483] Starting instance... {{(pid=63515) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 729.369232] env[63515]: DEBUG nova.compute.manager [req-c2b6a640-8446-45dc-8536-6c1c19bd5b73 req-525af8da-09ba-4a1b-9054-d9caf82da16d service nova] [instance: e91aa479-1540-4950-851b-b2409e5f89f1] Received event network-vif-plugged-fcab3239-8739-4ac7-9f35-4a121630aa6f {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 729.369527] env[63515]: DEBUG oslo_concurrency.lockutils [req-c2b6a640-8446-45dc-8536-6c1c19bd5b73 req-525af8da-09ba-4a1b-9054-d9caf82da16d service nova] Acquiring lock "e91aa479-1540-4950-851b-b2409e5f89f1-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 729.369796] env[63515]: DEBUG oslo_concurrency.lockutils [req-c2b6a640-8446-45dc-8536-6c1c19bd5b73 req-525af8da-09ba-4a1b-9054-d9caf82da16d service nova] Lock "e91aa479-1540-4950-851b-b2409e5f89f1-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 729.369977] env[63515]: DEBUG oslo_concurrency.lockutils [req-c2b6a640-8446-45dc-8536-6c1c19bd5b73 req-525af8da-09ba-4a1b-9054-d9caf82da16d service nova] Lock "e91aa479-1540-4950-851b-b2409e5f89f1-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 729.370238] env[63515]: DEBUG nova.compute.manager [req-c2b6a640-8446-45dc-8536-6c1c19bd5b73 req-525af8da-09ba-4a1b-9054-d9caf82da16d service nova] [instance: e91aa479-1540-4950-851b-b2409e5f89f1] No waiting events found dispatching network-vif-plugged-fcab3239-8739-4ac7-9f35-4a121630aa6f {{(pid=63515) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 729.370413] env[63515]: WARNING nova.compute.manager [req-c2b6a640-8446-45dc-8536-6c1c19bd5b73 req-525af8da-09ba-4a1b-9054-d9caf82da16d service nova] [instance: e91aa479-1540-4950-851b-b2409e5f89f1] Received unexpected event network-vif-plugged-fcab3239-8739-4ac7-9f35-4a121630aa6f for instance with vm_state building and task_state spawning. [ 729.370672] env[63515]: DEBUG nova.compute.manager [req-c2b6a640-8446-45dc-8536-6c1c19bd5b73 req-525af8da-09ba-4a1b-9054-d9caf82da16d service nova] [instance: e91aa479-1540-4950-851b-b2409e5f89f1] Received event network-changed-fcab3239-8739-4ac7-9f35-4a121630aa6f {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 729.371152] env[63515]: DEBUG nova.compute.manager [req-c2b6a640-8446-45dc-8536-6c1c19bd5b73 req-525af8da-09ba-4a1b-9054-d9caf82da16d service nova] [instance: e91aa479-1540-4950-851b-b2409e5f89f1] Refreshing instance network info cache due to event network-changed-fcab3239-8739-4ac7-9f35-4a121630aa6f. {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 729.371365] env[63515]: DEBUG oslo_concurrency.lockutils [req-c2b6a640-8446-45dc-8536-6c1c19bd5b73 req-525af8da-09ba-4a1b-9054-d9caf82da16d service nova] Acquiring lock "refresh_cache-e91aa479-1540-4950-851b-b2409e5f89f1" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 729.456910] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-996c1e42-eed8-414b-a716-3f08d65ec493 tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] [instance: b85e9a70-7f5b-4d32-b616-f2a97e3186c8] Powering off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 729.458274] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8a00b63a-47ce-4a3c-bc84-01bdb17a9a0b {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.464342] env[63515]: DEBUG oslo_vmware.api [None req-996c1e42-eed8-414b-a716-3f08d65ec493 tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Waiting for the task: (returnval){ [ 729.464342] env[63515]: value = "task-1111014" [ 729.464342] env[63515]: _type = "Task" [ 729.464342] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 729.473669] env[63515]: DEBUG oslo_vmware.api [None req-996c1e42-eed8-414b-a716-3f08d65ec493 tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Task: {'id': task-1111014, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 729.539199] env[63515]: DEBUG nova.network.neutron [None req-dcda8d78-de0d-44e9-b61f-e84b4a49533f tempest-ServersTestManualDisk-1114050206 tempest-ServersTestManualDisk-1114050206-project-member] [instance: e91aa479-1540-4950-851b-b2409e5f89f1] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 729.738236] env[63515]: DEBUG nova.network.neutron [None req-dcda8d78-de0d-44e9-b61f-e84b4a49533f tempest-ServersTestManualDisk-1114050206 tempest-ServersTestManualDisk-1114050206-project-member] [instance: e91aa479-1540-4950-851b-b2409e5f89f1] Updating instance_info_cache with network_info: [{"id": "fcab3239-8739-4ac7-9f35-4a121630aa6f", "address": "fa:16:3e:3d:c1:97", "network": {"id": "1604b180-4129-4de5-92e6-79f13694e373", "bridge": "br-int", "label": "tempest-ServersTestManualDisk-718841075-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "957bed1311824087944c5d19cfd0879a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8abee039-d93e-48a7-8911-6416a3e1ff30", "external-id": "nsx-vlan-transportzone-654", "segmentation_id": 654, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfcab3239-87", "ovs_interfaceid": "fcab3239-8739-4ac7-9f35-4a121630aa6f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 729.781592] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-537d3661-d7ce-4a17-942d-3579b595714a {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.800236] env[63515]: DEBUG oslo_vmware.api [None req-e71315b0-9587-4929-82a2-71bebc6fcbfc tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Task: {'id': task-1111013, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 729.803043] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7532cfa-c669-4392-8b88-206ada0b9a7a {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.835991] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee4d5eb8-6a9e-48a3-872e-c42336b5285b {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.848404] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2600e926-5fd5-428a-a305-a22c3b010829 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.862954] env[63515]: DEBUG nova.compute.provider_tree [None req-99f0df88-8ade-4c67-afd9-dbdc6b385015 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 729.864923] env[63515]: DEBUG oslo_concurrency.lockutils [None req-b8b2a7f6-808d-40d4-8de9-c8863e5b55c5 tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 729.975352] env[63515]: DEBUG oslo_vmware.api [None req-996c1e42-eed8-414b-a716-3f08d65ec493 tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Task: {'id': task-1111014, 'name': PowerOffVM_Task, 'duration_secs': 0.463639} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 729.975624] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-996c1e42-eed8-414b-a716-3f08d65ec493 tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] [instance: b85e9a70-7f5b-4d32-b616-f2a97e3186c8] Powered off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 729.975841] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-996c1e42-eed8-414b-a716-3f08d65ec493 tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] [instance: b85e9a70-7f5b-4d32-b616-f2a97e3186c8] Destroying instance {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 729.976613] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58547ec4-0517-403d-81e9-b344c8a83be6 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.983631] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-996c1e42-eed8-414b-a716-3f08d65ec493 tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] [instance: b85e9a70-7f5b-4d32-b616-f2a97e3186c8] Unregistering the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 729.983875] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c7bbc992-d6a0-4f64-ba7e-2acecadf8c34 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.011178] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-996c1e42-eed8-414b-a716-3f08d65ec493 tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] [instance: b85e9a70-7f5b-4d32-b616-f2a97e3186c8] Unregistered the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 730.011415] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-996c1e42-eed8-414b-a716-3f08d65ec493 tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] [instance: b85e9a70-7f5b-4d32-b616-f2a97e3186c8] Deleting contents of the VM from datastore datastore2 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 730.011600] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-996c1e42-eed8-414b-a716-3f08d65ec493 tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Deleting the datastore file [datastore2] b85e9a70-7f5b-4d32-b616-f2a97e3186c8 {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 730.011862] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c463b858-049c-4e8c-b0db-3b17a3802e88 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.018109] env[63515]: DEBUG oslo_vmware.api [None req-996c1e42-eed8-414b-a716-3f08d65ec493 tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Waiting for the task: (returnval){ [ 730.018109] env[63515]: value = "task-1111016" [ 730.018109] env[63515]: _type = "Task" [ 730.018109] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 730.025860] env[63515]: DEBUG oslo_vmware.api [None req-996c1e42-eed8-414b-a716-3f08d65ec493 tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Task: {'id': task-1111016, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 730.240492] env[63515]: DEBUG oslo_concurrency.lockutils [None req-dcda8d78-de0d-44e9-b61f-e84b4a49533f tempest-ServersTestManualDisk-1114050206 tempest-ServersTestManualDisk-1114050206-project-member] Releasing lock "refresh_cache-e91aa479-1540-4950-851b-b2409e5f89f1" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 730.240874] env[63515]: DEBUG nova.compute.manager [None req-dcda8d78-de0d-44e9-b61f-e84b4a49533f tempest-ServersTestManualDisk-1114050206 tempest-ServersTestManualDisk-1114050206-project-member] [instance: e91aa479-1540-4950-851b-b2409e5f89f1] Instance network_info: |[{"id": "fcab3239-8739-4ac7-9f35-4a121630aa6f", "address": "fa:16:3e:3d:c1:97", "network": {"id": "1604b180-4129-4de5-92e6-79f13694e373", "bridge": "br-int", "label": "tempest-ServersTestManualDisk-718841075-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "957bed1311824087944c5d19cfd0879a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8abee039-d93e-48a7-8911-6416a3e1ff30", "external-id": "nsx-vlan-transportzone-654", "segmentation_id": 654, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfcab3239-87", "ovs_interfaceid": "fcab3239-8739-4ac7-9f35-4a121630aa6f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 730.241195] env[63515]: DEBUG oslo_concurrency.lockutils [req-c2b6a640-8446-45dc-8536-6c1c19bd5b73 req-525af8da-09ba-4a1b-9054-d9caf82da16d service nova] Acquired lock "refresh_cache-e91aa479-1540-4950-851b-b2409e5f89f1" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 730.241371] env[63515]: DEBUG nova.network.neutron [req-c2b6a640-8446-45dc-8536-6c1c19bd5b73 req-525af8da-09ba-4a1b-9054-d9caf82da16d service nova] [instance: e91aa479-1540-4950-851b-b2409e5f89f1] Refreshing network info cache for port fcab3239-8739-4ac7-9f35-4a121630aa6f {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 730.242763] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-dcda8d78-de0d-44e9-b61f-e84b4a49533f tempest-ServersTestManualDisk-1114050206 tempest-ServersTestManualDisk-1114050206-project-member] [instance: e91aa479-1540-4950-851b-b2409e5f89f1] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:3d:c1:97', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '8abee039-d93e-48a7-8911-6416a3e1ff30', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'fcab3239-8739-4ac7-9f35-4a121630aa6f', 'vif_model': 'vmxnet3'}] {{(pid=63515) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 730.250635] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-dcda8d78-de0d-44e9-b61f-e84b4a49533f tempest-ServersTestManualDisk-1114050206 tempest-ServersTestManualDisk-1114050206-project-member] Creating folder: Project (957bed1311824087944c5d19cfd0879a). Parent ref: group-v243370. {{(pid=63515) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 730.250874] env[63515]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-9f3394b2-6775-4d6b-9e9d-fba2d2bfc369 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.264385] env[63515]: INFO nova.virt.vmwareapi.vm_util [None req-dcda8d78-de0d-44e9-b61f-e84b4a49533f tempest-ServersTestManualDisk-1114050206 tempest-ServersTestManualDisk-1114050206-project-member] Created folder: Project (957bed1311824087944c5d19cfd0879a) in parent group-v243370. [ 730.264385] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-dcda8d78-de0d-44e9-b61f-e84b4a49533f tempest-ServersTestManualDisk-1114050206 tempest-ServersTestManualDisk-1114050206-project-member] Creating folder: Instances. Parent ref: group-v243418. {{(pid=63515) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 730.264605] env[63515]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-265719da-f621-4d39-b4d3-58520e9d635d {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.275615] env[63515]: INFO nova.virt.vmwareapi.vm_util [None req-dcda8d78-de0d-44e9-b61f-e84b4a49533f tempest-ServersTestManualDisk-1114050206 tempest-ServersTestManualDisk-1114050206-project-member] Created folder: Instances in parent group-v243418. [ 730.275898] env[63515]: DEBUG oslo.service.loopingcall [None req-dcda8d78-de0d-44e9-b61f-e84b4a49533f tempest-ServersTestManualDisk-1114050206 tempest-ServersTestManualDisk-1114050206-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 730.276105] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e91aa479-1540-4950-851b-b2409e5f89f1] Creating VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 730.276339] env[63515]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3616bbdc-d74a-4f05-908f-fdc9a801dc14 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.295647] env[63515]: DEBUG nova.compute.manager [None req-4bc5b4e9-701c-413d-abd5-d300a7447cb8 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: ed550b10-d58f-45b8-b766-198f431c3788] Start spawning the instance on the hypervisor. {{(pid=63515) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 730.299345] env[63515]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 730.299345] env[63515]: value = "task-1111019" [ 730.299345] env[63515]: _type = "Task" [ 730.299345] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 730.306091] env[63515]: DEBUG oslo_vmware.api [None req-e71315b0-9587-4929-82a2-71bebc6fcbfc tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Task: {'id': task-1111013, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.570356} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 730.306491] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-e71315b0-9587-4929-82a2-71bebc6fcbfc tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk to [datastore1] d7e8dcdb-41cb-46fb-8b61-d251e7c2d372/d7e8dcdb-41cb-46fb-8b61-d251e7c2d372.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 730.306793] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-e71315b0-9587-4929-82a2-71bebc6fcbfc tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] [instance: d7e8dcdb-41cb-46fb-8b61-d251e7c2d372] Extending root virtual disk to 1048576 {{(pid=63515) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 730.307474] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-bc907bdc-a9ac-4415-bd79-63c888fdd3db {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.312771] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1111019, 'name': CreateVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 730.317084] env[63515]: DEBUG oslo_vmware.api [None req-e71315b0-9587-4929-82a2-71bebc6fcbfc tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Waiting for the task: (returnval){ [ 730.317084] env[63515]: value = "task-1111020" [ 730.317084] env[63515]: _type = "Task" [ 730.317084] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 730.328201] env[63515]: DEBUG oslo_vmware.api [None req-e71315b0-9587-4929-82a2-71bebc6fcbfc tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Task: {'id': task-1111020, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 730.330701] env[63515]: DEBUG nova.virt.hardware [None req-4bc5b4e9-701c-413d-abd5-d300a7447cb8 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T02:52:14Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T02:51:56Z,direct_url=,disk_format='vmdk',id=8a120570-cb06-4099-b262-554ca0ad15c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b569255cc43e42e7ae3f2b4ad37c6ef4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T02:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 730.330839] env[63515]: DEBUG nova.virt.hardware [None req-4bc5b4e9-701c-413d-abd5-d300a7447cb8 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Flavor limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 730.330999] env[63515]: DEBUG nova.virt.hardware [None req-4bc5b4e9-701c-413d-abd5-d300a7447cb8 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Image limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 730.331252] env[63515]: DEBUG nova.virt.hardware [None req-4bc5b4e9-701c-413d-abd5-d300a7447cb8 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Flavor pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 730.331413] env[63515]: DEBUG nova.virt.hardware [None req-4bc5b4e9-701c-413d-abd5-d300a7447cb8 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Image pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 730.331716] env[63515]: DEBUG nova.virt.hardware [None req-4bc5b4e9-701c-413d-abd5-d300a7447cb8 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 730.331767] env[63515]: DEBUG nova.virt.hardware [None req-4bc5b4e9-701c-413d-abd5-d300a7447cb8 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 730.331964] env[63515]: DEBUG nova.virt.hardware [None req-4bc5b4e9-701c-413d-abd5-d300a7447cb8 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 730.332069] env[63515]: DEBUG nova.virt.hardware [None req-4bc5b4e9-701c-413d-abd5-d300a7447cb8 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Got 1 possible topologies {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 730.332237] env[63515]: DEBUG nova.virt.hardware [None req-4bc5b4e9-701c-413d-abd5-d300a7447cb8 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 730.332407] env[63515]: DEBUG nova.virt.hardware [None req-4bc5b4e9-701c-413d-abd5-d300a7447cb8 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 730.333190] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b78fa4d4-eae8-49eb-acff-881f0e426692 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.341142] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5556017e-be95-46bf-8a3d-3420814dec22 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.366600] env[63515]: DEBUG nova.scheduler.client.report [None req-99f0df88-8ade-4c67-afd9-dbdc6b385015 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 730.528894] env[63515]: DEBUG oslo_vmware.api [None req-996c1e42-eed8-414b-a716-3f08d65ec493 tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Task: {'id': task-1111016, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.293841} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 730.529136] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-996c1e42-eed8-414b-a716-3f08d65ec493 tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Deleted the datastore file {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 730.529250] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-996c1e42-eed8-414b-a716-3f08d65ec493 tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] [instance: b85e9a70-7f5b-4d32-b616-f2a97e3186c8] Deleted contents of the VM from datastore datastore2 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 730.529425] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-996c1e42-eed8-414b-a716-3f08d65ec493 tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] [instance: b85e9a70-7f5b-4d32-b616-f2a97e3186c8] Instance destroyed {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 730.814943] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1111019, 'name': CreateVM_Task, 'duration_secs': 0.378799} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 730.814943] env[63515]: DEBUG nova.network.neutron [None req-4bc5b4e9-701c-413d-abd5-d300a7447cb8 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: ed550b10-d58f-45b8-b766-198f431c3788] Successfully updated port: ad9dda29-9326-472b-8c82-294a33a7faf0 {{(pid=63515) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 730.814943] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e91aa479-1540-4950-851b-b2409e5f89f1] Created VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 730.814943] env[63515]: DEBUG oslo_concurrency.lockutils [None req-dcda8d78-de0d-44e9-b61f-e84b4a49533f tempest-ServersTestManualDisk-1114050206 tempest-ServersTestManualDisk-1114050206-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 730.815360] env[63515]: DEBUG oslo_concurrency.lockutils [None req-dcda8d78-de0d-44e9-b61f-e84b4a49533f tempest-ServersTestManualDisk-1114050206 tempest-ServersTestManualDisk-1114050206-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 730.815638] env[63515]: DEBUG oslo_concurrency.lockutils [None req-dcda8d78-de0d-44e9-b61f-e84b4a49533f tempest-ServersTestManualDisk-1114050206 tempest-ServersTestManualDisk-1114050206-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 730.816086] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c388becc-ab3b-4181-b628-269ea1d45e55 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.824242] env[63515]: DEBUG oslo_vmware.api [None req-dcda8d78-de0d-44e9-b61f-e84b4a49533f tempest-ServersTestManualDisk-1114050206 tempest-ServersTestManualDisk-1114050206-project-member] Waiting for the task: (returnval){ [ 730.824242] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]52a99de3-fa8b-e940-7c60-c203753ee53e" [ 730.824242] env[63515]: _type = "Task" [ 730.824242] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 730.833245] env[63515]: DEBUG oslo_vmware.api [None req-e71315b0-9587-4929-82a2-71bebc6fcbfc tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Task: {'id': task-1111020, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.102615} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 730.833798] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-e71315b0-9587-4929-82a2-71bebc6fcbfc tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] [instance: d7e8dcdb-41cb-46fb-8b61-d251e7c2d372] Extended root virtual disk {{(pid=63515) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 730.834864] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4a06eca-83a9-4788-9482-ac799d94b2ac {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.840543] env[63515]: DEBUG oslo_vmware.api [None req-dcda8d78-de0d-44e9-b61f-e84b4a49533f tempest-ServersTestManualDisk-1114050206 tempest-ServersTestManualDisk-1114050206-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52a99de3-fa8b-e940-7c60-c203753ee53e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 730.860276] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-e71315b0-9587-4929-82a2-71bebc6fcbfc tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] [instance: d7e8dcdb-41cb-46fb-8b61-d251e7c2d372] Reconfiguring VM instance instance-0000002c to attach disk [datastore1] d7e8dcdb-41cb-46fb-8b61-d251e7c2d372/d7e8dcdb-41cb-46fb-8b61-d251e7c2d372.vmdk or device None with type sparse {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 730.866863] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d26093c6-c413-4283-9e94-738c8651a56f {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.878907] env[63515]: DEBUG oslo_concurrency.lockutils [None req-99f0df88-8ade-4c67-afd9-dbdc6b385015 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.622s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 730.879127] env[63515]: INFO nova.compute.manager [None req-99f0df88-8ade-4c67-afd9-dbdc6b385015 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] [instance: f1d01b75-ac9d-458d-8cc2-ae64cffca4e8] Migrating [ 730.879354] env[63515]: DEBUG oslo_concurrency.lockutils [None req-99f0df88-8ade-4c67-afd9-dbdc6b385015 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Acquiring lock "compute-rpcapi-router" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 730.879496] env[63515]: DEBUG oslo_concurrency.lockutils [None req-99f0df88-8ade-4c67-afd9-dbdc6b385015 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Acquired lock "compute-rpcapi-router" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 730.881637] env[63515]: DEBUG oslo_concurrency.lockutils [None req-3b658824-6e25-4b06-878c-99b3be1f1d31 tempest-ServerTagsTestJSON-1170819896 tempest-ServerTagsTestJSON-1170819896-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.463s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 730.883236] env[63515]: INFO nova.compute.claims [None req-3b658824-6e25-4b06-878c-99b3be1f1d31 tempest-ServerTagsTestJSON-1170819896 tempest-ServerTagsTestJSON-1170819896-project-member] [instance: de210780-5c0f-4fba-883c-13707566a2e1] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 730.891192] env[63515]: DEBUG oslo_vmware.api [None req-e71315b0-9587-4929-82a2-71bebc6fcbfc tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Waiting for the task: (returnval){ [ 730.891192] env[63515]: value = "task-1111021" [ 730.891192] env[63515]: _type = "Task" [ 730.891192] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 730.899919] env[63515]: DEBUG oslo_vmware.api [None req-e71315b0-9587-4929-82a2-71bebc6fcbfc tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Task: {'id': task-1111021, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 731.028984] env[63515]: DEBUG nova.network.neutron [req-c2b6a640-8446-45dc-8536-6c1c19bd5b73 req-525af8da-09ba-4a1b-9054-d9caf82da16d service nova] [instance: e91aa479-1540-4950-851b-b2409e5f89f1] Updated VIF entry in instance network info cache for port fcab3239-8739-4ac7-9f35-4a121630aa6f. {{(pid=63515) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 731.029387] env[63515]: DEBUG nova.network.neutron [req-c2b6a640-8446-45dc-8536-6c1c19bd5b73 req-525af8da-09ba-4a1b-9054-d9caf82da16d service nova] [instance: e91aa479-1540-4950-851b-b2409e5f89f1] Updating instance_info_cache with network_info: [{"id": "fcab3239-8739-4ac7-9f35-4a121630aa6f", "address": "fa:16:3e:3d:c1:97", "network": {"id": "1604b180-4129-4de5-92e6-79f13694e373", "bridge": "br-int", "label": "tempest-ServersTestManualDisk-718841075-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "957bed1311824087944c5d19cfd0879a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8abee039-d93e-48a7-8911-6416a3e1ff30", "external-id": "nsx-vlan-transportzone-654", "segmentation_id": 654, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfcab3239-87", "ovs_interfaceid": "fcab3239-8739-4ac7-9f35-4a121630aa6f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 731.318882] env[63515]: DEBUG oslo_concurrency.lockutils [None req-4bc5b4e9-701c-413d-abd5-d300a7447cb8 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Acquiring lock "refresh_cache-ed550b10-d58f-45b8-b766-198f431c3788" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 731.319093] env[63515]: DEBUG oslo_concurrency.lockutils [None req-4bc5b4e9-701c-413d-abd5-d300a7447cb8 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Acquired lock "refresh_cache-ed550b10-d58f-45b8-b766-198f431c3788" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 731.319281] env[63515]: DEBUG nova.network.neutron [None req-4bc5b4e9-701c-413d-abd5-d300a7447cb8 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: ed550b10-d58f-45b8-b766-198f431c3788] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 731.336659] env[63515]: DEBUG oslo_vmware.api [None req-dcda8d78-de0d-44e9-b61f-e84b4a49533f tempest-ServersTestManualDisk-1114050206 tempest-ServersTestManualDisk-1114050206-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52a99de3-fa8b-e940-7c60-c203753ee53e, 'name': SearchDatastore_Task, 'duration_secs': 0.020315} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 731.336659] env[63515]: DEBUG oslo_concurrency.lockutils [None req-dcda8d78-de0d-44e9-b61f-e84b4a49533f tempest-ServersTestManualDisk-1114050206 tempest-ServersTestManualDisk-1114050206-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 731.336873] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-dcda8d78-de0d-44e9-b61f-e84b4a49533f tempest-ServersTestManualDisk-1114050206 tempest-ServersTestManualDisk-1114050206-project-member] [instance: e91aa479-1540-4950-851b-b2409e5f89f1] Processing image 8a120570-cb06-4099-b262-554ca0ad15c5 {{(pid=63515) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 731.337024] env[63515]: DEBUG oslo_concurrency.lockutils [None req-dcda8d78-de0d-44e9-b61f-e84b4a49533f tempest-ServersTestManualDisk-1114050206 tempest-ServersTestManualDisk-1114050206-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 731.337115] env[63515]: DEBUG oslo_concurrency.lockutils [None req-dcda8d78-de0d-44e9-b61f-e84b4a49533f tempest-ServersTestManualDisk-1114050206 tempest-ServersTestManualDisk-1114050206-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 731.337295] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-dcda8d78-de0d-44e9-b61f-e84b4a49533f tempest-ServersTestManualDisk-1114050206 tempest-ServersTestManualDisk-1114050206-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 731.337559] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-64b5a62b-5e24-4b5a-b2a7-02f6ffb7b197 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.347261] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-dcda8d78-de0d-44e9-b61f-e84b4a49533f tempest-ServersTestManualDisk-1114050206 tempest-ServersTestManualDisk-1114050206-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 731.347523] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-dcda8d78-de0d-44e9-b61f-e84b4a49533f tempest-ServersTestManualDisk-1114050206 tempest-ServersTestManualDisk-1114050206-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63515) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 731.348803] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f423d73c-7a1a-48ac-91cf-0e631aea1088 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.354077] env[63515]: DEBUG oslo_vmware.api [None req-dcda8d78-de0d-44e9-b61f-e84b4a49533f tempest-ServersTestManualDisk-1114050206 tempest-ServersTestManualDisk-1114050206-project-member] Waiting for the task: (returnval){ [ 731.354077] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]522b0ceb-699e-2b7f-d88c-ff80312380e7" [ 731.354077] env[63515]: _type = "Task" [ 731.354077] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 731.361876] env[63515]: DEBUG oslo_vmware.api [None req-dcda8d78-de0d-44e9-b61f-e84b4a49533f tempest-ServersTestManualDisk-1114050206 tempest-ServersTestManualDisk-1114050206-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]522b0ceb-699e-2b7f-d88c-ff80312380e7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 731.382817] env[63515]: INFO nova.compute.rpcapi [None req-99f0df88-8ade-4c67-afd9-dbdc6b385015 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Automatically selected compute RPC version 6.3 from minimum service version 67 [ 731.383028] env[63515]: DEBUG oslo_concurrency.lockutils [None req-99f0df88-8ade-4c67-afd9-dbdc6b385015 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Releasing lock "compute-rpcapi-router" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 731.399695] env[63515]: DEBUG nova.compute.manager [req-9535ec3a-5c41-45ca-b2ee-433e366322e2 req-f2c6a066-fcd0-4f28-903a-0b02aa2dc5a2 service nova] [instance: ed550b10-d58f-45b8-b766-198f431c3788] Received event network-vif-plugged-ad9dda29-9326-472b-8c82-294a33a7faf0 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 731.399988] env[63515]: DEBUG oslo_concurrency.lockutils [req-9535ec3a-5c41-45ca-b2ee-433e366322e2 req-f2c6a066-fcd0-4f28-903a-0b02aa2dc5a2 service nova] Acquiring lock "ed550b10-d58f-45b8-b766-198f431c3788-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 731.400296] env[63515]: DEBUG oslo_concurrency.lockutils [req-9535ec3a-5c41-45ca-b2ee-433e366322e2 req-f2c6a066-fcd0-4f28-903a-0b02aa2dc5a2 service nova] Lock "ed550b10-d58f-45b8-b766-198f431c3788-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 731.400535] env[63515]: DEBUG oslo_concurrency.lockutils [req-9535ec3a-5c41-45ca-b2ee-433e366322e2 req-f2c6a066-fcd0-4f28-903a-0b02aa2dc5a2 service nova] Lock "ed550b10-d58f-45b8-b766-198f431c3788-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 731.400778] env[63515]: DEBUG nova.compute.manager [req-9535ec3a-5c41-45ca-b2ee-433e366322e2 req-f2c6a066-fcd0-4f28-903a-0b02aa2dc5a2 service nova] [instance: ed550b10-d58f-45b8-b766-198f431c3788] No waiting events found dispatching network-vif-plugged-ad9dda29-9326-472b-8c82-294a33a7faf0 {{(pid=63515) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 731.401310] env[63515]: WARNING nova.compute.manager [req-9535ec3a-5c41-45ca-b2ee-433e366322e2 req-f2c6a066-fcd0-4f28-903a-0b02aa2dc5a2 service nova] [instance: ed550b10-d58f-45b8-b766-198f431c3788] Received unexpected event network-vif-plugged-ad9dda29-9326-472b-8c82-294a33a7faf0 for instance with vm_state building and task_state spawning. [ 731.401310] env[63515]: DEBUG nova.compute.manager [req-9535ec3a-5c41-45ca-b2ee-433e366322e2 req-f2c6a066-fcd0-4f28-903a-0b02aa2dc5a2 service nova] [instance: ed550b10-d58f-45b8-b766-198f431c3788] Received event network-changed-ad9dda29-9326-472b-8c82-294a33a7faf0 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 731.401738] env[63515]: DEBUG nova.compute.manager [req-9535ec3a-5c41-45ca-b2ee-433e366322e2 req-f2c6a066-fcd0-4f28-903a-0b02aa2dc5a2 service nova] [instance: ed550b10-d58f-45b8-b766-198f431c3788] Refreshing instance network info cache due to event network-changed-ad9dda29-9326-472b-8c82-294a33a7faf0. {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 731.401738] env[63515]: DEBUG oslo_concurrency.lockutils [req-9535ec3a-5c41-45ca-b2ee-433e366322e2 req-f2c6a066-fcd0-4f28-903a-0b02aa2dc5a2 service nova] Acquiring lock "refresh_cache-ed550b10-d58f-45b8-b766-198f431c3788" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 731.426308] env[63515]: DEBUG oslo_vmware.api [None req-e71315b0-9587-4929-82a2-71bebc6fcbfc tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Task: {'id': task-1111021, 'name': ReconfigVM_Task, 'duration_secs': 0.301702} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 731.426467] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-e71315b0-9587-4929-82a2-71bebc6fcbfc tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] [instance: d7e8dcdb-41cb-46fb-8b61-d251e7c2d372] Reconfigured VM instance instance-0000002c to attach disk [datastore1] d7e8dcdb-41cb-46fb-8b61-d251e7c2d372/d7e8dcdb-41cb-46fb-8b61-d251e7c2d372.vmdk or device None with type sparse {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 731.427782] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-7537a167-1293-46d9-b2e7-0e5f45c04cde {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.436242] env[63515]: DEBUG oslo_vmware.api [None req-e71315b0-9587-4929-82a2-71bebc6fcbfc tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Waiting for the task: (returnval){ [ 731.436242] env[63515]: value = "task-1111022" [ 731.436242] env[63515]: _type = "Task" [ 731.436242] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 731.448121] env[63515]: DEBUG oslo_vmware.api [None req-e71315b0-9587-4929-82a2-71bebc6fcbfc tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Task: {'id': task-1111022, 'name': Rename_Task} progress is 5%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 731.531682] env[63515]: DEBUG oslo_concurrency.lockutils [req-c2b6a640-8446-45dc-8536-6c1c19bd5b73 req-525af8da-09ba-4a1b-9054-d9caf82da16d service nova] Releasing lock "refresh_cache-e91aa479-1540-4950-851b-b2409e5f89f1" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 731.565308] env[63515]: DEBUG nova.virt.hardware [None req-996c1e42-eed8-414b-a716-3f08d65ec493 tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T02:52:14Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T02:51:56Z,direct_url=,disk_format='vmdk',id=8a120570-cb06-4099-b262-554ca0ad15c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b569255cc43e42e7ae3f2b4ad37c6ef4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T02:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 731.565599] env[63515]: DEBUG nova.virt.hardware [None req-996c1e42-eed8-414b-a716-3f08d65ec493 tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Flavor limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 731.565779] env[63515]: DEBUG nova.virt.hardware [None req-996c1e42-eed8-414b-a716-3f08d65ec493 tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Image limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 731.565996] env[63515]: DEBUG nova.virt.hardware [None req-996c1e42-eed8-414b-a716-3f08d65ec493 tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Flavor pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 731.566120] env[63515]: DEBUG nova.virt.hardware [None req-996c1e42-eed8-414b-a716-3f08d65ec493 tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Image pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 731.566265] env[63515]: DEBUG nova.virt.hardware [None req-996c1e42-eed8-414b-a716-3f08d65ec493 tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 731.566467] env[63515]: DEBUG nova.virt.hardware [None req-996c1e42-eed8-414b-a716-3f08d65ec493 tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 731.566623] env[63515]: DEBUG nova.virt.hardware [None req-996c1e42-eed8-414b-a716-3f08d65ec493 tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 731.566789] env[63515]: DEBUG nova.virt.hardware [None req-996c1e42-eed8-414b-a716-3f08d65ec493 tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Got 1 possible topologies {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 731.566969] env[63515]: DEBUG nova.virt.hardware [None req-996c1e42-eed8-414b-a716-3f08d65ec493 tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 731.567174] env[63515]: DEBUG nova.virt.hardware [None req-996c1e42-eed8-414b-a716-3f08d65ec493 tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 731.568046] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-227a4aba-f014-47e2-b384-7570ee1c22ce {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.579939] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4446f02-438a-425e-81af-568dc561b786 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.594115] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-996c1e42-eed8-414b-a716-3f08d65ec493 tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] [instance: b85e9a70-7f5b-4d32-b616-f2a97e3186c8] Instance VIF info [] {{(pid=63515) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 731.599991] env[63515]: DEBUG oslo.service.loopingcall [None req-996c1e42-eed8-414b-a716-3f08d65ec493 tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 731.600249] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b85e9a70-7f5b-4d32-b616-f2a97e3186c8] Creating VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 731.600487] env[63515]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ce5280fc-a245-4c4a-801f-acd29ab09b3c {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.616812] env[63515]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 731.616812] env[63515]: value = "task-1111023" [ 731.616812] env[63515]: _type = "Task" [ 731.616812] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 731.624512] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1111023, 'name': CreateVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 731.853467] env[63515]: DEBUG nova.network.neutron [None req-4bc5b4e9-701c-413d-abd5-d300a7447cb8 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: ed550b10-d58f-45b8-b766-198f431c3788] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 731.864799] env[63515]: DEBUG oslo_vmware.api [None req-dcda8d78-de0d-44e9-b61f-e84b4a49533f tempest-ServersTestManualDisk-1114050206 tempest-ServersTestManualDisk-1114050206-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]522b0ceb-699e-2b7f-d88c-ff80312380e7, 'name': SearchDatastore_Task, 'duration_secs': 0.039713} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 731.865616] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3d928c58-e8bd-4b85-a23f-92ecaef8c886 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.871167] env[63515]: DEBUG oslo_vmware.api [None req-dcda8d78-de0d-44e9-b61f-e84b4a49533f tempest-ServersTestManualDisk-1114050206 tempest-ServersTestManualDisk-1114050206-project-member] Waiting for the task: (returnval){ [ 731.871167] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]521c1eb1-6bdb-aa6c-e868-ca3952de7132" [ 731.871167] env[63515]: _type = "Task" [ 731.871167] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 731.881693] env[63515]: DEBUG oslo_vmware.api [None req-dcda8d78-de0d-44e9-b61f-e84b4a49533f tempest-ServersTestManualDisk-1114050206 tempest-ServersTestManualDisk-1114050206-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]521c1eb1-6bdb-aa6c-e868-ca3952de7132, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 731.916094] env[63515]: DEBUG oslo_concurrency.lockutils [None req-99f0df88-8ade-4c67-afd9-dbdc6b385015 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Acquiring lock "refresh_cache-f1d01b75-ac9d-458d-8cc2-ae64cffca4e8" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 731.916348] env[63515]: DEBUG oslo_concurrency.lockutils [None req-99f0df88-8ade-4c67-afd9-dbdc6b385015 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Acquired lock "refresh_cache-f1d01b75-ac9d-458d-8cc2-ae64cffca4e8" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 731.916661] env[63515]: DEBUG nova.network.neutron [None req-99f0df88-8ade-4c67-afd9-dbdc6b385015 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] [instance: f1d01b75-ac9d-458d-8cc2-ae64cffca4e8] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 731.953037] env[63515]: DEBUG oslo_vmware.api [None req-e71315b0-9587-4929-82a2-71bebc6fcbfc tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Task: {'id': task-1111022, 'name': Rename_Task, 'duration_secs': 0.474763} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 731.953037] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-e71315b0-9587-4929-82a2-71bebc6fcbfc tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] [instance: d7e8dcdb-41cb-46fb-8b61-d251e7c2d372] Powering on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 731.953308] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2d594ff5-b205-4599-9c66-1c9c2447e2ef {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.960073] env[63515]: DEBUG oslo_vmware.api [None req-e71315b0-9587-4929-82a2-71bebc6fcbfc tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Waiting for the task: (returnval){ [ 731.960073] env[63515]: value = "task-1111024" [ 731.960073] env[63515]: _type = "Task" [ 731.960073] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 731.970919] env[63515]: DEBUG oslo_vmware.api [None req-e71315b0-9587-4929-82a2-71bebc6fcbfc tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Task: {'id': task-1111024, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 732.127789] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1111023, 'name': CreateVM_Task, 'duration_secs': 0.320568} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 732.127960] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b85e9a70-7f5b-4d32-b616-f2a97e3186c8] Created VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 732.128508] env[63515]: DEBUG oslo_concurrency.lockutils [None req-996c1e42-eed8-414b-a716-3f08d65ec493 tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 732.128669] env[63515]: DEBUG oslo_concurrency.lockutils [None req-996c1e42-eed8-414b-a716-3f08d65ec493 tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 732.129225] env[63515]: DEBUG oslo_concurrency.lockutils [None req-996c1e42-eed8-414b-a716-3f08d65ec493 tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 732.129477] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b87dad55-0264-4b49-915d-ea414ab7098f {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.136352] env[63515]: DEBUG oslo_vmware.api [None req-996c1e42-eed8-414b-a716-3f08d65ec493 tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Waiting for the task: (returnval){ [ 732.136352] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]529a47ee-367f-cc4e-567e-e196109eb843" [ 732.136352] env[63515]: _type = "Task" [ 732.136352] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 732.144708] env[63515]: DEBUG oslo_vmware.api [None req-996c1e42-eed8-414b-a716-3f08d65ec493 tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]529a47ee-367f-cc4e-567e-e196109eb843, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 732.223774] env[63515]: DEBUG nova.network.neutron [None req-4bc5b4e9-701c-413d-abd5-d300a7447cb8 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: ed550b10-d58f-45b8-b766-198f431c3788] Updating instance_info_cache with network_info: [{"id": "ad9dda29-9326-472b-8c82-294a33a7faf0", "address": "fa:16:3e:26:e3:72", "network": {"id": "88dfb29d-41e0-4df7-8f70-e52cda05620c", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1355578400-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7d4ada9ec35f42b19c6480a9101d21a4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "357d2811-e990-4985-9f9e-b158d10d3699", "external-id": "nsx-vlan-transportzone-641", "segmentation_id": 641, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapad9dda29-93", "ovs_interfaceid": "ad9dda29-9326-472b-8c82-294a33a7faf0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 732.331025] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7ded71a-ae55-49a4-a47c-f5bc32d82ad6 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.338660] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-baa877c7-6919-48e4-a0ca-25d0c0a624fa {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.372038] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9abb8334-a8bf-4056-be7e-e17d4a372dfb {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.384141] env[63515]: DEBUG oslo_vmware.api [None req-dcda8d78-de0d-44e9-b61f-e84b4a49533f tempest-ServersTestManualDisk-1114050206 tempest-ServersTestManualDisk-1114050206-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]521c1eb1-6bdb-aa6c-e868-ca3952de7132, 'name': SearchDatastore_Task, 'duration_secs': 0.011498} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 732.386236] env[63515]: DEBUG oslo_concurrency.lockutils [None req-dcda8d78-de0d-44e9-b61f-e84b4a49533f tempest-ServersTestManualDisk-1114050206 tempest-ServersTestManualDisk-1114050206-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 732.386510] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-dcda8d78-de0d-44e9-b61f-e84b4a49533f tempest-ServersTestManualDisk-1114050206 tempest-ServersTestManualDisk-1114050206-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk to [datastore1] e91aa479-1540-4950-851b-b2409e5f89f1/e91aa479-1540-4950-851b-b2409e5f89f1.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 732.386841] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-bdea7680-3313-4891-bbc4-a4f3413f65e4 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.389622] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f24edc2-0580-4b13-9cba-54a75fab6a2f {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.405030] env[63515]: DEBUG nova.compute.provider_tree [None req-3b658824-6e25-4b06-878c-99b3be1f1d31 tempest-ServerTagsTestJSON-1170819896 tempest-ServerTagsTestJSON-1170819896-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 732.405724] env[63515]: DEBUG oslo_vmware.api [None req-dcda8d78-de0d-44e9-b61f-e84b4a49533f tempest-ServersTestManualDisk-1114050206 tempest-ServersTestManualDisk-1114050206-project-member] Waiting for the task: (returnval){ [ 732.405724] env[63515]: value = "task-1111025" [ 732.405724] env[63515]: _type = "Task" [ 732.405724] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 732.414936] env[63515]: DEBUG oslo_vmware.api [None req-dcda8d78-de0d-44e9-b61f-e84b4a49533f tempest-ServersTestManualDisk-1114050206 tempest-ServersTestManualDisk-1114050206-project-member] Task: {'id': task-1111025, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 732.473457] env[63515]: DEBUG oslo_vmware.api [None req-e71315b0-9587-4929-82a2-71bebc6fcbfc tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Task: {'id': task-1111024, 'name': PowerOnVM_Task} progress is 100%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 732.647574] env[63515]: DEBUG oslo_vmware.api [None req-996c1e42-eed8-414b-a716-3f08d65ec493 tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]529a47ee-367f-cc4e-567e-e196109eb843, 'name': SearchDatastore_Task, 'duration_secs': 0.008554} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 732.647846] env[63515]: DEBUG oslo_concurrency.lockutils [None req-996c1e42-eed8-414b-a716-3f08d65ec493 tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 732.648097] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-996c1e42-eed8-414b-a716-3f08d65ec493 tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] [instance: b85e9a70-7f5b-4d32-b616-f2a97e3186c8] Processing image 8a120570-cb06-4099-b262-554ca0ad15c5 {{(pid=63515) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 732.648342] env[63515]: DEBUG oslo_concurrency.lockutils [None req-996c1e42-eed8-414b-a716-3f08d65ec493 tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 732.648484] env[63515]: DEBUG oslo_concurrency.lockutils [None req-996c1e42-eed8-414b-a716-3f08d65ec493 tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 732.648664] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-996c1e42-eed8-414b-a716-3f08d65ec493 tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 732.648936] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b8fdddd9-64a2-4a41-89a4-4d34cc29cf4e {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.659809] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-996c1e42-eed8-414b-a716-3f08d65ec493 tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 732.663021] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-996c1e42-eed8-414b-a716-3f08d65ec493 tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63515) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 732.663021] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e58ea099-7ec4-4687-95c1-7392360a5244 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.668138] env[63515]: DEBUG oslo_vmware.api [None req-996c1e42-eed8-414b-a716-3f08d65ec493 tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Waiting for the task: (returnval){ [ 732.668138] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]52b33920-53fb-22de-35e8-48638d4e3e90" [ 732.668138] env[63515]: _type = "Task" [ 732.668138] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 732.681293] env[63515]: DEBUG oslo_vmware.api [None req-996c1e42-eed8-414b-a716-3f08d65ec493 tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52b33920-53fb-22de-35e8-48638d4e3e90, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 732.722433] env[63515]: DEBUG nova.network.neutron [None req-99f0df88-8ade-4c67-afd9-dbdc6b385015 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] [instance: f1d01b75-ac9d-458d-8cc2-ae64cffca4e8] Updating instance_info_cache with network_info: [{"id": "48f0868f-94db-42f7-8153-d1a27fa02707", "address": "fa:16:3e:7a:74:da", "network": {"id": "b4abcbd0-0eac-4bde-ad11-7f93afcedcef", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.21", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "b569255cc43e42e7ae3f2b4ad37c6ef4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0cd5d325-3053-407e-a4ee-f627e82a23f9", "external-id": "nsx-vlan-transportzone-809", "segmentation_id": 809, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap48f0868f-94", "ovs_interfaceid": "48f0868f-94db-42f7-8153-d1a27fa02707", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 732.726850] env[63515]: DEBUG oslo_concurrency.lockutils [None req-4bc5b4e9-701c-413d-abd5-d300a7447cb8 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Releasing lock "refresh_cache-ed550b10-d58f-45b8-b766-198f431c3788" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 732.727240] env[63515]: DEBUG nova.compute.manager [None req-4bc5b4e9-701c-413d-abd5-d300a7447cb8 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: ed550b10-d58f-45b8-b766-198f431c3788] Instance network_info: |[{"id": "ad9dda29-9326-472b-8c82-294a33a7faf0", "address": "fa:16:3e:26:e3:72", "network": {"id": "88dfb29d-41e0-4df7-8f70-e52cda05620c", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1355578400-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7d4ada9ec35f42b19c6480a9101d21a4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "357d2811-e990-4985-9f9e-b158d10d3699", "external-id": "nsx-vlan-transportzone-641", "segmentation_id": 641, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapad9dda29-93", "ovs_interfaceid": "ad9dda29-9326-472b-8c82-294a33a7faf0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 732.727617] env[63515]: DEBUG oslo_concurrency.lockutils [req-9535ec3a-5c41-45ca-b2ee-433e366322e2 req-f2c6a066-fcd0-4f28-903a-0b02aa2dc5a2 service nova] Acquired lock "refresh_cache-ed550b10-d58f-45b8-b766-198f431c3788" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 732.727847] env[63515]: DEBUG nova.network.neutron [req-9535ec3a-5c41-45ca-b2ee-433e366322e2 req-f2c6a066-fcd0-4f28-903a-0b02aa2dc5a2 service nova] [instance: ed550b10-d58f-45b8-b766-198f431c3788] Refreshing network info cache for port ad9dda29-9326-472b-8c82-294a33a7faf0 {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 732.729237] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-4bc5b4e9-701c-413d-abd5-d300a7447cb8 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: ed550b10-d58f-45b8-b766-198f431c3788] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:26:e3:72', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '357d2811-e990-4985-9f9e-b158d10d3699', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ad9dda29-9326-472b-8c82-294a33a7faf0', 'vif_model': 'vmxnet3'}] {{(pid=63515) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 732.739637] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-4bc5b4e9-701c-413d-abd5-d300a7447cb8 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Creating folder: Project (7d4ada9ec35f42b19c6480a9101d21a4). Parent ref: group-v243370. {{(pid=63515) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 732.743616] env[63515]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d66c0d53-7465-46a7-a568-c1db7f4470d6 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.756982] env[63515]: INFO nova.virt.vmwareapi.vm_util [None req-4bc5b4e9-701c-413d-abd5-d300a7447cb8 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Created folder: Project (7d4ada9ec35f42b19c6480a9101d21a4) in parent group-v243370. [ 732.756982] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-4bc5b4e9-701c-413d-abd5-d300a7447cb8 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Creating folder: Instances. Parent ref: group-v243422. {{(pid=63515) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 732.756982] env[63515]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e5dfe560-e53c-4848-b9de-b6887c41ad21 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.768191] env[63515]: INFO nova.virt.vmwareapi.vm_util [None req-4bc5b4e9-701c-413d-abd5-d300a7447cb8 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Created folder: Instances in parent group-v243422. [ 732.768516] env[63515]: DEBUG oslo.service.loopingcall [None req-4bc5b4e9-701c-413d-abd5-d300a7447cb8 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 732.768766] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ed550b10-d58f-45b8-b766-198f431c3788] Creating VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 732.769115] env[63515]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-32d5726b-1441-42cf-bdb7-713dae52ec39 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.795934] env[63515]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 732.795934] env[63515]: value = "task-1111028" [ 732.795934] env[63515]: _type = "Task" [ 732.795934] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 732.806291] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1111028, 'name': CreateVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 732.909921] env[63515]: DEBUG nova.scheduler.client.report [None req-3b658824-6e25-4b06-878c-99b3be1f1d31 tempest-ServerTagsTestJSON-1170819896 tempest-ServerTagsTestJSON-1170819896-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 732.923844] env[63515]: DEBUG oslo_vmware.api [None req-dcda8d78-de0d-44e9-b61f-e84b4a49533f tempest-ServersTestManualDisk-1114050206 tempest-ServersTestManualDisk-1114050206-project-member] Task: {'id': task-1111025, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.467772} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 732.926089] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-dcda8d78-de0d-44e9-b61f-e84b4a49533f tempest-ServersTestManualDisk-1114050206 tempest-ServersTestManualDisk-1114050206-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk to [datastore1] e91aa479-1540-4950-851b-b2409e5f89f1/e91aa479-1540-4950-851b-b2409e5f89f1.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 732.926089] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-dcda8d78-de0d-44e9-b61f-e84b4a49533f tempest-ServersTestManualDisk-1114050206 tempest-ServersTestManualDisk-1114050206-project-member] [instance: e91aa479-1540-4950-851b-b2409e5f89f1] Extending root virtual disk to 1048576 {{(pid=63515) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 732.926089] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-4b53b577-e0b5-471d-a502-2666b7858b86 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.931127] env[63515]: DEBUG oslo_vmware.api [None req-dcda8d78-de0d-44e9-b61f-e84b4a49533f tempest-ServersTestManualDisk-1114050206 tempest-ServersTestManualDisk-1114050206-project-member] Waiting for the task: (returnval){ [ 732.931127] env[63515]: value = "task-1111029" [ 732.931127] env[63515]: _type = "Task" [ 732.931127] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 732.939482] env[63515]: DEBUG oslo_vmware.api [None req-dcda8d78-de0d-44e9-b61f-e84b4a49533f tempest-ServersTestManualDisk-1114050206 tempest-ServersTestManualDisk-1114050206-project-member] Task: {'id': task-1111029, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 732.973771] env[63515]: DEBUG oslo_vmware.api [None req-e71315b0-9587-4929-82a2-71bebc6fcbfc tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Task: {'id': task-1111024, 'name': PowerOnVM_Task, 'duration_secs': 0.514443} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 732.974060] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-e71315b0-9587-4929-82a2-71bebc6fcbfc tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] [instance: d7e8dcdb-41cb-46fb-8b61-d251e7c2d372] Powered on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 732.974261] env[63515]: INFO nova.compute.manager [None req-e71315b0-9587-4929-82a2-71bebc6fcbfc tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] [instance: d7e8dcdb-41cb-46fb-8b61-d251e7c2d372] Took 7.92 seconds to spawn the instance on the hypervisor. [ 732.974507] env[63515]: DEBUG nova.compute.manager [None req-e71315b0-9587-4929-82a2-71bebc6fcbfc tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] [instance: d7e8dcdb-41cb-46fb-8b61-d251e7c2d372] Checking state {{(pid=63515) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 732.975337] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b603659f-c33e-408f-9d0c-5e9e009dadff {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.027885] env[63515]: DEBUG nova.network.neutron [req-9535ec3a-5c41-45ca-b2ee-433e366322e2 req-f2c6a066-fcd0-4f28-903a-0b02aa2dc5a2 service nova] [instance: ed550b10-d58f-45b8-b766-198f431c3788] Updated VIF entry in instance network info cache for port ad9dda29-9326-472b-8c82-294a33a7faf0. {{(pid=63515) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 733.028243] env[63515]: DEBUG nova.network.neutron [req-9535ec3a-5c41-45ca-b2ee-433e366322e2 req-f2c6a066-fcd0-4f28-903a-0b02aa2dc5a2 service nova] [instance: ed550b10-d58f-45b8-b766-198f431c3788] Updating instance_info_cache with network_info: [{"id": "ad9dda29-9326-472b-8c82-294a33a7faf0", "address": "fa:16:3e:26:e3:72", "network": {"id": "88dfb29d-41e0-4df7-8f70-e52cda05620c", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1355578400-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7d4ada9ec35f42b19c6480a9101d21a4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "357d2811-e990-4985-9f9e-b158d10d3699", "external-id": "nsx-vlan-transportzone-641", "segmentation_id": 641, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapad9dda29-93", "ovs_interfaceid": "ad9dda29-9326-472b-8c82-294a33a7faf0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 733.182026] env[63515]: DEBUG oslo_vmware.api [None req-996c1e42-eed8-414b-a716-3f08d65ec493 tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52b33920-53fb-22de-35e8-48638d4e3e90, 'name': SearchDatastore_Task, 'duration_secs': 0.023962} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 733.182026] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-429e08b7-7af5-4b91-9ca5-308d61549153 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.187644] env[63515]: DEBUG oslo_vmware.api [None req-996c1e42-eed8-414b-a716-3f08d65ec493 tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Waiting for the task: (returnval){ [ 733.187644] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]521e3488-34c4-bcc5-b94a-65c415932bba" [ 733.187644] env[63515]: _type = "Task" [ 733.187644] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 733.195123] env[63515]: DEBUG oslo_vmware.api [None req-996c1e42-eed8-414b-a716-3f08d65ec493 tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]521e3488-34c4-bcc5-b94a-65c415932bba, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 733.225047] env[63515]: DEBUG oslo_concurrency.lockutils [None req-99f0df88-8ade-4c67-afd9-dbdc6b385015 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Releasing lock "refresh_cache-f1d01b75-ac9d-458d-8cc2-ae64cffca4e8" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 733.306074] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1111028, 'name': CreateVM_Task, 'duration_secs': 0.446662} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 733.306665] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ed550b10-d58f-45b8-b766-198f431c3788] Created VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 733.308228] env[63515]: DEBUG oslo_concurrency.lockutils [None req-4bc5b4e9-701c-413d-abd5-d300a7447cb8 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 733.308228] env[63515]: DEBUG oslo_concurrency.lockutils [None req-4bc5b4e9-701c-413d-abd5-d300a7447cb8 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 733.308228] env[63515]: DEBUG oslo_concurrency.lockutils [None req-4bc5b4e9-701c-413d-abd5-d300a7447cb8 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 733.308228] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-32ba1c73-6329-431e-aaa4-ea918d8029ad {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.312259] env[63515]: DEBUG oslo_vmware.api [None req-4bc5b4e9-701c-413d-abd5-d300a7447cb8 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Waiting for the task: (returnval){ [ 733.312259] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]529253da-32f1-e446-8719-db1b65527387" [ 733.312259] env[63515]: _type = "Task" [ 733.312259] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 733.321713] env[63515]: DEBUG oslo_vmware.api [None req-4bc5b4e9-701c-413d-abd5-d300a7447cb8 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]529253da-32f1-e446-8719-db1b65527387, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 733.420070] env[63515]: DEBUG oslo_concurrency.lockutils [None req-3b658824-6e25-4b06-878c-99b3be1f1d31 tempest-ServerTagsTestJSON-1170819896 tempest-ServerTagsTestJSON-1170819896-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.538s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 733.420684] env[63515]: DEBUG nova.compute.manager [None req-3b658824-6e25-4b06-878c-99b3be1f1d31 tempest-ServerTagsTestJSON-1170819896 tempest-ServerTagsTestJSON-1170819896-project-member] [instance: de210780-5c0f-4fba-883c-13707566a2e1] Start building networks asynchronously for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 733.423413] env[63515]: DEBUG oslo_concurrency.lockutils [None req-fd4a0cad-edc5-4c70-94f6-406aa5060c1e tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.360s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 733.424819] env[63515]: INFO nova.compute.claims [None req-fd4a0cad-edc5-4c70-94f6-406aa5060c1e tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: 001a48e1-b237-41f8-a6f4-2fe98d173512] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 733.442125] env[63515]: DEBUG oslo_vmware.api [None req-dcda8d78-de0d-44e9-b61f-e84b4a49533f tempest-ServersTestManualDisk-1114050206 tempest-ServersTestManualDisk-1114050206-project-member] Task: {'id': task-1111029, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.067778} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 733.442399] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-dcda8d78-de0d-44e9-b61f-e84b4a49533f tempest-ServersTestManualDisk-1114050206 tempest-ServersTestManualDisk-1114050206-project-member] [instance: e91aa479-1540-4950-851b-b2409e5f89f1] Extended root virtual disk {{(pid=63515) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 733.443262] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-439375e1-7a2c-4ea1-98b5-e9016c714817 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.466612] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-dcda8d78-de0d-44e9-b61f-e84b4a49533f tempest-ServersTestManualDisk-1114050206 tempest-ServersTestManualDisk-1114050206-project-member] [instance: e91aa479-1540-4950-851b-b2409e5f89f1] Reconfiguring VM instance instance-0000002d to attach disk [datastore1] e91aa479-1540-4950-851b-b2409e5f89f1/e91aa479-1540-4950-851b-b2409e5f89f1.vmdk or device None with type sparse {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 733.467620] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cead8642-748e-4821-b443-24f76a61f043 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.491032] env[63515]: DEBUG oslo_vmware.api [None req-dcda8d78-de0d-44e9-b61f-e84b4a49533f tempest-ServersTestManualDisk-1114050206 tempest-ServersTestManualDisk-1114050206-project-member] Waiting for the task: (returnval){ [ 733.491032] env[63515]: value = "task-1111030" [ 733.491032] env[63515]: _type = "Task" [ 733.491032] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 733.495323] env[63515]: INFO nova.compute.manager [None req-e71315b0-9587-4929-82a2-71bebc6fcbfc tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] [instance: d7e8dcdb-41cb-46fb-8b61-d251e7c2d372] Took 24.53 seconds to build instance. [ 733.501260] env[63515]: DEBUG oslo_vmware.api [None req-dcda8d78-de0d-44e9-b61f-e84b4a49533f tempest-ServersTestManualDisk-1114050206 tempest-ServersTestManualDisk-1114050206-project-member] Task: {'id': task-1111030, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 733.532128] env[63515]: DEBUG oslo_concurrency.lockutils [req-9535ec3a-5c41-45ca-b2ee-433e366322e2 req-f2c6a066-fcd0-4f28-903a-0b02aa2dc5a2 service nova] Releasing lock "refresh_cache-ed550b10-d58f-45b8-b766-198f431c3788" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 733.696724] env[63515]: DEBUG oslo_vmware.api [None req-996c1e42-eed8-414b-a716-3f08d65ec493 tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]521e3488-34c4-bcc5-b94a-65c415932bba, 'name': SearchDatastore_Task, 'duration_secs': 0.015165} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 733.696996] env[63515]: DEBUG oslo_concurrency.lockutils [None req-996c1e42-eed8-414b-a716-3f08d65ec493 tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 733.697308] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-996c1e42-eed8-414b-a716-3f08d65ec493 tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk to [datastore2] b85e9a70-7f5b-4d32-b616-f2a97e3186c8/b85e9a70-7f5b-4d32-b616-f2a97e3186c8.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 733.697566] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-34d9f171-1fef-4723-8887-fb884e8f7f14 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.704525] env[63515]: DEBUG oslo_vmware.api [None req-996c1e42-eed8-414b-a716-3f08d65ec493 tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Waiting for the task: (returnval){ [ 733.704525] env[63515]: value = "task-1111031" [ 733.704525] env[63515]: _type = "Task" [ 733.704525] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 733.712626] env[63515]: DEBUG oslo_vmware.api [None req-996c1e42-eed8-414b-a716-3f08d65ec493 tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Task: {'id': task-1111031, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 733.823291] env[63515]: DEBUG oslo_vmware.api [None req-4bc5b4e9-701c-413d-abd5-d300a7447cb8 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]529253da-32f1-e446-8719-db1b65527387, 'name': SearchDatastore_Task, 'duration_secs': 0.077204} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 733.823291] env[63515]: DEBUG oslo_concurrency.lockutils [None req-4bc5b4e9-701c-413d-abd5-d300a7447cb8 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 733.823566] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-4bc5b4e9-701c-413d-abd5-d300a7447cb8 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: ed550b10-d58f-45b8-b766-198f431c3788] Processing image 8a120570-cb06-4099-b262-554ca0ad15c5 {{(pid=63515) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 733.823765] env[63515]: DEBUG oslo_concurrency.lockutils [None req-4bc5b4e9-701c-413d-abd5-d300a7447cb8 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 733.823995] env[63515]: DEBUG oslo_concurrency.lockutils [None req-4bc5b4e9-701c-413d-abd5-d300a7447cb8 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 733.824097] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-4bc5b4e9-701c-413d-abd5-d300a7447cb8 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 733.824378] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-70114934-1eaa-4cbd-a5b8-b0755c577865 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.831788] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-4bc5b4e9-701c-413d-abd5-d300a7447cb8 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 733.832015] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-4bc5b4e9-701c-413d-abd5-d300a7447cb8 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63515) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 733.832767] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e9732b0e-f5f3-472c-8959-b76f165df709 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.838543] env[63515]: DEBUG oslo_vmware.api [None req-4bc5b4e9-701c-413d-abd5-d300a7447cb8 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Waiting for the task: (returnval){ [ 733.838543] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]52a723c4-11a8-625e-7045-fc1f6795614f" [ 733.838543] env[63515]: _type = "Task" [ 733.838543] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 733.846140] env[63515]: DEBUG oslo_vmware.api [None req-4bc5b4e9-701c-413d-abd5-d300a7447cb8 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52a723c4-11a8-625e-7045-fc1f6795614f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 733.929298] env[63515]: DEBUG nova.compute.utils [None req-3b658824-6e25-4b06-878c-99b3be1f1d31 tempest-ServerTagsTestJSON-1170819896 tempest-ServerTagsTestJSON-1170819896-project-member] Using /dev/sd instead of None {{(pid=63515) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 733.933225] env[63515]: DEBUG nova.compute.manager [None req-3b658824-6e25-4b06-878c-99b3be1f1d31 tempest-ServerTagsTestJSON-1170819896 tempest-ServerTagsTestJSON-1170819896-project-member] [instance: de210780-5c0f-4fba-883c-13707566a2e1] Allocating IP information in the background. {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 733.933512] env[63515]: DEBUG nova.network.neutron [None req-3b658824-6e25-4b06-878c-99b3be1f1d31 tempest-ServerTagsTestJSON-1170819896 tempest-ServerTagsTestJSON-1170819896-project-member] [instance: de210780-5c0f-4fba-883c-13707566a2e1] allocate_for_instance() {{(pid=63515) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 733.996919] env[63515]: DEBUG oslo_concurrency.lockutils [None req-e71315b0-9587-4929-82a2-71bebc6fcbfc tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Lock "d7e8dcdb-41cb-46fb-8b61-d251e7c2d372" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 110.760s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 734.008142] env[63515]: DEBUG nova.policy [None req-3b658824-6e25-4b06-878c-99b3be1f1d31 tempest-ServerTagsTestJSON-1170819896 tempest-ServerTagsTestJSON-1170819896-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8977a177450c4d45bb5882976e8ae965', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8fed3f874f794a3d9718b0665b2d40ae', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63515) authorize /opt/stack/nova/nova/policy.py:201}} [ 734.013423] env[63515]: DEBUG oslo_vmware.api [None req-dcda8d78-de0d-44e9-b61f-e84b4a49533f tempest-ServersTestManualDisk-1114050206 tempest-ServersTestManualDisk-1114050206-project-member] Task: {'id': task-1111030, 'name': ReconfigVM_Task, 'duration_secs': 0.26594} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 734.013423] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-dcda8d78-de0d-44e9-b61f-e84b4a49533f tempest-ServersTestManualDisk-1114050206 tempest-ServersTestManualDisk-1114050206-project-member] [instance: e91aa479-1540-4950-851b-b2409e5f89f1] Reconfigured VM instance instance-0000002d to attach disk [datastore1] e91aa479-1540-4950-851b-b2409e5f89f1/e91aa479-1540-4950-851b-b2409e5f89f1.vmdk or device None with type sparse {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 734.015564] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-9c5827f2-1404-43d7-a959-508edb74a5df {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.026097] env[63515]: DEBUG oslo_vmware.api [None req-dcda8d78-de0d-44e9-b61f-e84b4a49533f tempest-ServersTestManualDisk-1114050206 tempest-ServersTestManualDisk-1114050206-project-member] Waiting for the task: (returnval){ [ 734.026097] env[63515]: value = "task-1111032" [ 734.026097] env[63515]: _type = "Task" [ 734.026097] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 734.036862] env[63515]: DEBUG oslo_vmware.api [None req-dcda8d78-de0d-44e9-b61f-e84b4a49533f tempest-ServersTestManualDisk-1114050206 tempest-ServersTestManualDisk-1114050206-project-member] Task: {'id': task-1111032, 'name': Rename_Task} progress is 5%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 734.215396] env[63515]: DEBUG oslo_vmware.api [None req-996c1e42-eed8-414b-a716-3f08d65ec493 tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Task: {'id': task-1111031, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.49476} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 734.215802] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-996c1e42-eed8-414b-a716-3f08d65ec493 tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk to [datastore2] b85e9a70-7f5b-4d32-b616-f2a97e3186c8/b85e9a70-7f5b-4d32-b616-f2a97e3186c8.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 734.216218] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-996c1e42-eed8-414b-a716-3f08d65ec493 tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] [instance: b85e9a70-7f5b-4d32-b616-f2a97e3186c8] Extending root virtual disk to 1048576 {{(pid=63515) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 734.216534] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-1f395f52-2d9b-4e70-8b5f-18250e4ff37d {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.223106] env[63515]: DEBUG oslo_vmware.api [None req-996c1e42-eed8-414b-a716-3f08d65ec493 tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Waiting for the task: (returnval){ [ 734.223106] env[63515]: value = "task-1111033" [ 734.223106] env[63515]: _type = "Task" [ 734.223106] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 734.231153] env[63515]: DEBUG oslo_vmware.api [None req-996c1e42-eed8-414b-a716-3f08d65ec493 tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Task: {'id': task-1111033, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 734.354227] env[63515]: DEBUG oslo_vmware.api [None req-4bc5b4e9-701c-413d-abd5-d300a7447cb8 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52a723c4-11a8-625e-7045-fc1f6795614f, 'name': SearchDatastore_Task, 'duration_secs': 0.007916} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 734.354227] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-92d6d53c-808e-4e40-95e2-8934785fcf57 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.358468] env[63515]: DEBUG oslo_vmware.api [None req-4bc5b4e9-701c-413d-abd5-d300a7447cb8 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Waiting for the task: (returnval){ [ 734.358468] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]52fb4257-8194-3c47-6d03-725334e7ba0b" [ 734.358468] env[63515]: _type = "Task" [ 734.358468] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 734.364508] env[63515]: DEBUG nova.network.neutron [None req-3b658824-6e25-4b06-878c-99b3be1f1d31 tempest-ServerTagsTestJSON-1170819896 tempest-ServerTagsTestJSON-1170819896-project-member] [instance: de210780-5c0f-4fba-883c-13707566a2e1] Successfully created port: cd4f3cb8-f205-4ace-a17f-d54ed4f008bf {{(pid=63515) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 734.372506] env[63515]: DEBUG oslo_vmware.api [None req-4bc5b4e9-701c-413d-abd5-d300a7447cb8 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52fb4257-8194-3c47-6d03-725334e7ba0b, 'name': SearchDatastore_Task, 'duration_secs': 0.008534} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 734.372506] env[63515]: DEBUG oslo_concurrency.lockutils [None req-4bc5b4e9-701c-413d-abd5-d300a7447cb8 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 734.372506] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-4bc5b4e9-701c-413d-abd5-d300a7447cb8 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk to [datastore2] ed550b10-d58f-45b8-b766-198f431c3788/ed550b10-d58f-45b8-b766-198f431c3788.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 734.372891] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d14508ca-036c-4799-9a77-742b94c65a2c {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.379741] env[63515]: DEBUG oslo_vmware.api [None req-4bc5b4e9-701c-413d-abd5-d300a7447cb8 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Waiting for the task: (returnval){ [ 734.379741] env[63515]: value = "task-1111034" [ 734.379741] env[63515]: _type = "Task" [ 734.379741] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 734.389673] env[63515]: DEBUG oslo_vmware.api [None req-4bc5b4e9-701c-413d-abd5-d300a7447cb8 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Task: {'id': task-1111034, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 734.433638] env[63515]: DEBUG nova.compute.manager [None req-3b658824-6e25-4b06-878c-99b3be1f1d31 tempest-ServerTagsTestJSON-1170819896 tempest-ServerTagsTestJSON-1170819896-project-member] [instance: de210780-5c0f-4fba-883c-13707566a2e1] Start building block device mappings for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 734.502022] env[63515]: DEBUG nova.compute.manager [None req-8f7a1df0-5f0d-424a-ae8f-76bee2ee3dbb tempest-ServersAaction247Test-2019176257 tempest-ServersAaction247Test-2019176257-project-member] [instance: 5d50fbd6-a7fb-422c-9dd7-df7140fd0aae] Starting instance... {{(pid=63515) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 734.543119] env[63515]: DEBUG oslo_vmware.api [None req-dcda8d78-de0d-44e9-b61f-e84b4a49533f tempest-ServersTestManualDisk-1114050206 tempest-ServersTestManualDisk-1114050206-project-member] Task: {'id': task-1111032, 'name': Rename_Task, 'duration_secs': 0.167547} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 734.544763] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-dcda8d78-de0d-44e9-b61f-e84b4a49533f tempest-ServersTestManualDisk-1114050206 tempest-ServersTestManualDisk-1114050206-project-member] [instance: e91aa479-1540-4950-851b-b2409e5f89f1] Powering on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 734.546103] env[63515]: DEBUG nova.compute.manager [req-f34e2ff9-4f4f-44c8-99e5-29bcc5c8ce6e req-94b6a6e8-dbfd-4805-b234-01a503247617 service nova] [instance: d7e8dcdb-41cb-46fb-8b61-d251e7c2d372] Received event network-changed-9378f59d-f9bc-4c2e-8336-2c5710f987cf {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 734.546281] env[63515]: DEBUG nova.compute.manager [req-f34e2ff9-4f4f-44c8-99e5-29bcc5c8ce6e req-94b6a6e8-dbfd-4805-b234-01a503247617 service nova] [instance: d7e8dcdb-41cb-46fb-8b61-d251e7c2d372] Refreshing instance network info cache due to event network-changed-9378f59d-f9bc-4c2e-8336-2c5710f987cf. {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 734.546493] env[63515]: DEBUG oslo_concurrency.lockutils [req-f34e2ff9-4f4f-44c8-99e5-29bcc5c8ce6e req-94b6a6e8-dbfd-4805-b234-01a503247617 service nova] Acquiring lock "refresh_cache-d7e8dcdb-41cb-46fb-8b61-d251e7c2d372" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 734.546627] env[63515]: DEBUG oslo_concurrency.lockutils [req-f34e2ff9-4f4f-44c8-99e5-29bcc5c8ce6e req-94b6a6e8-dbfd-4805-b234-01a503247617 service nova] Acquired lock "refresh_cache-d7e8dcdb-41cb-46fb-8b61-d251e7c2d372" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 734.546930] env[63515]: DEBUG nova.network.neutron [req-f34e2ff9-4f4f-44c8-99e5-29bcc5c8ce6e req-94b6a6e8-dbfd-4805-b234-01a503247617 service nova] [instance: d7e8dcdb-41cb-46fb-8b61-d251e7c2d372] Refreshing network info cache for port 9378f59d-f9bc-4c2e-8336-2c5710f987cf {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 734.550807] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2a9102f2-6698-4fef-afbb-8cf00438f4f8 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.559522] env[63515]: DEBUG oslo_vmware.api [None req-dcda8d78-de0d-44e9-b61f-e84b4a49533f tempest-ServersTestManualDisk-1114050206 tempest-ServersTestManualDisk-1114050206-project-member] Waiting for the task: (returnval){ [ 734.559522] env[63515]: value = "task-1111035" [ 734.559522] env[63515]: _type = "Task" [ 734.559522] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 734.570966] env[63515]: DEBUG oslo_vmware.api [None req-dcda8d78-de0d-44e9-b61f-e84b4a49533f tempest-ServersTestManualDisk-1114050206 tempest-ServersTestManualDisk-1114050206-project-member] Task: {'id': task-1111035, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 734.740700] env[63515]: DEBUG oslo_vmware.api [None req-996c1e42-eed8-414b-a716-3f08d65ec493 tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Task: {'id': task-1111033, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.065675} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 734.743330] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-996c1e42-eed8-414b-a716-3f08d65ec493 tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] [instance: b85e9a70-7f5b-4d32-b616-f2a97e3186c8] Extended root virtual disk {{(pid=63515) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 734.744350] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de0a7ad8-54b3-45e1-a0b9-17343f059754 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.753173] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fbea38be-ef1f-41dc-a2e9-bfbc205aea61 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.772743] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-99f0df88-8ade-4c67-afd9-dbdc6b385015 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] [instance: f1d01b75-ac9d-458d-8cc2-ae64cffca4e8] Updating instance 'f1d01b75-ac9d-458d-8cc2-ae64cffca4e8' progress to 0 {{(pid=63515) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 734.800320] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-996c1e42-eed8-414b-a716-3f08d65ec493 tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] [instance: b85e9a70-7f5b-4d32-b616-f2a97e3186c8] Reconfiguring VM instance instance-0000002b to attach disk [datastore2] b85e9a70-7f5b-4d32-b616-f2a97e3186c8/b85e9a70-7f5b-4d32-b616-f2a97e3186c8.vmdk or device None with type sparse {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 734.801243] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3d409f89-2c75-4a28-a36e-0518afdc10aa {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.826558] env[63515]: DEBUG oslo_vmware.api [None req-996c1e42-eed8-414b-a716-3f08d65ec493 tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Waiting for the task: (returnval){ [ 734.826558] env[63515]: value = "task-1111036" [ 734.826558] env[63515]: _type = "Task" [ 734.826558] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 734.836074] env[63515]: DEBUG oslo_vmware.api [None req-996c1e42-eed8-414b-a716-3f08d65ec493 tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Task: {'id': task-1111036, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 734.889092] env[63515]: DEBUG oslo_vmware.api [None req-4bc5b4e9-701c-413d-abd5-d300a7447cb8 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Task: {'id': task-1111034, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.419844} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 734.891812] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-4bc5b4e9-701c-413d-abd5-d300a7447cb8 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk to [datastore2] ed550b10-d58f-45b8-b766-198f431c3788/ed550b10-d58f-45b8-b766-198f431c3788.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 734.892557] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-4bc5b4e9-701c-413d-abd5-d300a7447cb8 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: ed550b10-d58f-45b8-b766-198f431c3788] Extending root virtual disk to 1048576 {{(pid=63515) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 734.893403] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-9c087384-5ed5-4abc-818f-c3467bb5d3a6 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.903026] env[63515]: DEBUG oslo_vmware.api [None req-4bc5b4e9-701c-413d-abd5-d300a7447cb8 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Waiting for the task: (returnval){ [ 734.903026] env[63515]: value = "task-1111037" [ 734.903026] env[63515]: _type = "Task" [ 734.903026] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 734.914482] env[63515]: DEBUG oslo_vmware.api [None req-4bc5b4e9-701c-413d-abd5-d300a7447cb8 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Task: {'id': task-1111037, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 735.025853] env[63515]: DEBUG oslo_concurrency.lockutils [None req-8f7a1df0-5f0d-424a-ae8f-76bee2ee3dbb tempest-ServersAaction247Test-2019176257 tempest-ServersAaction247Test-2019176257-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 735.039125] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d881a6aa-87c8-499b-b1a8-ce0cf8b87404 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.047031] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be8173c8-57a1-4b9b-8d9f-5f6adb1b81ce {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.089636] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f240eee7-613d-4685-83a1-4ad052ea4bd7 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.098657] env[63515]: DEBUG oslo_vmware.api [None req-dcda8d78-de0d-44e9-b61f-e84b4a49533f tempest-ServersTestManualDisk-1114050206 tempest-ServersTestManualDisk-1114050206-project-member] Task: {'id': task-1111035, 'name': PowerOnVM_Task, 'duration_secs': 0.516507} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 735.100673] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-dcda8d78-de0d-44e9-b61f-e84b4a49533f tempest-ServersTestManualDisk-1114050206 tempest-ServersTestManualDisk-1114050206-project-member] [instance: e91aa479-1540-4950-851b-b2409e5f89f1] Powered on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 735.101216] env[63515]: INFO nova.compute.manager [None req-dcda8d78-de0d-44e9-b61f-e84b4a49533f tempest-ServersTestManualDisk-1114050206 tempest-ServersTestManualDisk-1114050206-project-member] [instance: e91aa479-1540-4950-851b-b2409e5f89f1] Took 7.45 seconds to spawn the instance on the hypervisor. [ 735.101216] env[63515]: DEBUG nova.compute.manager [None req-dcda8d78-de0d-44e9-b61f-e84b4a49533f tempest-ServersTestManualDisk-1114050206 tempest-ServersTestManualDisk-1114050206-project-member] [instance: e91aa479-1540-4950-851b-b2409e5f89f1] Checking state {{(pid=63515) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 735.101953] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1762914c-72ef-46b7-bfea-03e9fcc83a1d {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.105889] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-faba3755-9a89-4226-adca-256b2e8a3361 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.126729] env[63515]: DEBUG nova.compute.provider_tree [None req-fd4a0cad-edc5-4c70-94f6-406aa5060c1e tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 735.303817] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-99f0df88-8ade-4c67-afd9-dbdc6b385015 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] [instance: f1d01b75-ac9d-458d-8cc2-ae64cffca4e8] Powering off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 735.304184] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1e0944c0-2e26-4d88-acf6-fb21bd8513ac {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.311965] env[63515]: DEBUG oslo_vmware.api [None req-99f0df88-8ade-4c67-afd9-dbdc6b385015 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Waiting for the task: (returnval){ [ 735.311965] env[63515]: value = "task-1111038" [ 735.311965] env[63515]: _type = "Task" [ 735.311965] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 735.321025] env[63515]: DEBUG oslo_vmware.api [None req-99f0df88-8ade-4c67-afd9-dbdc6b385015 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Task: {'id': task-1111038, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 735.335966] env[63515]: DEBUG oslo_vmware.api [None req-996c1e42-eed8-414b-a716-3f08d65ec493 tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Task: {'id': task-1111036, 'name': ReconfigVM_Task, 'duration_secs': 0.426197} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 735.336300] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-996c1e42-eed8-414b-a716-3f08d65ec493 tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] [instance: b85e9a70-7f5b-4d32-b616-f2a97e3186c8] Reconfigured VM instance instance-0000002b to attach disk [datastore2] b85e9a70-7f5b-4d32-b616-f2a97e3186c8/b85e9a70-7f5b-4d32-b616-f2a97e3186c8.vmdk or device None with type sparse {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 735.336910] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-689df603-aaa5-474c-bb65-2b26d4d2303a {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.342854] env[63515]: DEBUG oslo_vmware.api [None req-996c1e42-eed8-414b-a716-3f08d65ec493 tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Waiting for the task: (returnval){ [ 735.342854] env[63515]: value = "task-1111039" [ 735.342854] env[63515]: _type = "Task" [ 735.342854] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 735.354106] env[63515]: DEBUG oslo_vmware.api [None req-996c1e42-eed8-414b-a716-3f08d65ec493 tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Task: {'id': task-1111039, 'name': Rename_Task} progress is 5%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 735.355052] env[63515]: DEBUG nova.network.neutron [req-f34e2ff9-4f4f-44c8-99e5-29bcc5c8ce6e req-94b6a6e8-dbfd-4805-b234-01a503247617 service nova] [instance: d7e8dcdb-41cb-46fb-8b61-d251e7c2d372] Updated VIF entry in instance network info cache for port 9378f59d-f9bc-4c2e-8336-2c5710f987cf. {{(pid=63515) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 735.355485] env[63515]: DEBUG nova.network.neutron [req-f34e2ff9-4f4f-44c8-99e5-29bcc5c8ce6e req-94b6a6e8-dbfd-4805-b234-01a503247617 service nova] [instance: d7e8dcdb-41cb-46fb-8b61-d251e7c2d372] Updating instance_info_cache with network_info: [{"id": "9378f59d-f9bc-4c2e-8336-2c5710f987cf", "address": "fa:16:3e:7b:d9:e1", "network": {"id": "57de0b07-78e5-453d-83ba-b12b7dcaf6ce", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-272799484-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.170", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "12a90929c98e40d9a1263929e2b97532", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a1895250-76cc-41f7-b7f8-2e5679494607", "external-id": "nsx-vlan-transportzone-785", "segmentation_id": 785, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9378f59d-f9", "ovs_interfaceid": "9378f59d-f9bc-4c2e-8336-2c5710f987cf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 735.410740] env[63515]: DEBUG oslo_vmware.api [None req-4bc5b4e9-701c-413d-abd5-d300a7447cb8 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Task: {'id': task-1111037, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.133074} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 735.411035] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-4bc5b4e9-701c-413d-abd5-d300a7447cb8 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: ed550b10-d58f-45b8-b766-198f431c3788] Extended root virtual disk {{(pid=63515) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 735.411814] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-687aaed2-c82e-4b36-9467-f5cd6588b445 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.435276] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-4bc5b4e9-701c-413d-abd5-d300a7447cb8 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: ed550b10-d58f-45b8-b766-198f431c3788] Reconfiguring VM instance instance-0000002e to attach disk [datastore2] ed550b10-d58f-45b8-b766-198f431c3788/ed550b10-d58f-45b8-b766-198f431c3788.vmdk or device None with type sparse {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 735.435626] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d36865f7-e9c7-42ac-a25d-8324f4a530fe {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.453329] env[63515]: DEBUG nova.compute.manager [None req-3b658824-6e25-4b06-878c-99b3be1f1d31 tempest-ServerTagsTestJSON-1170819896 tempest-ServerTagsTestJSON-1170819896-project-member] [instance: de210780-5c0f-4fba-883c-13707566a2e1] Start spawning the instance on the hypervisor. {{(pid=63515) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 735.461526] env[63515]: DEBUG oslo_vmware.api [None req-4bc5b4e9-701c-413d-abd5-d300a7447cb8 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Waiting for the task: (returnval){ [ 735.461526] env[63515]: value = "task-1111040" [ 735.461526] env[63515]: _type = "Task" [ 735.461526] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 735.472577] env[63515]: DEBUG oslo_vmware.api [None req-4bc5b4e9-701c-413d-abd5-d300a7447cb8 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Task: {'id': task-1111040, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 735.488665] env[63515]: DEBUG nova.virt.hardware [None req-3b658824-6e25-4b06-878c-99b3be1f1d31 tempest-ServerTagsTestJSON-1170819896 tempest-ServerTagsTestJSON-1170819896-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T02:52:14Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T02:51:56Z,direct_url=,disk_format='vmdk',id=8a120570-cb06-4099-b262-554ca0ad15c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b569255cc43e42e7ae3f2b4ad37c6ef4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T02:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 735.488971] env[63515]: DEBUG nova.virt.hardware [None req-3b658824-6e25-4b06-878c-99b3be1f1d31 tempest-ServerTagsTestJSON-1170819896 tempest-ServerTagsTestJSON-1170819896-project-member] Flavor limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 735.489155] env[63515]: DEBUG nova.virt.hardware [None req-3b658824-6e25-4b06-878c-99b3be1f1d31 tempest-ServerTagsTestJSON-1170819896 tempest-ServerTagsTestJSON-1170819896-project-member] Image limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 735.489339] env[63515]: DEBUG nova.virt.hardware [None req-3b658824-6e25-4b06-878c-99b3be1f1d31 tempest-ServerTagsTestJSON-1170819896 tempest-ServerTagsTestJSON-1170819896-project-member] Flavor pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 735.489485] env[63515]: DEBUG nova.virt.hardware [None req-3b658824-6e25-4b06-878c-99b3be1f1d31 tempest-ServerTagsTestJSON-1170819896 tempest-ServerTagsTestJSON-1170819896-project-member] Image pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 735.489631] env[63515]: DEBUG nova.virt.hardware [None req-3b658824-6e25-4b06-878c-99b3be1f1d31 tempest-ServerTagsTestJSON-1170819896 tempest-ServerTagsTestJSON-1170819896-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 735.489835] env[63515]: DEBUG nova.virt.hardware [None req-3b658824-6e25-4b06-878c-99b3be1f1d31 tempest-ServerTagsTestJSON-1170819896 tempest-ServerTagsTestJSON-1170819896-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 735.490033] env[63515]: DEBUG nova.virt.hardware [None req-3b658824-6e25-4b06-878c-99b3be1f1d31 tempest-ServerTagsTestJSON-1170819896 tempest-ServerTagsTestJSON-1170819896-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 735.490355] env[63515]: DEBUG nova.virt.hardware [None req-3b658824-6e25-4b06-878c-99b3be1f1d31 tempest-ServerTagsTestJSON-1170819896 tempest-ServerTagsTestJSON-1170819896-project-member] Got 1 possible topologies {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 735.490395] env[63515]: DEBUG nova.virt.hardware [None req-3b658824-6e25-4b06-878c-99b3be1f1d31 tempest-ServerTagsTestJSON-1170819896 tempest-ServerTagsTestJSON-1170819896-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 735.490545] env[63515]: DEBUG nova.virt.hardware [None req-3b658824-6e25-4b06-878c-99b3be1f1d31 tempest-ServerTagsTestJSON-1170819896 tempest-ServerTagsTestJSON-1170819896-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 735.491667] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab05909e-f438-41c1-a1d6-bac694173788 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.500150] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43e43308-e1ff-484b-a125-5afdfcf20b31 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.627920] env[63515]: INFO nova.compute.manager [None req-dcda8d78-de0d-44e9-b61f-e84b4a49533f tempest-ServersTestManualDisk-1114050206 tempest-ServersTestManualDisk-1114050206-project-member] [instance: e91aa479-1540-4950-851b-b2409e5f89f1] Took 25.29 seconds to build instance. [ 735.629267] env[63515]: DEBUG nova.scheduler.client.report [None req-fd4a0cad-edc5-4c70-94f6-406aa5060c1e tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 735.823044] env[63515]: DEBUG oslo_vmware.api [None req-99f0df88-8ade-4c67-afd9-dbdc6b385015 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Task: {'id': task-1111038, 'name': PowerOffVM_Task, 'duration_secs': 0.197204} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 735.823248] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-99f0df88-8ade-4c67-afd9-dbdc6b385015 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] [instance: f1d01b75-ac9d-458d-8cc2-ae64cffca4e8] Powered off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 735.823414] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-99f0df88-8ade-4c67-afd9-dbdc6b385015 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] [instance: f1d01b75-ac9d-458d-8cc2-ae64cffca4e8] Updating instance 'f1d01b75-ac9d-458d-8cc2-ae64cffca4e8' progress to 17 {{(pid=63515) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 735.854606] env[63515]: DEBUG oslo_vmware.api [None req-996c1e42-eed8-414b-a716-3f08d65ec493 tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Task: {'id': task-1111039, 'name': Rename_Task, 'duration_secs': 0.431129} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 735.854936] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-996c1e42-eed8-414b-a716-3f08d65ec493 tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] [instance: b85e9a70-7f5b-4d32-b616-f2a97e3186c8] Powering on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 735.855209] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c97cd9d8-c47c-4c3b-b8dc-c8e114c0485a {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.857918] env[63515]: DEBUG oslo_concurrency.lockutils [req-f34e2ff9-4f4f-44c8-99e5-29bcc5c8ce6e req-94b6a6e8-dbfd-4805-b234-01a503247617 service nova] Releasing lock "refresh_cache-d7e8dcdb-41cb-46fb-8b61-d251e7c2d372" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 735.863613] env[63515]: DEBUG oslo_vmware.api [None req-996c1e42-eed8-414b-a716-3f08d65ec493 tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Waiting for the task: (returnval){ [ 735.863613] env[63515]: value = "task-1111041" [ 735.863613] env[63515]: _type = "Task" [ 735.863613] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 735.872236] env[63515]: DEBUG oslo_vmware.api [None req-996c1e42-eed8-414b-a716-3f08d65ec493 tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Task: {'id': task-1111041, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 735.880719] env[63515]: DEBUG nova.compute.manager [req-3967a008-41be-4aa9-93f0-1f45e58a6e74 req-818a58f7-0db6-4472-a43d-40ebbda13676 service nova] [instance: de210780-5c0f-4fba-883c-13707566a2e1] Received event network-vif-plugged-cd4f3cb8-f205-4ace-a17f-d54ed4f008bf {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 735.880953] env[63515]: DEBUG oslo_concurrency.lockutils [req-3967a008-41be-4aa9-93f0-1f45e58a6e74 req-818a58f7-0db6-4472-a43d-40ebbda13676 service nova] Acquiring lock "de210780-5c0f-4fba-883c-13707566a2e1-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 735.881203] env[63515]: DEBUG oslo_concurrency.lockutils [req-3967a008-41be-4aa9-93f0-1f45e58a6e74 req-818a58f7-0db6-4472-a43d-40ebbda13676 service nova] Lock "de210780-5c0f-4fba-883c-13707566a2e1-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 735.881376] env[63515]: DEBUG oslo_concurrency.lockutils [req-3967a008-41be-4aa9-93f0-1f45e58a6e74 req-818a58f7-0db6-4472-a43d-40ebbda13676 service nova] Lock "de210780-5c0f-4fba-883c-13707566a2e1-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 735.881549] env[63515]: DEBUG nova.compute.manager [req-3967a008-41be-4aa9-93f0-1f45e58a6e74 req-818a58f7-0db6-4472-a43d-40ebbda13676 service nova] [instance: de210780-5c0f-4fba-883c-13707566a2e1] No waiting events found dispatching network-vif-plugged-cd4f3cb8-f205-4ace-a17f-d54ed4f008bf {{(pid=63515) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 735.881713] env[63515]: WARNING nova.compute.manager [req-3967a008-41be-4aa9-93f0-1f45e58a6e74 req-818a58f7-0db6-4472-a43d-40ebbda13676 service nova] [instance: de210780-5c0f-4fba-883c-13707566a2e1] Received unexpected event network-vif-plugged-cd4f3cb8-f205-4ace-a17f-d54ed4f008bf for instance with vm_state building and task_state spawning. [ 735.973786] env[63515]: DEBUG oslo_vmware.api [None req-4bc5b4e9-701c-413d-abd5-d300a7447cb8 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Task: {'id': task-1111040, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 735.997456] env[63515]: DEBUG nova.network.neutron [None req-3b658824-6e25-4b06-878c-99b3be1f1d31 tempest-ServerTagsTestJSON-1170819896 tempest-ServerTagsTestJSON-1170819896-project-member] [instance: de210780-5c0f-4fba-883c-13707566a2e1] Successfully updated port: cd4f3cb8-f205-4ace-a17f-d54ed4f008bf {{(pid=63515) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 736.130170] env[63515]: DEBUG oslo_concurrency.lockutils [None req-dcda8d78-de0d-44e9-b61f-e84b4a49533f tempest-ServersTestManualDisk-1114050206 tempest-ServersTestManualDisk-1114050206-project-member] Lock "e91aa479-1540-4950-851b-b2409e5f89f1" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 109.399s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 736.136103] env[63515]: DEBUG oslo_concurrency.lockutils [None req-fd4a0cad-edc5-4c70-94f6-406aa5060c1e tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.711s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 736.136103] env[63515]: DEBUG nova.compute.manager [None req-fd4a0cad-edc5-4c70-94f6-406aa5060c1e tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: 001a48e1-b237-41f8-a6f4-2fe98d173512] Start building networks asynchronously for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 736.138279] env[63515]: DEBUG oslo_concurrency.lockutils [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 18.280s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 736.138475] env[63515]: DEBUG oslo_concurrency.lockutils [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 736.138631] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=63515) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 736.138925] env[63515]: DEBUG oslo_concurrency.lockutils [None req-bb0fc58f-3c69-4616-a8cc-91b6b5f340c3 tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 15.889s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 736.139189] env[63515]: DEBUG nova.objects.instance [None req-bb0fc58f-3c69-4616-a8cc-91b6b5f340c3 tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] Lazy-loading 'resources' on Instance uuid 4087b2c4-6ed1-4b68-8b78-a36e34d935b1 {{(pid=63515) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 736.145777] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe4e377e-34b7-40a6-80fc-23f879ed4875 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.155640] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c23fa31-7e79-44aa-96be-153505c1e977 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.176024] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dec797bc-d314-4599-bd1c-6d57d9edadce {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.181000] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07b14005-6d61-4c6a-9fe1-e0c70c0cd8bf {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.211859] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181575MB free_disk=170GB free_vcpus=48 pci_devices=None {{(pid=63515) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 736.212045] env[63515]: DEBUG oslo_concurrency.lockutils [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 736.333047] env[63515]: DEBUG nova.virt.hardware [None req-99f0df88-8ade-4c67-afd9-dbdc6b385015 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T02:52:16Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=8a120570-cb06-4099-b262-554ca0ad15c5,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 736.333298] env[63515]: DEBUG nova.virt.hardware [None req-99f0df88-8ade-4c67-afd9-dbdc6b385015 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Flavor limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 736.333453] env[63515]: DEBUG nova.virt.hardware [None req-99f0df88-8ade-4c67-afd9-dbdc6b385015 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Image limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 736.333626] env[63515]: DEBUG nova.virt.hardware [None req-99f0df88-8ade-4c67-afd9-dbdc6b385015 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Flavor pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 736.333765] env[63515]: DEBUG nova.virt.hardware [None req-99f0df88-8ade-4c67-afd9-dbdc6b385015 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Image pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 736.333907] env[63515]: DEBUG nova.virt.hardware [None req-99f0df88-8ade-4c67-afd9-dbdc6b385015 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 736.334118] env[63515]: DEBUG nova.virt.hardware [None req-99f0df88-8ade-4c67-afd9-dbdc6b385015 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 736.334273] env[63515]: DEBUG nova.virt.hardware [None req-99f0df88-8ade-4c67-afd9-dbdc6b385015 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 736.334434] env[63515]: DEBUG nova.virt.hardware [None req-99f0df88-8ade-4c67-afd9-dbdc6b385015 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Got 1 possible topologies {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 736.334585] env[63515]: DEBUG nova.virt.hardware [None req-99f0df88-8ade-4c67-afd9-dbdc6b385015 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 736.334896] env[63515]: DEBUG nova.virt.hardware [None req-99f0df88-8ade-4c67-afd9-dbdc6b385015 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 736.342132] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f20f0adc-2801-49fb-bb74-671ab3a6e036 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.358600] env[63515]: DEBUG oslo_vmware.api [None req-99f0df88-8ade-4c67-afd9-dbdc6b385015 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Waiting for the task: (returnval){ [ 736.358600] env[63515]: value = "task-1111042" [ 736.358600] env[63515]: _type = "Task" [ 736.358600] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 736.367410] env[63515]: DEBUG oslo_vmware.api [None req-99f0df88-8ade-4c67-afd9-dbdc6b385015 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Task: {'id': task-1111042, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 736.377774] env[63515]: DEBUG oslo_vmware.api [None req-996c1e42-eed8-414b-a716-3f08d65ec493 tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Task: {'id': task-1111041, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 736.474430] env[63515]: DEBUG oslo_vmware.api [None req-4bc5b4e9-701c-413d-abd5-d300a7447cb8 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Task: {'id': task-1111040, 'name': ReconfigVM_Task, 'duration_secs': 0.757625} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 736.474722] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-4bc5b4e9-701c-413d-abd5-d300a7447cb8 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: ed550b10-d58f-45b8-b766-198f431c3788] Reconfigured VM instance instance-0000002e to attach disk [datastore2] ed550b10-d58f-45b8-b766-198f431c3788/ed550b10-d58f-45b8-b766-198f431c3788.vmdk or device None with type sparse {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 736.475408] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c7e65bc9-6d0d-4d7b-899b-056c9aab828a {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.483150] env[63515]: DEBUG oslo_vmware.api [None req-4bc5b4e9-701c-413d-abd5-d300a7447cb8 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Waiting for the task: (returnval){ [ 736.483150] env[63515]: value = "task-1111043" [ 736.483150] env[63515]: _type = "Task" [ 736.483150] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 736.491966] env[63515]: DEBUG oslo_vmware.api [None req-4bc5b4e9-701c-413d-abd5-d300a7447cb8 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Task: {'id': task-1111043, 'name': Rename_Task} progress is 5%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 736.500645] env[63515]: DEBUG oslo_concurrency.lockutils [None req-3b658824-6e25-4b06-878c-99b3be1f1d31 tempest-ServerTagsTestJSON-1170819896 tempest-ServerTagsTestJSON-1170819896-project-member] Acquiring lock "refresh_cache-de210780-5c0f-4fba-883c-13707566a2e1" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 736.501868] env[63515]: DEBUG oslo_concurrency.lockutils [None req-3b658824-6e25-4b06-878c-99b3be1f1d31 tempest-ServerTagsTestJSON-1170819896 tempest-ServerTagsTestJSON-1170819896-project-member] Acquired lock "refresh_cache-de210780-5c0f-4fba-883c-13707566a2e1" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 736.501868] env[63515]: DEBUG nova.network.neutron [None req-3b658824-6e25-4b06-878c-99b3be1f1d31 tempest-ServerTagsTestJSON-1170819896 tempest-ServerTagsTestJSON-1170819896-project-member] [instance: de210780-5c0f-4fba-883c-13707566a2e1] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 736.585724] env[63515]: DEBUG nova.compute.manager [req-8b21d566-2066-422a-b2ad-7ae20130e08a req-451c41fc-ed0e-4a7d-9bd5-5b022d7af73f service nova] [instance: e91aa479-1540-4950-851b-b2409e5f89f1] Received event network-changed-fcab3239-8739-4ac7-9f35-4a121630aa6f {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 736.588777] env[63515]: DEBUG nova.compute.manager [req-8b21d566-2066-422a-b2ad-7ae20130e08a req-451c41fc-ed0e-4a7d-9bd5-5b022d7af73f service nova] [instance: e91aa479-1540-4950-851b-b2409e5f89f1] Refreshing instance network info cache due to event network-changed-fcab3239-8739-4ac7-9f35-4a121630aa6f. {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 736.590038] env[63515]: DEBUG oslo_concurrency.lockutils [req-8b21d566-2066-422a-b2ad-7ae20130e08a req-451c41fc-ed0e-4a7d-9bd5-5b022d7af73f service nova] Acquiring lock "refresh_cache-e91aa479-1540-4950-851b-b2409e5f89f1" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 736.590038] env[63515]: DEBUG oslo_concurrency.lockutils [req-8b21d566-2066-422a-b2ad-7ae20130e08a req-451c41fc-ed0e-4a7d-9bd5-5b022d7af73f service nova] Acquired lock "refresh_cache-e91aa479-1540-4950-851b-b2409e5f89f1" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 736.590038] env[63515]: DEBUG nova.network.neutron [req-8b21d566-2066-422a-b2ad-7ae20130e08a req-451c41fc-ed0e-4a7d-9bd5-5b022d7af73f service nova] [instance: e91aa479-1540-4950-851b-b2409e5f89f1] Refreshing network info cache for port fcab3239-8739-4ac7-9f35-4a121630aa6f {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 736.636046] env[63515]: DEBUG nova.compute.manager [None req-549cc3f8-330f-4c6a-979d-35eedc6f8146 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: 84dbf321-2c70-4e08-b430-cb5a06fc6829] Starting instance... {{(pid=63515) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 736.644949] env[63515]: DEBUG nova.compute.utils [None req-fd4a0cad-edc5-4c70-94f6-406aa5060c1e tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Using /dev/sd instead of None {{(pid=63515) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 736.646278] env[63515]: DEBUG nova.compute.manager [None req-fd4a0cad-edc5-4c70-94f6-406aa5060c1e tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: 001a48e1-b237-41f8-a6f4-2fe98d173512] Allocating IP information in the background. {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 736.646436] env[63515]: DEBUG nova.network.neutron [None req-fd4a0cad-edc5-4c70-94f6-406aa5060c1e tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: 001a48e1-b237-41f8-a6f4-2fe98d173512] allocate_for_instance() {{(pid=63515) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 736.736964] env[63515]: DEBUG nova.policy [None req-fd4a0cad-edc5-4c70-94f6-406aa5060c1e tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b5de3e36ae88482eb795894592e76c7c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f9c98e9f6020475490aaa7e76d907ab7', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63515) authorize /opt/stack/nova/nova/policy.py:201}} [ 736.879048] env[63515]: DEBUG oslo_vmware.api [None req-99f0df88-8ade-4c67-afd9-dbdc6b385015 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Task: {'id': task-1111042, 'name': ReconfigVM_Task, 'duration_secs': 0.23911} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 736.880450] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-99f0df88-8ade-4c67-afd9-dbdc6b385015 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] [instance: f1d01b75-ac9d-458d-8cc2-ae64cffca4e8] Updating instance 'f1d01b75-ac9d-458d-8cc2-ae64cffca4e8' progress to 33 {{(pid=63515) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 736.890800] env[63515]: DEBUG oslo_vmware.api [None req-996c1e42-eed8-414b-a716-3f08d65ec493 tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Task: {'id': task-1111041, 'name': PowerOnVM_Task, 'duration_secs': 0.809176} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 736.890800] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-996c1e42-eed8-414b-a716-3f08d65ec493 tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] [instance: b85e9a70-7f5b-4d32-b616-f2a97e3186c8] Powered on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 736.890800] env[63515]: DEBUG nova.compute.manager [None req-996c1e42-eed8-414b-a716-3f08d65ec493 tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] [instance: b85e9a70-7f5b-4d32-b616-f2a97e3186c8] Checking state {{(pid=63515) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 736.890800] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dbc687ba-0695-4b03-bb07-c7851373b1fb {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.993555] env[63515]: DEBUG oslo_vmware.api [None req-4bc5b4e9-701c-413d-abd5-d300a7447cb8 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Task: {'id': task-1111043, 'name': Rename_Task, 'duration_secs': 0.210335} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 736.993555] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-4bc5b4e9-701c-413d-abd5-d300a7447cb8 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: ed550b10-d58f-45b8-b766-198f431c3788] Powering on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 736.994681] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1a1e6638-50b4-42cb-9700-95ee0efd211c {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.003390] env[63515]: DEBUG oslo_vmware.api [None req-4bc5b4e9-701c-413d-abd5-d300a7447cb8 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Waiting for the task: (returnval){ [ 737.003390] env[63515]: value = "task-1111044" [ 737.003390] env[63515]: _type = "Task" [ 737.003390] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 737.015740] env[63515]: DEBUG oslo_vmware.api [None req-4bc5b4e9-701c-413d-abd5-d300a7447cb8 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Task: {'id': task-1111044, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 737.041259] env[63515]: DEBUG nova.network.neutron [None req-3b658824-6e25-4b06-878c-99b3be1f1d31 tempest-ServerTagsTestJSON-1170819896 tempest-ServerTagsTestJSON-1170819896-project-member] [instance: de210780-5c0f-4fba-883c-13707566a2e1] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 737.055564] env[63515]: DEBUG nova.objects.instance [None req-36f9f908-0af4-4115-aa42-e2a81d5f503e tempest-AttachInterfacesUnderV243Test-1943594058 tempest-AttachInterfacesUnderV243Test-1943594058-project-member] Lazy-loading 'flavor' on Instance uuid b911a5b5-9617-4fb3-9b5e-fb8c492e4931 {{(pid=63515) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 737.137151] env[63515]: DEBUG nova.network.neutron [None req-fd4a0cad-edc5-4c70-94f6-406aa5060c1e tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: 001a48e1-b237-41f8-a6f4-2fe98d173512] Successfully created port: ed0b7b69-0679-4b35-af6f-44c60792b411 {{(pid=63515) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 737.153765] env[63515]: DEBUG nova.compute.manager [None req-fd4a0cad-edc5-4c70-94f6-406aa5060c1e tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: 001a48e1-b237-41f8-a6f4-2fe98d173512] Start building block device mappings for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 737.169993] env[63515]: DEBUG oslo_concurrency.lockutils [None req-549cc3f8-330f-4c6a-979d-35eedc6f8146 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 737.204280] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53790bbf-28cf-4a47-9e46-c6e579a0467b {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.214162] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f3c95ed-0bd7-4ff8-9c6d-b21533887509 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.250193] env[63515]: DEBUG nova.network.neutron [None req-3b658824-6e25-4b06-878c-99b3be1f1d31 tempest-ServerTagsTestJSON-1170819896 tempest-ServerTagsTestJSON-1170819896-project-member] [instance: de210780-5c0f-4fba-883c-13707566a2e1] Updating instance_info_cache with network_info: [{"id": "cd4f3cb8-f205-4ace-a17f-d54ed4f008bf", "address": "fa:16:3e:13:70:08", "network": {"id": "0a7e65e5-f19f-4bbb-a871-b419001d7a04", "bridge": "br-int", "label": "tempest-ServerTagsTestJSON-1046053312-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8fed3f874f794a3d9718b0665b2d40ae", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7514a465-f1a4-4a8b-b76b-726b1a9d7e2f", "external-id": "nsx-vlan-transportzone-36", "segmentation_id": 36, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcd4f3cb8-f2", "ovs_interfaceid": "cd4f3cb8-f205-4ace-a17f-d54ed4f008bf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 737.251419] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03d11e84-28de-436a-b350-2d58975b12ab {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.263282] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b52c6f96-1169-4fd9-b593-7c7fb8130681 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.278068] env[63515]: DEBUG nova.compute.provider_tree [None req-bb0fc58f-3c69-4616-a8cc-91b6b5f340c3 tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] Updating inventory in ProviderTree for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 737.390407] env[63515]: DEBUG nova.virt.hardware [None req-99f0df88-8ade-4c67-afd9-dbdc6b385015 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T02:54:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='1fae9a09-456d-4e70-9655-fc9d700f1547',id=38,is_public=True,memory_mb=192,name='tempest-test_resize_flavor_-1433437001',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=8a120570-cb06-4099-b262-554ca0ad15c5,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 737.390662] env[63515]: DEBUG nova.virt.hardware [None req-99f0df88-8ade-4c67-afd9-dbdc6b385015 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Flavor limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 737.390817] env[63515]: DEBUG nova.virt.hardware [None req-99f0df88-8ade-4c67-afd9-dbdc6b385015 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Image limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 737.391172] env[63515]: DEBUG nova.virt.hardware [None req-99f0df88-8ade-4c67-afd9-dbdc6b385015 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Flavor pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 737.391172] env[63515]: DEBUG nova.virt.hardware [None req-99f0df88-8ade-4c67-afd9-dbdc6b385015 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Image pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 737.391358] env[63515]: DEBUG nova.virt.hardware [None req-99f0df88-8ade-4c67-afd9-dbdc6b385015 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 737.391599] env[63515]: DEBUG nova.virt.hardware [None req-99f0df88-8ade-4c67-afd9-dbdc6b385015 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 737.391780] env[63515]: DEBUG nova.virt.hardware [None req-99f0df88-8ade-4c67-afd9-dbdc6b385015 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 737.391915] env[63515]: DEBUG nova.virt.hardware [None req-99f0df88-8ade-4c67-afd9-dbdc6b385015 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Got 1 possible topologies {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 737.392583] env[63515]: DEBUG nova.virt.hardware [None req-99f0df88-8ade-4c67-afd9-dbdc6b385015 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 737.392678] env[63515]: DEBUG nova.virt.hardware [None req-99f0df88-8ade-4c67-afd9-dbdc6b385015 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 737.400452] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-99f0df88-8ade-4c67-afd9-dbdc6b385015 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] [instance: f1d01b75-ac9d-458d-8cc2-ae64cffca4e8] Reconfiguring VM instance instance-00000022 to detach disk 2000 {{(pid=63515) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 737.400452] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-88db69f7-c892-4f8f-8634-04d0179e4b93 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.426476] env[63515]: DEBUG oslo_concurrency.lockutils [None req-996c1e42-eed8-414b-a716-3f08d65ec493 tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 737.426965] env[63515]: DEBUG oslo_vmware.api [None req-99f0df88-8ade-4c67-afd9-dbdc6b385015 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Waiting for the task: (returnval){ [ 737.426965] env[63515]: value = "task-1111045" [ 737.426965] env[63515]: _type = "Task" [ 737.426965] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 737.435084] env[63515]: DEBUG oslo_vmware.api [None req-99f0df88-8ade-4c67-afd9-dbdc6b385015 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Task: {'id': task-1111045, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 737.517326] env[63515]: DEBUG oslo_vmware.api [None req-4bc5b4e9-701c-413d-abd5-d300a7447cb8 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Task: {'id': task-1111044, 'name': PowerOnVM_Task} progress is 89%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 737.561263] env[63515]: DEBUG oslo_concurrency.lockutils [None req-36f9f908-0af4-4115-aa42-e2a81d5f503e tempest-AttachInterfacesUnderV243Test-1943594058 tempest-AttachInterfacesUnderV243Test-1943594058-project-member] Acquiring lock "refresh_cache-b911a5b5-9617-4fb3-9b5e-fb8c492e4931" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 737.561461] env[63515]: DEBUG oslo_concurrency.lockutils [None req-36f9f908-0af4-4115-aa42-e2a81d5f503e tempest-AttachInterfacesUnderV243Test-1943594058 tempest-AttachInterfacesUnderV243Test-1943594058-project-member] Acquired lock "refresh_cache-b911a5b5-9617-4fb3-9b5e-fb8c492e4931" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 737.628954] env[63515]: DEBUG nova.network.neutron [req-8b21d566-2066-422a-b2ad-7ae20130e08a req-451c41fc-ed0e-4a7d-9bd5-5b022d7af73f service nova] [instance: e91aa479-1540-4950-851b-b2409e5f89f1] Updated VIF entry in instance network info cache for port fcab3239-8739-4ac7-9f35-4a121630aa6f. {{(pid=63515) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 737.629376] env[63515]: DEBUG nova.network.neutron [req-8b21d566-2066-422a-b2ad-7ae20130e08a req-451c41fc-ed0e-4a7d-9bd5-5b022d7af73f service nova] [instance: e91aa479-1540-4950-851b-b2409e5f89f1] Updating instance_info_cache with network_info: [{"id": "fcab3239-8739-4ac7-9f35-4a121630aa6f", "address": "fa:16:3e:3d:c1:97", "network": {"id": "1604b180-4129-4de5-92e6-79f13694e373", "bridge": "br-int", "label": "tempest-ServersTestManualDisk-718841075-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.187", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "957bed1311824087944c5d19cfd0879a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8abee039-d93e-48a7-8911-6416a3e1ff30", "external-id": "nsx-vlan-transportzone-654", "segmentation_id": 654, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfcab3239-87", "ovs_interfaceid": "fcab3239-8739-4ac7-9f35-4a121630aa6f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 737.758497] env[63515]: DEBUG oslo_concurrency.lockutils [None req-3b658824-6e25-4b06-878c-99b3be1f1d31 tempest-ServerTagsTestJSON-1170819896 tempest-ServerTagsTestJSON-1170819896-project-member] Releasing lock "refresh_cache-de210780-5c0f-4fba-883c-13707566a2e1" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 737.758820] env[63515]: DEBUG nova.compute.manager [None req-3b658824-6e25-4b06-878c-99b3be1f1d31 tempest-ServerTagsTestJSON-1170819896 tempest-ServerTagsTestJSON-1170819896-project-member] [instance: de210780-5c0f-4fba-883c-13707566a2e1] Instance network_info: |[{"id": "cd4f3cb8-f205-4ace-a17f-d54ed4f008bf", "address": "fa:16:3e:13:70:08", "network": {"id": "0a7e65e5-f19f-4bbb-a871-b419001d7a04", "bridge": "br-int", "label": "tempest-ServerTagsTestJSON-1046053312-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8fed3f874f794a3d9718b0665b2d40ae", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7514a465-f1a4-4a8b-b76b-726b1a9d7e2f", "external-id": "nsx-vlan-transportzone-36", "segmentation_id": 36, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcd4f3cb8-f2", "ovs_interfaceid": "cd4f3cb8-f205-4ace-a17f-d54ed4f008bf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 737.759259] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-3b658824-6e25-4b06-878c-99b3be1f1d31 tempest-ServerTagsTestJSON-1170819896 tempest-ServerTagsTestJSON-1170819896-project-member] [instance: de210780-5c0f-4fba-883c-13707566a2e1] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:13:70:08', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '7514a465-f1a4-4a8b-b76b-726b1a9d7e2f', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'cd4f3cb8-f205-4ace-a17f-d54ed4f008bf', 'vif_model': 'vmxnet3'}] {{(pid=63515) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 737.769393] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-3b658824-6e25-4b06-878c-99b3be1f1d31 tempest-ServerTagsTestJSON-1170819896 tempest-ServerTagsTestJSON-1170819896-project-member] Creating folder: Project (8fed3f874f794a3d9718b0665b2d40ae). Parent ref: group-v243370. {{(pid=63515) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 737.769918] env[63515]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-be100ce0-a0e0-4501-9b0f-87ca7dfd3cec {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.784795] env[63515]: INFO nova.virt.vmwareapi.vm_util [None req-3b658824-6e25-4b06-878c-99b3be1f1d31 tempest-ServerTagsTestJSON-1170819896 tempest-ServerTagsTestJSON-1170819896-project-member] Created folder: Project (8fed3f874f794a3d9718b0665b2d40ae) in parent group-v243370. [ 737.784795] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-3b658824-6e25-4b06-878c-99b3be1f1d31 tempest-ServerTagsTestJSON-1170819896 tempest-ServerTagsTestJSON-1170819896-project-member] Creating folder: Instances. Parent ref: group-v243425. {{(pid=63515) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 737.784795] env[63515]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-1ebdecda-4a67-4252-80e9-dbd58434eda4 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.796020] env[63515]: INFO nova.virt.vmwareapi.vm_util [None req-3b658824-6e25-4b06-878c-99b3be1f1d31 tempest-ServerTagsTestJSON-1170819896 tempest-ServerTagsTestJSON-1170819896-project-member] Created folder: Instances in parent group-v243425. [ 737.796020] env[63515]: DEBUG oslo.service.loopingcall [None req-3b658824-6e25-4b06-878c-99b3be1f1d31 tempest-ServerTagsTestJSON-1170819896 tempest-ServerTagsTestJSON-1170819896-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 737.796020] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: de210780-5c0f-4fba-883c-13707566a2e1] Creating VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 737.796020] env[63515]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3c372d31-1eb9-4b62-9530-bddae0bf55b3 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.812518] env[63515]: ERROR nova.scheduler.client.report [None req-bb0fc58f-3c69-4616-a8cc-91b6b5f340c3 tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] [req-7ba2e119-858c-48d7-90ed-0fabbef25c08] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 2bd86232-2b6f-44d5-9057-1a3a6b27185a. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-7ba2e119-858c-48d7-90ed-0fabbef25c08"}]} [ 737.822239] env[63515]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 737.822239] env[63515]: value = "task-1111048" [ 737.822239] env[63515]: _type = "Task" [ 737.822239] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 737.829507] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1111048, 'name': CreateVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 737.833714] env[63515]: DEBUG nova.scheduler.client.report [None req-bb0fc58f-3c69-4616-a8cc-91b6b5f340c3 tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] Refreshing inventories for resource provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 737.851921] env[63515]: DEBUG nova.scheduler.client.report [None req-bb0fc58f-3c69-4616-a8cc-91b6b5f340c3 tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] Updating ProviderTree inventory for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 737.852108] env[63515]: DEBUG nova.compute.provider_tree [None req-bb0fc58f-3c69-4616-a8cc-91b6b5f340c3 tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] Updating inventory in ProviderTree for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 737.865927] env[63515]: DEBUG nova.scheduler.client.report [None req-bb0fc58f-3c69-4616-a8cc-91b6b5f340c3 tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] Refreshing aggregate associations for resource provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a, aggregates: None {{(pid=63515) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 737.890514] env[63515]: DEBUG nova.scheduler.client.report [None req-bb0fc58f-3c69-4616-a8cc-91b6b5f340c3 tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] Refreshing trait associations for resource provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a, traits: COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=63515) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 737.939032] env[63515]: DEBUG oslo_vmware.api [None req-99f0df88-8ade-4c67-afd9-dbdc6b385015 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Task: {'id': task-1111045, 'name': ReconfigVM_Task, 'duration_secs': 0.278435} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 737.939032] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-99f0df88-8ade-4c67-afd9-dbdc6b385015 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] [instance: f1d01b75-ac9d-458d-8cc2-ae64cffca4e8] Reconfigured VM instance instance-00000022 to detach disk 2000 {{(pid=63515) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 737.939032] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7d93032-1ce7-4126-ab93-b4c0c1e3106e {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.967807] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-99f0df88-8ade-4c67-afd9-dbdc6b385015 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] [instance: f1d01b75-ac9d-458d-8cc2-ae64cffca4e8] Reconfiguring VM instance instance-00000022 to attach disk [datastore2] f1d01b75-ac9d-458d-8cc2-ae64cffca4e8/f1d01b75-ac9d-458d-8cc2-ae64cffca4e8.vmdk or device None with type thin {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 737.968378] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e9139905-9634-4b69-b3c1-302a2194559a {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.990469] env[63515]: DEBUG oslo_vmware.api [None req-99f0df88-8ade-4c67-afd9-dbdc6b385015 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Waiting for the task: (returnval){ [ 737.990469] env[63515]: value = "task-1111049" [ 737.990469] env[63515]: _type = "Task" [ 737.990469] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 738.000285] env[63515]: DEBUG oslo_vmware.api [None req-99f0df88-8ade-4c67-afd9-dbdc6b385015 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Task: {'id': task-1111049, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 738.008488] env[63515]: DEBUG nova.compute.manager [req-0a455758-60c3-4284-94b8-bcc64f4f6f31 req-661e7f94-82b3-4d28-85d7-96c992bbe56b service nova] [instance: de210780-5c0f-4fba-883c-13707566a2e1] Received event network-changed-cd4f3cb8-f205-4ace-a17f-d54ed4f008bf {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 738.008834] env[63515]: DEBUG nova.compute.manager [req-0a455758-60c3-4284-94b8-bcc64f4f6f31 req-661e7f94-82b3-4d28-85d7-96c992bbe56b service nova] [instance: de210780-5c0f-4fba-883c-13707566a2e1] Refreshing instance network info cache due to event network-changed-cd4f3cb8-f205-4ace-a17f-d54ed4f008bf. {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 738.009189] env[63515]: DEBUG oslo_concurrency.lockutils [req-0a455758-60c3-4284-94b8-bcc64f4f6f31 req-661e7f94-82b3-4d28-85d7-96c992bbe56b service nova] Acquiring lock "refresh_cache-de210780-5c0f-4fba-883c-13707566a2e1" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 738.009730] env[63515]: DEBUG oslo_concurrency.lockutils [req-0a455758-60c3-4284-94b8-bcc64f4f6f31 req-661e7f94-82b3-4d28-85d7-96c992bbe56b service nova] Acquired lock "refresh_cache-de210780-5c0f-4fba-883c-13707566a2e1" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 738.009730] env[63515]: DEBUG nova.network.neutron [req-0a455758-60c3-4284-94b8-bcc64f4f6f31 req-661e7f94-82b3-4d28-85d7-96c992bbe56b service nova] [instance: de210780-5c0f-4fba-883c-13707566a2e1] Refreshing network info cache for port cd4f3cb8-f205-4ace-a17f-d54ed4f008bf {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 738.039148] env[63515]: DEBUG oslo_vmware.api [None req-4bc5b4e9-701c-413d-abd5-d300a7447cb8 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Task: {'id': task-1111044, 'name': PowerOnVM_Task, 'duration_secs': 0.645782} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 738.041441] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-4bc5b4e9-701c-413d-abd5-d300a7447cb8 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: ed550b10-d58f-45b8-b766-198f431c3788] Powered on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 738.042009] env[63515]: INFO nova.compute.manager [None req-4bc5b4e9-701c-413d-abd5-d300a7447cb8 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: ed550b10-d58f-45b8-b766-198f431c3788] Took 7.75 seconds to spawn the instance on the hypervisor. [ 738.042009] env[63515]: DEBUG nova.compute.manager [None req-4bc5b4e9-701c-413d-abd5-d300a7447cb8 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: ed550b10-d58f-45b8-b766-198f431c3788] Checking state {{(pid=63515) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 738.042974] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4eae476-127c-42e0-9564-5a0b0bfd8fa0 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.067617] env[63515]: DEBUG nova.network.neutron [None req-36f9f908-0af4-4115-aa42-e2a81d5f503e tempest-AttachInterfacesUnderV243Test-1943594058 tempest-AttachInterfacesUnderV243Test-1943594058-project-member] [instance: b911a5b5-9617-4fb3-9b5e-fb8c492e4931] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 738.134498] env[63515]: DEBUG oslo_concurrency.lockutils [req-8b21d566-2066-422a-b2ad-7ae20130e08a req-451c41fc-ed0e-4a7d-9bd5-5b022d7af73f service nova] Releasing lock "refresh_cache-e91aa479-1540-4950-851b-b2409e5f89f1" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 738.167911] env[63515]: DEBUG nova.compute.manager [None req-fd4a0cad-edc5-4c70-94f6-406aa5060c1e tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: 001a48e1-b237-41f8-a6f4-2fe98d173512] Start spawning the instance on the hypervisor. {{(pid=63515) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 738.203798] env[63515]: DEBUG nova.virt.hardware [None req-fd4a0cad-edc5-4c70-94f6-406aa5060c1e tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T02:52:14Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T02:51:56Z,direct_url=,disk_format='vmdk',id=8a120570-cb06-4099-b262-554ca0ad15c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b569255cc43e42e7ae3f2b4ad37c6ef4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T02:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 738.204308] env[63515]: DEBUG nova.virt.hardware [None req-fd4a0cad-edc5-4c70-94f6-406aa5060c1e tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Flavor limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 738.204308] env[63515]: DEBUG nova.virt.hardware [None req-fd4a0cad-edc5-4c70-94f6-406aa5060c1e tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Image limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 738.204510] env[63515]: DEBUG nova.virt.hardware [None req-fd4a0cad-edc5-4c70-94f6-406aa5060c1e tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Flavor pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 738.204661] env[63515]: DEBUG nova.virt.hardware [None req-fd4a0cad-edc5-4c70-94f6-406aa5060c1e tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Image pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 738.205022] env[63515]: DEBUG nova.virt.hardware [None req-fd4a0cad-edc5-4c70-94f6-406aa5060c1e tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 738.205277] env[63515]: DEBUG nova.virt.hardware [None req-fd4a0cad-edc5-4c70-94f6-406aa5060c1e tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 738.205480] env[63515]: DEBUG nova.virt.hardware [None req-fd4a0cad-edc5-4c70-94f6-406aa5060c1e tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 738.205750] env[63515]: DEBUG nova.virt.hardware [None req-fd4a0cad-edc5-4c70-94f6-406aa5060c1e tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Got 1 possible topologies {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 738.205851] env[63515]: DEBUG nova.virt.hardware [None req-fd4a0cad-edc5-4c70-94f6-406aa5060c1e tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 738.206103] env[63515]: DEBUG nova.virt.hardware [None req-fd4a0cad-edc5-4c70-94f6-406aa5060c1e tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 738.207029] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37d499a1-2d23-43fa-a0f4-9a36719e2deb {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.218019] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-282f7dec-f649-4b3f-9111-4096d22ec831 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.307372] env[63515]: DEBUG nova.network.neutron [req-0a455758-60c3-4284-94b8-bcc64f4f6f31 req-661e7f94-82b3-4d28-85d7-96c992bbe56b service nova] [instance: de210780-5c0f-4fba-883c-13707566a2e1] Updated VIF entry in instance network info cache for port cd4f3cb8-f205-4ace-a17f-d54ed4f008bf. {{(pid=63515) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 738.307372] env[63515]: DEBUG nova.network.neutron [req-0a455758-60c3-4284-94b8-bcc64f4f6f31 req-661e7f94-82b3-4d28-85d7-96c992bbe56b service nova] [instance: de210780-5c0f-4fba-883c-13707566a2e1] Updating instance_info_cache with network_info: [{"id": "cd4f3cb8-f205-4ace-a17f-d54ed4f008bf", "address": "fa:16:3e:13:70:08", "network": {"id": "0a7e65e5-f19f-4bbb-a871-b419001d7a04", "bridge": "br-int", "label": "tempest-ServerTagsTestJSON-1046053312-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8fed3f874f794a3d9718b0665b2d40ae", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7514a465-f1a4-4a8b-b76b-726b1a9d7e2f", "external-id": "nsx-vlan-transportzone-36", "segmentation_id": 36, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcd4f3cb8-f2", "ovs_interfaceid": "cd4f3cb8-f205-4ace-a17f-d54ed4f008bf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 738.332942] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1111048, 'name': CreateVM_Task} progress is 99%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 738.372323] env[63515]: DEBUG oslo_concurrency.lockutils [None req-d4ac6a62-1537-4962-916c-41ec7cc28737 tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Acquiring lock "b85e9a70-7f5b-4d32-b616-f2a97e3186c8" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 738.372576] env[63515]: DEBUG oslo_concurrency.lockutils [None req-d4ac6a62-1537-4962-916c-41ec7cc28737 tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Lock "b85e9a70-7f5b-4d32-b616-f2a97e3186c8" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 738.372791] env[63515]: DEBUG oslo_concurrency.lockutils [None req-d4ac6a62-1537-4962-916c-41ec7cc28737 tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Acquiring lock "b85e9a70-7f5b-4d32-b616-f2a97e3186c8-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 738.373058] env[63515]: DEBUG oslo_concurrency.lockutils [None req-d4ac6a62-1537-4962-916c-41ec7cc28737 tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Lock "b85e9a70-7f5b-4d32-b616-f2a97e3186c8-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 738.373244] env[63515]: DEBUG oslo_concurrency.lockutils [None req-d4ac6a62-1537-4962-916c-41ec7cc28737 tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Lock "b85e9a70-7f5b-4d32-b616-f2a97e3186c8-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 738.375487] env[63515]: INFO nova.compute.manager [None req-d4ac6a62-1537-4962-916c-41ec7cc28737 tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] [instance: b85e9a70-7f5b-4d32-b616-f2a97e3186c8] Terminating instance [ 738.377367] env[63515]: DEBUG oslo_concurrency.lockutils [None req-d4ac6a62-1537-4962-916c-41ec7cc28737 tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Acquiring lock "refresh_cache-b85e9a70-7f5b-4d32-b616-f2a97e3186c8" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 738.377610] env[63515]: DEBUG oslo_concurrency.lockutils [None req-d4ac6a62-1537-4962-916c-41ec7cc28737 tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Acquired lock "refresh_cache-b85e9a70-7f5b-4d32-b616-f2a97e3186c8" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 738.377742] env[63515]: DEBUG nova.network.neutron [None req-d4ac6a62-1537-4962-916c-41ec7cc28737 tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] [instance: b85e9a70-7f5b-4d32-b616-f2a97e3186c8] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 738.449725] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-088fa517-6a52-4577-b04c-7b10370ef0b9 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.457415] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2c2e8a9-1a98-4399-a60d-04eb16d2dc6a {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.488157] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0177c894-f501-4abd-a0bc-6de789822d6b {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.501327] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85e83c1a-9727-46ba-9907-63c93b94bd9c {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.505097] env[63515]: DEBUG oslo_vmware.api [None req-99f0df88-8ade-4c67-afd9-dbdc6b385015 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Task: {'id': task-1111049, 'name': ReconfigVM_Task, 'duration_secs': 0.286369} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 738.505365] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-99f0df88-8ade-4c67-afd9-dbdc6b385015 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] [instance: f1d01b75-ac9d-458d-8cc2-ae64cffca4e8] Reconfigured VM instance instance-00000022 to attach disk [datastore2] f1d01b75-ac9d-458d-8cc2-ae64cffca4e8/f1d01b75-ac9d-458d-8cc2-ae64cffca4e8.vmdk or device None with type thin {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 738.505636] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-99f0df88-8ade-4c67-afd9-dbdc6b385015 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] [instance: f1d01b75-ac9d-458d-8cc2-ae64cffca4e8] Updating instance 'f1d01b75-ac9d-458d-8cc2-ae64cffca4e8' progress to 50 {{(pid=63515) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 738.522560] env[63515]: DEBUG nova.compute.provider_tree [None req-bb0fc58f-3c69-4616-a8cc-91b6b5f340c3 tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] Updating inventory in ProviderTree for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 738.560663] env[63515]: INFO nova.compute.manager [None req-4bc5b4e9-701c-413d-abd5-d300a7447cb8 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: ed550b10-d58f-45b8-b766-198f431c3788] Took 26.89 seconds to build instance. [ 738.611414] env[63515]: DEBUG nova.compute.manager [req-4051955c-a19d-4456-9a8c-3b4f9aafb30e req-5fe698ed-4e2d-4e58-84de-95ddc9ebdc4b service nova] [instance: b911a5b5-9617-4fb3-9b5e-fb8c492e4931] Received event network-changed-9bf84cb2-7053-4129-8d8a-dc09e80b384e {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 738.611414] env[63515]: DEBUG nova.compute.manager [req-4051955c-a19d-4456-9a8c-3b4f9aafb30e req-5fe698ed-4e2d-4e58-84de-95ddc9ebdc4b service nova] [instance: b911a5b5-9617-4fb3-9b5e-fb8c492e4931] Refreshing instance network info cache due to event network-changed-9bf84cb2-7053-4129-8d8a-dc09e80b384e. {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 738.611414] env[63515]: DEBUG oslo_concurrency.lockutils [req-4051955c-a19d-4456-9a8c-3b4f9aafb30e req-5fe698ed-4e2d-4e58-84de-95ddc9ebdc4b service nova] Acquiring lock "refresh_cache-b911a5b5-9617-4fb3-9b5e-fb8c492e4931" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 738.814617] env[63515]: DEBUG oslo_concurrency.lockutils [req-0a455758-60c3-4284-94b8-bcc64f4f6f31 req-661e7f94-82b3-4d28-85d7-96c992bbe56b service nova] Releasing lock "refresh_cache-de210780-5c0f-4fba-883c-13707566a2e1" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 738.842025] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1111048, 'name': CreateVM_Task} progress is 99%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 738.902947] env[63515]: DEBUG nova.network.neutron [None req-d4ac6a62-1537-4962-916c-41ec7cc28737 tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] [instance: b85e9a70-7f5b-4d32-b616-f2a97e3186c8] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 738.931480] env[63515]: DEBUG nova.network.neutron [None req-36f9f908-0af4-4115-aa42-e2a81d5f503e tempest-AttachInterfacesUnderV243Test-1943594058 tempest-AttachInterfacesUnderV243Test-1943594058-project-member] [instance: b911a5b5-9617-4fb3-9b5e-fb8c492e4931] Updating instance_info_cache with network_info: [{"id": "9bf84cb2-7053-4129-8d8a-dc09e80b384e", "address": "fa:16:3e:c6:e8:2f", "network": {"id": "22a0cdee-4685-4b2b-9df7-2506b2373a19", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-878688764-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.147", "type": "floating", "version": 4, "meta": {}}]}, {"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5086c79c531e4d039896a24a9ea29f73", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "78b49840-c3fc-455c-8491-a253ccd92bb5", "external-id": "nsx-vlan-transportzone-779", "segmentation_id": 779, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9bf84cb2-70", "ovs_interfaceid": "9bf84cb2-7053-4129-8d8a-dc09e80b384e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 738.984484] env[63515]: DEBUG nova.network.neutron [None req-d4ac6a62-1537-4962-916c-41ec7cc28737 tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] [instance: b85e9a70-7f5b-4d32-b616-f2a97e3186c8] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 739.015934] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5cfcdd15-f9a7-4c97-8490-d6277f2d9860 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.042749] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33630a49-2996-401a-881c-92d4a7f4db60 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.065383] env[63515]: DEBUG oslo_concurrency.lockutils [None req-4bc5b4e9-701c-413d-abd5-d300a7447cb8 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Lock "ed550b10-d58f-45b8-b766-198f431c3788" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 106.567s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 739.065745] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-99f0df88-8ade-4c67-afd9-dbdc6b385015 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] [instance: f1d01b75-ac9d-458d-8cc2-ae64cffca4e8] Updating instance 'f1d01b75-ac9d-458d-8cc2-ae64cffca4e8' progress to 67 {{(pid=63515) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 739.086686] env[63515]: DEBUG nova.scheduler.client.report [None req-bb0fc58f-3c69-4616-a8cc-91b6b5f340c3 tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] Updated inventory for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a with generation 71 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 739.086811] env[63515]: DEBUG nova.compute.provider_tree [None req-bb0fc58f-3c69-4616-a8cc-91b6b5f340c3 tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] Updating resource provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a generation from 71 to 72 during operation: update_inventory {{(pid=63515) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 739.087129] env[63515]: DEBUG nova.compute.provider_tree [None req-bb0fc58f-3c69-4616-a8cc-91b6b5f340c3 tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] Updating inventory in ProviderTree for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 739.136727] env[63515]: DEBUG nova.network.neutron [None req-fd4a0cad-edc5-4c70-94f6-406aa5060c1e tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: 001a48e1-b237-41f8-a6f4-2fe98d173512] Successfully updated port: ed0b7b69-0679-4b35-af6f-44c60792b411 {{(pid=63515) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 739.332743] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1111048, 'name': CreateVM_Task, 'duration_secs': 1.472272} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 739.333102] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: de210780-5c0f-4fba-883c-13707566a2e1] Created VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 739.333702] env[63515]: DEBUG oslo_concurrency.lockutils [None req-3b658824-6e25-4b06-878c-99b3be1f1d31 tempest-ServerTagsTestJSON-1170819896 tempest-ServerTagsTestJSON-1170819896-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 739.333861] env[63515]: DEBUG oslo_concurrency.lockutils [None req-3b658824-6e25-4b06-878c-99b3be1f1d31 tempest-ServerTagsTestJSON-1170819896 tempest-ServerTagsTestJSON-1170819896-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 739.334200] env[63515]: DEBUG oslo_concurrency.lockutils [None req-3b658824-6e25-4b06-878c-99b3be1f1d31 tempest-ServerTagsTestJSON-1170819896 tempest-ServerTagsTestJSON-1170819896-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 739.334454] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4f2c55f2-4a8d-41fd-863c-eefa869b40d7 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.339468] env[63515]: DEBUG oslo_vmware.api [None req-3b658824-6e25-4b06-878c-99b3be1f1d31 tempest-ServerTagsTestJSON-1170819896 tempest-ServerTagsTestJSON-1170819896-project-member] Waiting for the task: (returnval){ [ 739.339468] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]529d50de-8d37-6347-0d00-c50fc4f57ec7" [ 739.339468] env[63515]: _type = "Task" [ 739.339468] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 739.347495] env[63515]: DEBUG oslo_vmware.api [None req-3b658824-6e25-4b06-878c-99b3be1f1d31 tempest-ServerTagsTestJSON-1170819896 tempest-ServerTagsTestJSON-1170819896-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]529d50de-8d37-6347-0d00-c50fc4f57ec7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 739.437489] env[63515]: DEBUG oslo_concurrency.lockutils [None req-36f9f908-0af4-4115-aa42-e2a81d5f503e tempest-AttachInterfacesUnderV243Test-1943594058 tempest-AttachInterfacesUnderV243Test-1943594058-project-member] Releasing lock "refresh_cache-b911a5b5-9617-4fb3-9b5e-fb8c492e4931" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 739.438024] env[63515]: DEBUG nova.compute.manager [None req-36f9f908-0af4-4115-aa42-e2a81d5f503e tempest-AttachInterfacesUnderV243Test-1943594058 tempest-AttachInterfacesUnderV243Test-1943594058-project-member] [instance: b911a5b5-9617-4fb3-9b5e-fb8c492e4931] Inject network info {{(pid=63515) _inject_network_info /opt/stack/nova/nova/compute/manager.py:7219}} [ 739.438337] env[63515]: DEBUG nova.compute.manager [None req-36f9f908-0af4-4115-aa42-e2a81d5f503e tempest-AttachInterfacesUnderV243Test-1943594058 tempest-AttachInterfacesUnderV243Test-1943594058-project-member] [instance: b911a5b5-9617-4fb3-9b5e-fb8c492e4931] network_info to inject: |[{"id": "9bf84cb2-7053-4129-8d8a-dc09e80b384e", "address": "fa:16:3e:c6:e8:2f", "network": {"id": "22a0cdee-4685-4b2b-9df7-2506b2373a19", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-878688764-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.147", "type": "floating", "version": 4, "meta": {}}]}, {"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5086c79c531e4d039896a24a9ea29f73", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "78b49840-c3fc-455c-8491-a253ccd92bb5", "external-id": "nsx-vlan-transportzone-779", "segmentation_id": 779, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9bf84cb2-70", "ovs_interfaceid": "9bf84cb2-7053-4129-8d8a-dc09e80b384e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63515) _inject_network_info /opt/stack/nova/nova/compute/manager.py:7220}} [ 739.448232] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-36f9f908-0af4-4115-aa42-e2a81d5f503e tempest-AttachInterfacesUnderV243Test-1943594058 tempest-AttachInterfacesUnderV243Test-1943594058-project-member] [instance: b911a5b5-9617-4fb3-9b5e-fb8c492e4931] Reconfiguring VM instance to set the machine id {{(pid=63515) _set_machine_id /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1796}} [ 739.448232] env[63515]: DEBUG oslo_concurrency.lockutils [req-4051955c-a19d-4456-9a8c-3b4f9aafb30e req-5fe698ed-4e2d-4e58-84de-95ddc9ebdc4b service nova] Acquired lock "refresh_cache-b911a5b5-9617-4fb3-9b5e-fb8c492e4931" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 739.448232] env[63515]: DEBUG nova.network.neutron [req-4051955c-a19d-4456-9a8c-3b4f9aafb30e req-5fe698ed-4e2d-4e58-84de-95ddc9ebdc4b service nova] [instance: b911a5b5-9617-4fb3-9b5e-fb8c492e4931] Refreshing network info cache for port 9bf84cb2-7053-4129-8d8a-dc09e80b384e {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 739.448654] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-785a2c61-40c0-4cfb-a553-f8bc31b607b2 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.468950] env[63515]: DEBUG nova.objects.instance [None req-91af7cb6-fa08-44e9-a8d2-9d871a7103c3 tempest-AttachInterfacesUnderV243Test-1943594058 tempest-AttachInterfacesUnderV243Test-1943594058-project-member] Lazy-loading 'flavor' on Instance uuid b911a5b5-9617-4fb3-9b5e-fb8c492e4931 {{(pid=63515) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 739.472321] env[63515]: DEBUG oslo_vmware.api [None req-36f9f908-0af4-4115-aa42-e2a81d5f503e tempest-AttachInterfacesUnderV243Test-1943594058 tempest-AttachInterfacesUnderV243Test-1943594058-project-member] Waiting for the task: (returnval){ [ 739.472321] env[63515]: value = "task-1111050" [ 739.472321] env[63515]: _type = "Task" [ 739.472321] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 739.485278] env[63515]: DEBUG oslo_vmware.api [None req-36f9f908-0af4-4115-aa42-e2a81d5f503e tempest-AttachInterfacesUnderV243Test-1943594058 tempest-AttachInterfacesUnderV243Test-1943594058-project-member] Task: {'id': task-1111050, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 739.488240] env[63515]: DEBUG oslo_concurrency.lockutils [None req-d4ac6a62-1537-4962-916c-41ec7cc28737 tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Releasing lock "refresh_cache-b85e9a70-7f5b-4d32-b616-f2a97e3186c8" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 739.488655] env[63515]: DEBUG nova.compute.manager [None req-d4ac6a62-1537-4962-916c-41ec7cc28737 tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] [instance: b85e9a70-7f5b-4d32-b616-f2a97e3186c8] Start destroying the instance on the hypervisor. {{(pid=63515) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 739.488896] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-d4ac6a62-1537-4962-916c-41ec7cc28737 tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] [instance: b85e9a70-7f5b-4d32-b616-f2a97e3186c8] Destroying instance {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 739.490177] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27d0c23e-eb09-4a50-a568-f75714d4a988 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.500982] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-d4ac6a62-1537-4962-916c-41ec7cc28737 tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] [instance: b85e9a70-7f5b-4d32-b616-f2a97e3186c8] Powering off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 739.501413] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-807aa9b2-7914-4c17-a153-2f3d918712da {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.507815] env[63515]: DEBUG oslo_vmware.api [None req-d4ac6a62-1537-4962-916c-41ec7cc28737 tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Waiting for the task: (returnval){ [ 739.507815] env[63515]: value = "task-1111051" [ 739.507815] env[63515]: _type = "Task" [ 739.507815] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 739.524020] env[63515]: DEBUG oslo_vmware.api [None req-d4ac6a62-1537-4962-916c-41ec7cc28737 tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Task: {'id': task-1111051, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 739.578288] env[63515]: DEBUG nova.compute.manager [None req-2786351a-303c-45a8-b63b-65e218eb6ef6 tempest-ServersTestJSON-324454119 tempest-ServersTestJSON-324454119-project-member] [instance: 70ed982f-affd-4dd1-bc90-c64e7c6d49d2] Starting instance... {{(pid=63515) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 739.592598] env[63515]: DEBUG oslo_concurrency.lockutils [None req-bb0fc58f-3c69-4616-a8cc-91b6b5f340c3 tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 3.453s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 739.595309] env[63515]: DEBUG oslo_concurrency.lockutils [None req-99ae4d48-1fae-4ace-913c-d249ca72bbeb tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.753s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 739.596945] env[63515]: INFO nova.compute.claims [None req-99ae4d48-1fae-4ace-913c-d249ca72bbeb tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] [instance: d2a4f66a-f1a2-4f0d-826d-0b7c4f04bd23] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 739.639111] env[63515]: INFO nova.scheduler.client.report [None req-bb0fc58f-3c69-4616-a8cc-91b6b5f340c3 tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] Deleted allocations for instance 4087b2c4-6ed1-4b68-8b78-a36e34d935b1 [ 739.644053] env[63515]: DEBUG oslo_concurrency.lockutils [None req-fd4a0cad-edc5-4c70-94f6-406aa5060c1e tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Acquiring lock "refresh_cache-001a48e1-b237-41f8-a6f4-2fe98d173512" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 739.644053] env[63515]: DEBUG oslo_concurrency.lockutils [None req-fd4a0cad-edc5-4c70-94f6-406aa5060c1e tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Acquired lock "refresh_cache-001a48e1-b237-41f8-a6f4-2fe98d173512" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 739.644053] env[63515]: DEBUG nova.network.neutron [None req-fd4a0cad-edc5-4c70-94f6-406aa5060c1e tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: 001a48e1-b237-41f8-a6f4-2fe98d173512] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 739.679339] env[63515]: DEBUG nova.network.neutron [None req-99f0df88-8ade-4c67-afd9-dbdc6b385015 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] [instance: f1d01b75-ac9d-458d-8cc2-ae64cffca4e8] Port 48f0868f-94db-42f7-8153-d1a27fa02707 binding to destination host cpu-1 is already ACTIVE {{(pid=63515) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 739.726882] env[63515]: DEBUG nova.network.neutron [None req-fd4a0cad-edc5-4c70-94f6-406aa5060c1e tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: 001a48e1-b237-41f8-a6f4-2fe98d173512] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 739.854187] env[63515]: DEBUG oslo_vmware.api [None req-3b658824-6e25-4b06-878c-99b3be1f1d31 tempest-ServerTagsTestJSON-1170819896 tempest-ServerTagsTestJSON-1170819896-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]529d50de-8d37-6347-0d00-c50fc4f57ec7, 'name': SearchDatastore_Task, 'duration_secs': 0.009933} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 739.854187] env[63515]: DEBUG oslo_concurrency.lockutils [None req-3b658824-6e25-4b06-878c-99b3be1f1d31 tempest-ServerTagsTestJSON-1170819896 tempest-ServerTagsTestJSON-1170819896-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 739.854187] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-3b658824-6e25-4b06-878c-99b3be1f1d31 tempest-ServerTagsTestJSON-1170819896 tempest-ServerTagsTestJSON-1170819896-project-member] [instance: de210780-5c0f-4fba-883c-13707566a2e1] Processing image 8a120570-cb06-4099-b262-554ca0ad15c5 {{(pid=63515) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 739.854187] env[63515]: DEBUG oslo_concurrency.lockutils [None req-3b658824-6e25-4b06-878c-99b3be1f1d31 tempest-ServerTagsTestJSON-1170819896 tempest-ServerTagsTestJSON-1170819896-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 739.854520] env[63515]: DEBUG oslo_concurrency.lockutils [None req-3b658824-6e25-4b06-878c-99b3be1f1d31 tempest-ServerTagsTestJSON-1170819896 tempest-ServerTagsTestJSON-1170819896-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 739.854675] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-3b658824-6e25-4b06-878c-99b3be1f1d31 tempest-ServerTagsTestJSON-1170819896 tempest-ServerTagsTestJSON-1170819896-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 739.854988] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-8f0ea96d-2184-468b-afae-031e3e0f78f3 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.870444] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-3b658824-6e25-4b06-878c-99b3be1f1d31 tempest-ServerTagsTestJSON-1170819896 tempest-ServerTagsTestJSON-1170819896-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 739.870759] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-3b658824-6e25-4b06-878c-99b3be1f1d31 tempest-ServerTagsTestJSON-1170819896 tempest-ServerTagsTestJSON-1170819896-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63515) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 739.871630] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9508f668-e360-43e2-8570-8e27851c8645 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.877065] env[63515]: DEBUG oslo_vmware.api [None req-3b658824-6e25-4b06-878c-99b3be1f1d31 tempest-ServerTagsTestJSON-1170819896 tempest-ServerTagsTestJSON-1170819896-project-member] Waiting for the task: (returnval){ [ 739.877065] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]52a1cbb3-a1cd-22cf-5769-1898dbe8b5be" [ 739.877065] env[63515]: _type = "Task" [ 739.877065] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 739.884532] env[63515]: DEBUG oslo_vmware.api [None req-3b658824-6e25-4b06-878c-99b3be1f1d31 tempest-ServerTagsTestJSON-1170819896 tempest-ServerTagsTestJSON-1170819896-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52a1cbb3-a1cd-22cf-5769-1898dbe8b5be, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 739.979654] env[63515]: DEBUG oslo_concurrency.lockutils [None req-91af7cb6-fa08-44e9-a8d2-9d871a7103c3 tempest-AttachInterfacesUnderV243Test-1943594058 tempest-AttachInterfacesUnderV243Test-1943594058-project-member] Acquiring lock "refresh_cache-b911a5b5-9617-4fb3-9b5e-fb8c492e4931" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 739.986182] env[63515]: DEBUG oslo_vmware.api [None req-36f9f908-0af4-4115-aa42-e2a81d5f503e tempest-AttachInterfacesUnderV243Test-1943594058 tempest-AttachInterfacesUnderV243Test-1943594058-project-member] Task: {'id': task-1111050, 'name': ReconfigVM_Task, 'duration_secs': 0.151041} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 739.986182] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-36f9f908-0af4-4115-aa42-e2a81d5f503e tempest-AttachInterfacesUnderV243Test-1943594058 tempest-AttachInterfacesUnderV243Test-1943594058-project-member] [instance: b911a5b5-9617-4fb3-9b5e-fb8c492e4931] Reconfigured VM instance to set the machine id {{(pid=63515) _set_machine_id /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1799}} [ 740.019717] env[63515]: DEBUG oslo_vmware.api [None req-d4ac6a62-1537-4962-916c-41ec7cc28737 tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Task: {'id': task-1111051, 'name': PowerOffVM_Task, 'duration_secs': 0.200782} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 740.020168] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-d4ac6a62-1537-4962-916c-41ec7cc28737 tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] [instance: b85e9a70-7f5b-4d32-b616-f2a97e3186c8] Powered off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 740.020495] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-d4ac6a62-1537-4962-916c-41ec7cc28737 tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] [instance: b85e9a70-7f5b-4d32-b616-f2a97e3186c8] Unregistering the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 740.023354] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b808c2b1-9024-418b-8907-8df2e4314e57 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.052256] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-d4ac6a62-1537-4962-916c-41ec7cc28737 tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] [instance: b85e9a70-7f5b-4d32-b616-f2a97e3186c8] Unregistered the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 740.053026] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-d4ac6a62-1537-4962-916c-41ec7cc28737 tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] [instance: b85e9a70-7f5b-4d32-b616-f2a97e3186c8] Deleting contents of the VM from datastore datastore2 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 740.053026] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-d4ac6a62-1537-4962-916c-41ec7cc28737 tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Deleting the datastore file [datastore2] b85e9a70-7f5b-4d32-b616-f2a97e3186c8 {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 740.053026] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-14fe4d8e-645e-4d18-a921-7eeeed58cf24 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.060811] env[63515]: DEBUG oslo_vmware.api [None req-d4ac6a62-1537-4962-916c-41ec7cc28737 tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Waiting for the task: (returnval){ [ 740.060811] env[63515]: value = "task-1111053" [ 740.060811] env[63515]: _type = "Task" [ 740.060811] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 740.070540] env[63515]: DEBUG oslo_vmware.api [None req-d4ac6a62-1537-4962-916c-41ec7cc28737 tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Task: {'id': task-1111053, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 740.109115] env[63515]: DEBUG nova.network.neutron [req-4051955c-a19d-4456-9a8c-3b4f9aafb30e req-5fe698ed-4e2d-4e58-84de-95ddc9ebdc4b service nova] [instance: b911a5b5-9617-4fb3-9b5e-fb8c492e4931] Updated VIF entry in instance network info cache for port 9bf84cb2-7053-4129-8d8a-dc09e80b384e. {{(pid=63515) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 740.109637] env[63515]: DEBUG nova.network.neutron [req-4051955c-a19d-4456-9a8c-3b4f9aafb30e req-5fe698ed-4e2d-4e58-84de-95ddc9ebdc4b service nova] [instance: b911a5b5-9617-4fb3-9b5e-fb8c492e4931] Updating instance_info_cache with network_info: [{"id": "9bf84cb2-7053-4129-8d8a-dc09e80b384e", "address": "fa:16:3e:c6:e8:2f", "network": {"id": "22a0cdee-4685-4b2b-9df7-2506b2373a19", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-878688764-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.147", "type": "floating", "version": 4, "meta": {}}]}, {"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5086c79c531e4d039896a24a9ea29f73", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "78b49840-c3fc-455c-8491-a253ccd92bb5", "external-id": "nsx-vlan-transportzone-779", "segmentation_id": 779, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9bf84cb2-70", "ovs_interfaceid": "9bf84cb2-7053-4129-8d8a-dc09e80b384e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 740.117879] env[63515]: DEBUG oslo_concurrency.lockutils [None req-2786351a-303c-45a8-b63b-65e218eb6ef6 tempest-ServersTestJSON-324454119 tempest-ServersTestJSON-324454119-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 740.159038] env[63515]: DEBUG nova.network.neutron [None req-fd4a0cad-edc5-4c70-94f6-406aa5060c1e tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: 001a48e1-b237-41f8-a6f4-2fe98d173512] Updating instance_info_cache with network_info: [{"id": "ed0b7b69-0679-4b35-af6f-44c60792b411", "address": "fa:16:3e:6d:04:dc", "network": {"id": "15fdaabe-6793-400f-9737-2cdff8f07238", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1175204866-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f9c98e9f6020475490aaa7e76d907ab7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7d377d75-3add-4a15-8691-74b2eb010924", "external-id": "nsx-vlan-transportzone-71", "segmentation_id": 71, "bound_drivers": {"0": "nsxv3"}}, "devname": "taped0b7b69-06", "ovs_interfaceid": "ed0b7b69-0679-4b35-af6f-44c60792b411", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 740.167316] env[63515]: DEBUG nova.compute.manager [req-576e7d56-32cf-46d7-8935-a2303c8a4317 req-c2a1b848-f900-4926-9de4-e151f3d970e5 service nova] [instance: 001a48e1-b237-41f8-a6f4-2fe98d173512] Received event network-vif-plugged-ed0b7b69-0679-4b35-af6f-44c60792b411 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 740.167571] env[63515]: DEBUG oslo_concurrency.lockutils [req-576e7d56-32cf-46d7-8935-a2303c8a4317 req-c2a1b848-f900-4926-9de4-e151f3d970e5 service nova] Acquiring lock "001a48e1-b237-41f8-a6f4-2fe98d173512-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 740.167825] env[63515]: DEBUG oslo_concurrency.lockutils [req-576e7d56-32cf-46d7-8935-a2303c8a4317 req-c2a1b848-f900-4926-9de4-e151f3d970e5 service nova] Lock "001a48e1-b237-41f8-a6f4-2fe98d173512-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 740.168035] env[63515]: DEBUG oslo_concurrency.lockutils [req-576e7d56-32cf-46d7-8935-a2303c8a4317 req-c2a1b848-f900-4926-9de4-e151f3d970e5 service nova] Lock "001a48e1-b237-41f8-a6f4-2fe98d173512-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 740.168286] env[63515]: DEBUG nova.compute.manager [req-576e7d56-32cf-46d7-8935-a2303c8a4317 req-c2a1b848-f900-4926-9de4-e151f3d970e5 service nova] [instance: 001a48e1-b237-41f8-a6f4-2fe98d173512] No waiting events found dispatching network-vif-plugged-ed0b7b69-0679-4b35-af6f-44c60792b411 {{(pid=63515) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 740.168432] env[63515]: WARNING nova.compute.manager [req-576e7d56-32cf-46d7-8935-a2303c8a4317 req-c2a1b848-f900-4926-9de4-e151f3d970e5 service nova] [instance: 001a48e1-b237-41f8-a6f4-2fe98d173512] Received unexpected event network-vif-plugged-ed0b7b69-0679-4b35-af6f-44c60792b411 for instance with vm_state building and task_state spawning. [ 740.168615] env[63515]: DEBUG nova.compute.manager [req-576e7d56-32cf-46d7-8935-a2303c8a4317 req-c2a1b848-f900-4926-9de4-e151f3d970e5 service nova] [instance: 001a48e1-b237-41f8-a6f4-2fe98d173512] Received event network-changed-ed0b7b69-0679-4b35-af6f-44c60792b411 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 740.168837] env[63515]: DEBUG nova.compute.manager [req-576e7d56-32cf-46d7-8935-a2303c8a4317 req-c2a1b848-f900-4926-9de4-e151f3d970e5 service nova] [instance: 001a48e1-b237-41f8-a6f4-2fe98d173512] Refreshing instance network info cache due to event network-changed-ed0b7b69-0679-4b35-af6f-44c60792b411. {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 740.169767] env[63515]: DEBUG oslo_concurrency.lockutils [req-576e7d56-32cf-46d7-8935-a2303c8a4317 req-c2a1b848-f900-4926-9de4-e151f3d970e5 service nova] Acquiring lock "refresh_cache-001a48e1-b237-41f8-a6f4-2fe98d173512" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 740.170295] env[63515]: DEBUG oslo_concurrency.lockutils [None req-bb0fc58f-3c69-4616-a8cc-91b6b5f340c3 tempest-VolumesAdminNegativeTest-1091772657 tempest-VolumesAdminNegativeTest-1091772657-project-member] Lock "4087b2c4-6ed1-4b68-8b78-a36e34d935b1" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 23.642s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 740.389175] env[63515]: DEBUG oslo_vmware.api [None req-3b658824-6e25-4b06-878c-99b3be1f1d31 tempest-ServerTagsTestJSON-1170819896 tempest-ServerTagsTestJSON-1170819896-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52a1cbb3-a1cd-22cf-5769-1898dbe8b5be, 'name': SearchDatastore_Task, 'duration_secs': 0.012854} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 740.390504] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cc3b55e0-5fba-49fd-8ab4-51308ee18898 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.398074] env[63515]: DEBUG oslo_vmware.api [None req-3b658824-6e25-4b06-878c-99b3be1f1d31 tempest-ServerTagsTestJSON-1170819896 tempest-ServerTagsTestJSON-1170819896-project-member] Waiting for the task: (returnval){ [ 740.398074] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]52d29fb3-b492-5846-61ed-cfa205952800" [ 740.398074] env[63515]: _type = "Task" [ 740.398074] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 740.405969] env[63515]: DEBUG oslo_vmware.api [None req-3b658824-6e25-4b06-878c-99b3be1f1d31 tempest-ServerTagsTestJSON-1170819896 tempest-ServerTagsTestJSON-1170819896-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52d29fb3-b492-5846-61ed-cfa205952800, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 740.570978] env[63515]: DEBUG oslo_vmware.api [None req-d4ac6a62-1537-4962-916c-41ec7cc28737 tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Task: {'id': task-1111053, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.12048} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 740.571329] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-d4ac6a62-1537-4962-916c-41ec7cc28737 tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Deleted the datastore file {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 740.571548] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-d4ac6a62-1537-4962-916c-41ec7cc28737 tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] [instance: b85e9a70-7f5b-4d32-b616-f2a97e3186c8] Deleted contents of the VM from datastore datastore2 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 740.571910] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-d4ac6a62-1537-4962-916c-41ec7cc28737 tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] [instance: b85e9a70-7f5b-4d32-b616-f2a97e3186c8] Instance destroyed {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 740.571910] env[63515]: INFO nova.compute.manager [None req-d4ac6a62-1537-4962-916c-41ec7cc28737 tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] [instance: b85e9a70-7f5b-4d32-b616-f2a97e3186c8] Took 1.08 seconds to destroy the instance on the hypervisor. [ 740.572167] env[63515]: DEBUG oslo.service.loopingcall [None req-d4ac6a62-1537-4962-916c-41ec7cc28737 tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 740.572385] env[63515]: DEBUG nova.compute.manager [-] [instance: b85e9a70-7f5b-4d32-b616-f2a97e3186c8] Deallocating network for instance {{(pid=63515) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 740.572450] env[63515]: DEBUG nova.network.neutron [-] [instance: b85e9a70-7f5b-4d32-b616-f2a97e3186c8] deallocate_for_instance() {{(pid=63515) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 740.587860] env[63515]: DEBUG nova.network.neutron [-] [instance: b85e9a70-7f5b-4d32-b616-f2a97e3186c8] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 740.613845] env[63515]: DEBUG oslo_concurrency.lockutils [req-4051955c-a19d-4456-9a8c-3b4f9aafb30e req-5fe698ed-4e2d-4e58-84de-95ddc9ebdc4b service nova] Releasing lock "refresh_cache-b911a5b5-9617-4fb3-9b5e-fb8c492e4931" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 740.613845] env[63515]: DEBUG oslo_concurrency.lockutils [None req-91af7cb6-fa08-44e9-a8d2-9d871a7103c3 tempest-AttachInterfacesUnderV243Test-1943594058 tempest-AttachInterfacesUnderV243Test-1943594058-project-member] Acquired lock "refresh_cache-b911a5b5-9617-4fb3-9b5e-fb8c492e4931" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 740.635119] env[63515]: DEBUG nova.compute.manager [req-9d5311a7-6352-4add-8c5f-cd143886549f req-f5461eeb-e453-4d96-9b37-21ecaffcf957 service nova] [instance: ed550b10-d58f-45b8-b766-198f431c3788] Received event network-changed-ad9dda29-9326-472b-8c82-294a33a7faf0 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 740.635317] env[63515]: DEBUG nova.compute.manager [req-9d5311a7-6352-4add-8c5f-cd143886549f req-f5461eeb-e453-4d96-9b37-21ecaffcf957 service nova] [instance: ed550b10-d58f-45b8-b766-198f431c3788] Refreshing instance network info cache due to event network-changed-ad9dda29-9326-472b-8c82-294a33a7faf0. {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 740.635415] env[63515]: DEBUG oslo_concurrency.lockutils [req-9d5311a7-6352-4add-8c5f-cd143886549f req-f5461eeb-e453-4d96-9b37-21ecaffcf957 service nova] Acquiring lock "refresh_cache-ed550b10-d58f-45b8-b766-198f431c3788" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 740.635543] env[63515]: DEBUG oslo_concurrency.lockutils [req-9d5311a7-6352-4add-8c5f-cd143886549f req-f5461eeb-e453-4d96-9b37-21ecaffcf957 service nova] Acquired lock "refresh_cache-ed550b10-d58f-45b8-b766-198f431c3788" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 740.635696] env[63515]: DEBUG nova.network.neutron [req-9d5311a7-6352-4add-8c5f-cd143886549f req-f5461eeb-e453-4d96-9b37-21ecaffcf957 service nova] [instance: ed550b10-d58f-45b8-b766-198f431c3788] Refreshing network info cache for port ad9dda29-9326-472b-8c82-294a33a7faf0 {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 740.662926] env[63515]: DEBUG oslo_concurrency.lockutils [None req-fd4a0cad-edc5-4c70-94f6-406aa5060c1e tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Releasing lock "refresh_cache-001a48e1-b237-41f8-a6f4-2fe98d173512" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 740.662926] env[63515]: DEBUG nova.compute.manager [None req-fd4a0cad-edc5-4c70-94f6-406aa5060c1e tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: 001a48e1-b237-41f8-a6f4-2fe98d173512] Instance network_info: |[{"id": "ed0b7b69-0679-4b35-af6f-44c60792b411", "address": "fa:16:3e:6d:04:dc", "network": {"id": "15fdaabe-6793-400f-9737-2cdff8f07238", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1175204866-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f9c98e9f6020475490aaa7e76d907ab7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7d377d75-3add-4a15-8691-74b2eb010924", "external-id": "nsx-vlan-transportzone-71", "segmentation_id": 71, "bound_drivers": {"0": "nsxv3"}}, "devname": "taped0b7b69-06", "ovs_interfaceid": "ed0b7b69-0679-4b35-af6f-44c60792b411", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 740.665577] env[63515]: DEBUG oslo_concurrency.lockutils [req-576e7d56-32cf-46d7-8935-a2303c8a4317 req-c2a1b848-f900-4926-9de4-e151f3d970e5 service nova] Acquired lock "refresh_cache-001a48e1-b237-41f8-a6f4-2fe98d173512" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 740.665748] env[63515]: DEBUG nova.network.neutron [req-576e7d56-32cf-46d7-8935-a2303c8a4317 req-c2a1b848-f900-4926-9de4-e151f3d970e5 service nova] [instance: 001a48e1-b237-41f8-a6f4-2fe98d173512] Refreshing network info cache for port ed0b7b69-0679-4b35-af6f-44c60792b411 {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 740.666864] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-fd4a0cad-edc5-4c70-94f6-406aa5060c1e tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: 001a48e1-b237-41f8-a6f4-2fe98d173512] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:6d:04:dc', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '7d377d75-3add-4a15-8691-74b2eb010924', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ed0b7b69-0679-4b35-af6f-44c60792b411', 'vif_model': 'vmxnet3'}] {{(pid=63515) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 740.674502] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-fd4a0cad-edc5-4c70-94f6-406aa5060c1e tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Creating folder: Project (f9c98e9f6020475490aaa7e76d907ab7). Parent ref: group-v243370. {{(pid=63515) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 740.677756] env[63515]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a09824be-55f3-4095-bdd8-301a28eb4e8b {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.690031] env[63515]: INFO nova.virt.vmwareapi.vm_util [None req-fd4a0cad-edc5-4c70-94f6-406aa5060c1e tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Created folder: Project (f9c98e9f6020475490aaa7e76d907ab7) in parent group-v243370. [ 740.690031] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-fd4a0cad-edc5-4c70-94f6-406aa5060c1e tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Creating folder: Instances. Parent ref: group-v243428. {{(pid=63515) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 740.697012] env[63515]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a4fd4d1d-0ace-49ca-accd-56b07ef6d446 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.708031] env[63515]: DEBUG oslo_concurrency.lockutils [None req-99f0df88-8ade-4c67-afd9-dbdc6b385015 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Acquiring lock "f1d01b75-ac9d-458d-8cc2-ae64cffca4e8-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 740.708031] env[63515]: DEBUG oslo_concurrency.lockutils [None req-99f0df88-8ade-4c67-afd9-dbdc6b385015 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Lock "f1d01b75-ac9d-458d-8cc2-ae64cffca4e8-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 740.708031] env[63515]: DEBUG oslo_concurrency.lockutils [None req-99f0df88-8ade-4c67-afd9-dbdc6b385015 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Lock "f1d01b75-ac9d-458d-8cc2-ae64cffca4e8-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 740.717371] env[63515]: INFO nova.virt.vmwareapi.vm_util [None req-fd4a0cad-edc5-4c70-94f6-406aa5060c1e tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Created folder: Instances in parent group-v243428. [ 740.717628] env[63515]: DEBUG oslo.service.loopingcall [None req-fd4a0cad-edc5-4c70-94f6-406aa5060c1e tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 740.720223] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 001a48e1-b237-41f8-a6f4-2fe98d173512] Creating VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 740.720888] env[63515]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-060ce7ad-a99f-4937-97b1-738ec0315246 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.744246] env[63515]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 740.744246] env[63515]: value = "task-1111056" [ 740.744246] env[63515]: _type = "Task" [ 740.744246] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 740.754450] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1111056, 'name': CreateVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 740.910433] env[63515]: DEBUG oslo_vmware.api [None req-3b658824-6e25-4b06-878c-99b3be1f1d31 tempest-ServerTagsTestJSON-1170819896 tempest-ServerTagsTestJSON-1170819896-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52d29fb3-b492-5846-61ed-cfa205952800, 'name': SearchDatastore_Task, 'duration_secs': 0.009264} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 740.910696] env[63515]: DEBUG oslo_concurrency.lockutils [None req-3b658824-6e25-4b06-878c-99b3be1f1d31 tempest-ServerTagsTestJSON-1170819896 tempest-ServerTagsTestJSON-1170819896-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 740.910947] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-3b658824-6e25-4b06-878c-99b3be1f1d31 tempest-ServerTagsTestJSON-1170819896 tempest-ServerTagsTestJSON-1170819896-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk to [datastore2] de210780-5c0f-4fba-883c-13707566a2e1/de210780-5c0f-4fba-883c-13707566a2e1.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 740.911326] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2b371df2-2263-463a-8f5c-f1145e7431ab {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.920328] env[63515]: DEBUG oslo_vmware.api [None req-3b658824-6e25-4b06-878c-99b3be1f1d31 tempest-ServerTagsTestJSON-1170819896 tempest-ServerTagsTestJSON-1170819896-project-member] Waiting for the task: (returnval){ [ 740.920328] env[63515]: value = "task-1111057" [ 740.920328] env[63515]: _type = "Task" [ 740.920328] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 740.928579] env[63515]: DEBUG oslo_vmware.api [None req-3b658824-6e25-4b06-878c-99b3be1f1d31 tempest-ServerTagsTestJSON-1170819896 tempest-ServerTagsTestJSON-1170819896-project-member] Task: {'id': task-1111057, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 741.094545] env[63515]: DEBUG nova.network.neutron [None req-91af7cb6-fa08-44e9-a8d2-9d871a7103c3 tempest-AttachInterfacesUnderV243Test-1943594058 tempest-AttachInterfacesUnderV243Test-1943594058-project-member] [instance: b911a5b5-9617-4fb3-9b5e-fb8c492e4931] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 741.094545] env[63515]: DEBUG nova.network.neutron [-] [instance: b85e9a70-7f5b-4d32-b616-f2a97e3186c8] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 741.102484] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-932cbd45-ddc2-4d83-924d-8e18c0884643 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.113913] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae82cb69-e521-4cdd-a1ae-92c2e7c6011b {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.151896] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2ddd507-8d47-4e6c-adc0-ad7df4afeaf7 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.160465] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1211d86-4929-402c-879b-e7d9b56e3566 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.179201] env[63515]: DEBUG nova.compute.provider_tree [None req-99ae4d48-1fae-4ace-913c-d249ca72bbeb tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 741.257894] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1111056, 'name': CreateVM_Task} progress is 99%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 741.432038] env[63515]: DEBUG oslo_vmware.api [None req-3b658824-6e25-4b06-878c-99b3be1f1d31 tempest-ServerTagsTestJSON-1170819896 tempest-ServerTagsTestJSON-1170819896-project-member] Task: {'id': task-1111057, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.464255} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 741.432348] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-3b658824-6e25-4b06-878c-99b3be1f1d31 tempest-ServerTagsTestJSON-1170819896 tempest-ServerTagsTestJSON-1170819896-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk to [datastore2] de210780-5c0f-4fba-883c-13707566a2e1/de210780-5c0f-4fba-883c-13707566a2e1.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 741.432490] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-3b658824-6e25-4b06-878c-99b3be1f1d31 tempest-ServerTagsTestJSON-1170819896 tempest-ServerTagsTestJSON-1170819896-project-member] [instance: de210780-5c0f-4fba-883c-13707566a2e1] Extending root virtual disk to 1048576 {{(pid=63515) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 741.432732] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-5b9ed771-bbb8-4c12-baf0-0d59d72dd2c8 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.438749] env[63515]: DEBUG oslo_vmware.api [None req-3b658824-6e25-4b06-878c-99b3be1f1d31 tempest-ServerTagsTestJSON-1170819896 tempest-ServerTagsTestJSON-1170819896-project-member] Waiting for the task: (returnval){ [ 741.438749] env[63515]: value = "task-1111058" [ 741.438749] env[63515]: _type = "Task" [ 741.438749] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 741.446181] env[63515]: DEBUG oslo_vmware.api [None req-3b658824-6e25-4b06-878c-99b3be1f1d31 tempest-ServerTagsTestJSON-1170819896 tempest-ServerTagsTestJSON-1170819896-project-member] Task: {'id': task-1111058, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 741.602924] env[63515]: INFO nova.compute.manager [-] [instance: b85e9a70-7f5b-4d32-b616-f2a97e3186c8] Took 1.03 seconds to deallocate network for instance. [ 741.614076] env[63515]: DEBUG nova.network.neutron [req-9d5311a7-6352-4add-8c5f-cd143886549f req-f5461eeb-e453-4d96-9b37-21ecaffcf957 service nova] [instance: ed550b10-d58f-45b8-b766-198f431c3788] Updated VIF entry in instance network info cache for port ad9dda29-9326-472b-8c82-294a33a7faf0. {{(pid=63515) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 741.614561] env[63515]: DEBUG nova.network.neutron [req-9d5311a7-6352-4add-8c5f-cd143886549f req-f5461eeb-e453-4d96-9b37-21ecaffcf957 service nova] [instance: ed550b10-d58f-45b8-b766-198f431c3788] Updating instance_info_cache with network_info: [{"id": "ad9dda29-9326-472b-8c82-294a33a7faf0", "address": "fa:16:3e:26:e3:72", "network": {"id": "88dfb29d-41e0-4df7-8f70-e52cda05620c", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1355578400-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.251", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7d4ada9ec35f42b19c6480a9101d21a4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "357d2811-e990-4985-9f9e-b158d10d3699", "external-id": "nsx-vlan-transportzone-641", "segmentation_id": 641, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapad9dda29-93", "ovs_interfaceid": "ad9dda29-9326-472b-8c82-294a33a7faf0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 741.620442] env[63515]: DEBUG nova.network.neutron [req-576e7d56-32cf-46d7-8935-a2303c8a4317 req-c2a1b848-f900-4926-9de4-e151f3d970e5 service nova] [instance: 001a48e1-b237-41f8-a6f4-2fe98d173512] Updated VIF entry in instance network info cache for port ed0b7b69-0679-4b35-af6f-44c60792b411. {{(pid=63515) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 741.620883] env[63515]: DEBUG nova.network.neutron [req-576e7d56-32cf-46d7-8935-a2303c8a4317 req-c2a1b848-f900-4926-9de4-e151f3d970e5 service nova] [instance: 001a48e1-b237-41f8-a6f4-2fe98d173512] Updating instance_info_cache with network_info: [{"id": "ed0b7b69-0679-4b35-af6f-44c60792b411", "address": "fa:16:3e:6d:04:dc", "network": {"id": "15fdaabe-6793-400f-9737-2cdff8f07238", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1175204866-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f9c98e9f6020475490aaa7e76d907ab7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7d377d75-3add-4a15-8691-74b2eb010924", "external-id": "nsx-vlan-transportzone-71", "segmentation_id": 71, "bound_drivers": {"0": "nsxv3"}}, "devname": "taped0b7b69-06", "ovs_interfaceid": "ed0b7b69-0679-4b35-af6f-44c60792b411", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 741.683347] env[63515]: DEBUG nova.scheduler.client.report [None req-99ae4d48-1fae-4ace-913c-d249ca72bbeb tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 741.756829] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1111056, 'name': CreateVM_Task} progress is 99%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 741.761360] env[63515]: DEBUG oslo_concurrency.lockutils [None req-99f0df88-8ade-4c67-afd9-dbdc6b385015 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Acquiring lock "refresh_cache-f1d01b75-ac9d-458d-8cc2-ae64cffca4e8" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 741.761360] env[63515]: DEBUG oslo_concurrency.lockutils [None req-99f0df88-8ade-4c67-afd9-dbdc6b385015 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Acquired lock "refresh_cache-f1d01b75-ac9d-458d-8cc2-ae64cffca4e8" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 741.761360] env[63515]: DEBUG nova.network.neutron [None req-99f0df88-8ade-4c67-afd9-dbdc6b385015 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] [instance: f1d01b75-ac9d-458d-8cc2-ae64cffca4e8] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 741.950364] env[63515]: DEBUG oslo_vmware.api [None req-3b658824-6e25-4b06-878c-99b3be1f1d31 tempest-ServerTagsTestJSON-1170819896 tempest-ServerTagsTestJSON-1170819896-project-member] Task: {'id': task-1111058, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.074444} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 741.951091] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-3b658824-6e25-4b06-878c-99b3be1f1d31 tempest-ServerTagsTestJSON-1170819896 tempest-ServerTagsTestJSON-1170819896-project-member] [instance: de210780-5c0f-4fba-883c-13707566a2e1] Extended root virtual disk {{(pid=63515) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 741.952390] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43db1ee5-1835-41e8-85ca-ad85e5840925 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.977867] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-3b658824-6e25-4b06-878c-99b3be1f1d31 tempest-ServerTagsTestJSON-1170819896 tempest-ServerTagsTestJSON-1170819896-project-member] [instance: de210780-5c0f-4fba-883c-13707566a2e1] Reconfiguring VM instance instance-0000002f to attach disk [datastore2] de210780-5c0f-4fba-883c-13707566a2e1/de210780-5c0f-4fba-883c-13707566a2e1.vmdk or device None with type sparse {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 741.980837] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6175094e-bd3d-4eb5-b4d1-90b2a19f729b {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.004105] env[63515]: DEBUG oslo_vmware.api [None req-3b658824-6e25-4b06-878c-99b3be1f1d31 tempest-ServerTagsTestJSON-1170819896 tempest-ServerTagsTestJSON-1170819896-project-member] Waiting for the task: (returnval){ [ 742.004105] env[63515]: value = "task-1111059" [ 742.004105] env[63515]: _type = "Task" [ 742.004105] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 742.011346] env[63515]: DEBUG oslo_vmware.api [None req-3b658824-6e25-4b06-878c-99b3be1f1d31 tempest-ServerTagsTestJSON-1170819896 tempest-ServerTagsTestJSON-1170819896-project-member] Task: {'id': task-1111059, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 742.110451] env[63515]: DEBUG oslo_concurrency.lockutils [None req-d4ac6a62-1537-4962-916c-41ec7cc28737 tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 742.118407] env[63515]: DEBUG oslo_concurrency.lockutils [req-9d5311a7-6352-4add-8c5f-cd143886549f req-f5461eeb-e453-4d96-9b37-21ecaffcf957 service nova] Releasing lock "refresh_cache-ed550b10-d58f-45b8-b766-198f431c3788" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 742.124705] env[63515]: DEBUG oslo_concurrency.lockutils [req-576e7d56-32cf-46d7-8935-a2303c8a4317 req-c2a1b848-f900-4926-9de4-e151f3d970e5 service nova] Releasing lock "refresh_cache-001a48e1-b237-41f8-a6f4-2fe98d173512" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 742.190075] env[63515]: DEBUG oslo_concurrency.lockutils [None req-99ae4d48-1fae-4ace-913c-d249ca72bbeb tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.593s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 742.190075] env[63515]: DEBUG nova.compute.manager [None req-99ae4d48-1fae-4ace-913c-d249ca72bbeb tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] [instance: d2a4f66a-f1a2-4f0d-826d-0b7c4f04bd23] Start building networks asynchronously for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 742.194994] env[63515]: DEBUG oslo_concurrency.lockutils [None req-5bd41396-218d-4ce7-9765-053dcc695301 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.989s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 742.197043] env[63515]: INFO nova.compute.claims [None req-5bd41396-218d-4ce7-9765-053dcc695301 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] [instance: 6688cdaa-29ba-413a-8131-4f834cdb70e4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 742.239564] env[63515]: DEBUG nova.compute.manager [req-2c42f0b9-e954-49f7-8eb9-d613e5f5ead1 req-9488c0d0-0db0-4965-adb9-cc4d6f9dd345 service nova] [instance: b911a5b5-9617-4fb3-9b5e-fb8c492e4931] Received event network-changed-9bf84cb2-7053-4129-8d8a-dc09e80b384e {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 742.240948] env[63515]: DEBUG nova.compute.manager [req-2c42f0b9-e954-49f7-8eb9-d613e5f5ead1 req-9488c0d0-0db0-4965-adb9-cc4d6f9dd345 service nova] [instance: b911a5b5-9617-4fb3-9b5e-fb8c492e4931] Refreshing instance network info cache due to event network-changed-9bf84cb2-7053-4129-8d8a-dc09e80b384e. {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 742.240948] env[63515]: DEBUG oslo_concurrency.lockutils [req-2c42f0b9-e954-49f7-8eb9-d613e5f5ead1 req-9488c0d0-0db0-4965-adb9-cc4d6f9dd345 service nova] Acquiring lock "refresh_cache-b911a5b5-9617-4fb3-9b5e-fb8c492e4931" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 742.259724] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1111056, 'name': CreateVM_Task, 'duration_secs': 1.384616} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 742.259724] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 001a48e1-b237-41f8-a6f4-2fe98d173512] Created VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 742.259724] env[63515]: DEBUG oslo_concurrency.lockutils [None req-fd4a0cad-edc5-4c70-94f6-406aa5060c1e tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 742.259724] env[63515]: DEBUG oslo_concurrency.lockutils [None req-fd4a0cad-edc5-4c70-94f6-406aa5060c1e tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 742.259724] env[63515]: DEBUG oslo_concurrency.lockutils [None req-fd4a0cad-edc5-4c70-94f6-406aa5060c1e tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 742.259724] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-64aea5b3-ebf3-4803-a054-f7e357ac34a7 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.264343] env[63515]: DEBUG oslo_vmware.api [None req-fd4a0cad-edc5-4c70-94f6-406aa5060c1e tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Waiting for the task: (returnval){ [ 742.264343] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]5295464f-20fd-c293-a9aa-c35edc5b736b" [ 742.264343] env[63515]: _type = "Task" [ 742.264343] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 742.270416] env[63515]: DEBUG nova.network.neutron [None req-91af7cb6-fa08-44e9-a8d2-9d871a7103c3 tempest-AttachInterfacesUnderV243Test-1943594058 tempest-AttachInterfacesUnderV243Test-1943594058-project-member] [instance: b911a5b5-9617-4fb3-9b5e-fb8c492e4931] Updating instance_info_cache with network_info: [{"id": "9bf84cb2-7053-4129-8d8a-dc09e80b384e", "address": "fa:16:3e:c6:e8:2f", "network": {"id": "22a0cdee-4685-4b2b-9df7-2506b2373a19", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-878688764-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.147", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5086c79c531e4d039896a24a9ea29f73", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "78b49840-c3fc-455c-8491-a253ccd92bb5", "external-id": "nsx-vlan-transportzone-779", "segmentation_id": 779, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9bf84cb2-70", "ovs_interfaceid": "9bf84cb2-7053-4129-8d8a-dc09e80b384e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 742.276218] env[63515]: DEBUG oslo_vmware.api [None req-fd4a0cad-edc5-4c70-94f6-406aa5060c1e tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]5295464f-20fd-c293-a9aa-c35edc5b736b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 742.515456] env[63515]: DEBUG oslo_vmware.api [None req-3b658824-6e25-4b06-878c-99b3be1f1d31 tempest-ServerTagsTestJSON-1170819896 tempest-ServerTagsTestJSON-1170819896-project-member] Task: {'id': task-1111059, 'name': ReconfigVM_Task, 'duration_secs': 0.335641} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 742.515776] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-3b658824-6e25-4b06-878c-99b3be1f1d31 tempest-ServerTagsTestJSON-1170819896 tempest-ServerTagsTestJSON-1170819896-project-member] [instance: de210780-5c0f-4fba-883c-13707566a2e1] Reconfigured VM instance instance-0000002f to attach disk [datastore2] de210780-5c0f-4fba-883c-13707566a2e1/de210780-5c0f-4fba-883c-13707566a2e1.vmdk or device None with type sparse {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 742.516476] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-75533b1d-bf47-4c32-81f9-b220127985cf {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.522522] env[63515]: DEBUG oslo_vmware.api [None req-3b658824-6e25-4b06-878c-99b3be1f1d31 tempest-ServerTagsTestJSON-1170819896 tempest-ServerTagsTestJSON-1170819896-project-member] Waiting for the task: (returnval){ [ 742.522522] env[63515]: value = "task-1111060" [ 742.522522] env[63515]: _type = "Task" [ 742.522522] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 742.534133] env[63515]: DEBUG oslo_vmware.api [None req-3b658824-6e25-4b06-878c-99b3be1f1d31 tempest-ServerTagsTestJSON-1170819896 tempest-ServerTagsTestJSON-1170819896-project-member] Task: {'id': task-1111060, 'name': Rename_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 742.552289] env[63515]: DEBUG nova.network.neutron [None req-99f0df88-8ade-4c67-afd9-dbdc6b385015 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] [instance: f1d01b75-ac9d-458d-8cc2-ae64cffca4e8] Updating instance_info_cache with network_info: [{"id": "48f0868f-94db-42f7-8153-d1a27fa02707", "address": "fa:16:3e:7a:74:da", "network": {"id": "b4abcbd0-0eac-4bde-ad11-7f93afcedcef", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.21", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "b569255cc43e42e7ae3f2b4ad37c6ef4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0cd5d325-3053-407e-a4ee-f627e82a23f9", "external-id": "nsx-vlan-transportzone-809", "segmentation_id": 809, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap48f0868f-94", "ovs_interfaceid": "48f0868f-94db-42f7-8153-d1a27fa02707", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 742.703487] env[63515]: DEBUG nova.compute.utils [None req-99ae4d48-1fae-4ace-913c-d249ca72bbeb tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Using /dev/sd instead of None {{(pid=63515) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 742.706315] env[63515]: DEBUG nova.compute.manager [None req-99ae4d48-1fae-4ace-913c-d249ca72bbeb tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] [instance: d2a4f66a-f1a2-4f0d-826d-0b7c4f04bd23] Allocating IP information in the background. {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 742.708559] env[63515]: DEBUG nova.network.neutron [None req-99ae4d48-1fae-4ace-913c-d249ca72bbeb tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] [instance: d2a4f66a-f1a2-4f0d-826d-0b7c4f04bd23] allocate_for_instance() {{(pid=63515) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 742.752268] env[63515]: DEBUG nova.policy [None req-99ae4d48-1fae-4ace-913c-d249ca72bbeb tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'eeb19c6fd65a41568e8138b3acb57bb8', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6f501ed813754f759464e338884993c1', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63515) authorize /opt/stack/nova/nova/policy.py:201}} [ 742.772966] env[63515]: DEBUG oslo_concurrency.lockutils [None req-91af7cb6-fa08-44e9-a8d2-9d871a7103c3 tempest-AttachInterfacesUnderV243Test-1943594058 tempest-AttachInterfacesUnderV243Test-1943594058-project-member] Releasing lock "refresh_cache-b911a5b5-9617-4fb3-9b5e-fb8c492e4931" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 742.773305] env[63515]: DEBUG nova.compute.manager [None req-91af7cb6-fa08-44e9-a8d2-9d871a7103c3 tempest-AttachInterfacesUnderV243Test-1943594058 tempest-AttachInterfacesUnderV243Test-1943594058-project-member] [instance: b911a5b5-9617-4fb3-9b5e-fb8c492e4931] Inject network info {{(pid=63515) _inject_network_info /opt/stack/nova/nova/compute/manager.py:7219}} [ 742.773905] env[63515]: DEBUG nova.compute.manager [None req-91af7cb6-fa08-44e9-a8d2-9d871a7103c3 tempest-AttachInterfacesUnderV243Test-1943594058 tempest-AttachInterfacesUnderV243Test-1943594058-project-member] [instance: b911a5b5-9617-4fb3-9b5e-fb8c492e4931] network_info to inject: |[{"id": "9bf84cb2-7053-4129-8d8a-dc09e80b384e", "address": "fa:16:3e:c6:e8:2f", "network": {"id": "22a0cdee-4685-4b2b-9df7-2506b2373a19", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-878688764-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.147", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5086c79c531e4d039896a24a9ea29f73", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "78b49840-c3fc-455c-8491-a253ccd92bb5", "external-id": "nsx-vlan-transportzone-779", "segmentation_id": 779, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9bf84cb2-70", "ovs_interfaceid": "9bf84cb2-7053-4129-8d8a-dc09e80b384e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63515) _inject_network_info /opt/stack/nova/nova/compute/manager.py:7220}} [ 742.781797] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-91af7cb6-fa08-44e9-a8d2-9d871a7103c3 tempest-AttachInterfacesUnderV243Test-1943594058 tempest-AttachInterfacesUnderV243Test-1943594058-project-member] [instance: b911a5b5-9617-4fb3-9b5e-fb8c492e4931] Reconfiguring VM instance to set the machine id {{(pid=63515) _set_machine_id /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1796}} [ 742.788198] env[63515]: DEBUG oslo_concurrency.lockutils [req-2c42f0b9-e954-49f7-8eb9-d613e5f5ead1 req-9488c0d0-0db0-4965-adb9-cc4d6f9dd345 service nova] Acquired lock "refresh_cache-b911a5b5-9617-4fb3-9b5e-fb8c492e4931" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 742.788198] env[63515]: DEBUG nova.network.neutron [req-2c42f0b9-e954-49f7-8eb9-d613e5f5ead1 req-9488c0d0-0db0-4965-adb9-cc4d6f9dd345 service nova] [instance: b911a5b5-9617-4fb3-9b5e-fb8c492e4931] Refreshing network info cache for port 9bf84cb2-7053-4129-8d8a-dc09e80b384e {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 742.789480] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5966ee01-213d-4f54-a33f-d7d55aa579ef {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.803261] env[63515]: DEBUG oslo_vmware.api [None req-fd4a0cad-edc5-4c70-94f6-406aa5060c1e tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]5295464f-20fd-c293-a9aa-c35edc5b736b, 'name': SearchDatastore_Task, 'duration_secs': 0.02038} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 742.808098] env[63515]: DEBUG oslo_concurrency.lockutils [None req-fd4a0cad-edc5-4c70-94f6-406aa5060c1e tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 742.808499] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-fd4a0cad-edc5-4c70-94f6-406aa5060c1e tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: 001a48e1-b237-41f8-a6f4-2fe98d173512] Processing image 8a120570-cb06-4099-b262-554ca0ad15c5 {{(pid=63515) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 742.809265] env[63515]: DEBUG oslo_concurrency.lockutils [None req-fd4a0cad-edc5-4c70-94f6-406aa5060c1e tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 742.809265] env[63515]: DEBUG oslo_concurrency.lockutils [None req-fd4a0cad-edc5-4c70-94f6-406aa5060c1e tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 742.809265] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-fd4a0cad-edc5-4c70-94f6-406aa5060c1e tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 742.813497] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-76a6ceb5-4c73-4690-8d31-6a8b54dba76a {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.818897] env[63515]: DEBUG oslo_vmware.api [None req-91af7cb6-fa08-44e9-a8d2-9d871a7103c3 tempest-AttachInterfacesUnderV243Test-1943594058 tempest-AttachInterfacesUnderV243Test-1943594058-project-member] Waiting for the task: (returnval){ [ 742.818897] env[63515]: value = "task-1111061" [ 742.818897] env[63515]: _type = "Task" [ 742.818897] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 742.826565] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-fd4a0cad-edc5-4c70-94f6-406aa5060c1e tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 742.826750] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-fd4a0cad-edc5-4c70-94f6-406aa5060c1e tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63515) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 742.830844] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e730bf4b-c364-4387-b1df-9e953d355c30 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.833722] env[63515]: DEBUG oslo_vmware.api [None req-91af7cb6-fa08-44e9-a8d2-9d871a7103c3 tempest-AttachInterfacesUnderV243Test-1943594058 tempest-AttachInterfacesUnderV243Test-1943594058-project-member] Task: {'id': task-1111061, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 742.836701] env[63515]: DEBUG oslo_vmware.api [None req-fd4a0cad-edc5-4c70-94f6-406aa5060c1e tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Waiting for the task: (returnval){ [ 742.836701] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]528b1a0b-122e-d2ea-f7fd-e2894f4b3534" [ 742.836701] env[63515]: _type = "Task" [ 742.836701] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 742.844569] env[63515]: DEBUG oslo_vmware.api [None req-fd4a0cad-edc5-4c70-94f6-406aa5060c1e tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]528b1a0b-122e-d2ea-f7fd-e2894f4b3534, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 743.036834] env[63515]: DEBUG oslo_vmware.api [None req-3b658824-6e25-4b06-878c-99b3be1f1d31 tempest-ServerTagsTestJSON-1170819896 tempest-ServerTagsTestJSON-1170819896-project-member] Task: {'id': task-1111060, 'name': Rename_Task, 'duration_secs': 0.144033} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 743.039701] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-3b658824-6e25-4b06-878c-99b3be1f1d31 tempest-ServerTagsTestJSON-1170819896 tempest-ServerTagsTestJSON-1170819896-project-member] [instance: de210780-5c0f-4fba-883c-13707566a2e1] Powering on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 743.040118] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0e0c9f11-b8ac-4c99-9e50-7cfef63833ac {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.049295] env[63515]: DEBUG oslo_vmware.api [None req-3b658824-6e25-4b06-878c-99b3be1f1d31 tempest-ServerTagsTestJSON-1170819896 tempest-ServerTagsTestJSON-1170819896-project-member] Waiting for the task: (returnval){ [ 743.049295] env[63515]: value = "task-1111062" [ 743.049295] env[63515]: _type = "Task" [ 743.049295] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 743.055513] env[63515]: DEBUG oslo_concurrency.lockutils [None req-99f0df88-8ade-4c67-afd9-dbdc6b385015 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Releasing lock "refresh_cache-f1d01b75-ac9d-458d-8cc2-ae64cffca4e8" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 743.069296] env[63515]: DEBUG oslo_vmware.api [None req-3b658824-6e25-4b06-878c-99b3be1f1d31 tempest-ServerTagsTestJSON-1170819896 tempest-ServerTagsTestJSON-1170819896-project-member] Task: {'id': task-1111062, 'name': PowerOnVM_Task} progress is 33%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 743.155237] env[63515]: DEBUG nova.network.neutron [None req-99ae4d48-1fae-4ace-913c-d249ca72bbeb tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] [instance: d2a4f66a-f1a2-4f0d-826d-0b7c4f04bd23] Successfully created port: 3bca52da-5931-4d22-bdf1-33072a3aca76 {{(pid=63515) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 743.212444] env[63515]: DEBUG nova.compute.manager [None req-99ae4d48-1fae-4ace-913c-d249ca72bbeb tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] [instance: d2a4f66a-f1a2-4f0d-826d-0b7c4f04bd23] Start building block device mappings for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 743.243034] env[63515]: DEBUG nova.network.neutron [req-2c42f0b9-e954-49f7-8eb9-d613e5f5ead1 req-9488c0d0-0db0-4965-adb9-cc4d6f9dd345 service nova] [instance: b911a5b5-9617-4fb3-9b5e-fb8c492e4931] Updated VIF entry in instance network info cache for port 9bf84cb2-7053-4129-8d8a-dc09e80b384e. {{(pid=63515) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 743.243438] env[63515]: DEBUG nova.network.neutron [req-2c42f0b9-e954-49f7-8eb9-d613e5f5ead1 req-9488c0d0-0db0-4965-adb9-cc4d6f9dd345 service nova] [instance: b911a5b5-9617-4fb3-9b5e-fb8c492e4931] Updating instance_info_cache with network_info: [{"id": "9bf84cb2-7053-4129-8d8a-dc09e80b384e", "address": "fa:16:3e:c6:e8:2f", "network": {"id": "22a0cdee-4685-4b2b-9df7-2506b2373a19", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-878688764-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.147", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5086c79c531e4d039896a24a9ea29f73", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "78b49840-c3fc-455c-8491-a253ccd92bb5", "external-id": "nsx-vlan-transportzone-779", "segmentation_id": 779, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9bf84cb2-70", "ovs_interfaceid": "9bf84cb2-7053-4129-8d8a-dc09e80b384e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 743.323357] env[63515]: DEBUG oslo_concurrency.lockutils [None req-dac2c5cc-03a7-4df4-a3f6-d707671aaebe tempest-AttachInterfacesUnderV243Test-1943594058 tempest-AttachInterfacesUnderV243Test-1943594058-project-member] Acquiring lock "b911a5b5-9617-4fb3-9b5e-fb8c492e4931" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 743.323357] env[63515]: DEBUG oslo_concurrency.lockutils [None req-dac2c5cc-03a7-4df4-a3f6-d707671aaebe tempest-AttachInterfacesUnderV243Test-1943594058 tempest-AttachInterfacesUnderV243Test-1943594058-project-member] Lock "b911a5b5-9617-4fb3-9b5e-fb8c492e4931" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 743.323357] env[63515]: DEBUG oslo_concurrency.lockutils [None req-dac2c5cc-03a7-4df4-a3f6-d707671aaebe tempest-AttachInterfacesUnderV243Test-1943594058 tempest-AttachInterfacesUnderV243Test-1943594058-project-member] Acquiring lock "b911a5b5-9617-4fb3-9b5e-fb8c492e4931-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 743.323357] env[63515]: DEBUG oslo_concurrency.lockutils [None req-dac2c5cc-03a7-4df4-a3f6-d707671aaebe tempest-AttachInterfacesUnderV243Test-1943594058 tempest-AttachInterfacesUnderV243Test-1943594058-project-member] Lock "b911a5b5-9617-4fb3-9b5e-fb8c492e4931-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 743.323893] env[63515]: DEBUG oslo_concurrency.lockutils [None req-dac2c5cc-03a7-4df4-a3f6-d707671aaebe tempest-AttachInterfacesUnderV243Test-1943594058 tempest-AttachInterfacesUnderV243Test-1943594058-project-member] Lock "b911a5b5-9617-4fb3-9b5e-fb8c492e4931-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 743.329015] env[63515]: INFO nova.compute.manager [None req-dac2c5cc-03a7-4df4-a3f6-d707671aaebe tempest-AttachInterfacesUnderV243Test-1943594058 tempest-AttachInterfacesUnderV243Test-1943594058-project-member] [instance: b911a5b5-9617-4fb3-9b5e-fb8c492e4931] Terminating instance [ 743.331406] env[63515]: DEBUG nova.compute.manager [None req-dac2c5cc-03a7-4df4-a3f6-d707671aaebe tempest-AttachInterfacesUnderV243Test-1943594058 tempest-AttachInterfacesUnderV243Test-1943594058-project-member] [instance: b911a5b5-9617-4fb3-9b5e-fb8c492e4931] Start destroying the instance on the hypervisor. {{(pid=63515) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 743.331748] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-dac2c5cc-03a7-4df4-a3f6-d707671aaebe tempest-AttachInterfacesUnderV243Test-1943594058 tempest-AttachInterfacesUnderV243Test-1943594058-project-member] [instance: b911a5b5-9617-4fb3-9b5e-fb8c492e4931] Destroying instance {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 743.332687] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78fefca5-f703-4b12-9526-0926612bd06a {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.339910] env[63515]: DEBUG oslo_vmware.api [None req-91af7cb6-fa08-44e9-a8d2-9d871a7103c3 tempest-AttachInterfacesUnderV243Test-1943594058 tempest-AttachInterfacesUnderV243Test-1943594058-project-member] Task: {'id': task-1111061, 'name': ReconfigVM_Task, 'duration_secs': 0.14506} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 743.347120] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-91af7cb6-fa08-44e9-a8d2-9d871a7103c3 tempest-AttachInterfacesUnderV243Test-1943594058 tempest-AttachInterfacesUnderV243Test-1943594058-project-member] [instance: b911a5b5-9617-4fb3-9b5e-fb8c492e4931] Reconfigured VM instance to set the machine id {{(pid=63515) _set_machine_id /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1799}} [ 743.353163] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-dac2c5cc-03a7-4df4-a3f6-d707671aaebe tempest-AttachInterfacesUnderV243Test-1943594058 tempest-AttachInterfacesUnderV243Test-1943594058-project-member] [instance: b911a5b5-9617-4fb3-9b5e-fb8c492e4931] Powering off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 743.353163] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b788021e-78da-47eb-a984-a20a9819bf3d {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.358195] env[63515]: DEBUG oslo_vmware.api [None req-fd4a0cad-edc5-4c70-94f6-406aa5060c1e tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]528b1a0b-122e-d2ea-f7fd-e2894f4b3534, 'name': SearchDatastore_Task, 'duration_secs': 0.010225} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 743.359912] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-78c160da-b460-4170-b05d-67b7fb0dfd90 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.364470] env[63515]: DEBUG oslo_vmware.api [None req-dac2c5cc-03a7-4df4-a3f6-d707671aaebe tempest-AttachInterfacesUnderV243Test-1943594058 tempest-AttachInterfacesUnderV243Test-1943594058-project-member] Waiting for the task: (returnval){ [ 743.364470] env[63515]: value = "task-1111063" [ 743.364470] env[63515]: _type = "Task" [ 743.364470] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 743.373434] env[63515]: DEBUG oslo_vmware.api [None req-fd4a0cad-edc5-4c70-94f6-406aa5060c1e tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Waiting for the task: (returnval){ [ 743.373434] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]525da070-62e2-f658-54da-cac78c6a17a5" [ 743.373434] env[63515]: _type = "Task" [ 743.373434] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 743.383320] env[63515]: DEBUG oslo_vmware.api [None req-dac2c5cc-03a7-4df4-a3f6-d707671aaebe tempest-AttachInterfacesUnderV243Test-1943594058 tempest-AttachInterfacesUnderV243Test-1943594058-project-member] Task: {'id': task-1111063, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 743.390678] env[63515]: DEBUG oslo_vmware.api [None req-fd4a0cad-edc5-4c70-94f6-406aa5060c1e tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]525da070-62e2-f658-54da-cac78c6a17a5, 'name': SearchDatastore_Task, 'duration_secs': 0.018481} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 743.392020] env[63515]: DEBUG oslo_concurrency.lockutils [None req-fd4a0cad-edc5-4c70-94f6-406aa5060c1e tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 743.392661] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-fd4a0cad-edc5-4c70-94f6-406aa5060c1e tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk to [datastore2] 001a48e1-b237-41f8-a6f4-2fe98d173512/001a48e1-b237-41f8-a6f4-2fe98d173512.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 743.395888] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-bffe5097-6843-4728-b863-782c2a0d3a0c {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.403962] env[63515]: DEBUG oslo_vmware.api [None req-fd4a0cad-edc5-4c70-94f6-406aa5060c1e tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Waiting for the task: (returnval){ [ 743.403962] env[63515]: value = "task-1111064" [ 743.403962] env[63515]: _type = "Task" [ 743.403962] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 743.412629] env[63515]: DEBUG oslo_vmware.api [None req-fd4a0cad-edc5-4c70-94f6-406aa5060c1e tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Task: {'id': task-1111064, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 743.561419] env[63515]: DEBUG oslo_vmware.api [None req-3b658824-6e25-4b06-878c-99b3be1f1d31 tempest-ServerTagsTestJSON-1170819896 tempest-ServerTagsTestJSON-1170819896-project-member] Task: {'id': task-1111062, 'name': PowerOnVM_Task, 'duration_secs': 0.463438} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 743.563940] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-3b658824-6e25-4b06-878c-99b3be1f1d31 tempest-ServerTagsTestJSON-1170819896 tempest-ServerTagsTestJSON-1170819896-project-member] [instance: de210780-5c0f-4fba-883c-13707566a2e1] Powered on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 743.566020] env[63515]: INFO nova.compute.manager [None req-3b658824-6e25-4b06-878c-99b3be1f1d31 tempest-ServerTagsTestJSON-1170819896 tempest-ServerTagsTestJSON-1170819896-project-member] [instance: de210780-5c0f-4fba-883c-13707566a2e1] Took 8.11 seconds to spawn the instance on the hypervisor. [ 743.566020] env[63515]: DEBUG nova.compute.manager [None req-3b658824-6e25-4b06-878c-99b3be1f1d31 tempest-ServerTagsTestJSON-1170819896 tempest-ServerTagsTestJSON-1170819896-project-member] [instance: de210780-5c0f-4fba-883c-13707566a2e1] Checking state {{(pid=63515) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 743.566020] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1893ddf-ca29-470a-9df2-a3eb54ae1dc8 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.591175] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-205bb18d-8f07-4d3a-b447-41cfce54f679 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.623842] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6de0af2a-31d5-4371-955d-879926341c52 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.632582] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-99f0df88-8ade-4c67-afd9-dbdc6b385015 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] [instance: f1d01b75-ac9d-458d-8cc2-ae64cffca4e8] Updating instance 'f1d01b75-ac9d-458d-8cc2-ae64cffca4e8' progress to 83 {{(pid=63515) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 743.747682] env[63515]: DEBUG oslo_concurrency.lockutils [req-2c42f0b9-e954-49f7-8eb9-d613e5f5ead1 req-9488c0d0-0db0-4965-adb9-cc4d6f9dd345 service nova] Releasing lock "refresh_cache-b911a5b5-9617-4fb3-9b5e-fb8c492e4931" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 743.757833] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2dd472b0-40dc-488f-bd3c-e7bea44666db {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.767683] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ea1c9f3-94cc-4f90-bd71-d987977a93ec {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.806714] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-abb8fb64-f941-4b36-a08e-edc623890dbf {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.816058] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eed74917-46fe-4d4f-8fbd-ebb6d805029a {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.832080] env[63515]: DEBUG nova.compute.provider_tree [None req-5bd41396-218d-4ce7-9765-053dcc695301 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 743.878611] env[63515]: DEBUG oslo_vmware.api [None req-dac2c5cc-03a7-4df4-a3f6-d707671aaebe tempest-AttachInterfacesUnderV243Test-1943594058 tempest-AttachInterfacesUnderV243Test-1943594058-project-member] Task: {'id': task-1111063, 'name': PowerOffVM_Task, 'duration_secs': 0.164802} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 743.878893] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-dac2c5cc-03a7-4df4-a3f6-d707671aaebe tempest-AttachInterfacesUnderV243Test-1943594058 tempest-AttachInterfacesUnderV243Test-1943594058-project-member] [instance: b911a5b5-9617-4fb3-9b5e-fb8c492e4931] Powered off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 743.879076] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-dac2c5cc-03a7-4df4-a3f6-d707671aaebe tempest-AttachInterfacesUnderV243Test-1943594058 tempest-AttachInterfacesUnderV243Test-1943594058-project-member] [instance: b911a5b5-9617-4fb3-9b5e-fb8c492e4931] Unregistering the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 743.879332] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-db0e8e73-0dce-494b-b091-752bd75cec97 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.913720] env[63515]: DEBUG oslo_vmware.api [None req-fd4a0cad-edc5-4c70-94f6-406aa5060c1e tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Task: {'id': task-1111064, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 743.947716] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-dac2c5cc-03a7-4df4-a3f6-d707671aaebe tempest-AttachInterfacesUnderV243Test-1943594058 tempest-AttachInterfacesUnderV243Test-1943594058-project-member] [instance: b911a5b5-9617-4fb3-9b5e-fb8c492e4931] Unregistered the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 743.947932] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-dac2c5cc-03a7-4df4-a3f6-d707671aaebe tempest-AttachInterfacesUnderV243Test-1943594058 tempest-AttachInterfacesUnderV243Test-1943594058-project-member] [instance: b911a5b5-9617-4fb3-9b5e-fb8c492e4931] Deleting contents of the VM from datastore datastore1 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 743.948127] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-dac2c5cc-03a7-4df4-a3f6-d707671aaebe tempest-AttachInterfacesUnderV243Test-1943594058 tempest-AttachInterfacesUnderV243Test-1943594058-project-member] Deleting the datastore file [datastore1] b911a5b5-9617-4fb3-9b5e-fb8c492e4931 {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 743.949030] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-50c35249-588b-439a-a6ed-bc9ac15018d7 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.954938] env[63515]: DEBUG oslo_vmware.api [None req-dac2c5cc-03a7-4df4-a3f6-d707671aaebe tempest-AttachInterfacesUnderV243Test-1943594058 tempest-AttachInterfacesUnderV243Test-1943594058-project-member] Waiting for the task: (returnval){ [ 743.954938] env[63515]: value = "task-1111066" [ 743.954938] env[63515]: _type = "Task" [ 743.954938] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 743.963131] env[63515]: DEBUG oslo_vmware.api [None req-dac2c5cc-03a7-4df4-a3f6-d707671aaebe tempest-AttachInterfacesUnderV243Test-1943594058 tempest-AttachInterfacesUnderV243Test-1943594058-project-member] Task: {'id': task-1111066, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 744.087583] env[63515]: INFO nova.compute.manager [None req-3b658824-6e25-4b06-878c-99b3be1f1d31 tempest-ServerTagsTestJSON-1170819896 tempest-ServerTagsTestJSON-1170819896-project-member] [instance: de210780-5c0f-4fba-883c-13707566a2e1] Took 29.69 seconds to build instance. [ 744.141393] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-99f0df88-8ade-4c67-afd9-dbdc6b385015 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] [instance: f1d01b75-ac9d-458d-8cc2-ae64cffca4e8] Powering on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 744.141685] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0e0a777d-bebe-4d71-8d3e-7ab50cf6b576 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.148841] env[63515]: DEBUG oslo_vmware.api [None req-99f0df88-8ade-4c67-afd9-dbdc6b385015 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Waiting for the task: (returnval){ [ 744.148841] env[63515]: value = "task-1111067" [ 744.148841] env[63515]: _type = "Task" [ 744.148841] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 744.156941] env[63515]: DEBUG oslo_vmware.api [None req-99f0df88-8ade-4c67-afd9-dbdc6b385015 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Task: {'id': task-1111067, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 744.232015] env[63515]: DEBUG nova.compute.manager [None req-99ae4d48-1fae-4ace-913c-d249ca72bbeb tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] [instance: d2a4f66a-f1a2-4f0d-826d-0b7c4f04bd23] Start spawning the instance on the hypervisor. {{(pid=63515) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 744.257748] env[63515]: DEBUG nova.virt.hardware [None req-99ae4d48-1fae-4ace-913c-d249ca72bbeb tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T02:52:14Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T02:51:56Z,direct_url=,disk_format='vmdk',id=8a120570-cb06-4099-b262-554ca0ad15c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b569255cc43e42e7ae3f2b4ad37c6ef4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T02:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 744.258137] env[63515]: DEBUG nova.virt.hardware [None req-99ae4d48-1fae-4ace-913c-d249ca72bbeb tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Flavor limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 744.258358] env[63515]: DEBUG nova.virt.hardware [None req-99ae4d48-1fae-4ace-913c-d249ca72bbeb tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Image limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 744.258595] env[63515]: DEBUG nova.virt.hardware [None req-99ae4d48-1fae-4ace-913c-d249ca72bbeb tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Flavor pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 744.258786] env[63515]: DEBUG nova.virt.hardware [None req-99ae4d48-1fae-4ace-913c-d249ca72bbeb tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Image pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 744.258950] env[63515]: DEBUG nova.virt.hardware [None req-99ae4d48-1fae-4ace-913c-d249ca72bbeb tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 744.259168] env[63515]: DEBUG nova.virt.hardware [None req-99ae4d48-1fae-4ace-913c-d249ca72bbeb tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 744.259327] env[63515]: DEBUG nova.virt.hardware [None req-99ae4d48-1fae-4ace-913c-d249ca72bbeb tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 744.259487] env[63515]: DEBUG nova.virt.hardware [None req-99ae4d48-1fae-4ace-913c-d249ca72bbeb tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Got 1 possible topologies {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 744.259645] env[63515]: DEBUG nova.virt.hardware [None req-99ae4d48-1fae-4ace-913c-d249ca72bbeb tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 744.259814] env[63515]: DEBUG nova.virt.hardware [None req-99ae4d48-1fae-4ace-913c-d249ca72bbeb tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 744.260664] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0bd30730-111e-40c2-90c7-5bcff06a2d1d {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.269286] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ec13861-8df3-477c-a46e-694657d993c6 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.335280] env[63515]: DEBUG nova.scheduler.client.report [None req-5bd41396-218d-4ce7-9765-053dcc695301 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 744.415469] env[63515]: DEBUG oslo_vmware.api [None req-fd4a0cad-edc5-4c70-94f6-406aa5060c1e tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Task: {'id': task-1111064, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.545879} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 744.416117] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-fd4a0cad-edc5-4c70-94f6-406aa5060c1e tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk to [datastore2] 001a48e1-b237-41f8-a6f4-2fe98d173512/001a48e1-b237-41f8-a6f4-2fe98d173512.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 744.417129] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-fd4a0cad-edc5-4c70-94f6-406aa5060c1e tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: 001a48e1-b237-41f8-a6f4-2fe98d173512] Extending root virtual disk to 1048576 {{(pid=63515) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 744.417129] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-3892941a-c7f6-4455-bd9f-734372dec459 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.423365] env[63515]: DEBUG oslo_vmware.api [None req-fd4a0cad-edc5-4c70-94f6-406aa5060c1e tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Waiting for the task: (returnval){ [ 744.423365] env[63515]: value = "task-1111068" [ 744.423365] env[63515]: _type = "Task" [ 744.423365] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 744.435190] env[63515]: DEBUG oslo_vmware.api [None req-fd4a0cad-edc5-4c70-94f6-406aa5060c1e tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Task: {'id': task-1111068, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 744.469883] env[63515]: DEBUG oslo_vmware.api [None req-dac2c5cc-03a7-4df4-a3f6-d707671aaebe tempest-AttachInterfacesUnderV243Test-1943594058 tempest-AttachInterfacesUnderV243Test-1943594058-project-member] Task: {'id': task-1111066, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.140207} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 744.470243] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-dac2c5cc-03a7-4df4-a3f6-d707671aaebe tempest-AttachInterfacesUnderV243Test-1943594058 tempest-AttachInterfacesUnderV243Test-1943594058-project-member] Deleted the datastore file {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 744.470369] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-dac2c5cc-03a7-4df4-a3f6-d707671aaebe tempest-AttachInterfacesUnderV243Test-1943594058 tempest-AttachInterfacesUnderV243Test-1943594058-project-member] [instance: b911a5b5-9617-4fb3-9b5e-fb8c492e4931] Deleted contents of the VM from datastore datastore1 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 744.471011] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-dac2c5cc-03a7-4df4-a3f6-d707671aaebe tempest-AttachInterfacesUnderV243Test-1943594058 tempest-AttachInterfacesUnderV243Test-1943594058-project-member] [instance: b911a5b5-9617-4fb3-9b5e-fb8c492e4931] Instance destroyed {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 744.471011] env[63515]: INFO nova.compute.manager [None req-dac2c5cc-03a7-4df4-a3f6-d707671aaebe tempest-AttachInterfacesUnderV243Test-1943594058 tempest-AttachInterfacesUnderV243Test-1943594058-project-member] [instance: b911a5b5-9617-4fb3-9b5e-fb8c492e4931] Took 1.14 seconds to destroy the instance on the hypervisor. [ 744.471150] env[63515]: DEBUG oslo.service.loopingcall [None req-dac2c5cc-03a7-4df4-a3f6-d707671aaebe tempest-AttachInterfacesUnderV243Test-1943594058 tempest-AttachInterfacesUnderV243Test-1943594058-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 744.471292] env[63515]: DEBUG nova.compute.manager [-] [instance: b911a5b5-9617-4fb3-9b5e-fb8c492e4931] Deallocating network for instance {{(pid=63515) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 744.471391] env[63515]: DEBUG nova.network.neutron [-] [instance: b911a5b5-9617-4fb3-9b5e-fb8c492e4931] deallocate_for_instance() {{(pid=63515) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 744.590179] env[63515]: DEBUG oslo_concurrency.lockutils [None req-3b658824-6e25-4b06-878c-99b3be1f1d31 tempest-ServerTagsTestJSON-1170819896 tempest-ServerTagsTestJSON-1170819896-project-member] Lock "de210780-5c0f-4fba-883c-13707566a2e1" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 111.638s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 744.660832] env[63515]: DEBUG oslo_vmware.api [None req-99f0df88-8ade-4c67-afd9-dbdc6b385015 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Task: {'id': task-1111067, 'name': PowerOnVM_Task} progress is 88%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 744.841214] env[63515]: DEBUG oslo_concurrency.lockutils [None req-5bd41396-218d-4ce7-9765-053dcc695301 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.645s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 744.841214] env[63515]: DEBUG nova.compute.manager [None req-5bd41396-218d-4ce7-9765-053dcc695301 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] [instance: 6688cdaa-29ba-413a-8131-4f834cdb70e4] Start building networks asynchronously for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 744.847672] env[63515]: DEBUG oslo_concurrency.lockutils [None req-dfa187f2-adf0-4e24-9cf4-7df6d0016e12 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 20.588s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 744.851325] env[63515]: INFO nova.compute.claims [None req-dfa187f2-adf0-4e24-9cf4-7df6d0016e12 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] [instance: 38d6d030-06b0-4185-904d-44a038b3a752] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 744.937279] env[63515]: DEBUG oslo_vmware.api [None req-fd4a0cad-edc5-4c70-94f6-406aa5060c1e tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Task: {'id': task-1111068, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.264229} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 744.937610] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-fd4a0cad-edc5-4c70-94f6-406aa5060c1e tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: 001a48e1-b237-41f8-a6f4-2fe98d173512] Extended root virtual disk {{(pid=63515) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 744.938343] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58fd4a2b-9b42-491a-88ec-4858168a4ae6 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.964419] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-fd4a0cad-edc5-4c70-94f6-406aa5060c1e tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: 001a48e1-b237-41f8-a6f4-2fe98d173512] Reconfiguring VM instance instance-00000030 to attach disk [datastore2] 001a48e1-b237-41f8-a6f4-2fe98d173512/001a48e1-b237-41f8-a6f4-2fe98d173512.vmdk or device None with type sparse {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 744.964419] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-66064e02-1edd-46c8-9b4d-fdc48fcdc319 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.989357] env[63515]: DEBUG oslo_vmware.api [None req-fd4a0cad-edc5-4c70-94f6-406aa5060c1e tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Waiting for the task: (returnval){ [ 744.989357] env[63515]: value = "task-1111069" [ 744.989357] env[63515]: _type = "Task" [ 744.989357] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 744.997132] env[63515]: DEBUG oslo_vmware.api [None req-fd4a0cad-edc5-4c70-94f6-406aa5060c1e tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Task: {'id': task-1111069, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 745.033657] env[63515]: DEBUG nova.compute.manager [req-1af86298-ca52-43de-bb77-f2eb958d6f88 req-e93a7167-af70-42b0-8d32-5b20109dd643 service nova] [instance: d2a4f66a-f1a2-4f0d-826d-0b7c4f04bd23] Received event network-vif-plugged-3bca52da-5931-4d22-bdf1-33072a3aca76 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 745.033862] env[63515]: DEBUG oslo_concurrency.lockutils [req-1af86298-ca52-43de-bb77-f2eb958d6f88 req-e93a7167-af70-42b0-8d32-5b20109dd643 service nova] Acquiring lock "d2a4f66a-f1a2-4f0d-826d-0b7c4f04bd23-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 745.034297] env[63515]: DEBUG oslo_concurrency.lockutils [req-1af86298-ca52-43de-bb77-f2eb958d6f88 req-e93a7167-af70-42b0-8d32-5b20109dd643 service nova] Lock "d2a4f66a-f1a2-4f0d-826d-0b7c4f04bd23-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 745.034490] env[63515]: DEBUG oslo_concurrency.lockutils [req-1af86298-ca52-43de-bb77-f2eb958d6f88 req-e93a7167-af70-42b0-8d32-5b20109dd643 service nova] Lock "d2a4f66a-f1a2-4f0d-826d-0b7c4f04bd23-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 745.034660] env[63515]: DEBUG nova.compute.manager [req-1af86298-ca52-43de-bb77-f2eb958d6f88 req-e93a7167-af70-42b0-8d32-5b20109dd643 service nova] [instance: d2a4f66a-f1a2-4f0d-826d-0b7c4f04bd23] No waiting events found dispatching network-vif-plugged-3bca52da-5931-4d22-bdf1-33072a3aca76 {{(pid=63515) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 745.034950] env[63515]: WARNING nova.compute.manager [req-1af86298-ca52-43de-bb77-f2eb958d6f88 req-e93a7167-af70-42b0-8d32-5b20109dd643 service nova] [instance: d2a4f66a-f1a2-4f0d-826d-0b7c4f04bd23] Received unexpected event network-vif-plugged-3bca52da-5931-4d22-bdf1-33072a3aca76 for instance with vm_state building and task_state spawning. [ 745.051116] env[63515]: DEBUG nova.network.neutron [None req-99ae4d48-1fae-4ace-913c-d249ca72bbeb tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] [instance: d2a4f66a-f1a2-4f0d-826d-0b7c4f04bd23] Successfully updated port: 3bca52da-5931-4d22-bdf1-33072a3aca76 {{(pid=63515) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 745.093384] env[63515]: DEBUG nova.compute.manager [None req-8407c401-88c7-47ad-b91e-e7693f87e0f1 tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] [instance: 192137b0-03e5-4bc4-b911-4b4f1a874f74] Starting instance... {{(pid=63515) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 745.164479] env[63515]: DEBUG oslo_vmware.api [None req-99f0df88-8ade-4c67-afd9-dbdc6b385015 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Task: {'id': task-1111067, 'name': PowerOnVM_Task, 'duration_secs': 0.777493} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 745.165109] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-99f0df88-8ade-4c67-afd9-dbdc6b385015 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] [instance: f1d01b75-ac9d-458d-8cc2-ae64cffca4e8] Powered on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 745.165295] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-99f0df88-8ade-4c67-afd9-dbdc6b385015 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] [instance: f1d01b75-ac9d-458d-8cc2-ae64cffca4e8] Updating instance 'f1d01b75-ac9d-458d-8cc2-ae64cffca4e8' progress to 100 {{(pid=63515) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 745.357521] env[63515]: DEBUG nova.compute.utils [None req-5bd41396-218d-4ce7-9765-053dcc695301 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Using /dev/sd instead of None {{(pid=63515) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 745.361060] env[63515]: DEBUG nova.compute.manager [None req-5bd41396-218d-4ce7-9765-053dcc695301 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] [instance: 6688cdaa-29ba-413a-8131-4f834cdb70e4] Allocating IP information in the background. {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 745.361243] env[63515]: DEBUG nova.network.neutron [None req-5bd41396-218d-4ce7-9765-053dcc695301 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] [instance: 6688cdaa-29ba-413a-8131-4f834cdb70e4] allocate_for_instance() {{(pid=63515) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 745.366841] env[63515]: DEBUG nova.compute.manager [req-aeb31f2e-9e00-4957-9765-05da7c05a84d req-5ade7412-7fcc-47a4-aa37-2ef110f97ccc service nova] [instance: b911a5b5-9617-4fb3-9b5e-fb8c492e4931] Received event network-vif-deleted-9bf84cb2-7053-4129-8d8a-dc09e80b384e {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 745.366841] env[63515]: INFO nova.compute.manager [req-aeb31f2e-9e00-4957-9765-05da7c05a84d req-5ade7412-7fcc-47a4-aa37-2ef110f97ccc service nova] [instance: b911a5b5-9617-4fb3-9b5e-fb8c492e4931] Neutron deleted interface 9bf84cb2-7053-4129-8d8a-dc09e80b384e; detaching it from the instance and deleting it from the info cache [ 745.366932] env[63515]: DEBUG nova.network.neutron [req-aeb31f2e-9e00-4957-9765-05da7c05a84d req-5ade7412-7fcc-47a4-aa37-2ef110f97ccc service nova] [instance: b911a5b5-9617-4fb3-9b5e-fb8c492e4931] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 745.407021] env[63515]: DEBUG nova.policy [None req-5bd41396-218d-4ce7-9765-053dcc695301 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'eeb19c6fd65a41568e8138b3acb57bb8', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6f501ed813754f759464e338884993c1', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63515) authorize /opt/stack/nova/nova/policy.py:201}} [ 745.504239] env[63515]: DEBUG oslo_vmware.api [None req-fd4a0cad-edc5-4c70-94f6-406aa5060c1e tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Task: {'id': task-1111069, 'name': ReconfigVM_Task, 'duration_secs': 0.329528} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 745.504239] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-fd4a0cad-edc5-4c70-94f6-406aa5060c1e tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: 001a48e1-b237-41f8-a6f4-2fe98d173512] Reconfigured VM instance instance-00000030 to attach disk [datastore2] 001a48e1-b237-41f8-a6f4-2fe98d173512/001a48e1-b237-41f8-a6f4-2fe98d173512.vmdk or device None with type sparse {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 745.504239] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-06975547-122a-482d-a4a8-8e09e402b7ee {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.511102] env[63515]: DEBUG oslo_vmware.api [None req-fd4a0cad-edc5-4c70-94f6-406aa5060c1e tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Waiting for the task: (returnval){ [ 745.511102] env[63515]: value = "task-1111070" [ 745.511102] env[63515]: _type = "Task" [ 745.511102] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 745.519652] env[63515]: DEBUG oslo_vmware.api [None req-fd4a0cad-edc5-4c70-94f6-406aa5060c1e tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Task: {'id': task-1111070, 'name': Rename_Task} progress is 5%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 745.551832] env[63515]: DEBUG oslo_concurrency.lockutils [None req-99ae4d48-1fae-4ace-913c-d249ca72bbeb tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Acquiring lock "refresh_cache-d2a4f66a-f1a2-4f0d-826d-0b7c4f04bd23" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 745.551970] env[63515]: DEBUG oslo_concurrency.lockutils [None req-99ae4d48-1fae-4ace-913c-d249ca72bbeb tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Acquired lock "refresh_cache-d2a4f66a-f1a2-4f0d-826d-0b7c4f04bd23" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 745.552215] env[63515]: DEBUG nova.network.neutron [None req-99ae4d48-1fae-4ace-913c-d249ca72bbeb tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] [instance: d2a4f66a-f1a2-4f0d-826d-0b7c4f04bd23] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 745.621804] env[63515]: DEBUG oslo_concurrency.lockutils [None req-8407c401-88c7-47ad-b91e-e7693f87e0f1 tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 745.705039] env[63515]: DEBUG nova.network.neutron [None req-5bd41396-218d-4ce7-9765-053dcc695301 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] [instance: 6688cdaa-29ba-413a-8131-4f834cdb70e4] Successfully created port: 6ccc2ab3-1e4f-484b-984b-5e0ebb251bd4 {{(pid=63515) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 745.839902] env[63515]: DEBUG nova.network.neutron [-] [instance: b911a5b5-9617-4fb3-9b5e-fb8c492e4931] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 745.863032] env[63515]: DEBUG nova.compute.manager [None req-5bd41396-218d-4ce7-9765-053dcc695301 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] [instance: 6688cdaa-29ba-413a-8131-4f834cdb70e4] Start building block device mappings for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 745.879025] env[63515]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-5d340985-7fed-4972-903a-6354249151f0 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.888066] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a21eefb1-69cd-4c4a-9da9-e69453860c8f {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.922055] env[63515]: DEBUG nova.compute.manager [req-aeb31f2e-9e00-4957-9765-05da7c05a84d req-5ade7412-7fcc-47a4-aa37-2ef110f97ccc service nova] [instance: b911a5b5-9617-4fb3-9b5e-fb8c492e4931] Detach interface failed, port_id=9bf84cb2-7053-4129-8d8a-dc09e80b384e, reason: Instance b911a5b5-9617-4fb3-9b5e-fb8c492e4931 could not be found. {{(pid=63515) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 746.026767] env[63515]: DEBUG oslo_vmware.api [None req-fd4a0cad-edc5-4c70-94f6-406aa5060c1e tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Task: {'id': task-1111070, 'name': Rename_Task, 'duration_secs': 0.159466} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 746.027616] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-fd4a0cad-edc5-4c70-94f6-406aa5060c1e tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: 001a48e1-b237-41f8-a6f4-2fe98d173512] Powering on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 746.027616] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3936d5ce-6a76-4f29-b334-0dd483688b7f {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.038897] env[63515]: DEBUG oslo_vmware.api [None req-fd4a0cad-edc5-4c70-94f6-406aa5060c1e tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Waiting for the task: (returnval){ [ 746.038897] env[63515]: value = "task-1111071" [ 746.038897] env[63515]: _type = "Task" [ 746.038897] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 746.048016] env[63515]: DEBUG oslo_vmware.api [None req-fd4a0cad-edc5-4c70-94f6-406aa5060c1e tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Task: {'id': task-1111071, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 746.088503] env[63515]: DEBUG nova.network.neutron [None req-99ae4d48-1fae-4ace-913c-d249ca72bbeb tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] [instance: d2a4f66a-f1a2-4f0d-826d-0b7c4f04bd23] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 746.323153] env[63515]: DEBUG nova.network.neutron [None req-99ae4d48-1fae-4ace-913c-d249ca72bbeb tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] [instance: d2a4f66a-f1a2-4f0d-826d-0b7c4f04bd23] Updating instance_info_cache with network_info: [{"id": "3bca52da-5931-4d22-bdf1-33072a3aca76", "address": "fa:16:3e:4e:8c:2d", "network": {"id": "d2aef558-9770-46ad-b51c-61bccb0de7ec", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-2087172546-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6f501ed813754f759464e338884993c1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6d4ef133-b6f3-41d1-add4-92a1482195cf", "external-id": "nsx-vlan-transportzone-446", "segmentation_id": 446, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3bca52da-59", "ovs_interfaceid": "3bca52da-5931-4d22-bdf1-33072a3aca76", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 746.346937] env[63515]: INFO nova.compute.manager [-] [instance: b911a5b5-9617-4fb3-9b5e-fb8c492e4931] Took 1.88 seconds to deallocate network for instance. [ 746.412235] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e69426d7-debb-4deb-b49b-4288ff3a7ac9 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.420563] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69309f80-c7d6-4f16-9de0-89a3d13cff75 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.455636] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c667787-9b0e-4a8c-863f-caaf6c14ae70 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.463441] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-845ad9be-3ff2-45c8-a425-bf53dcab86cb {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.477006] env[63515]: DEBUG nova.compute.provider_tree [None req-dfa187f2-adf0-4e24-9cf4-7df6d0016e12 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Updating inventory in ProviderTree for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 746.550928] env[63515]: DEBUG oslo_vmware.api [None req-fd4a0cad-edc5-4c70-94f6-406aa5060c1e tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Task: {'id': task-1111071, 'name': PowerOnVM_Task} progress is 88%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 746.828307] env[63515]: DEBUG oslo_concurrency.lockutils [None req-99ae4d48-1fae-4ace-913c-d249ca72bbeb tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Releasing lock "refresh_cache-d2a4f66a-f1a2-4f0d-826d-0b7c4f04bd23" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 746.828640] env[63515]: DEBUG nova.compute.manager [None req-99ae4d48-1fae-4ace-913c-d249ca72bbeb tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] [instance: d2a4f66a-f1a2-4f0d-826d-0b7c4f04bd23] Instance network_info: |[{"id": "3bca52da-5931-4d22-bdf1-33072a3aca76", "address": "fa:16:3e:4e:8c:2d", "network": {"id": "d2aef558-9770-46ad-b51c-61bccb0de7ec", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-2087172546-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6f501ed813754f759464e338884993c1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6d4ef133-b6f3-41d1-add4-92a1482195cf", "external-id": "nsx-vlan-transportzone-446", "segmentation_id": 446, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3bca52da-59", "ovs_interfaceid": "3bca52da-5931-4d22-bdf1-33072a3aca76", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 746.829257] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-99ae4d48-1fae-4ace-913c-d249ca72bbeb tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] [instance: d2a4f66a-f1a2-4f0d-826d-0b7c4f04bd23] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:4e:8c:2d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '6d4ef133-b6f3-41d1-add4-92a1482195cf', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '3bca52da-5931-4d22-bdf1-33072a3aca76', 'vif_model': 'vmxnet3'}] {{(pid=63515) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 746.836970] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-99ae4d48-1fae-4ace-913c-d249ca72bbeb tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Creating folder: Project (6f501ed813754f759464e338884993c1). Parent ref: group-v243370. {{(pid=63515) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 746.837597] env[63515]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-7cccd4bd-5464-4f4b-af51-44fa17149ff0 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.848760] env[63515]: INFO nova.virt.vmwareapi.vm_util [None req-99ae4d48-1fae-4ace-913c-d249ca72bbeb tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Created folder: Project (6f501ed813754f759464e338884993c1) in parent group-v243370. [ 746.848966] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-99ae4d48-1fae-4ace-913c-d249ca72bbeb tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Creating folder: Instances. Parent ref: group-v243431. {{(pid=63515) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 746.849224] env[63515]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-2cd75851-18a3-488e-a51c-6d09c54c740f {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.853903] env[63515]: DEBUG oslo_concurrency.lockutils [None req-dac2c5cc-03a7-4df4-a3f6-d707671aaebe tempest-AttachInterfacesUnderV243Test-1943594058 tempest-AttachInterfacesUnderV243Test-1943594058-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 746.858600] env[63515]: INFO nova.virt.vmwareapi.vm_util [None req-99ae4d48-1fae-4ace-913c-d249ca72bbeb tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Created folder: Instances in parent group-v243431. [ 746.858694] env[63515]: DEBUG oslo.service.loopingcall [None req-99ae4d48-1fae-4ace-913c-d249ca72bbeb tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 746.858840] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d2a4f66a-f1a2-4f0d-826d-0b7c4f04bd23] Creating VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 746.859089] env[63515]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-9bc7daa5-5e9f-4631-9b6c-b35f2d91636f {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.878683] env[63515]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 746.878683] env[63515]: value = "task-1111074" [ 746.878683] env[63515]: _type = "Task" [ 746.878683] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 746.882855] env[63515]: DEBUG nova.compute.manager [None req-5bd41396-218d-4ce7-9765-053dcc695301 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] [instance: 6688cdaa-29ba-413a-8131-4f834cdb70e4] Start spawning the instance on the hypervisor. {{(pid=63515) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 746.889619] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1111074, 'name': CreateVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 746.908448] env[63515]: DEBUG nova.virt.hardware [None req-5bd41396-218d-4ce7-9765-053dcc695301 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T02:52:14Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T02:51:56Z,direct_url=,disk_format='vmdk',id=8a120570-cb06-4099-b262-554ca0ad15c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b569255cc43e42e7ae3f2b4ad37c6ef4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T02:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 746.908696] env[63515]: DEBUG nova.virt.hardware [None req-5bd41396-218d-4ce7-9765-053dcc695301 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Flavor limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 746.908856] env[63515]: DEBUG nova.virt.hardware [None req-5bd41396-218d-4ce7-9765-053dcc695301 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Image limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 746.909104] env[63515]: DEBUG nova.virt.hardware [None req-5bd41396-218d-4ce7-9765-053dcc695301 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Flavor pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 746.909303] env[63515]: DEBUG nova.virt.hardware [None req-5bd41396-218d-4ce7-9765-053dcc695301 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Image pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 746.909462] env[63515]: DEBUG nova.virt.hardware [None req-5bd41396-218d-4ce7-9765-053dcc695301 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 746.909671] env[63515]: DEBUG nova.virt.hardware [None req-5bd41396-218d-4ce7-9765-053dcc695301 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 746.909848] env[63515]: DEBUG nova.virt.hardware [None req-5bd41396-218d-4ce7-9765-053dcc695301 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 746.910031] env[63515]: DEBUG nova.virt.hardware [None req-5bd41396-218d-4ce7-9765-053dcc695301 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Got 1 possible topologies {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 746.910199] env[63515]: DEBUG nova.virt.hardware [None req-5bd41396-218d-4ce7-9765-053dcc695301 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 746.910368] env[63515]: DEBUG nova.virt.hardware [None req-5bd41396-218d-4ce7-9765-053dcc695301 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 746.911275] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-914078d0-4945-47e2-b68f-2a77fbd5e282 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.919292] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e06f0cdd-1cb6-4dd8-a693-299237afdc02 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.024434] env[63515]: DEBUG nova.scheduler.client.report [None req-dfa187f2-adf0-4e24-9cf4-7df6d0016e12 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Updated inventory for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a with generation 72 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 747.024697] env[63515]: DEBUG nova.compute.provider_tree [None req-dfa187f2-adf0-4e24-9cf4-7df6d0016e12 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Updating resource provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a generation from 72 to 73 during operation: update_inventory {{(pid=63515) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 747.024906] env[63515]: DEBUG nova.compute.provider_tree [None req-dfa187f2-adf0-4e24-9cf4-7df6d0016e12 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Updating inventory in ProviderTree for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 747.033031] env[63515]: DEBUG oslo_concurrency.lockutils [None req-711301e3-5f69-4e98-bb19-e93c4ab2a0db tempest-ServerTagsTestJSON-1170819896 tempest-ServerTagsTestJSON-1170819896-project-member] Acquiring lock "de210780-5c0f-4fba-883c-13707566a2e1" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 747.033031] env[63515]: DEBUG oslo_concurrency.lockutils [None req-711301e3-5f69-4e98-bb19-e93c4ab2a0db tempest-ServerTagsTestJSON-1170819896 tempest-ServerTagsTestJSON-1170819896-project-member] Lock "de210780-5c0f-4fba-883c-13707566a2e1" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 747.033326] env[63515]: DEBUG oslo_concurrency.lockutils [None req-711301e3-5f69-4e98-bb19-e93c4ab2a0db tempest-ServerTagsTestJSON-1170819896 tempest-ServerTagsTestJSON-1170819896-project-member] Acquiring lock "de210780-5c0f-4fba-883c-13707566a2e1-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 747.033398] env[63515]: DEBUG oslo_concurrency.lockutils [None req-711301e3-5f69-4e98-bb19-e93c4ab2a0db tempest-ServerTagsTestJSON-1170819896 tempest-ServerTagsTestJSON-1170819896-project-member] Lock "de210780-5c0f-4fba-883c-13707566a2e1-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 747.033597] env[63515]: DEBUG oslo_concurrency.lockutils [None req-711301e3-5f69-4e98-bb19-e93c4ab2a0db tempest-ServerTagsTestJSON-1170819896 tempest-ServerTagsTestJSON-1170819896-project-member] Lock "de210780-5c0f-4fba-883c-13707566a2e1-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 747.037317] env[63515]: INFO nova.compute.manager [None req-711301e3-5f69-4e98-bb19-e93c4ab2a0db tempest-ServerTagsTestJSON-1170819896 tempest-ServerTagsTestJSON-1170819896-project-member] [instance: de210780-5c0f-4fba-883c-13707566a2e1] Terminating instance [ 747.040108] env[63515]: DEBUG nova.compute.manager [None req-711301e3-5f69-4e98-bb19-e93c4ab2a0db tempest-ServerTagsTestJSON-1170819896 tempest-ServerTagsTestJSON-1170819896-project-member] [instance: de210780-5c0f-4fba-883c-13707566a2e1] Start destroying the instance on the hypervisor. {{(pid=63515) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 747.040234] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-711301e3-5f69-4e98-bb19-e93c4ab2a0db tempest-ServerTagsTestJSON-1170819896 tempest-ServerTagsTestJSON-1170819896-project-member] [instance: de210780-5c0f-4fba-883c-13707566a2e1] Destroying instance {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 747.043281] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ccd81933-5473-47a6-9583-c5efd5fab51e {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.057915] env[63515]: DEBUG oslo_vmware.api [None req-fd4a0cad-edc5-4c70-94f6-406aa5060c1e tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Task: {'id': task-1111071, 'name': PowerOnVM_Task} progress is 100%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 747.060375] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-711301e3-5f69-4e98-bb19-e93c4ab2a0db tempest-ServerTagsTestJSON-1170819896 tempest-ServerTagsTestJSON-1170819896-project-member] [instance: de210780-5c0f-4fba-883c-13707566a2e1] Powering off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 747.060508] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ebe5284d-b5bf-4ca1-a953-8f737b6bca8f {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.068703] env[63515]: DEBUG oslo_vmware.api [None req-711301e3-5f69-4e98-bb19-e93c4ab2a0db tempest-ServerTagsTestJSON-1170819896 tempest-ServerTagsTestJSON-1170819896-project-member] Waiting for the task: (returnval){ [ 747.068703] env[63515]: value = "task-1111075" [ 747.068703] env[63515]: _type = "Task" [ 747.068703] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 747.077769] env[63515]: DEBUG oslo_vmware.api [None req-711301e3-5f69-4e98-bb19-e93c4ab2a0db tempest-ServerTagsTestJSON-1170819896 tempest-ServerTagsTestJSON-1170819896-project-member] Task: {'id': task-1111075, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 747.119622] env[63515]: DEBUG nova.compute.manager [req-8db4d88d-d5ab-46fa-b516-fa249bd6caaf req-bb14f3ce-a15f-42d2-ad0b-f33279f615ac service nova] [instance: d2a4f66a-f1a2-4f0d-826d-0b7c4f04bd23] Received event network-changed-3bca52da-5931-4d22-bdf1-33072a3aca76 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 747.119622] env[63515]: DEBUG nova.compute.manager [req-8db4d88d-d5ab-46fa-b516-fa249bd6caaf req-bb14f3ce-a15f-42d2-ad0b-f33279f615ac service nova] [instance: d2a4f66a-f1a2-4f0d-826d-0b7c4f04bd23] Refreshing instance network info cache due to event network-changed-3bca52da-5931-4d22-bdf1-33072a3aca76. {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 747.119622] env[63515]: DEBUG oslo_concurrency.lockutils [req-8db4d88d-d5ab-46fa-b516-fa249bd6caaf req-bb14f3ce-a15f-42d2-ad0b-f33279f615ac service nova] Acquiring lock "refresh_cache-d2a4f66a-f1a2-4f0d-826d-0b7c4f04bd23" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 747.119622] env[63515]: DEBUG oslo_concurrency.lockutils [req-8db4d88d-d5ab-46fa-b516-fa249bd6caaf req-bb14f3ce-a15f-42d2-ad0b-f33279f615ac service nova] Acquired lock "refresh_cache-d2a4f66a-f1a2-4f0d-826d-0b7c4f04bd23" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 747.120274] env[63515]: DEBUG nova.network.neutron [req-8db4d88d-d5ab-46fa-b516-fa249bd6caaf req-bb14f3ce-a15f-42d2-ad0b-f33279f615ac service nova] [instance: d2a4f66a-f1a2-4f0d-826d-0b7c4f04bd23] Refreshing network info cache for port 3bca52da-5931-4d22-bdf1-33072a3aca76 {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 747.390121] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1111074, 'name': CreateVM_Task, 'duration_secs': 0.40665} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 747.390666] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d2a4f66a-f1a2-4f0d-826d-0b7c4f04bd23] Created VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 747.391938] env[63515]: DEBUG oslo_concurrency.lockutils [None req-99ae4d48-1fae-4ace-913c-d249ca72bbeb tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 747.392462] env[63515]: DEBUG oslo_concurrency.lockutils [None req-99ae4d48-1fae-4ace-913c-d249ca72bbeb tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 747.394163] env[63515]: DEBUG oslo_concurrency.lockutils [None req-99ae4d48-1fae-4ace-913c-d249ca72bbeb tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 747.394163] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3c7ad16c-3950-44fc-a72b-7d46c772e0d7 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.399761] env[63515]: DEBUG oslo_vmware.api [None req-99ae4d48-1fae-4ace-913c-d249ca72bbeb tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Waiting for the task: (returnval){ [ 747.399761] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]529cb459-0e76-df90-1275-1998a3847895" [ 747.399761] env[63515]: _type = "Task" [ 747.399761] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 747.411240] env[63515]: DEBUG oslo_vmware.api [None req-99ae4d48-1fae-4ace-913c-d249ca72bbeb tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]529cb459-0e76-df90-1275-1998a3847895, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 747.532426] env[63515]: DEBUG oslo_concurrency.lockutils [None req-dfa187f2-adf0-4e24-9cf4-7df6d0016e12 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.686s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 747.532969] env[63515]: DEBUG nova.compute.manager [None req-dfa187f2-adf0-4e24-9cf4-7df6d0016e12 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] [instance: 38d6d030-06b0-4185-904d-44a038b3a752] Start building networks asynchronously for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 747.536184] env[63515]: DEBUG nova.compute.manager [req-b0aff6a9-9032-46e8-a537-00f34cf188a6 req-9c1afe2c-9d56-47ce-bc0d-1990685be068 service nova] [instance: 6688cdaa-29ba-413a-8131-4f834cdb70e4] Received event network-vif-plugged-6ccc2ab3-1e4f-484b-984b-5e0ebb251bd4 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 747.536400] env[63515]: DEBUG oslo_concurrency.lockutils [req-b0aff6a9-9032-46e8-a537-00f34cf188a6 req-9c1afe2c-9d56-47ce-bc0d-1990685be068 service nova] Acquiring lock "6688cdaa-29ba-413a-8131-4f834cdb70e4-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 747.536621] env[63515]: DEBUG oslo_concurrency.lockutils [req-b0aff6a9-9032-46e8-a537-00f34cf188a6 req-9c1afe2c-9d56-47ce-bc0d-1990685be068 service nova] Lock "6688cdaa-29ba-413a-8131-4f834cdb70e4-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 747.536787] env[63515]: DEBUG oslo_concurrency.lockutils [req-b0aff6a9-9032-46e8-a537-00f34cf188a6 req-9c1afe2c-9d56-47ce-bc0d-1990685be068 service nova] Lock "6688cdaa-29ba-413a-8131-4f834cdb70e4-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 747.536963] env[63515]: DEBUG nova.compute.manager [req-b0aff6a9-9032-46e8-a537-00f34cf188a6 req-9c1afe2c-9d56-47ce-bc0d-1990685be068 service nova] [instance: 6688cdaa-29ba-413a-8131-4f834cdb70e4] No waiting events found dispatching network-vif-plugged-6ccc2ab3-1e4f-484b-984b-5e0ebb251bd4 {{(pid=63515) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 747.537149] env[63515]: WARNING nova.compute.manager [req-b0aff6a9-9032-46e8-a537-00f34cf188a6 req-9c1afe2c-9d56-47ce-bc0d-1990685be068 service nova] [instance: 6688cdaa-29ba-413a-8131-4f834cdb70e4] Received unexpected event network-vif-plugged-6ccc2ab3-1e4f-484b-984b-5e0ebb251bd4 for instance with vm_state building and task_state spawning. [ 747.540192] env[63515]: DEBUG oslo_concurrency.lockutils [None req-750d4038-2a95-4b14-b9d7-c9c6803c64b6 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 22.280s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 747.541493] env[63515]: INFO nova.compute.claims [None req-750d4038-2a95-4b14-b9d7-c9c6803c64b6 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] [instance: 8eb17506-ff93-4d25-b9af-ec5886569e65] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 747.554149] env[63515]: DEBUG oslo_vmware.api [None req-fd4a0cad-edc5-4c70-94f6-406aa5060c1e tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Task: {'id': task-1111071, 'name': PowerOnVM_Task, 'duration_secs': 1.035313} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 747.554405] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-fd4a0cad-edc5-4c70-94f6-406aa5060c1e tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: 001a48e1-b237-41f8-a6f4-2fe98d173512] Powered on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 747.554844] env[63515]: INFO nova.compute.manager [None req-fd4a0cad-edc5-4c70-94f6-406aa5060c1e tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: 001a48e1-b237-41f8-a6f4-2fe98d173512] Took 9.39 seconds to spawn the instance on the hypervisor. [ 747.555146] env[63515]: DEBUG nova.compute.manager [None req-fd4a0cad-edc5-4c70-94f6-406aa5060c1e tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: 001a48e1-b237-41f8-a6f4-2fe98d173512] Checking state {{(pid=63515) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 747.555958] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87128bb9-9cdd-47d7-a245-23bfc3e20ec5 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.577998] env[63515]: DEBUG oslo_vmware.api [None req-711301e3-5f69-4e98-bb19-e93c4ab2a0db tempest-ServerTagsTestJSON-1170819896 tempest-ServerTagsTestJSON-1170819896-project-member] Task: {'id': task-1111075, 'name': PowerOffVM_Task, 'duration_secs': 0.247053} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 747.580056] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-711301e3-5f69-4e98-bb19-e93c4ab2a0db tempest-ServerTagsTestJSON-1170819896 tempest-ServerTagsTestJSON-1170819896-project-member] [instance: de210780-5c0f-4fba-883c-13707566a2e1] Powered off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 747.580056] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-711301e3-5f69-4e98-bb19-e93c4ab2a0db tempest-ServerTagsTestJSON-1170819896 tempest-ServerTagsTestJSON-1170819896-project-member] [instance: de210780-5c0f-4fba-883c-13707566a2e1] Unregistering the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 747.580056] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-0596d44e-9369-4414-8fa7-f34f3212a303 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.635694] env[63515]: DEBUG nova.network.neutron [None req-5bd41396-218d-4ce7-9765-053dcc695301 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] [instance: 6688cdaa-29ba-413a-8131-4f834cdb70e4] Successfully updated port: 6ccc2ab3-1e4f-484b-984b-5e0ebb251bd4 {{(pid=63515) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 747.656683] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-711301e3-5f69-4e98-bb19-e93c4ab2a0db tempest-ServerTagsTestJSON-1170819896 tempest-ServerTagsTestJSON-1170819896-project-member] [instance: de210780-5c0f-4fba-883c-13707566a2e1] Unregistered the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 747.660535] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-711301e3-5f69-4e98-bb19-e93c4ab2a0db tempest-ServerTagsTestJSON-1170819896 tempest-ServerTagsTestJSON-1170819896-project-member] [instance: de210780-5c0f-4fba-883c-13707566a2e1] Deleting contents of the VM from datastore datastore2 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 747.660535] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-711301e3-5f69-4e98-bb19-e93c4ab2a0db tempest-ServerTagsTestJSON-1170819896 tempest-ServerTagsTestJSON-1170819896-project-member] Deleting the datastore file [datastore2] de210780-5c0f-4fba-883c-13707566a2e1 {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 747.660535] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4c4f241a-9256-49ea-9088-de4f4ce6b59c {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.666924] env[63515]: DEBUG oslo_vmware.api [None req-711301e3-5f69-4e98-bb19-e93c4ab2a0db tempest-ServerTagsTestJSON-1170819896 tempest-ServerTagsTestJSON-1170819896-project-member] Waiting for the task: (returnval){ [ 747.666924] env[63515]: value = "task-1111077" [ 747.666924] env[63515]: _type = "Task" [ 747.666924] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 747.677017] env[63515]: DEBUG oslo_vmware.api [None req-711301e3-5f69-4e98-bb19-e93c4ab2a0db tempest-ServerTagsTestJSON-1170819896 tempest-ServerTagsTestJSON-1170819896-project-member] Task: {'id': task-1111077, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 747.703390] env[63515]: DEBUG nova.network.neutron [None req-ffbb9c80-abfa-457c-b756-9497c4635a40 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] [instance: f1d01b75-ac9d-458d-8cc2-ae64cffca4e8] Port 48f0868f-94db-42f7-8153-d1a27fa02707 binding to destination host cpu-1 is already ACTIVE {{(pid=63515) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 747.703809] env[63515]: DEBUG oslo_concurrency.lockutils [None req-ffbb9c80-abfa-457c-b756-9497c4635a40 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Acquiring lock "refresh_cache-f1d01b75-ac9d-458d-8cc2-ae64cffca4e8" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 747.704658] env[63515]: DEBUG oslo_concurrency.lockutils [None req-ffbb9c80-abfa-457c-b756-9497c4635a40 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Acquired lock "refresh_cache-f1d01b75-ac9d-458d-8cc2-ae64cffca4e8" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 747.704658] env[63515]: DEBUG nova.network.neutron [None req-ffbb9c80-abfa-457c-b756-9497c4635a40 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] [instance: f1d01b75-ac9d-458d-8cc2-ae64cffca4e8] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 747.877511] env[63515]: DEBUG nova.network.neutron [req-8db4d88d-d5ab-46fa-b516-fa249bd6caaf req-bb14f3ce-a15f-42d2-ad0b-f33279f615ac service nova] [instance: d2a4f66a-f1a2-4f0d-826d-0b7c4f04bd23] Updated VIF entry in instance network info cache for port 3bca52da-5931-4d22-bdf1-33072a3aca76. {{(pid=63515) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 747.879387] env[63515]: DEBUG nova.network.neutron [req-8db4d88d-d5ab-46fa-b516-fa249bd6caaf req-bb14f3ce-a15f-42d2-ad0b-f33279f615ac service nova] [instance: d2a4f66a-f1a2-4f0d-826d-0b7c4f04bd23] Updating instance_info_cache with network_info: [{"id": "3bca52da-5931-4d22-bdf1-33072a3aca76", "address": "fa:16:3e:4e:8c:2d", "network": {"id": "d2aef558-9770-46ad-b51c-61bccb0de7ec", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-2087172546-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6f501ed813754f759464e338884993c1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6d4ef133-b6f3-41d1-add4-92a1482195cf", "external-id": "nsx-vlan-transportzone-446", "segmentation_id": 446, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3bca52da-59", "ovs_interfaceid": "3bca52da-5931-4d22-bdf1-33072a3aca76", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 747.912948] env[63515]: DEBUG oslo_vmware.api [None req-99ae4d48-1fae-4ace-913c-d249ca72bbeb tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]529cb459-0e76-df90-1275-1998a3847895, 'name': SearchDatastore_Task, 'duration_secs': 0.027617} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 747.913281] env[63515]: DEBUG oslo_concurrency.lockutils [None req-99ae4d48-1fae-4ace-913c-d249ca72bbeb tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 747.913512] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-99ae4d48-1fae-4ace-913c-d249ca72bbeb tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] [instance: d2a4f66a-f1a2-4f0d-826d-0b7c4f04bd23] Processing image 8a120570-cb06-4099-b262-554ca0ad15c5 {{(pid=63515) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 747.913740] env[63515]: DEBUG oslo_concurrency.lockutils [None req-99ae4d48-1fae-4ace-913c-d249ca72bbeb tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 747.913883] env[63515]: DEBUG oslo_concurrency.lockutils [None req-99ae4d48-1fae-4ace-913c-d249ca72bbeb tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 747.914071] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-99ae4d48-1fae-4ace-913c-d249ca72bbeb tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 747.914327] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-4e0a0371-5401-4f32-9857-f8012fce547a {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.924957] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-99ae4d48-1fae-4ace-913c-d249ca72bbeb tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 747.925164] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-99ae4d48-1fae-4ace-913c-d249ca72bbeb tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63515) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 747.925903] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0f1c03a3-5b39-423e-be2c-c4e164b0c1b7 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.931770] env[63515]: DEBUG oslo_vmware.api [None req-99ae4d48-1fae-4ace-913c-d249ca72bbeb tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Waiting for the task: (returnval){ [ 747.931770] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]5208de1b-7eae-e042-36d4-2e010eae0f7d" [ 747.931770] env[63515]: _type = "Task" [ 747.931770] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 747.939628] env[63515]: DEBUG oslo_vmware.api [None req-99ae4d48-1fae-4ace-913c-d249ca72bbeb tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]5208de1b-7eae-e042-36d4-2e010eae0f7d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 748.039110] env[63515]: DEBUG nova.compute.utils [None req-dfa187f2-adf0-4e24-9cf4-7df6d0016e12 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Using /dev/sd instead of None {{(pid=63515) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 748.040543] env[63515]: DEBUG nova.compute.manager [None req-dfa187f2-adf0-4e24-9cf4-7df6d0016e12 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] [instance: 38d6d030-06b0-4185-904d-44a038b3a752] Allocating IP information in the background. {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 748.040714] env[63515]: DEBUG nova.network.neutron [None req-dfa187f2-adf0-4e24-9cf4-7df6d0016e12 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] [instance: 38d6d030-06b0-4185-904d-44a038b3a752] allocate_for_instance() {{(pid=63515) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 748.076181] env[63515]: INFO nova.compute.manager [None req-fd4a0cad-edc5-4c70-94f6-406aa5060c1e tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: 001a48e1-b237-41f8-a6f4-2fe98d173512] Took 32.05 seconds to build instance. [ 748.089562] env[63515]: DEBUG nova.policy [None req-dfa187f2-adf0-4e24-9cf4-7df6d0016e12 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'eeb19c6fd65a41568e8138b3acb57bb8', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6f501ed813754f759464e338884993c1', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63515) authorize /opt/stack/nova/nova/policy.py:201}} [ 748.141417] env[63515]: DEBUG oslo_concurrency.lockutils [None req-5bd41396-218d-4ce7-9765-053dcc695301 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Acquiring lock "refresh_cache-6688cdaa-29ba-413a-8131-4f834cdb70e4" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 748.141957] env[63515]: DEBUG oslo_concurrency.lockutils [None req-5bd41396-218d-4ce7-9765-053dcc695301 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Acquired lock "refresh_cache-6688cdaa-29ba-413a-8131-4f834cdb70e4" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 748.142180] env[63515]: DEBUG nova.network.neutron [None req-5bd41396-218d-4ce7-9765-053dcc695301 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] [instance: 6688cdaa-29ba-413a-8131-4f834cdb70e4] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 748.176651] env[63515]: DEBUG oslo_vmware.api [None req-711301e3-5f69-4e98-bb19-e93c4ab2a0db tempest-ServerTagsTestJSON-1170819896 tempest-ServerTagsTestJSON-1170819896-project-member] Task: {'id': task-1111077, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.457208} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 748.176917] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-711301e3-5f69-4e98-bb19-e93c4ab2a0db tempest-ServerTagsTestJSON-1170819896 tempest-ServerTagsTestJSON-1170819896-project-member] Deleted the datastore file {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 748.177157] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-711301e3-5f69-4e98-bb19-e93c4ab2a0db tempest-ServerTagsTestJSON-1170819896 tempest-ServerTagsTestJSON-1170819896-project-member] [instance: de210780-5c0f-4fba-883c-13707566a2e1] Deleted contents of the VM from datastore datastore2 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 748.177366] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-711301e3-5f69-4e98-bb19-e93c4ab2a0db tempest-ServerTagsTestJSON-1170819896 tempest-ServerTagsTestJSON-1170819896-project-member] [instance: de210780-5c0f-4fba-883c-13707566a2e1] Instance destroyed {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 748.177546] env[63515]: INFO nova.compute.manager [None req-711301e3-5f69-4e98-bb19-e93c4ab2a0db tempest-ServerTagsTestJSON-1170819896 tempest-ServerTagsTestJSON-1170819896-project-member] [instance: de210780-5c0f-4fba-883c-13707566a2e1] Took 1.14 seconds to destroy the instance on the hypervisor. [ 748.177783] env[63515]: DEBUG oslo.service.loopingcall [None req-711301e3-5f69-4e98-bb19-e93c4ab2a0db tempest-ServerTagsTestJSON-1170819896 tempest-ServerTagsTestJSON-1170819896-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 748.177966] env[63515]: DEBUG nova.compute.manager [-] [instance: de210780-5c0f-4fba-883c-13707566a2e1] Deallocating network for instance {{(pid=63515) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 748.178084] env[63515]: DEBUG nova.network.neutron [-] [instance: de210780-5c0f-4fba-883c-13707566a2e1] deallocate_for_instance() {{(pid=63515) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 748.382834] env[63515]: DEBUG oslo_concurrency.lockutils [req-8db4d88d-d5ab-46fa-b516-fa249bd6caaf req-bb14f3ce-a15f-42d2-ad0b-f33279f615ac service nova] Releasing lock "refresh_cache-d2a4f66a-f1a2-4f0d-826d-0b7c4f04bd23" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 748.442988] env[63515]: DEBUG oslo_vmware.api [None req-99ae4d48-1fae-4ace-913c-d249ca72bbeb tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]5208de1b-7eae-e042-36d4-2e010eae0f7d, 'name': SearchDatastore_Task, 'duration_secs': 0.018314} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 748.443964] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-36d61c79-3d3f-4573-9e7d-66c9211919c2 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.450647] env[63515]: DEBUG oslo_vmware.api [None req-99ae4d48-1fae-4ace-913c-d249ca72bbeb tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Waiting for the task: (returnval){ [ 748.450647] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]5201e0fa-35c4-6d6e-ac9d-6d3483440f82" [ 748.450647] env[63515]: _type = "Task" [ 748.450647] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 748.459868] env[63515]: DEBUG oslo_vmware.api [None req-99ae4d48-1fae-4ace-913c-d249ca72bbeb tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]5201e0fa-35c4-6d6e-ac9d-6d3483440f82, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 748.470812] env[63515]: DEBUG nova.network.neutron [None req-ffbb9c80-abfa-457c-b756-9497c4635a40 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] [instance: f1d01b75-ac9d-458d-8cc2-ae64cffca4e8] Updating instance_info_cache with network_info: [{"id": "48f0868f-94db-42f7-8153-d1a27fa02707", "address": "fa:16:3e:7a:74:da", "network": {"id": "b4abcbd0-0eac-4bde-ad11-7f93afcedcef", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.21", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "b569255cc43e42e7ae3f2b4ad37c6ef4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0cd5d325-3053-407e-a4ee-f627e82a23f9", "external-id": "nsx-vlan-transportzone-809", "segmentation_id": 809, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap48f0868f-94", "ovs_interfaceid": "48f0868f-94db-42f7-8153-d1a27fa02707", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 748.545030] env[63515]: DEBUG nova.compute.manager [None req-dfa187f2-adf0-4e24-9cf4-7df6d0016e12 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] [instance: 38d6d030-06b0-4185-904d-44a038b3a752] Start building block device mappings for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 748.581592] env[63515]: DEBUG oslo_concurrency.lockutils [None req-fd4a0cad-edc5-4c70-94f6-406aa5060c1e tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Lock "001a48e1-b237-41f8-a6f4-2fe98d173512" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 111.417s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 748.631843] env[63515]: DEBUG nova.network.neutron [None req-dfa187f2-adf0-4e24-9cf4-7df6d0016e12 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] [instance: 38d6d030-06b0-4185-904d-44a038b3a752] Successfully created port: 2242e364-37c0-4750-8755-d807d47e899d {{(pid=63515) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 748.701123] env[63515]: DEBUG nova.network.neutron [None req-5bd41396-218d-4ce7-9765-053dcc695301 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] [instance: 6688cdaa-29ba-413a-8131-4f834cdb70e4] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 748.879218] env[63515]: DEBUG nova.network.neutron [None req-5bd41396-218d-4ce7-9765-053dcc695301 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] [instance: 6688cdaa-29ba-413a-8131-4f834cdb70e4] Updating instance_info_cache with network_info: [{"id": "6ccc2ab3-1e4f-484b-984b-5e0ebb251bd4", "address": "fa:16:3e:e0:db:2c", "network": {"id": "d2aef558-9770-46ad-b51c-61bccb0de7ec", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-2087172546-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6f501ed813754f759464e338884993c1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6d4ef133-b6f3-41d1-add4-92a1482195cf", "external-id": "nsx-vlan-transportzone-446", "segmentation_id": 446, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6ccc2ab3-1e", "ovs_interfaceid": "6ccc2ab3-1e4f-484b-984b-5e0ebb251bd4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 748.966870] env[63515]: DEBUG oslo_vmware.api [None req-99ae4d48-1fae-4ace-913c-d249ca72bbeb tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]5201e0fa-35c4-6d6e-ac9d-6d3483440f82, 'name': SearchDatastore_Task, 'duration_secs': 0.043767} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 748.966870] env[63515]: DEBUG oslo_concurrency.lockutils [None req-99ae4d48-1fae-4ace-913c-d249ca72bbeb tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 748.966870] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-99ae4d48-1fae-4ace-913c-d249ca72bbeb tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk to [datastore1] d2a4f66a-f1a2-4f0d-826d-0b7c4f04bd23/d2a4f66a-f1a2-4f0d-826d-0b7c4f04bd23.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 748.966870] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-762ce425-c40d-460d-8605-b34a6f07edd1 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.972609] env[63515]: DEBUG oslo_vmware.api [None req-99ae4d48-1fae-4ace-913c-d249ca72bbeb tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Waiting for the task: (returnval){ [ 748.972609] env[63515]: value = "task-1111078" [ 748.972609] env[63515]: _type = "Task" [ 748.972609] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 748.973397] env[63515]: DEBUG oslo_concurrency.lockutils [None req-ffbb9c80-abfa-457c-b756-9497c4635a40 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Releasing lock "refresh_cache-f1d01b75-ac9d-458d-8cc2-ae64cffca4e8" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 748.989247] env[63515]: DEBUG oslo_vmware.api [None req-99ae4d48-1fae-4ace-913c-d249ca72bbeb tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Task: {'id': task-1111078, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 749.007599] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e65e3c26-6df2-43c6-8dae-bad82bc1d895 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.015302] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2a9d966-d281-407a-ad97-a396186812f8 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.045949] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eadfb848-a3bc-4e05-a022-bb5bc7aeed2a {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.056254] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06c0b2e8-ff38-4255-b205-6907e032370d {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.074517] env[63515]: DEBUG nova.compute.provider_tree [None req-750d4038-2a95-4b14-b9d7-c9c6803c64b6 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 749.084059] env[63515]: DEBUG nova.compute.manager [None req-83fbd51b-fc37-49eb-8423-cdaa019d3b0c tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] [instance: 5cbce760-0163-4b27-8ae3-e46c926c8916] Starting instance... {{(pid=63515) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 749.148763] env[63515]: DEBUG nova.compute.manager [req-f1bf6141-436a-484f-b066-984dda728a23 req-2fad47e0-675a-4928-9013-7d1d1524584a service nova] [instance: de210780-5c0f-4fba-883c-13707566a2e1] Received event network-vif-deleted-cd4f3cb8-f205-4ace-a17f-d54ed4f008bf {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 749.149350] env[63515]: INFO nova.compute.manager [req-f1bf6141-436a-484f-b066-984dda728a23 req-2fad47e0-675a-4928-9013-7d1d1524584a service nova] [instance: de210780-5c0f-4fba-883c-13707566a2e1] Neutron deleted interface cd4f3cb8-f205-4ace-a17f-d54ed4f008bf; detaching it from the instance and deleting it from the info cache [ 749.149350] env[63515]: DEBUG nova.network.neutron [req-f1bf6141-436a-484f-b066-984dda728a23 req-2fad47e0-675a-4928-9013-7d1d1524584a service nova] [instance: de210780-5c0f-4fba-883c-13707566a2e1] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 749.193906] env[63515]: DEBUG nova.network.neutron [-] [instance: de210780-5c0f-4fba-883c-13707566a2e1] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 749.384610] env[63515]: DEBUG oslo_concurrency.lockutils [None req-5bd41396-218d-4ce7-9765-053dcc695301 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Releasing lock "refresh_cache-6688cdaa-29ba-413a-8131-4f834cdb70e4" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 749.384979] env[63515]: DEBUG nova.compute.manager [None req-5bd41396-218d-4ce7-9765-053dcc695301 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] [instance: 6688cdaa-29ba-413a-8131-4f834cdb70e4] Instance network_info: |[{"id": "6ccc2ab3-1e4f-484b-984b-5e0ebb251bd4", "address": "fa:16:3e:e0:db:2c", "network": {"id": "d2aef558-9770-46ad-b51c-61bccb0de7ec", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-2087172546-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6f501ed813754f759464e338884993c1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6d4ef133-b6f3-41d1-add4-92a1482195cf", "external-id": "nsx-vlan-transportzone-446", "segmentation_id": 446, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6ccc2ab3-1e", "ovs_interfaceid": "6ccc2ab3-1e4f-484b-984b-5e0ebb251bd4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 749.385482] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-5bd41396-218d-4ce7-9765-053dcc695301 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] [instance: 6688cdaa-29ba-413a-8131-4f834cdb70e4] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:e0:db:2c', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '6d4ef133-b6f3-41d1-add4-92a1482195cf', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '6ccc2ab3-1e4f-484b-984b-5e0ebb251bd4', 'vif_model': 'vmxnet3'}] {{(pid=63515) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 749.396163] env[63515]: DEBUG oslo.service.loopingcall [None req-5bd41396-218d-4ce7-9765-053dcc695301 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 749.396163] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6688cdaa-29ba-413a-8131-4f834cdb70e4] Creating VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 749.396378] env[63515]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-dc155321-ecb9-4aac-9d6c-efd4cb639da3 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.418715] env[63515]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 749.418715] env[63515]: value = "task-1111079" [ 749.418715] env[63515]: _type = "Task" [ 749.418715] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 749.427242] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1111079, 'name': CreateVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 749.481620] env[63515]: DEBUG nova.compute.manager [None req-ffbb9c80-abfa-457c-b756-9497c4635a40 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] [instance: f1d01b75-ac9d-458d-8cc2-ae64cffca4e8] Hypervisor driver does not support instance shared storage check, assuming it's not on shared storage {{(pid=63515) _is_instance_storage_shared /opt/stack/nova/nova/compute/manager.py:897}} [ 749.481963] env[63515]: DEBUG oslo_concurrency.lockutils [None req-ffbb9c80-abfa-457c-b756-9497c4635a40 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 749.489222] env[63515]: DEBUG oslo_vmware.api [None req-99ae4d48-1fae-4ace-913c-d249ca72bbeb tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Task: {'id': task-1111078, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 749.554790] env[63515]: DEBUG nova.compute.manager [None req-dfa187f2-adf0-4e24-9cf4-7df6d0016e12 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] [instance: 38d6d030-06b0-4185-904d-44a038b3a752] Start spawning the instance on the hypervisor. {{(pid=63515) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 749.578467] env[63515]: DEBUG nova.virt.hardware [None req-dfa187f2-adf0-4e24-9cf4-7df6d0016e12 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T02:52:16Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T02:51:56Z,direct_url=,disk_format='vmdk',id=8a120570-cb06-4099-b262-554ca0ad15c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b569255cc43e42e7ae3f2b4ad37c6ef4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T02:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 749.578697] env[63515]: DEBUG nova.virt.hardware [None req-dfa187f2-adf0-4e24-9cf4-7df6d0016e12 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Flavor limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 749.578888] env[63515]: DEBUG nova.virt.hardware [None req-dfa187f2-adf0-4e24-9cf4-7df6d0016e12 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Image limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 749.579141] env[63515]: DEBUG nova.virt.hardware [None req-dfa187f2-adf0-4e24-9cf4-7df6d0016e12 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Flavor pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 749.579303] env[63515]: DEBUG nova.virt.hardware [None req-dfa187f2-adf0-4e24-9cf4-7df6d0016e12 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Image pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 749.579449] env[63515]: DEBUG nova.virt.hardware [None req-dfa187f2-adf0-4e24-9cf4-7df6d0016e12 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 749.579699] env[63515]: DEBUG nova.virt.hardware [None req-dfa187f2-adf0-4e24-9cf4-7df6d0016e12 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 749.579803] env[63515]: DEBUG nova.virt.hardware [None req-dfa187f2-adf0-4e24-9cf4-7df6d0016e12 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 749.580131] env[63515]: DEBUG nova.virt.hardware [None req-dfa187f2-adf0-4e24-9cf4-7df6d0016e12 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Got 1 possible topologies {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 749.580216] env[63515]: DEBUG nova.virt.hardware [None req-dfa187f2-adf0-4e24-9cf4-7df6d0016e12 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 749.580309] env[63515]: DEBUG nova.virt.hardware [None req-dfa187f2-adf0-4e24-9cf4-7df6d0016e12 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 749.581228] env[63515]: DEBUG nova.scheduler.client.report [None req-750d4038-2a95-4b14-b9d7-c9c6803c64b6 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 749.585149] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c52adb04-3a60-4917-a5e9-6ece4579632d {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.597422] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31608ffd-69ae-48d9-a419-688e5ea77ad0 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.613014] env[63515]: DEBUG oslo_concurrency.lockutils [None req-83fbd51b-fc37-49eb-8423-cdaa019d3b0c tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 749.652402] env[63515]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-0d24ba40-b60e-48b9-9007-505447495a34 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.663018] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c7217d1-1e3e-4dd7-b839-626a0215fc4e {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.699027] env[63515]: INFO nova.compute.manager [-] [instance: de210780-5c0f-4fba-883c-13707566a2e1] Took 1.52 seconds to deallocate network for instance. [ 749.699451] env[63515]: DEBUG nova.compute.manager [req-f1bf6141-436a-484f-b066-984dda728a23 req-2fad47e0-675a-4928-9013-7d1d1524584a service nova] [instance: de210780-5c0f-4fba-883c-13707566a2e1] Detach interface failed, port_id=cd4f3cb8-f205-4ace-a17f-d54ed4f008bf, reason: Instance de210780-5c0f-4fba-883c-13707566a2e1 could not be found. {{(pid=63515) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 749.703131] env[63515]: DEBUG nova.compute.manager [req-3a1680d3-877c-4a88-a414-564445271df9 req-c30453f9-4a58-4daa-a933-3f51ba81af45 service nova] [instance: 6688cdaa-29ba-413a-8131-4f834cdb70e4] Received event network-changed-6ccc2ab3-1e4f-484b-984b-5e0ebb251bd4 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 749.704041] env[63515]: DEBUG nova.compute.manager [req-3a1680d3-877c-4a88-a414-564445271df9 req-c30453f9-4a58-4daa-a933-3f51ba81af45 service nova] [instance: 6688cdaa-29ba-413a-8131-4f834cdb70e4] Refreshing instance network info cache due to event network-changed-6ccc2ab3-1e4f-484b-984b-5e0ebb251bd4. {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 749.704041] env[63515]: DEBUG oslo_concurrency.lockutils [req-3a1680d3-877c-4a88-a414-564445271df9 req-c30453f9-4a58-4daa-a933-3f51ba81af45 service nova] Acquiring lock "refresh_cache-6688cdaa-29ba-413a-8131-4f834cdb70e4" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 749.704041] env[63515]: DEBUG oslo_concurrency.lockutils [req-3a1680d3-877c-4a88-a414-564445271df9 req-c30453f9-4a58-4daa-a933-3f51ba81af45 service nova] Acquired lock "refresh_cache-6688cdaa-29ba-413a-8131-4f834cdb70e4" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 749.704041] env[63515]: DEBUG nova.network.neutron [req-3a1680d3-877c-4a88-a414-564445271df9 req-c30453f9-4a58-4daa-a933-3f51ba81af45 service nova] [instance: 6688cdaa-29ba-413a-8131-4f834cdb70e4] Refreshing network info cache for port 6ccc2ab3-1e4f-484b-984b-5e0ebb251bd4 {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 750.518635] env[63515]: DEBUG nova.network.neutron [None req-dfa187f2-adf0-4e24-9cf4-7df6d0016e12 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] [instance: 38d6d030-06b0-4185-904d-44a038b3a752] Successfully updated port: 2242e364-37c0-4750-8755-d807d47e899d {{(pid=63515) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 750.520243] env[63515]: INFO nova.compute.manager [None req-0e5835c9-1a06-4538-91e4-48a9dfc2854d tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: 001a48e1-b237-41f8-a6f4-2fe98d173512] Rebuilding instance [ 750.522518] env[63515]: DEBUG oslo_concurrency.lockutils [None req-750d4038-2a95-4b14-b9d7-c9c6803c64b6 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.982s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 750.523079] env[63515]: DEBUG nova.compute.manager [None req-750d4038-2a95-4b14-b9d7-c9c6803c64b6 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] [instance: 8eb17506-ff93-4d25-b9af-ec5886569e65] Start building networks asynchronously for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 750.528186] env[63515]: DEBUG oslo_concurrency.lockutils [None req-711301e3-5f69-4e98-bb19-e93c4ab2a0db tempest-ServerTagsTestJSON-1170819896 tempest-ServerTagsTestJSON-1170819896-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 750.533847] env[63515]: DEBUG oslo_concurrency.lockutils [None req-9695633a-0f62-476f-8466-deb3365d2b47 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 24.023s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 750.534111] env[63515]: DEBUG nova.objects.instance [None req-9695633a-0f62-476f-8466-deb3365d2b47 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Lazy-loading 'resources' on Instance uuid 48668736-df27-4f2a-94d9-132f5b49701b {{(pid=63515) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 750.544298] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1111079, 'name': CreateVM_Task, 'duration_secs': 0.453281} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 750.544517] env[63515]: DEBUG oslo_vmware.api [None req-99ae4d48-1fae-4ace-913c-d249ca72bbeb tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Task: {'id': task-1111078, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.589604} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 750.544657] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6688cdaa-29ba-413a-8131-4f834cdb70e4] Created VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 750.544899] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-99ae4d48-1fae-4ace-913c-d249ca72bbeb tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk to [datastore1] d2a4f66a-f1a2-4f0d-826d-0b7c4f04bd23/d2a4f66a-f1a2-4f0d-826d-0b7c4f04bd23.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 750.546069] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-99ae4d48-1fae-4ace-913c-d249ca72bbeb tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] [instance: d2a4f66a-f1a2-4f0d-826d-0b7c4f04bd23] Extending root virtual disk to 1048576 {{(pid=63515) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 750.546069] env[63515]: DEBUG oslo_concurrency.lockutils [None req-5bd41396-218d-4ce7-9765-053dcc695301 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 750.546069] env[63515]: DEBUG oslo_concurrency.lockutils [None req-5bd41396-218d-4ce7-9765-053dcc695301 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 750.546268] env[63515]: DEBUG oslo_concurrency.lockutils [None req-5bd41396-218d-4ce7-9765-053dcc695301 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 750.546726] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-f0b30c73-58ca-4ba6-a24f-1f64c42a8ce6 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.548752] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-79e3026e-a9d6-42ff-a3b7-b45fffc45051 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.553992] env[63515]: DEBUG oslo_vmware.api [None req-5bd41396-218d-4ce7-9765-053dcc695301 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Waiting for the task: (returnval){ [ 750.553992] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]52efcb73-f195-474c-6551-c74cc8a2a79e" [ 750.553992] env[63515]: _type = "Task" [ 750.553992] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 750.558357] env[63515]: DEBUG oslo_vmware.api [None req-99ae4d48-1fae-4ace-913c-d249ca72bbeb tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Waiting for the task: (returnval){ [ 750.558357] env[63515]: value = "task-1111080" [ 750.558357] env[63515]: _type = "Task" [ 750.558357] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 750.568767] env[63515]: DEBUG oslo_vmware.api [None req-5bd41396-218d-4ce7-9765-053dcc695301 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52efcb73-f195-474c-6551-c74cc8a2a79e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 750.573715] env[63515]: DEBUG oslo_vmware.api [None req-99ae4d48-1fae-4ace-913c-d249ca72bbeb tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Task: {'id': task-1111080, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 750.587703] env[63515]: DEBUG nova.compute.manager [None req-0e5835c9-1a06-4538-91e4-48a9dfc2854d tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: 001a48e1-b237-41f8-a6f4-2fe98d173512] Checking state {{(pid=63515) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 750.588584] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3566032-f0b8-4128-a1ce-f7b68776079f {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.845254] env[63515]: DEBUG nova.network.neutron [req-3a1680d3-877c-4a88-a414-564445271df9 req-c30453f9-4a58-4daa-a933-3f51ba81af45 service nova] [instance: 6688cdaa-29ba-413a-8131-4f834cdb70e4] Updated VIF entry in instance network info cache for port 6ccc2ab3-1e4f-484b-984b-5e0ebb251bd4. {{(pid=63515) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 750.845254] env[63515]: DEBUG nova.network.neutron [req-3a1680d3-877c-4a88-a414-564445271df9 req-c30453f9-4a58-4daa-a933-3f51ba81af45 service nova] [instance: 6688cdaa-29ba-413a-8131-4f834cdb70e4] Updating instance_info_cache with network_info: [{"id": "6ccc2ab3-1e4f-484b-984b-5e0ebb251bd4", "address": "fa:16:3e:e0:db:2c", "network": {"id": "d2aef558-9770-46ad-b51c-61bccb0de7ec", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-2087172546-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6f501ed813754f759464e338884993c1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6d4ef133-b6f3-41d1-add4-92a1482195cf", "external-id": "nsx-vlan-transportzone-446", "segmentation_id": 446, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6ccc2ab3-1e", "ovs_interfaceid": "6ccc2ab3-1e4f-484b-984b-5e0ebb251bd4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 751.029874] env[63515]: DEBUG nova.compute.utils [None req-750d4038-2a95-4b14-b9d7-c9c6803c64b6 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] Using /dev/sd instead of None {{(pid=63515) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 751.032079] env[63515]: DEBUG oslo_concurrency.lockutils [None req-dfa187f2-adf0-4e24-9cf4-7df6d0016e12 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Acquiring lock "refresh_cache-38d6d030-06b0-4185-904d-44a038b3a752" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 751.033011] env[63515]: DEBUG oslo_concurrency.lockutils [None req-dfa187f2-adf0-4e24-9cf4-7df6d0016e12 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Acquired lock "refresh_cache-38d6d030-06b0-4185-904d-44a038b3a752" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 751.033011] env[63515]: DEBUG nova.network.neutron [None req-dfa187f2-adf0-4e24-9cf4-7df6d0016e12 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] [instance: 38d6d030-06b0-4185-904d-44a038b3a752] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 751.033719] env[63515]: DEBUG nova.compute.manager [None req-750d4038-2a95-4b14-b9d7-c9c6803c64b6 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] [instance: 8eb17506-ff93-4d25-b9af-ec5886569e65] Not allocating networking since 'none' was specified. {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 751.068786] env[63515]: DEBUG oslo_vmware.api [None req-5bd41396-218d-4ce7-9765-053dcc695301 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52efcb73-f195-474c-6551-c74cc8a2a79e, 'name': SearchDatastore_Task, 'duration_secs': 0.025273} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 751.071576] env[63515]: DEBUG oslo_concurrency.lockutils [None req-5bd41396-218d-4ce7-9765-053dcc695301 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 751.071818] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-5bd41396-218d-4ce7-9765-053dcc695301 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] [instance: 6688cdaa-29ba-413a-8131-4f834cdb70e4] Processing image 8a120570-cb06-4099-b262-554ca0ad15c5 {{(pid=63515) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 751.072069] env[63515]: DEBUG oslo_concurrency.lockutils [None req-5bd41396-218d-4ce7-9765-053dcc695301 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 751.072411] env[63515]: DEBUG oslo_concurrency.lockutils [None req-5bd41396-218d-4ce7-9765-053dcc695301 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 751.072411] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-5bd41396-218d-4ce7-9765-053dcc695301 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 751.072838] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f5f9e8cc-f064-4e07-854d-e69353b3532c {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.077450] env[63515]: DEBUG oslo_vmware.api [None req-99ae4d48-1fae-4ace-913c-d249ca72bbeb tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Task: {'id': task-1111080, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.076017} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 751.078021] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-99ae4d48-1fae-4ace-913c-d249ca72bbeb tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] [instance: d2a4f66a-f1a2-4f0d-826d-0b7c4f04bd23] Extended root virtual disk {{(pid=63515) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 751.078905] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4756585-1428-469b-86af-3c31869b15a2 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.086061] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-5bd41396-218d-4ce7-9765-053dcc695301 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 751.086237] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-5bd41396-218d-4ce7-9765-053dcc695301 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63515) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 751.086993] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-31e1a374-3f7a-4f19-9533-9833fb3aa2ee {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.100504] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-0e5835c9-1a06-4538-91e4-48a9dfc2854d tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: 001a48e1-b237-41f8-a6f4-2fe98d173512] Powering off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 751.109759] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-99ae4d48-1fae-4ace-913c-d249ca72bbeb tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] [instance: d2a4f66a-f1a2-4f0d-826d-0b7c4f04bd23] Reconfiguring VM instance instance-00000031 to attach disk [datastore1] d2a4f66a-f1a2-4f0d-826d-0b7c4f04bd23/d2a4f66a-f1a2-4f0d-826d-0b7c4f04bd23.vmdk or device None with type sparse {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 751.112791] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7b521e78-25dc-4fe9-91ae-ba163f35367d {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.114608] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-93c33e31-8cdf-43a4-9105-3bcf31f079e2 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.130014] env[63515]: DEBUG oslo_vmware.api [None req-5bd41396-218d-4ce7-9765-053dcc695301 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Waiting for the task: (returnval){ [ 751.130014] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]52065cbb-0eff-c3f1-ad23-f621d3393936" [ 751.130014] env[63515]: _type = "Task" [ 751.130014] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 751.135372] env[63515]: DEBUG oslo_vmware.api [None req-0e5835c9-1a06-4538-91e4-48a9dfc2854d tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Waiting for the task: (returnval){ [ 751.135372] env[63515]: value = "task-1111081" [ 751.135372] env[63515]: _type = "Task" [ 751.135372] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 751.139318] env[63515]: DEBUG oslo_vmware.api [None req-99ae4d48-1fae-4ace-913c-d249ca72bbeb tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Waiting for the task: (returnval){ [ 751.139318] env[63515]: value = "task-1111082" [ 751.139318] env[63515]: _type = "Task" [ 751.139318] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 751.139533] env[63515]: DEBUG oslo_vmware.api [None req-5bd41396-218d-4ce7-9765-053dcc695301 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52065cbb-0eff-c3f1-ad23-f621d3393936, 'name': SearchDatastore_Task, 'duration_secs': 0.010555} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 751.148838] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-084ddfd6-2102-4c2d-8977-3d630f8a2a3c {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.160022] env[63515]: DEBUG oslo_vmware.api [None req-99ae4d48-1fae-4ace-913c-d249ca72bbeb tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Task: {'id': task-1111082, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 751.160477] env[63515]: DEBUG oslo_vmware.api [None req-0e5835c9-1a06-4538-91e4-48a9dfc2854d tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Task: {'id': task-1111081, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 751.160730] env[63515]: DEBUG oslo_vmware.api [None req-5bd41396-218d-4ce7-9765-053dcc695301 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Waiting for the task: (returnval){ [ 751.160730] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]52d95952-b00f-66ae-fed1-31c45ea4000f" [ 751.160730] env[63515]: _type = "Task" [ 751.160730] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 751.170691] env[63515]: DEBUG oslo_vmware.api [None req-5bd41396-218d-4ce7-9765-053dcc695301 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52d95952-b00f-66ae-fed1-31c45ea4000f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 751.179462] env[63515]: DEBUG nova.compute.manager [req-231a3298-d6aa-4848-bca0-d930d8df5447 req-5f760e0d-5cd9-4673-93d9-ebaa0a57541b service nova] [instance: 38d6d030-06b0-4185-904d-44a038b3a752] Received event network-vif-plugged-2242e364-37c0-4750-8755-d807d47e899d {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 751.179686] env[63515]: DEBUG oslo_concurrency.lockutils [req-231a3298-d6aa-4848-bca0-d930d8df5447 req-5f760e0d-5cd9-4673-93d9-ebaa0a57541b service nova] Acquiring lock "38d6d030-06b0-4185-904d-44a038b3a752-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 751.179960] env[63515]: DEBUG oslo_concurrency.lockutils [req-231a3298-d6aa-4848-bca0-d930d8df5447 req-5f760e0d-5cd9-4673-93d9-ebaa0a57541b service nova] Lock "38d6d030-06b0-4185-904d-44a038b3a752-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 751.180069] env[63515]: DEBUG oslo_concurrency.lockutils [req-231a3298-d6aa-4848-bca0-d930d8df5447 req-5f760e0d-5cd9-4673-93d9-ebaa0a57541b service nova] Lock "38d6d030-06b0-4185-904d-44a038b3a752-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 751.180264] env[63515]: DEBUG nova.compute.manager [req-231a3298-d6aa-4848-bca0-d930d8df5447 req-5f760e0d-5cd9-4673-93d9-ebaa0a57541b service nova] [instance: 38d6d030-06b0-4185-904d-44a038b3a752] No waiting events found dispatching network-vif-plugged-2242e364-37c0-4750-8755-d807d47e899d {{(pid=63515) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 751.180443] env[63515]: WARNING nova.compute.manager [req-231a3298-d6aa-4848-bca0-d930d8df5447 req-5f760e0d-5cd9-4673-93d9-ebaa0a57541b service nova] [instance: 38d6d030-06b0-4185-904d-44a038b3a752] Received unexpected event network-vif-plugged-2242e364-37c0-4750-8755-d807d47e899d for instance with vm_state building and task_state spawning. [ 751.180604] env[63515]: DEBUG nova.compute.manager [req-231a3298-d6aa-4848-bca0-d930d8df5447 req-5f760e0d-5cd9-4673-93d9-ebaa0a57541b service nova] [instance: 38d6d030-06b0-4185-904d-44a038b3a752] Received event network-changed-2242e364-37c0-4750-8755-d807d47e899d {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 751.180755] env[63515]: DEBUG nova.compute.manager [req-231a3298-d6aa-4848-bca0-d930d8df5447 req-5f760e0d-5cd9-4673-93d9-ebaa0a57541b service nova] [instance: 38d6d030-06b0-4185-904d-44a038b3a752] Refreshing instance network info cache due to event network-changed-2242e364-37c0-4750-8755-d807d47e899d. {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 751.180922] env[63515]: DEBUG oslo_concurrency.lockutils [req-231a3298-d6aa-4848-bca0-d930d8df5447 req-5f760e0d-5cd9-4673-93d9-ebaa0a57541b service nova] Acquiring lock "refresh_cache-38d6d030-06b0-4185-904d-44a038b3a752" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 751.349636] env[63515]: DEBUG oslo_concurrency.lockutils [req-3a1680d3-877c-4a88-a414-564445271df9 req-c30453f9-4a58-4daa-a933-3f51ba81af45 service nova] Releasing lock "refresh_cache-6688cdaa-29ba-413a-8131-4f834cdb70e4" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 751.517995] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f7bc68b-61d5-4c8f-8d77-f80bd532d10e {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.524801] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d4e7ac8-ff11-46d9-abb3-d31d32745b53 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.552427] env[63515]: DEBUG nova.compute.manager [None req-750d4038-2a95-4b14-b9d7-c9c6803c64b6 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] [instance: 8eb17506-ff93-4d25-b9af-ec5886569e65] Start building block device mappings for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 751.557401] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10d32b28-336b-4e55-9b54-258b7cf7329d {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.564784] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2d8a500-f017-4229-baea-add6bdd0e0a5 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.578951] env[63515]: DEBUG nova.compute.provider_tree [None req-9695633a-0f62-476f-8466-deb3365d2b47 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Updating inventory in ProviderTree for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 751.586881] env[63515]: DEBUG nova.network.neutron [None req-dfa187f2-adf0-4e24-9cf4-7df6d0016e12 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] [instance: 38d6d030-06b0-4185-904d-44a038b3a752] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 751.649854] env[63515]: DEBUG oslo_vmware.api [None req-0e5835c9-1a06-4538-91e4-48a9dfc2854d tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Task: {'id': task-1111081, 'name': PowerOffVM_Task, 'duration_secs': 0.20787} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 751.650661] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-0e5835c9-1a06-4538-91e4-48a9dfc2854d tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: 001a48e1-b237-41f8-a6f4-2fe98d173512] Powered off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 751.650661] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-0e5835c9-1a06-4538-91e4-48a9dfc2854d tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: 001a48e1-b237-41f8-a6f4-2fe98d173512] Destroying instance {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 751.651378] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec6bf4e0-7edf-48f1-a850-5760aecb838b {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.658873] env[63515]: DEBUG oslo_vmware.api [None req-99ae4d48-1fae-4ace-913c-d249ca72bbeb tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Task: {'id': task-1111082, 'name': ReconfigVM_Task, 'duration_secs': 0.330576} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 751.659468] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-99ae4d48-1fae-4ace-913c-d249ca72bbeb tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] [instance: d2a4f66a-f1a2-4f0d-826d-0b7c4f04bd23] Reconfigured VM instance instance-00000031 to attach disk [datastore1] d2a4f66a-f1a2-4f0d-826d-0b7c4f04bd23/d2a4f66a-f1a2-4f0d-826d-0b7c4f04bd23.vmdk or device None with type sparse {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 751.660311] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-7a2de0c0-a62f-4fb5-9d8f-229c709025fa {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.663558] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-0e5835c9-1a06-4538-91e4-48a9dfc2854d tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: 001a48e1-b237-41f8-a6f4-2fe98d173512] Unregistering the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 751.666480] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-6553e7b9-a94e-4dea-85c4-8c5cdd2ef93d {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.668904] env[63515]: DEBUG oslo_vmware.api [None req-99ae4d48-1fae-4ace-913c-d249ca72bbeb tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Waiting for the task: (returnval){ [ 751.668904] env[63515]: value = "task-1111083" [ 751.668904] env[63515]: _type = "Task" [ 751.668904] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 751.675043] env[63515]: DEBUG oslo_vmware.api [None req-5bd41396-218d-4ce7-9765-053dcc695301 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52d95952-b00f-66ae-fed1-31c45ea4000f, 'name': SearchDatastore_Task, 'duration_secs': 0.018972} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 751.675578] env[63515]: DEBUG oslo_concurrency.lockutils [None req-5bd41396-218d-4ce7-9765-053dcc695301 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 751.675836] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-5bd41396-218d-4ce7-9765-053dcc695301 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk to [datastore1] 6688cdaa-29ba-413a-8131-4f834cdb70e4/6688cdaa-29ba-413a-8131-4f834cdb70e4.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 751.676105] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7d5f1e67-ffd8-4a9e-998e-5cd80565189f {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.680694] env[63515]: DEBUG oslo_vmware.api [None req-99ae4d48-1fae-4ace-913c-d249ca72bbeb tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Task: {'id': task-1111083, 'name': Rename_Task} progress is 5%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 751.684632] env[63515]: DEBUG oslo_vmware.api [None req-5bd41396-218d-4ce7-9765-053dcc695301 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Waiting for the task: (returnval){ [ 751.684632] env[63515]: value = "task-1111085" [ 751.684632] env[63515]: _type = "Task" [ 751.684632] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 751.691754] env[63515]: DEBUG oslo_vmware.api [None req-5bd41396-218d-4ce7-9765-053dcc695301 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Task: {'id': task-1111085, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 751.728372] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-0e5835c9-1a06-4538-91e4-48a9dfc2854d tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: 001a48e1-b237-41f8-a6f4-2fe98d173512] Unregistered the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 751.728575] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-0e5835c9-1a06-4538-91e4-48a9dfc2854d tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: 001a48e1-b237-41f8-a6f4-2fe98d173512] Deleting contents of the VM from datastore datastore2 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 751.728757] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-0e5835c9-1a06-4538-91e4-48a9dfc2854d tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Deleting the datastore file [datastore2] 001a48e1-b237-41f8-a6f4-2fe98d173512 {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 751.729108] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7372c193-c97d-49eb-b4f3-e1cccf5556ac {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.735396] env[63515]: DEBUG oslo_vmware.api [None req-0e5835c9-1a06-4538-91e4-48a9dfc2854d tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Waiting for the task: (returnval){ [ 751.735396] env[63515]: value = "task-1111086" [ 751.735396] env[63515]: _type = "Task" [ 751.735396] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 751.736298] env[63515]: DEBUG nova.network.neutron [None req-dfa187f2-adf0-4e24-9cf4-7df6d0016e12 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] [instance: 38d6d030-06b0-4185-904d-44a038b3a752] Updating instance_info_cache with network_info: [{"id": "2242e364-37c0-4750-8755-d807d47e899d", "address": "fa:16:3e:d4:3c:7a", "network": {"id": "d2aef558-9770-46ad-b51c-61bccb0de7ec", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-2087172546-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6f501ed813754f759464e338884993c1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6d4ef133-b6f3-41d1-add4-92a1482195cf", "external-id": "nsx-vlan-transportzone-446", "segmentation_id": 446, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2242e364-37", "ovs_interfaceid": "2242e364-37c0-4750-8755-d807d47e899d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 751.745490] env[63515]: DEBUG oslo_vmware.api [None req-0e5835c9-1a06-4538-91e4-48a9dfc2854d tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Task: {'id': task-1111086, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 752.120181] env[63515]: DEBUG nova.scheduler.client.report [None req-9695633a-0f62-476f-8466-deb3365d2b47 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Updated inventory for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a with generation 73 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 752.120495] env[63515]: DEBUG nova.compute.provider_tree [None req-9695633a-0f62-476f-8466-deb3365d2b47 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Updating resource provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a generation from 73 to 74 during operation: update_inventory {{(pid=63515) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 752.120677] env[63515]: DEBUG nova.compute.provider_tree [None req-9695633a-0f62-476f-8466-deb3365d2b47 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Updating inventory in ProviderTree for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 752.181556] env[63515]: DEBUG oslo_vmware.api [None req-99ae4d48-1fae-4ace-913c-d249ca72bbeb tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Task: {'id': task-1111083, 'name': Rename_Task, 'duration_secs': 0.255492} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 752.181875] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-99ae4d48-1fae-4ace-913c-d249ca72bbeb tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] [instance: d2a4f66a-f1a2-4f0d-826d-0b7c4f04bd23] Powering on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 752.182825] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-27a999a6-bbb6-40b1-8e24-357dcadda70d {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.191352] env[63515]: DEBUG oslo_vmware.api [None req-99ae4d48-1fae-4ace-913c-d249ca72bbeb tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Waiting for the task: (returnval){ [ 752.191352] env[63515]: value = "task-1111087" [ 752.191352] env[63515]: _type = "Task" [ 752.191352] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 752.198464] env[63515]: DEBUG oslo_vmware.api [None req-5bd41396-218d-4ce7-9765-053dcc695301 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Task: {'id': task-1111085, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 752.203554] env[63515]: DEBUG oslo_vmware.api [None req-99ae4d48-1fae-4ace-913c-d249ca72bbeb tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Task: {'id': task-1111087, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 752.242346] env[63515]: DEBUG oslo_concurrency.lockutils [None req-dfa187f2-adf0-4e24-9cf4-7df6d0016e12 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Releasing lock "refresh_cache-38d6d030-06b0-4185-904d-44a038b3a752" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 752.242673] env[63515]: DEBUG nova.compute.manager [None req-dfa187f2-adf0-4e24-9cf4-7df6d0016e12 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] [instance: 38d6d030-06b0-4185-904d-44a038b3a752] Instance network_info: |[{"id": "2242e364-37c0-4750-8755-d807d47e899d", "address": "fa:16:3e:d4:3c:7a", "network": {"id": "d2aef558-9770-46ad-b51c-61bccb0de7ec", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-2087172546-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6f501ed813754f759464e338884993c1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6d4ef133-b6f3-41d1-add4-92a1482195cf", "external-id": "nsx-vlan-transportzone-446", "segmentation_id": 446, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2242e364-37", "ovs_interfaceid": "2242e364-37c0-4750-8755-d807d47e899d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 752.243091] env[63515]: DEBUG oslo_concurrency.lockutils [req-231a3298-d6aa-4848-bca0-d930d8df5447 req-5f760e0d-5cd9-4673-93d9-ebaa0a57541b service nova] Acquired lock "refresh_cache-38d6d030-06b0-4185-904d-44a038b3a752" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 752.243287] env[63515]: DEBUG nova.network.neutron [req-231a3298-d6aa-4848-bca0-d930d8df5447 req-5f760e0d-5cd9-4673-93d9-ebaa0a57541b service nova] [instance: 38d6d030-06b0-4185-904d-44a038b3a752] Refreshing network info cache for port 2242e364-37c0-4750-8755-d807d47e899d {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 752.244629] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-dfa187f2-adf0-4e24-9cf4-7df6d0016e12 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] [instance: 38d6d030-06b0-4185-904d-44a038b3a752] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:d4:3c:7a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '6d4ef133-b6f3-41d1-add4-92a1482195cf', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '2242e364-37c0-4750-8755-d807d47e899d', 'vif_model': 'vmxnet3'}] {{(pid=63515) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 752.253093] env[63515]: DEBUG oslo.service.loopingcall [None req-dfa187f2-adf0-4e24-9cf4-7df6d0016e12 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 752.253980] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 38d6d030-06b0-4185-904d-44a038b3a752] Creating VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 752.254238] env[63515]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f77a108f-ef53-4ada-b9a6-734db4e88322 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.272571] env[63515]: DEBUG oslo_vmware.api [None req-0e5835c9-1a06-4538-91e4-48a9dfc2854d tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Task: {'id': task-1111086, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.230627} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 752.273261] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-0e5835c9-1a06-4538-91e4-48a9dfc2854d tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Deleted the datastore file {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 752.273485] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-0e5835c9-1a06-4538-91e4-48a9dfc2854d tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: 001a48e1-b237-41f8-a6f4-2fe98d173512] Deleted contents of the VM from datastore datastore2 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 752.273662] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-0e5835c9-1a06-4538-91e4-48a9dfc2854d tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: 001a48e1-b237-41f8-a6f4-2fe98d173512] Instance destroyed {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 752.277421] env[63515]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 752.277421] env[63515]: value = "task-1111088" [ 752.277421] env[63515]: _type = "Task" [ 752.277421] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 752.287250] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1111088, 'name': CreateVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 752.564258] env[63515]: DEBUG nova.compute.manager [None req-750d4038-2a95-4b14-b9d7-c9c6803c64b6 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] [instance: 8eb17506-ff93-4d25-b9af-ec5886569e65] Start spawning the instance on the hypervisor. {{(pid=63515) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 752.590799] env[63515]: DEBUG nova.virt.hardware [None req-750d4038-2a95-4b14-b9d7-c9c6803c64b6 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T02:52:14Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T02:51:56Z,direct_url=,disk_format='vmdk',id=8a120570-cb06-4099-b262-554ca0ad15c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b569255cc43e42e7ae3f2b4ad37c6ef4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T02:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 752.591178] env[63515]: DEBUG nova.virt.hardware [None req-750d4038-2a95-4b14-b9d7-c9c6803c64b6 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] Flavor limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 752.591362] env[63515]: DEBUG nova.virt.hardware [None req-750d4038-2a95-4b14-b9d7-c9c6803c64b6 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] Image limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 752.591586] env[63515]: DEBUG nova.virt.hardware [None req-750d4038-2a95-4b14-b9d7-c9c6803c64b6 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] Flavor pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 752.591760] env[63515]: DEBUG nova.virt.hardware [None req-750d4038-2a95-4b14-b9d7-c9c6803c64b6 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] Image pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 752.591948] env[63515]: DEBUG nova.virt.hardware [None req-750d4038-2a95-4b14-b9d7-c9c6803c64b6 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 752.592253] env[63515]: DEBUG nova.virt.hardware [None req-750d4038-2a95-4b14-b9d7-c9c6803c64b6 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 752.592442] env[63515]: DEBUG nova.virt.hardware [None req-750d4038-2a95-4b14-b9d7-c9c6803c64b6 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 752.592664] env[63515]: DEBUG nova.virt.hardware [None req-750d4038-2a95-4b14-b9d7-c9c6803c64b6 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] Got 1 possible topologies {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 752.592852] env[63515]: DEBUG nova.virt.hardware [None req-750d4038-2a95-4b14-b9d7-c9c6803c64b6 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 752.593063] env[63515]: DEBUG nova.virt.hardware [None req-750d4038-2a95-4b14-b9d7-c9c6803c64b6 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 752.594059] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d525d09-417f-4687-9faa-2cc56495814f {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.603078] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95108040-200e-4f17-92f7-3a5c19712fee {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.617131] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-750d4038-2a95-4b14-b9d7-c9c6803c64b6 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] [instance: 8eb17506-ff93-4d25-b9af-ec5886569e65] Instance VIF info [] {{(pid=63515) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 752.622421] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-750d4038-2a95-4b14-b9d7-c9c6803c64b6 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] Creating folder: Project (a3c63227e9174c7599ce0940759544d8). Parent ref: group-v243370. {{(pid=63515) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 752.622688] env[63515]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-c35236b4-b0d2-431f-a7f0-08c449e4dbbb {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.627307] env[63515]: DEBUG oslo_concurrency.lockutils [None req-9695633a-0f62-476f-8466-deb3365d2b47 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.093s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 752.630230] env[63515]: DEBUG oslo_concurrency.lockutils [None req-b8b2a7f6-808d-40d4-8de9-c8863e5b55c5 tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 22.765s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 752.631685] env[63515]: INFO nova.compute.claims [None req-b8b2a7f6-808d-40d4-8de9-c8863e5b55c5 tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] [instance: 30b8b872-40ca-4297-b98e-a64c83a75483] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 752.635560] env[63515]: INFO nova.virt.vmwareapi.vm_util [None req-750d4038-2a95-4b14-b9d7-c9c6803c64b6 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] Created folder: Project (a3c63227e9174c7599ce0940759544d8) in parent group-v243370. [ 752.635785] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-750d4038-2a95-4b14-b9d7-c9c6803c64b6 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] Creating folder: Instances. Parent ref: group-v243436. {{(pid=63515) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 752.636276] env[63515]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-4d7242bb-7a4d-44df-aef5-9c7715cf404d {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.645256] env[63515]: INFO nova.virt.vmwareapi.vm_util [None req-750d4038-2a95-4b14-b9d7-c9c6803c64b6 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] Created folder: Instances in parent group-v243436. [ 752.646332] env[63515]: DEBUG oslo.service.loopingcall [None req-750d4038-2a95-4b14-b9d7-c9c6803c64b6 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 752.646332] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8eb17506-ff93-4d25-b9af-ec5886569e65] Creating VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 752.646332] env[63515]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-9d261451-b923-407c-bde4-d5a998f9ee8c {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.658320] env[63515]: INFO nova.scheduler.client.report [None req-9695633a-0f62-476f-8466-deb3365d2b47 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Deleted allocations for instance 48668736-df27-4f2a-94d9-132f5b49701b [ 752.664399] env[63515]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 752.664399] env[63515]: value = "task-1111091" [ 752.664399] env[63515]: _type = "Task" [ 752.664399] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 752.672012] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1111091, 'name': CreateVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 752.698277] env[63515]: DEBUG oslo_vmware.api [None req-5bd41396-218d-4ce7-9765-053dcc695301 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Task: {'id': task-1111085, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.602238} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 752.698459] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-5bd41396-218d-4ce7-9765-053dcc695301 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk to [datastore1] 6688cdaa-29ba-413a-8131-4f834cdb70e4/6688cdaa-29ba-413a-8131-4f834cdb70e4.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 752.698673] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-5bd41396-218d-4ce7-9765-053dcc695301 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] [instance: 6688cdaa-29ba-413a-8131-4f834cdb70e4] Extending root virtual disk to 1048576 {{(pid=63515) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 752.699349] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-5af505cf-e822-4ed2-91cc-de8e878579cd {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.704494] env[63515]: DEBUG oslo_vmware.api [None req-99ae4d48-1fae-4ace-913c-d249ca72bbeb tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Task: {'id': task-1111087, 'name': PowerOnVM_Task} progress is 100%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 752.709830] env[63515]: DEBUG oslo_vmware.api [None req-5bd41396-218d-4ce7-9765-053dcc695301 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Waiting for the task: (returnval){ [ 752.709830] env[63515]: value = "task-1111092" [ 752.709830] env[63515]: _type = "Task" [ 752.709830] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 752.718135] env[63515]: DEBUG oslo_vmware.api [None req-5bd41396-218d-4ce7-9765-053dcc695301 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Task: {'id': task-1111092, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 752.793062] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1111088, 'name': CreateVM_Task, 'duration_secs': 0.335577} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 752.793348] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 38d6d030-06b0-4185-904d-44a038b3a752] Created VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 752.794092] env[63515]: DEBUG oslo_concurrency.lockutils [None req-dfa187f2-adf0-4e24-9cf4-7df6d0016e12 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 752.794298] env[63515]: DEBUG oslo_concurrency.lockutils [None req-dfa187f2-adf0-4e24-9cf4-7df6d0016e12 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 752.794944] env[63515]: DEBUG oslo_concurrency.lockutils [None req-dfa187f2-adf0-4e24-9cf4-7df6d0016e12 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 752.794944] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-749c0204-b03e-4638-9a31-52d6c31136bb {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.799354] env[63515]: DEBUG oslo_vmware.api [None req-dfa187f2-adf0-4e24-9cf4-7df6d0016e12 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Waiting for the task: (returnval){ [ 752.799354] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]5250ed57-15b1-dda3-76fa-28825ee1f8c6" [ 752.799354] env[63515]: _type = "Task" [ 752.799354] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 752.807489] env[63515]: DEBUG oslo_vmware.api [None req-dfa187f2-adf0-4e24-9cf4-7df6d0016e12 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]5250ed57-15b1-dda3-76fa-28825ee1f8c6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 752.975258] env[63515]: DEBUG nova.network.neutron [req-231a3298-d6aa-4848-bca0-d930d8df5447 req-5f760e0d-5cd9-4673-93d9-ebaa0a57541b service nova] [instance: 38d6d030-06b0-4185-904d-44a038b3a752] Updated VIF entry in instance network info cache for port 2242e364-37c0-4750-8755-d807d47e899d. {{(pid=63515) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 752.975637] env[63515]: DEBUG nova.network.neutron [req-231a3298-d6aa-4848-bca0-d930d8df5447 req-5f760e0d-5cd9-4673-93d9-ebaa0a57541b service nova] [instance: 38d6d030-06b0-4185-904d-44a038b3a752] Updating instance_info_cache with network_info: [{"id": "2242e364-37c0-4750-8755-d807d47e899d", "address": "fa:16:3e:d4:3c:7a", "network": {"id": "d2aef558-9770-46ad-b51c-61bccb0de7ec", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-2087172546-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6f501ed813754f759464e338884993c1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6d4ef133-b6f3-41d1-add4-92a1482195cf", "external-id": "nsx-vlan-transportzone-446", "segmentation_id": 446, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2242e364-37", "ovs_interfaceid": "2242e364-37c0-4750-8755-d807d47e899d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 753.170052] env[63515]: DEBUG oslo_concurrency.lockutils [None req-9695633a-0f62-476f-8466-deb3365d2b47 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Lock "48668736-df27-4f2a-94d9-132f5b49701b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 29.498s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 753.177574] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1111091, 'name': CreateVM_Task} progress is 99%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 753.201581] env[63515]: DEBUG oslo_vmware.api [None req-99ae4d48-1fae-4ace-913c-d249ca72bbeb tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Task: {'id': task-1111087, 'name': PowerOnVM_Task, 'duration_secs': 0.536337} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 753.202114] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-99ae4d48-1fae-4ace-913c-d249ca72bbeb tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] [instance: d2a4f66a-f1a2-4f0d-826d-0b7c4f04bd23] Powered on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 753.202114] env[63515]: INFO nova.compute.manager [None req-99ae4d48-1fae-4ace-913c-d249ca72bbeb tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] [instance: d2a4f66a-f1a2-4f0d-826d-0b7c4f04bd23] Took 8.97 seconds to spawn the instance on the hypervisor. [ 753.202264] env[63515]: DEBUG nova.compute.manager [None req-99ae4d48-1fae-4ace-913c-d249ca72bbeb tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] [instance: d2a4f66a-f1a2-4f0d-826d-0b7c4f04bd23] Checking state {{(pid=63515) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 753.203094] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fcff02c9-d45e-4738-ae75-aa5fc960d5aa {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.219916] env[63515]: DEBUG oslo_vmware.api [None req-5bd41396-218d-4ce7-9765-053dcc695301 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Task: {'id': task-1111092, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.070608} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 753.220190] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-5bd41396-218d-4ce7-9765-053dcc695301 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] [instance: 6688cdaa-29ba-413a-8131-4f834cdb70e4] Extended root virtual disk {{(pid=63515) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 753.220943] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca6c674f-cb09-499e-89eb-658d6c701f42 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.243239] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-5bd41396-218d-4ce7-9765-053dcc695301 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] [instance: 6688cdaa-29ba-413a-8131-4f834cdb70e4] Reconfiguring VM instance instance-00000032 to attach disk [datastore1] 6688cdaa-29ba-413a-8131-4f834cdb70e4/6688cdaa-29ba-413a-8131-4f834cdb70e4.vmdk or device None with type sparse {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 753.243551] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5c84558f-ce72-4b77-a369-e053c1d35fee {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.264994] env[63515]: DEBUG oslo_vmware.api [None req-5bd41396-218d-4ce7-9765-053dcc695301 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Waiting for the task: (returnval){ [ 753.264994] env[63515]: value = "task-1111093" [ 753.264994] env[63515]: _type = "Task" [ 753.264994] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 753.273305] env[63515]: DEBUG oslo_vmware.api [None req-5bd41396-218d-4ce7-9765-053dcc695301 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Task: {'id': task-1111093, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 753.310196] env[63515]: DEBUG oslo_vmware.api [None req-dfa187f2-adf0-4e24-9cf4-7df6d0016e12 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]5250ed57-15b1-dda3-76fa-28825ee1f8c6, 'name': SearchDatastore_Task, 'duration_secs': 0.015564} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 753.312441] env[63515]: DEBUG oslo_concurrency.lockutils [None req-dfa187f2-adf0-4e24-9cf4-7df6d0016e12 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 753.312685] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-dfa187f2-adf0-4e24-9cf4-7df6d0016e12 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] [instance: 38d6d030-06b0-4185-904d-44a038b3a752] Processing image 8a120570-cb06-4099-b262-554ca0ad15c5 {{(pid=63515) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 753.312921] env[63515]: DEBUG oslo_concurrency.lockutils [None req-dfa187f2-adf0-4e24-9cf4-7df6d0016e12 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 753.313086] env[63515]: DEBUG oslo_concurrency.lockutils [None req-dfa187f2-adf0-4e24-9cf4-7df6d0016e12 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 753.313267] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-dfa187f2-adf0-4e24-9cf4-7df6d0016e12 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 753.313775] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c48eb1ed-e5c9-4b4a-8c43-421ed41b1e17 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.323849] env[63515]: DEBUG nova.virt.hardware [None req-0e5835c9-1a06-4538-91e4-48a9dfc2854d tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T02:52:14Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T02:51:56Z,direct_url=,disk_format='vmdk',id=8a120570-cb06-4099-b262-554ca0ad15c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b569255cc43e42e7ae3f2b4ad37c6ef4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T02:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 753.324108] env[63515]: DEBUG nova.virt.hardware [None req-0e5835c9-1a06-4538-91e4-48a9dfc2854d tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Flavor limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 753.324268] env[63515]: DEBUG nova.virt.hardware [None req-0e5835c9-1a06-4538-91e4-48a9dfc2854d tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Image limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 753.324450] env[63515]: DEBUG nova.virt.hardware [None req-0e5835c9-1a06-4538-91e4-48a9dfc2854d tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Flavor pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 753.324592] env[63515]: DEBUG nova.virt.hardware [None req-0e5835c9-1a06-4538-91e4-48a9dfc2854d tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Image pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 753.324735] env[63515]: DEBUG nova.virt.hardware [None req-0e5835c9-1a06-4538-91e4-48a9dfc2854d tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 753.324968] env[63515]: DEBUG nova.virt.hardware [None req-0e5835c9-1a06-4538-91e4-48a9dfc2854d tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 753.325197] env[63515]: DEBUG nova.virt.hardware [None req-0e5835c9-1a06-4538-91e4-48a9dfc2854d tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 753.325436] env[63515]: DEBUG nova.virt.hardware [None req-0e5835c9-1a06-4538-91e4-48a9dfc2854d tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Got 1 possible topologies {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 753.325737] env[63515]: DEBUG nova.virt.hardware [None req-0e5835c9-1a06-4538-91e4-48a9dfc2854d tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 753.325933] env[63515]: DEBUG nova.virt.hardware [None req-0e5835c9-1a06-4538-91e4-48a9dfc2854d tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 753.327365] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e97277b-f3b1-4b95-865f-df516756599d {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.331151] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-dfa187f2-adf0-4e24-9cf4-7df6d0016e12 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 753.331358] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-dfa187f2-adf0-4e24-9cf4-7df6d0016e12 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63515) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 753.332406] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e5412b11-3b5e-424b-a90d-eccac4ee0df9 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.339108] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3095fb54-8428-4a51-9f69-ebfe4f5b29aa {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.344634] env[63515]: DEBUG oslo_vmware.api [None req-dfa187f2-adf0-4e24-9cf4-7df6d0016e12 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Waiting for the task: (returnval){ [ 753.344634] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]528f1886-a82c-b1c9-ec1d-f97c6057118c" [ 753.344634] env[63515]: _type = "Task" [ 753.344634] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 753.356741] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-0e5835c9-1a06-4538-91e4-48a9dfc2854d tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: 001a48e1-b237-41f8-a6f4-2fe98d173512] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:6d:04:dc', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '7d377d75-3add-4a15-8691-74b2eb010924', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ed0b7b69-0679-4b35-af6f-44c60792b411', 'vif_model': 'vmxnet3'}] {{(pid=63515) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 753.364363] env[63515]: DEBUG oslo.service.loopingcall [None req-0e5835c9-1a06-4538-91e4-48a9dfc2854d tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 753.365282] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 001a48e1-b237-41f8-a6f4-2fe98d173512] Creating VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 753.365498] env[63515]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-fc021235-3b91-4087-bab4-462b4c046c78 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.385475] env[63515]: DEBUG oslo_vmware.api [None req-dfa187f2-adf0-4e24-9cf4-7df6d0016e12 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]528f1886-a82c-b1c9-ec1d-f97c6057118c, 'name': SearchDatastore_Task, 'duration_secs': 0.025651} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 753.386789] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e63ffac5-fc22-4395-a4cc-fc6570171de6 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.390676] env[63515]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 753.390676] env[63515]: value = "task-1111094" [ 753.390676] env[63515]: _type = "Task" [ 753.390676] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 753.395267] env[63515]: DEBUG oslo_vmware.api [None req-dfa187f2-adf0-4e24-9cf4-7df6d0016e12 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Waiting for the task: (returnval){ [ 753.395267] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]521ad43b-71f4-1478-785b-9ec0deb36db0" [ 753.395267] env[63515]: _type = "Task" [ 753.395267] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 753.401113] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1111094, 'name': CreateVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 753.405906] env[63515]: DEBUG oslo_vmware.api [None req-dfa187f2-adf0-4e24-9cf4-7df6d0016e12 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]521ad43b-71f4-1478-785b-9ec0deb36db0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 753.479097] env[63515]: DEBUG oslo_concurrency.lockutils [req-231a3298-d6aa-4848-bca0-d930d8df5447 req-5f760e0d-5cd9-4673-93d9-ebaa0a57541b service nova] Releasing lock "refresh_cache-38d6d030-06b0-4185-904d-44a038b3a752" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 753.621022] env[63515]: DEBUG oslo_concurrency.lockutils [None req-78f15907-fa94-44dd-a2b3-ccdc7dec0a0f tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Acquiring lock "87c468d9-9594-4804-b461-527f01f6118f" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 753.621022] env[63515]: DEBUG oslo_concurrency.lockutils [None req-78f15907-fa94-44dd-a2b3-ccdc7dec0a0f tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Lock "87c468d9-9594-4804-b461-527f01f6118f" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 753.621022] env[63515]: INFO nova.compute.manager [None req-78f15907-fa94-44dd-a2b3-ccdc7dec0a0f tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] [instance: 87c468d9-9594-4804-b461-527f01f6118f] Shelving [ 753.677957] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1111091, 'name': CreateVM_Task, 'duration_secs': 0.705755} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 753.678364] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8eb17506-ff93-4d25-b9af-ec5886569e65] Created VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 753.678836] env[63515]: DEBUG oslo_concurrency.lockutils [None req-750d4038-2a95-4b14-b9d7-c9c6803c64b6 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 753.678999] env[63515]: DEBUG oslo_concurrency.lockutils [None req-750d4038-2a95-4b14-b9d7-c9c6803c64b6 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 753.679987] env[63515]: DEBUG oslo_concurrency.lockutils [None req-750d4038-2a95-4b14-b9d7-c9c6803c64b6 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 753.682017] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-363ad665-ccf5-49a5-a3df-53cef4984b9c {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.687288] env[63515]: DEBUG oslo_vmware.api [None req-750d4038-2a95-4b14-b9d7-c9c6803c64b6 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] Waiting for the task: (returnval){ [ 753.687288] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]529d5798-6a53-ecdb-bfa3-40e9597b28d3" [ 753.687288] env[63515]: _type = "Task" [ 753.687288] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 753.698613] env[63515]: DEBUG oslo_vmware.api [None req-750d4038-2a95-4b14-b9d7-c9c6803c64b6 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]529d5798-6a53-ecdb-bfa3-40e9597b28d3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 753.722553] env[63515]: INFO nova.compute.manager [None req-99ae4d48-1fae-4ace-913c-d249ca72bbeb tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] [instance: d2a4f66a-f1a2-4f0d-826d-0b7c4f04bd23] Took 31.90 seconds to build instance. [ 753.776941] env[63515]: DEBUG oslo_vmware.api [None req-5bd41396-218d-4ce7-9765-053dcc695301 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Task: {'id': task-1111093, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 753.908785] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1111094, 'name': CreateVM_Task} progress is 99%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 753.909094] env[63515]: DEBUG oslo_vmware.api [None req-dfa187f2-adf0-4e24-9cf4-7df6d0016e12 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]521ad43b-71f4-1478-785b-9ec0deb36db0, 'name': SearchDatastore_Task, 'duration_secs': 0.020934} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 753.911759] env[63515]: DEBUG oslo_concurrency.lockutils [None req-dfa187f2-adf0-4e24-9cf4-7df6d0016e12 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 753.912381] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-dfa187f2-adf0-4e24-9cf4-7df6d0016e12 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk to [datastore2] 38d6d030-06b0-4185-904d-44a038b3a752/38d6d030-06b0-4185-904d-44a038b3a752.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 753.912526] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f427a978-5064-4fcb-8f09-ceba29742231 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.918759] env[63515]: DEBUG oslo_vmware.api [None req-dfa187f2-adf0-4e24-9cf4-7df6d0016e12 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Waiting for the task: (returnval){ [ 753.918759] env[63515]: value = "task-1111095" [ 753.918759] env[63515]: _type = "Task" [ 753.918759] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 753.929855] env[63515]: DEBUG oslo_vmware.api [None req-dfa187f2-adf0-4e24-9cf4-7df6d0016e12 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Task: {'id': task-1111095, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 754.068150] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f052697-bd0d-44fb-a6fc-3922c59f0fc8 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.076565] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9d609ba-cd88-4ca5-8053-9fe830a68d14 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.109807] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9286dbe-5615-4d49-9c51-f8dc11333410 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.117243] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d8ae939-b54c-4086-a0b7-7d700df6da89 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.132449] env[63515]: DEBUG nova.compute.provider_tree [None req-b8b2a7f6-808d-40d4-8de9-c8863e5b55c5 tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 754.135792] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-78f15907-fa94-44dd-a2b3-ccdc7dec0a0f tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] [instance: 87c468d9-9594-4804-b461-527f01f6118f] Powering off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 754.136051] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-352711d8-5121-4bc2-841d-3f12584f87a1 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.141969] env[63515]: DEBUG oslo_vmware.api [None req-78f15907-fa94-44dd-a2b3-ccdc7dec0a0f tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Waiting for the task: (returnval){ [ 754.141969] env[63515]: value = "task-1111096" [ 754.141969] env[63515]: _type = "Task" [ 754.141969] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 754.151283] env[63515]: DEBUG oslo_vmware.api [None req-78f15907-fa94-44dd-a2b3-ccdc7dec0a0f tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Task: {'id': task-1111096, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 754.197920] env[63515]: DEBUG oslo_vmware.api [None req-750d4038-2a95-4b14-b9d7-c9c6803c64b6 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]529d5798-6a53-ecdb-bfa3-40e9597b28d3, 'name': SearchDatastore_Task, 'duration_secs': 0.02456} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 754.198221] env[63515]: DEBUG oslo_concurrency.lockutils [None req-750d4038-2a95-4b14-b9d7-c9c6803c64b6 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 754.198450] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-750d4038-2a95-4b14-b9d7-c9c6803c64b6 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] [instance: 8eb17506-ff93-4d25-b9af-ec5886569e65] Processing image 8a120570-cb06-4099-b262-554ca0ad15c5 {{(pid=63515) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 754.198683] env[63515]: DEBUG oslo_concurrency.lockutils [None req-750d4038-2a95-4b14-b9d7-c9c6803c64b6 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 754.198828] env[63515]: DEBUG oslo_concurrency.lockutils [None req-750d4038-2a95-4b14-b9d7-c9c6803c64b6 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 754.199009] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-750d4038-2a95-4b14-b9d7-c9c6803c64b6 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 754.199279] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-43e6cdc5-215e-4502-bc66-ce8727481cc7 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.207857] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-750d4038-2a95-4b14-b9d7-c9c6803c64b6 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 754.207857] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-750d4038-2a95-4b14-b9d7-c9c6803c64b6 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63515) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 754.208471] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6454173a-c900-40f4-9e59-a7e487566a20 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.214878] env[63515]: DEBUG oslo_vmware.api [None req-750d4038-2a95-4b14-b9d7-c9c6803c64b6 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] Waiting for the task: (returnval){ [ 754.214878] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]522930ca-15ad-1916-b533-c7e2feb88773" [ 754.214878] env[63515]: _type = "Task" [ 754.214878] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 754.224845] env[63515]: DEBUG oslo_vmware.api [None req-750d4038-2a95-4b14-b9d7-c9c6803c64b6 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]522930ca-15ad-1916-b533-c7e2feb88773, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 754.226873] env[63515]: DEBUG oslo_concurrency.lockutils [None req-99ae4d48-1fae-4ace-913c-d249ca72bbeb tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Lock "d2a4f66a-f1a2-4f0d-826d-0b7c4f04bd23" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 115.212s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 754.283069] env[63515]: DEBUG oslo_vmware.api [None req-5bd41396-218d-4ce7-9765-053dcc695301 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Task: {'id': task-1111093, 'name': ReconfigVM_Task, 'duration_secs': 0.895859} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 754.283570] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-5bd41396-218d-4ce7-9765-053dcc695301 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] [instance: 6688cdaa-29ba-413a-8131-4f834cdb70e4] Reconfigured VM instance instance-00000032 to attach disk [datastore1] 6688cdaa-29ba-413a-8131-4f834cdb70e4/6688cdaa-29ba-413a-8131-4f834cdb70e4.vmdk or device None with type sparse {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 754.284505] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-2420ead3-1ee5-449b-9995-8bb2f2466e5a {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.295475] env[63515]: DEBUG oslo_vmware.api [None req-5bd41396-218d-4ce7-9765-053dcc695301 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Waiting for the task: (returnval){ [ 754.295475] env[63515]: value = "task-1111097" [ 754.295475] env[63515]: _type = "Task" [ 754.295475] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 754.307619] env[63515]: DEBUG oslo_vmware.api [None req-5bd41396-218d-4ce7-9765-053dcc695301 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Task: {'id': task-1111097, 'name': Rename_Task} progress is 6%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 754.404583] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1111094, 'name': CreateVM_Task, 'duration_secs': 0.891207} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 754.404771] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 001a48e1-b237-41f8-a6f4-2fe98d173512] Created VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 754.405529] env[63515]: DEBUG oslo_concurrency.lockutils [None req-0e5835c9-1a06-4538-91e4-48a9dfc2854d tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 754.405694] env[63515]: DEBUG oslo_concurrency.lockutils [None req-0e5835c9-1a06-4538-91e4-48a9dfc2854d tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 754.406039] env[63515]: DEBUG oslo_concurrency.lockutils [None req-0e5835c9-1a06-4538-91e4-48a9dfc2854d tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 754.406301] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-acc7ff33-5601-4333-950c-520fc2746dbe {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.413429] env[63515]: DEBUG oslo_vmware.api [None req-0e5835c9-1a06-4538-91e4-48a9dfc2854d tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Waiting for the task: (returnval){ [ 754.413429] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]520c6603-3a5f-9f0e-68ef-e7efd3ccd4ae" [ 754.413429] env[63515]: _type = "Task" [ 754.413429] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 754.422421] env[63515]: DEBUG oslo_vmware.api [None req-0e5835c9-1a06-4538-91e4-48a9dfc2854d tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]520c6603-3a5f-9f0e-68ef-e7efd3ccd4ae, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 754.432460] env[63515]: DEBUG oslo_vmware.api [None req-dfa187f2-adf0-4e24-9cf4-7df6d0016e12 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Task: {'id': task-1111095, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 754.636564] env[63515]: DEBUG nova.scheduler.client.report [None req-b8b2a7f6-808d-40d4-8de9-c8863e5b55c5 tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 754.651088] env[63515]: DEBUG oslo_vmware.api [None req-78f15907-fa94-44dd-a2b3-ccdc7dec0a0f tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Task: {'id': task-1111096, 'name': PowerOffVM_Task, 'duration_secs': 0.228589} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 754.651373] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-78f15907-fa94-44dd-a2b3-ccdc7dec0a0f tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] [instance: 87c468d9-9594-4804-b461-527f01f6118f] Powered off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 754.653629] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c9a3a2d-ab1c-4157-964d-fc5538099c04 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.672156] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c3b736c-4f18-4f64-b899-8e713583678f {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.725905] env[63515]: DEBUG oslo_vmware.api [None req-750d4038-2a95-4b14-b9d7-c9c6803c64b6 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]522930ca-15ad-1916-b533-c7e2feb88773, 'name': SearchDatastore_Task, 'duration_secs': 0.044836} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 754.726969] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ff311816-a0d5-4264-be3c-61bd7365c741 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.729897] env[63515]: DEBUG nova.compute.manager [None req-fa0cfe6d-3452-46d4-9987-2cb422956d62 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] [instance: 63ce797d-7180-4209-ac2c-81978bf7607a] Starting instance... {{(pid=63515) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 754.735529] env[63515]: DEBUG oslo_vmware.api [None req-750d4038-2a95-4b14-b9d7-c9c6803c64b6 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] Waiting for the task: (returnval){ [ 754.735529] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]526a0ce0-a54a-383e-38bd-48714ead2a77" [ 754.735529] env[63515]: _type = "Task" [ 754.735529] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 754.743244] env[63515]: DEBUG oslo_vmware.api [None req-750d4038-2a95-4b14-b9d7-c9c6803c64b6 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]526a0ce0-a54a-383e-38bd-48714ead2a77, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 754.807920] env[63515]: DEBUG oslo_vmware.api [None req-5bd41396-218d-4ce7-9765-053dcc695301 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Task: {'id': task-1111097, 'name': Rename_Task, 'duration_secs': 0.290488} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 754.808744] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-5bd41396-218d-4ce7-9765-053dcc695301 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] [instance: 6688cdaa-29ba-413a-8131-4f834cdb70e4] Powering on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 754.809136] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-99e89428-59d3-4087-8412-fe157fa43106 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.817778] env[63515]: DEBUG oslo_vmware.api [None req-5bd41396-218d-4ce7-9765-053dcc695301 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Waiting for the task: (returnval){ [ 754.817778] env[63515]: value = "task-1111098" [ 754.817778] env[63515]: _type = "Task" [ 754.817778] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 754.829278] env[63515]: DEBUG oslo_vmware.api [None req-5bd41396-218d-4ce7-9765-053dcc695301 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Task: {'id': task-1111098, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 754.924713] env[63515]: DEBUG oslo_vmware.api [None req-0e5835c9-1a06-4538-91e4-48a9dfc2854d tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]520c6603-3a5f-9f0e-68ef-e7efd3ccd4ae, 'name': SearchDatastore_Task, 'duration_secs': 0.057103} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 754.927512] env[63515]: DEBUG oslo_concurrency.lockutils [None req-0e5835c9-1a06-4538-91e4-48a9dfc2854d tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 754.928263] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-0e5835c9-1a06-4538-91e4-48a9dfc2854d tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: 001a48e1-b237-41f8-a6f4-2fe98d173512] Processing image 8a120570-cb06-4099-b262-554ca0ad15c5 {{(pid=63515) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 754.928263] env[63515]: DEBUG oslo_concurrency.lockutils [None req-0e5835c9-1a06-4538-91e4-48a9dfc2854d tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 754.933026] env[63515]: DEBUG oslo_vmware.api [None req-dfa187f2-adf0-4e24-9cf4-7df6d0016e12 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Task: {'id': task-1111095, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.634374} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 754.933264] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-dfa187f2-adf0-4e24-9cf4-7df6d0016e12 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk to [datastore2] 38d6d030-06b0-4185-904d-44a038b3a752/38d6d030-06b0-4185-904d-44a038b3a752.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 754.933376] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-dfa187f2-adf0-4e24-9cf4-7df6d0016e12 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] [instance: 38d6d030-06b0-4185-904d-44a038b3a752] Extending root virtual disk to 1048576 {{(pid=63515) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 754.933618] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-3f9861b2-0ed4-4e62-b40b-9ea6cf87cc96 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.939813] env[63515]: DEBUG oslo_vmware.api [None req-dfa187f2-adf0-4e24-9cf4-7df6d0016e12 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Waiting for the task: (returnval){ [ 754.939813] env[63515]: value = "task-1111099" [ 754.939813] env[63515]: _type = "Task" [ 754.939813] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 754.947880] env[63515]: DEBUG oslo_vmware.api [None req-dfa187f2-adf0-4e24-9cf4-7df6d0016e12 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Task: {'id': task-1111099, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 755.141280] env[63515]: DEBUG oslo_concurrency.lockutils [None req-b8b2a7f6-808d-40d4-8de9-c8863e5b55c5 tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.511s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 755.141809] env[63515]: DEBUG nova.compute.manager [None req-b8b2a7f6-808d-40d4-8de9-c8863e5b55c5 tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] [instance: 30b8b872-40ca-4297-b98e-a64c83a75483] Start building networks asynchronously for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 755.144430] env[63515]: DEBUG oslo_concurrency.lockutils [None req-8f7a1df0-5f0d-424a-ae8f-76bee2ee3dbb tempest-ServersAaction247Test-2019176257 tempest-ServersAaction247Test-2019176257-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 20.119s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 755.145758] env[63515]: INFO nova.compute.claims [None req-8f7a1df0-5f0d-424a-ae8f-76bee2ee3dbb tempest-ServersAaction247Test-2019176257 tempest-ServersAaction247Test-2019176257-project-member] [instance: 5d50fbd6-a7fb-422c-9dd7-df7140fd0aae] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 755.182561] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-78f15907-fa94-44dd-a2b3-ccdc7dec0a0f tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] [instance: 87c468d9-9594-4804-b461-527f01f6118f] Creating Snapshot of the VM instance {{(pid=63515) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 755.182889] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-db6ea45b-e4c0-4248-b997-25731934e6aa {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.190078] env[63515]: DEBUG oslo_concurrency.lockutils [None req-64ffe526-19dd-467f-8550-ae608ae5d75b tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Acquiring lock "452f3ca2-6141-43b2-a77a-c9ab5754192d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 755.190397] env[63515]: DEBUG oslo_concurrency.lockutils [None req-64ffe526-19dd-467f-8550-ae608ae5d75b tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Lock "452f3ca2-6141-43b2-a77a-c9ab5754192d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 755.190651] env[63515]: DEBUG oslo_concurrency.lockutils [None req-64ffe526-19dd-467f-8550-ae608ae5d75b tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Acquiring lock "452f3ca2-6141-43b2-a77a-c9ab5754192d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 755.190904] env[63515]: DEBUG oslo_concurrency.lockutils [None req-64ffe526-19dd-467f-8550-ae608ae5d75b tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Lock "452f3ca2-6141-43b2-a77a-c9ab5754192d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 755.191104] env[63515]: DEBUG oslo_concurrency.lockutils [None req-64ffe526-19dd-467f-8550-ae608ae5d75b tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Lock "452f3ca2-6141-43b2-a77a-c9ab5754192d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 755.194531] env[63515]: INFO nova.compute.manager [None req-64ffe526-19dd-467f-8550-ae608ae5d75b tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] [instance: 452f3ca2-6141-43b2-a77a-c9ab5754192d] Terminating instance [ 755.196019] env[63515]: DEBUG oslo_vmware.api [None req-78f15907-fa94-44dd-a2b3-ccdc7dec0a0f tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Waiting for the task: (returnval){ [ 755.196019] env[63515]: value = "task-1111100" [ 755.196019] env[63515]: _type = "Task" [ 755.196019] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 755.196519] env[63515]: DEBUG nova.compute.manager [None req-64ffe526-19dd-467f-8550-ae608ae5d75b tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] [instance: 452f3ca2-6141-43b2-a77a-c9ab5754192d] Start destroying the instance on the hypervisor. {{(pid=63515) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 755.196746] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-64ffe526-19dd-467f-8550-ae608ae5d75b tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] [instance: 452f3ca2-6141-43b2-a77a-c9ab5754192d] Destroying instance {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 755.197641] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ed0e9f2-1539-4371-bf6b-e79be7b00649 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.210447] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-64ffe526-19dd-467f-8550-ae608ae5d75b tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] [instance: 452f3ca2-6141-43b2-a77a-c9ab5754192d] Powering off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 755.213949] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4d3e18f3-f9f1-4823-b9ff-d396744746b3 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.215433] env[63515]: DEBUG oslo_vmware.api [None req-78f15907-fa94-44dd-a2b3-ccdc7dec0a0f tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Task: {'id': task-1111100, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 755.221350] env[63515]: DEBUG oslo_vmware.api [None req-64ffe526-19dd-467f-8550-ae608ae5d75b tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Waiting for the task: (returnval){ [ 755.221350] env[63515]: value = "task-1111101" [ 755.221350] env[63515]: _type = "Task" [ 755.221350] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 755.229649] env[63515]: DEBUG oslo_vmware.api [None req-64ffe526-19dd-467f-8550-ae608ae5d75b tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Task: {'id': task-1111101, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 755.248074] env[63515]: DEBUG oslo_vmware.api [None req-750d4038-2a95-4b14-b9d7-c9c6803c64b6 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]526a0ce0-a54a-383e-38bd-48714ead2a77, 'name': SearchDatastore_Task, 'duration_secs': 0.037838} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 755.248439] env[63515]: DEBUG oslo_concurrency.lockutils [None req-750d4038-2a95-4b14-b9d7-c9c6803c64b6 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 755.248770] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-750d4038-2a95-4b14-b9d7-c9c6803c64b6 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk to [datastore2] 8eb17506-ff93-4d25-b9af-ec5886569e65/8eb17506-ff93-4d25-b9af-ec5886569e65.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 755.249195] env[63515]: DEBUG oslo_concurrency.lockutils [None req-0e5835c9-1a06-4538-91e4-48a9dfc2854d tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 755.249440] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-0e5835c9-1a06-4538-91e4-48a9dfc2854d tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 755.249709] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-32f2cd15-6ef0-4c39-bfad-b09481b16369 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.252054] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f8a8bc3f-4a40-4198-879e-08400dc6d22d {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.257146] env[63515]: DEBUG oslo_concurrency.lockutils [None req-fa0cfe6d-3452-46d4-9987-2cb422956d62 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 755.258866] env[63515]: DEBUG oslo_vmware.api [None req-750d4038-2a95-4b14-b9d7-c9c6803c64b6 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] Waiting for the task: (returnval){ [ 755.258866] env[63515]: value = "task-1111102" [ 755.258866] env[63515]: _type = "Task" [ 755.258866] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 755.263306] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-0e5835c9-1a06-4538-91e4-48a9dfc2854d tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 755.263499] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-0e5835c9-1a06-4538-91e4-48a9dfc2854d tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63515) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 755.264542] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3e96372f-d0e1-4e7e-b788-3851cbcc13ba {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.269851] env[63515]: DEBUG oslo_vmware.api [None req-750d4038-2a95-4b14-b9d7-c9c6803c64b6 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] Task: {'id': task-1111102, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 755.272653] env[63515]: DEBUG oslo_vmware.api [None req-0e5835c9-1a06-4538-91e4-48a9dfc2854d tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Waiting for the task: (returnval){ [ 755.272653] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]52f7df3a-ebb7-9579-b850-7bd0d82c11fb" [ 755.272653] env[63515]: _type = "Task" [ 755.272653] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 755.280064] env[63515]: DEBUG oslo_vmware.api [None req-0e5835c9-1a06-4538-91e4-48a9dfc2854d tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52f7df3a-ebb7-9579-b850-7bd0d82c11fb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 755.328456] env[63515]: DEBUG oslo_vmware.api [None req-5bd41396-218d-4ce7-9765-053dcc695301 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Task: {'id': task-1111098, 'name': PowerOnVM_Task} progress is 87%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 755.375386] env[63515]: DEBUG oslo_concurrency.lockutils [None req-938cccb1-5a46-437d-9e93-3058a8134cd5 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Acquiring lock "37e6c27e-317b-45d2-bd55-2fd78ccf009f" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 755.375770] env[63515]: DEBUG oslo_concurrency.lockutils [None req-938cccb1-5a46-437d-9e93-3058a8134cd5 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Lock "37e6c27e-317b-45d2-bd55-2fd78ccf009f" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 755.376140] env[63515]: DEBUG oslo_concurrency.lockutils [None req-938cccb1-5a46-437d-9e93-3058a8134cd5 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Acquiring lock "37e6c27e-317b-45d2-bd55-2fd78ccf009f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 755.376524] env[63515]: DEBUG oslo_concurrency.lockutils [None req-938cccb1-5a46-437d-9e93-3058a8134cd5 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Lock "37e6c27e-317b-45d2-bd55-2fd78ccf009f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 755.376838] env[63515]: DEBUG oslo_concurrency.lockutils [None req-938cccb1-5a46-437d-9e93-3058a8134cd5 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Lock "37e6c27e-317b-45d2-bd55-2fd78ccf009f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 755.379535] env[63515]: INFO nova.compute.manager [None req-938cccb1-5a46-437d-9e93-3058a8134cd5 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] [instance: 37e6c27e-317b-45d2-bd55-2fd78ccf009f] Terminating instance [ 755.381812] env[63515]: DEBUG nova.compute.manager [None req-938cccb1-5a46-437d-9e93-3058a8134cd5 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] [instance: 37e6c27e-317b-45d2-bd55-2fd78ccf009f] Start destroying the instance on the hypervisor. {{(pid=63515) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 755.382047] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-938cccb1-5a46-437d-9e93-3058a8134cd5 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] [instance: 37e6c27e-317b-45d2-bd55-2fd78ccf009f] Destroying instance {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 755.383164] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ad17cdb-af0b-4fbe-a4b4-58ab4805b7c1 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.391830] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-938cccb1-5a46-437d-9e93-3058a8134cd5 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] [instance: 37e6c27e-317b-45d2-bd55-2fd78ccf009f] Powering off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 755.391977] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a15c369b-33d8-4d78-b44d-164dc3cf9eaa {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.399443] env[63515]: DEBUG oslo_vmware.api [None req-938cccb1-5a46-437d-9e93-3058a8134cd5 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Waiting for the task: (returnval){ [ 755.399443] env[63515]: value = "task-1111103" [ 755.399443] env[63515]: _type = "Task" [ 755.399443] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 755.408565] env[63515]: DEBUG oslo_vmware.api [None req-938cccb1-5a46-437d-9e93-3058a8134cd5 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Task: {'id': task-1111103, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 755.451430] env[63515]: DEBUG oslo_vmware.api [None req-dfa187f2-adf0-4e24-9cf4-7df6d0016e12 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Task: {'id': task-1111099, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.066012} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 755.451648] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-dfa187f2-adf0-4e24-9cf4-7df6d0016e12 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] [instance: 38d6d030-06b0-4185-904d-44a038b3a752] Extended root virtual disk {{(pid=63515) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 755.452569] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61355c30-f2a7-4694-bdfe-1a97e2d703cc {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.476490] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-dfa187f2-adf0-4e24-9cf4-7df6d0016e12 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] [instance: 38d6d030-06b0-4185-904d-44a038b3a752] Reconfiguring VM instance instance-00000033 to attach disk [datastore2] 38d6d030-06b0-4185-904d-44a038b3a752/38d6d030-06b0-4185-904d-44a038b3a752.vmdk or device None with type sparse {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 755.479026] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-692e04a4-a8b7-4ae2-8487-848ac43fc37a {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.500817] env[63515]: DEBUG oslo_vmware.api [None req-dfa187f2-adf0-4e24-9cf4-7df6d0016e12 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Waiting for the task: (returnval){ [ 755.500817] env[63515]: value = "task-1111104" [ 755.500817] env[63515]: _type = "Task" [ 755.500817] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 755.511822] env[63515]: DEBUG oslo_vmware.api [None req-dfa187f2-adf0-4e24-9cf4-7df6d0016e12 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Task: {'id': task-1111104, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 755.650289] env[63515]: DEBUG nova.compute.utils [None req-b8b2a7f6-808d-40d4-8de9-c8863e5b55c5 tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] Using /dev/sd instead of None {{(pid=63515) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 755.655193] env[63515]: DEBUG nova.compute.manager [None req-b8b2a7f6-808d-40d4-8de9-c8863e5b55c5 tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] [instance: 30b8b872-40ca-4297-b98e-a64c83a75483] Allocating IP information in the background. {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 755.655491] env[63515]: DEBUG nova.network.neutron [None req-b8b2a7f6-808d-40d4-8de9-c8863e5b55c5 tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] [instance: 30b8b872-40ca-4297-b98e-a64c83a75483] allocate_for_instance() {{(pid=63515) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 755.710053] env[63515]: DEBUG oslo_vmware.api [None req-78f15907-fa94-44dd-a2b3-ccdc7dec0a0f tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Task: {'id': task-1111100, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 755.716982] env[63515]: DEBUG nova.policy [None req-b8b2a7f6-808d-40d4-8de9-c8863e5b55c5 tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '03a2c6b25cd6405fb5cba6858d83e309', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '72b65c09ed6d4492864022352180ff49', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63515) authorize /opt/stack/nova/nova/policy.py:201}} [ 755.732280] env[63515]: DEBUG oslo_vmware.api [None req-64ffe526-19dd-467f-8550-ae608ae5d75b tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Task: {'id': task-1111101, 'name': PowerOffVM_Task, 'duration_secs': 0.242033} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 755.732676] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-64ffe526-19dd-467f-8550-ae608ae5d75b tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] [instance: 452f3ca2-6141-43b2-a77a-c9ab5754192d] Powered off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 755.732923] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-64ffe526-19dd-467f-8550-ae608ae5d75b tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] [instance: 452f3ca2-6141-43b2-a77a-c9ab5754192d] Unregistering the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 755.733262] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-8e705ac9-974b-410c-8ee4-4edddea10e5a {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.769935] env[63515]: DEBUG oslo_vmware.api [None req-750d4038-2a95-4b14-b9d7-c9c6803c64b6 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] Task: {'id': task-1111102, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 755.783282] env[63515]: DEBUG oslo_vmware.api [None req-0e5835c9-1a06-4538-91e4-48a9dfc2854d tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52f7df3a-ebb7-9579-b850-7bd0d82c11fb, 'name': SearchDatastore_Task, 'duration_secs': 0.011087} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 755.783510] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-587ac809-e92e-4708-bf25-e4ac3c5c44d7 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.789515] env[63515]: DEBUG oslo_vmware.api [None req-0e5835c9-1a06-4538-91e4-48a9dfc2854d tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Waiting for the task: (returnval){ [ 755.789515] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]52fee063-d231-4d81-cb28-b0b333ae53a0" [ 755.789515] env[63515]: _type = "Task" [ 755.789515] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 755.799926] env[63515]: DEBUG oslo_vmware.api [None req-0e5835c9-1a06-4538-91e4-48a9dfc2854d tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52fee063-d231-4d81-cb28-b0b333ae53a0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 755.807570] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-64ffe526-19dd-467f-8550-ae608ae5d75b tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] [instance: 452f3ca2-6141-43b2-a77a-c9ab5754192d] Unregistered the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 755.807784] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-64ffe526-19dd-467f-8550-ae608ae5d75b tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] [instance: 452f3ca2-6141-43b2-a77a-c9ab5754192d] Deleting contents of the VM from datastore datastore1 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 755.807964] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-64ffe526-19dd-467f-8550-ae608ae5d75b tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Deleting the datastore file [datastore1] 452f3ca2-6141-43b2-a77a-c9ab5754192d {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 755.808283] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a067deb9-13b8-4283-9d09-4f61cb42de44 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.815980] env[63515]: DEBUG oslo_vmware.api [None req-64ffe526-19dd-467f-8550-ae608ae5d75b tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Waiting for the task: (returnval){ [ 755.815980] env[63515]: value = "task-1111106" [ 755.815980] env[63515]: _type = "Task" [ 755.815980] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 755.826628] env[63515]: DEBUG oslo_vmware.api [None req-64ffe526-19dd-467f-8550-ae608ae5d75b tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Task: {'id': task-1111106, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 755.830089] env[63515]: DEBUG oslo_vmware.api [None req-5bd41396-218d-4ce7-9765-053dcc695301 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Task: {'id': task-1111098, 'name': PowerOnVM_Task, 'duration_secs': 0.690589} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 755.830356] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-5bd41396-218d-4ce7-9765-053dcc695301 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] [instance: 6688cdaa-29ba-413a-8131-4f834cdb70e4] Powered on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 755.830558] env[63515]: INFO nova.compute.manager [None req-5bd41396-218d-4ce7-9765-053dcc695301 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] [instance: 6688cdaa-29ba-413a-8131-4f834cdb70e4] Took 8.95 seconds to spawn the instance on the hypervisor. [ 755.830754] env[63515]: DEBUG nova.compute.manager [None req-5bd41396-218d-4ce7-9765-053dcc695301 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] [instance: 6688cdaa-29ba-413a-8131-4f834cdb70e4] Checking state {{(pid=63515) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 755.833521] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3ef2095-5756-4c2f-9883-6ae70d950002 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.909803] env[63515]: DEBUG oslo_vmware.api [None req-938cccb1-5a46-437d-9e93-3058a8134cd5 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Task: {'id': task-1111103, 'name': PowerOffVM_Task, 'duration_secs': 0.225605} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 755.910106] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-938cccb1-5a46-437d-9e93-3058a8134cd5 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] [instance: 37e6c27e-317b-45d2-bd55-2fd78ccf009f] Powered off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 755.910301] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-938cccb1-5a46-437d-9e93-3058a8134cd5 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] [instance: 37e6c27e-317b-45d2-bd55-2fd78ccf009f] Unregistering the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 755.910582] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ea45d2ac-7d7a-4ac8-a337-d238e65b84c5 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.011722] env[63515]: DEBUG oslo_vmware.api [None req-dfa187f2-adf0-4e24-9cf4-7df6d0016e12 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Task: {'id': task-1111104, 'name': ReconfigVM_Task, 'duration_secs': 0.443951} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 756.011998] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-dfa187f2-adf0-4e24-9cf4-7df6d0016e12 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] [instance: 38d6d030-06b0-4185-904d-44a038b3a752] Reconfigured VM instance instance-00000033 to attach disk [datastore2] 38d6d030-06b0-4185-904d-44a038b3a752/38d6d030-06b0-4185-904d-44a038b3a752.vmdk or device None with type sparse {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 756.012671] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-02346a3f-f9ca-4a30-b2c1-f97955af0276 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.019280] env[63515]: DEBUG oslo_vmware.api [None req-dfa187f2-adf0-4e24-9cf4-7df6d0016e12 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Waiting for the task: (returnval){ [ 756.019280] env[63515]: value = "task-1111108" [ 756.019280] env[63515]: _type = "Task" [ 756.019280] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 756.027740] env[63515]: DEBUG oslo_vmware.api [None req-dfa187f2-adf0-4e24-9cf4-7df6d0016e12 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Task: {'id': task-1111108, 'name': Rename_Task} progress is 5%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 756.028971] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-938cccb1-5a46-437d-9e93-3058a8134cd5 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] [instance: 37e6c27e-317b-45d2-bd55-2fd78ccf009f] Unregistered the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 756.029180] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-938cccb1-5a46-437d-9e93-3058a8134cd5 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] [instance: 37e6c27e-317b-45d2-bd55-2fd78ccf009f] Deleting contents of the VM from datastore datastore1 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 756.029353] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-938cccb1-5a46-437d-9e93-3058a8134cd5 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Deleting the datastore file [datastore1] 37e6c27e-317b-45d2-bd55-2fd78ccf009f {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 756.029858] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-447ec495-e29c-48d5-a1bc-e53dc4f5989c {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.037740] env[63515]: DEBUG oslo_vmware.api [None req-938cccb1-5a46-437d-9e93-3058a8134cd5 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Waiting for the task: (returnval){ [ 756.037740] env[63515]: value = "task-1111109" [ 756.037740] env[63515]: _type = "Task" [ 756.037740] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 756.045829] env[63515]: DEBUG oslo_vmware.api [None req-938cccb1-5a46-437d-9e93-3058a8134cd5 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Task: {'id': task-1111109, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 756.089431] env[63515]: DEBUG nova.network.neutron [None req-b8b2a7f6-808d-40d4-8de9-c8863e5b55c5 tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] [instance: 30b8b872-40ca-4297-b98e-a64c83a75483] Successfully created port: b8fe94ee-46b7-489a-819d-0699bc02bbb7 {{(pid=63515) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 756.155638] env[63515]: DEBUG nova.compute.manager [None req-b8b2a7f6-808d-40d4-8de9-c8863e5b55c5 tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] [instance: 30b8b872-40ca-4297-b98e-a64c83a75483] Start building block device mappings for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 756.208326] env[63515]: DEBUG oslo_vmware.api [None req-78f15907-fa94-44dd-a2b3-ccdc7dec0a0f tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Task: {'id': task-1111100, 'name': CreateSnapshot_Task, 'duration_secs': 0.904091} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 756.208591] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-78f15907-fa94-44dd-a2b3-ccdc7dec0a0f tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] [instance: 87c468d9-9594-4804-b461-527f01f6118f] Created Snapshot of the VM instance {{(pid=63515) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 756.209330] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25624029-30fd-44f7-8677-2b6bb2b0eb14 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.270198] env[63515]: DEBUG oslo_vmware.api [None req-750d4038-2a95-4b14-b9d7-c9c6803c64b6 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] Task: {'id': task-1111102, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.54679} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 756.272630] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-750d4038-2a95-4b14-b9d7-c9c6803c64b6 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk to [datastore2] 8eb17506-ff93-4d25-b9af-ec5886569e65/8eb17506-ff93-4d25-b9af-ec5886569e65.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 756.272850] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-750d4038-2a95-4b14-b9d7-c9c6803c64b6 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] [instance: 8eb17506-ff93-4d25-b9af-ec5886569e65] Extending root virtual disk to 1048576 {{(pid=63515) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 756.273298] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-589fb550-d8c7-4185-a062-37d82bd00a3f {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.279383] env[63515]: DEBUG oslo_vmware.api [None req-750d4038-2a95-4b14-b9d7-c9c6803c64b6 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] Waiting for the task: (returnval){ [ 756.279383] env[63515]: value = "task-1111110" [ 756.279383] env[63515]: _type = "Task" [ 756.279383] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 756.291476] env[63515]: DEBUG oslo_vmware.api [None req-750d4038-2a95-4b14-b9d7-c9c6803c64b6 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] Task: {'id': task-1111110, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 756.304642] env[63515]: DEBUG oslo_vmware.api [None req-0e5835c9-1a06-4538-91e4-48a9dfc2854d tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52fee063-d231-4d81-cb28-b0b333ae53a0, 'name': SearchDatastore_Task, 'duration_secs': 0.015332} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 756.304982] env[63515]: DEBUG oslo_concurrency.lockutils [None req-0e5835c9-1a06-4538-91e4-48a9dfc2854d tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 756.305801] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-0e5835c9-1a06-4538-91e4-48a9dfc2854d tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk to [datastore2] 001a48e1-b237-41f8-a6f4-2fe98d173512/001a48e1-b237-41f8-a6f4-2fe98d173512.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 756.305801] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c76a09a0-83e0-4951-b71b-2bb744374b66 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.312177] env[63515]: DEBUG oslo_vmware.api [None req-0e5835c9-1a06-4538-91e4-48a9dfc2854d tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Waiting for the task: (returnval){ [ 756.312177] env[63515]: value = "task-1111111" [ 756.312177] env[63515]: _type = "Task" [ 756.312177] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 756.323892] env[63515]: DEBUG oslo_vmware.api [None req-0e5835c9-1a06-4538-91e4-48a9dfc2854d tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Task: {'id': task-1111111, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 756.328849] env[63515]: DEBUG oslo_vmware.api [None req-64ffe526-19dd-467f-8550-ae608ae5d75b tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Task: {'id': task-1111106, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.469056} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 756.331359] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-64ffe526-19dd-467f-8550-ae608ae5d75b tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Deleted the datastore file {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 756.331563] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-64ffe526-19dd-467f-8550-ae608ae5d75b tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] [instance: 452f3ca2-6141-43b2-a77a-c9ab5754192d] Deleted contents of the VM from datastore datastore1 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 756.331739] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-64ffe526-19dd-467f-8550-ae608ae5d75b tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] [instance: 452f3ca2-6141-43b2-a77a-c9ab5754192d] Instance destroyed {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 756.331907] env[63515]: INFO nova.compute.manager [None req-64ffe526-19dd-467f-8550-ae608ae5d75b tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] [instance: 452f3ca2-6141-43b2-a77a-c9ab5754192d] Took 1.14 seconds to destroy the instance on the hypervisor. [ 756.332163] env[63515]: DEBUG oslo.service.loopingcall [None req-64ffe526-19dd-467f-8550-ae608ae5d75b tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 756.333499] env[63515]: DEBUG nova.compute.manager [-] [instance: 452f3ca2-6141-43b2-a77a-c9ab5754192d] Deallocating network for instance {{(pid=63515) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 756.333499] env[63515]: DEBUG nova.network.neutron [-] [instance: 452f3ca2-6141-43b2-a77a-c9ab5754192d] deallocate_for_instance() {{(pid=63515) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 756.352023] env[63515]: INFO nova.compute.manager [None req-5bd41396-218d-4ce7-9765-053dcc695301 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] [instance: 6688cdaa-29ba-413a-8131-4f834cdb70e4] Took 32.16 seconds to build instance. [ 756.531109] env[63515]: DEBUG oslo_vmware.api [None req-dfa187f2-adf0-4e24-9cf4-7df6d0016e12 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Task: {'id': task-1111108, 'name': Rename_Task, 'duration_secs': 0.143395} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 756.531378] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-dfa187f2-adf0-4e24-9cf4-7df6d0016e12 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] [instance: 38d6d030-06b0-4185-904d-44a038b3a752] Powering on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 756.532601] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9c5ea482-d7bb-43e5-b886-68af7c87418b {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.544602] env[63515]: DEBUG oslo_vmware.api [None req-dfa187f2-adf0-4e24-9cf4-7df6d0016e12 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Waiting for the task: (returnval){ [ 756.544602] env[63515]: value = "task-1111112" [ 756.544602] env[63515]: _type = "Task" [ 756.544602] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 756.551862] env[63515]: DEBUG oslo_vmware.api [None req-938cccb1-5a46-437d-9e93-3058a8134cd5 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Task: {'id': task-1111109, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 756.558697] env[63515]: DEBUG oslo_vmware.api [None req-dfa187f2-adf0-4e24-9cf4-7df6d0016e12 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Task: {'id': task-1111112, 'name': PowerOnVM_Task} progress is 33%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 756.612279] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57b95485-1bf4-44f3-8325-4c921ccce7f2 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.616977] env[63515]: DEBUG nova.compute.manager [req-b386db57-1c4f-4d78-88f4-319718c1c0b9 req-65b353c9-6a38-472c-9e99-18c1d1800124 service nova] [instance: 452f3ca2-6141-43b2-a77a-c9ab5754192d] Received event network-vif-deleted-26b2403b-166a-432e-9692-27511a1a4ae6 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 756.617301] env[63515]: INFO nova.compute.manager [req-b386db57-1c4f-4d78-88f4-319718c1c0b9 req-65b353c9-6a38-472c-9e99-18c1d1800124 service nova] [instance: 452f3ca2-6141-43b2-a77a-c9ab5754192d] Neutron deleted interface 26b2403b-166a-432e-9692-27511a1a4ae6; detaching it from the instance and deleting it from the info cache [ 756.617395] env[63515]: DEBUG nova.network.neutron [req-b386db57-1c4f-4d78-88f4-319718c1c0b9 req-65b353c9-6a38-472c-9e99-18c1d1800124 service nova] [instance: 452f3ca2-6141-43b2-a77a-c9ab5754192d] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 756.623935] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd1b66d3-ce2f-466d-b8c4-aca5f663c6c0 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.656808] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b0bfd06-e796-4244-ba10-c10fbeaa5709 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.664931] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6af3a705-75d2-4eeb-b7b6-bfea03a65992 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.683079] env[63515]: DEBUG nova.compute.provider_tree [None req-8f7a1df0-5f0d-424a-ae8f-76bee2ee3dbb tempest-ServersAaction247Test-2019176257 tempest-ServersAaction247Test-2019176257-project-member] Updating inventory in ProviderTree for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 756.729861] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-78f15907-fa94-44dd-a2b3-ccdc7dec0a0f tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] [instance: 87c468d9-9594-4804-b461-527f01f6118f] Creating linked-clone VM from snapshot {{(pid=63515) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 756.729983] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-d00a8dc7-3450-4386-af8b-aa3e91b382d4 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.739394] env[63515]: DEBUG oslo_vmware.api [None req-78f15907-fa94-44dd-a2b3-ccdc7dec0a0f tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Waiting for the task: (returnval){ [ 756.739394] env[63515]: value = "task-1111113" [ 756.739394] env[63515]: _type = "Task" [ 756.739394] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 756.747695] env[63515]: DEBUG oslo_vmware.api [None req-78f15907-fa94-44dd-a2b3-ccdc7dec0a0f tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Task: {'id': task-1111113, 'name': CloneVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 756.788914] env[63515]: DEBUG oslo_vmware.api [None req-750d4038-2a95-4b14-b9d7-c9c6803c64b6 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] Task: {'id': task-1111110, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.067898} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 756.789188] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-750d4038-2a95-4b14-b9d7-c9c6803c64b6 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] [instance: 8eb17506-ff93-4d25-b9af-ec5886569e65] Extended root virtual disk {{(pid=63515) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 756.789976] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76ad2267-0de3-418b-a5d6-cb2332f9742c {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.810052] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-750d4038-2a95-4b14-b9d7-c9c6803c64b6 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] [instance: 8eb17506-ff93-4d25-b9af-ec5886569e65] Reconfiguring VM instance instance-00000034 to attach disk [datastore2] 8eb17506-ff93-4d25-b9af-ec5886569e65/8eb17506-ff93-4d25-b9af-ec5886569e65.vmdk or device None with type sparse {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 756.810427] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d97c11e4-9bc6-47db-b5b0-a906d689f0f1 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.834207] env[63515]: DEBUG oslo_vmware.api [None req-0e5835c9-1a06-4538-91e4-48a9dfc2854d tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Task: {'id': task-1111111, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 756.835476] env[63515]: DEBUG oslo_vmware.api [None req-750d4038-2a95-4b14-b9d7-c9c6803c64b6 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] Waiting for the task: (returnval){ [ 756.835476] env[63515]: value = "task-1111114" [ 756.835476] env[63515]: _type = "Task" [ 756.835476] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 756.842808] env[63515]: DEBUG oslo_vmware.api [None req-750d4038-2a95-4b14-b9d7-c9c6803c64b6 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] Task: {'id': task-1111114, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 756.853900] env[63515]: DEBUG oslo_concurrency.lockutils [None req-5bd41396-218d-4ce7-9765-053dcc695301 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Lock "6688cdaa-29ba-413a-8131-4f834cdb70e4" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 117.516s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 757.052821] env[63515]: DEBUG oslo_vmware.api [None req-938cccb1-5a46-437d-9e93-3058a8134cd5 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Task: {'id': task-1111109, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.608977} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 757.053602] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-938cccb1-5a46-437d-9e93-3058a8134cd5 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Deleted the datastore file {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 757.053806] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-938cccb1-5a46-437d-9e93-3058a8134cd5 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] [instance: 37e6c27e-317b-45d2-bd55-2fd78ccf009f] Deleted contents of the VM from datastore datastore1 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 757.054075] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-938cccb1-5a46-437d-9e93-3058a8134cd5 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] [instance: 37e6c27e-317b-45d2-bd55-2fd78ccf009f] Instance destroyed {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 757.054219] env[63515]: INFO nova.compute.manager [None req-938cccb1-5a46-437d-9e93-3058a8134cd5 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] [instance: 37e6c27e-317b-45d2-bd55-2fd78ccf009f] Took 1.67 seconds to destroy the instance on the hypervisor. [ 757.054515] env[63515]: DEBUG oslo.service.loopingcall [None req-938cccb1-5a46-437d-9e93-3058a8134cd5 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 757.054736] env[63515]: DEBUG nova.compute.manager [-] [instance: 37e6c27e-317b-45d2-bd55-2fd78ccf009f] Deallocating network for instance {{(pid=63515) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 757.054807] env[63515]: DEBUG nova.network.neutron [-] [instance: 37e6c27e-317b-45d2-bd55-2fd78ccf009f] deallocate_for_instance() {{(pid=63515) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 757.061032] env[63515]: DEBUG oslo_vmware.api [None req-dfa187f2-adf0-4e24-9cf4-7df6d0016e12 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Task: {'id': task-1111112, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 757.091851] env[63515]: DEBUG nova.network.neutron [-] [instance: 452f3ca2-6141-43b2-a77a-c9ab5754192d] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 757.120821] env[63515]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-93befd15-a48a-43f4-9bd5-43e3c7ac6bb3 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.133652] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8137e387-7ee7-4ba2-a811-d0e589b09590 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.169564] env[63515]: DEBUG nova.compute.manager [req-b386db57-1c4f-4d78-88f4-319718c1c0b9 req-65b353c9-6a38-472c-9e99-18c1d1800124 service nova] [instance: 452f3ca2-6141-43b2-a77a-c9ab5754192d] Detach interface failed, port_id=26b2403b-166a-432e-9692-27511a1a4ae6, reason: Instance 452f3ca2-6141-43b2-a77a-c9ab5754192d could not be found. {{(pid=63515) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 757.175392] env[63515]: DEBUG nova.compute.manager [None req-b8b2a7f6-808d-40d4-8de9-c8863e5b55c5 tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] [instance: 30b8b872-40ca-4297-b98e-a64c83a75483] Start spawning the instance on the hypervisor. {{(pid=63515) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 757.212628] env[63515]: DEBUG nova.virt.hardware [None req-b8b2a7f6-808d-40d4-8de9-c8863e5b55c5 tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T02:52:14Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T02:51:56Z,direct_url=,disk_format='vmdk',id=8a120570-cb06-4099-b262-554ca0ad15c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b569255cc43e42e7ae3f2b4ad37c6ef4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T02:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 757.213112] env[63515]: DEBUG nova.virt.hardware [None req-b8b2a7f6-808d-40d4-8de9-c8863e5b55c5 tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] Flavor limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 757.213343] env[63515]: DEBUG nova.virt.hardware [None req-b8b2a7f6-808d-40d4-8de9-c8863e5b55c5 tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] Image limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 757.213656] env[63515]: DEBUG nova.virt.hardware [None req-b8b2a7f6-808d-40d4-8de9-c8863e5b55c5 tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] Flavor pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 757.213940] env[63515]: DEBUG nova.virt.hardware [None req-b8b2a7f6-808d-40d4-8de9-c8863e5b55c5 tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] Image pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 757.214187] env[63515]: DEBUG nova.virt.hardware [None req-b8b2a7f6-808d-40d4-8de9-c8863e5b55c5 tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 757.214470] env[63515]: DEBUG nova.virt.hardware [None req-b8b2a7f6-808d-40d4-8de9-c8863e5b55c5 tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 757.214735] env[63515]: DEBUG nova.virt.hardware [None req-b8b2a7f6-808d-40d4-8de9-c8863e5b55c5 tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 757.215099] env[63515]: DEBUG nova.virt.hardware [None req-b8b2a7f6-808d-40d4-8de9-c8863e5b55c5 tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] Got 1 possible topologies {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 757.215355] env[63515]: DEBUG nova.virt.hardware [None req-b8b2a7f6-808d-40d4-8de9-c8863e5b55c5 tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 757.215631] env[63515]: DEBUG nova.virt.hardware [None req-b8b2a7f6-808d-40d4-8de9-c8863e5b55c5 tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 757.216774] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f051c0ff-67c0-4307-bb89-5529c2950883 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.221984] env[63515]: DEBUG nova.scheduler.client.report [None req-8f7a1df0-5f0d-424a-ae8f-76bee2ee3dbb tempest-ServersAaction247Test-2019176257 tempest-ServersAaction247Test-2019176257-project-member] Updated inventory for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a with generation 74 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 757.222346] env[63515]: DEBUG nova.compute.provider_tree [None req-8f7a1df0-5f0d-424a-ae8f-76bee2ee3dbb tempest-ServersAaction247Test-2019176257 tempest-ServersAaction247Test-2019176257-project-member] Updating resource provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a generation from 74 to 75 during operation: update_inventory {{(pid=63515) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 757.222494] env[63515]: DEBUG nova.compute.provider_tree [None req-8f7a1df0-5f0d-424a-ae8f-76bee2ee3dbb tempest-ServersAaction247Test-2019176257 tempest-ServersAaction247Test-2019176257-project-member] Updating inventory in ProviderTree for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 757.230674] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c52467e2-7f3f-43ab-ba2a-60d003b94d46 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.259384] env[63515]: DEBUG oslo_vmware.api [None req-78f15907-fa94-44dd-a2b3-ccdc7dec0a0f tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Task: {'id': task-1111113, 'name': CloneVM_Task} progress is 94%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 757.334703] env[63515]: DEBUG oslo_vmware.api [None req-0e5835c9-1a06-4538-91e4-48a9dfc2854d tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Task: {'id': task-1111111, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.974571} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 757.335064] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-0e5835c9-1a06-4538-91e4-48a9dfc2854d tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk to [datastore2] 001a48e1-b237-41f8-a6f4-2fe98d173512/001a48e1-b237-41f8-a6f4-2fe98d173512.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 757.335299] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-0e5835c9-1a06-4538-91e4-48a9dfc2854d tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: 001a48e1-b237-41f8-a6f4-2fe98d173512] Extending root virtual disk to 1048576 {{(pid=63515) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 757.335658] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-430b69f8-13ef-4933-be82-a5c1e1e72dd6 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.346328] env[63515]: DEBUG oslo_vmware.api [None req-750d4038-2a95-4b14-b9d7-c9c6803c64b6 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] Task: {'id': task-1111114, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 757.347865] env[63515]: DEBUG oslo_vmware.api [None req-0e5835c9-1a06-4538-91e4-48a9dfc2854d tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Waiting for the task: (returnval){ [ 757.347865] env[63515]: value = "task-1111115" [ 757.347865] env[63515]: _type = "Task" [ 757.347865] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 757.355998] env[63515]: DEBUG nova.compute.manager [None req-4b5baf79-d4c3-467e-88d4-a131ea6620d7 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] [instance: e666825c-ff4e-4a0e-93c0-43c00f167bbb] Starting instance... {{(pid=63515) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 757.358458] env[63515]: DEBUG oslo_vmware.api [None req-0e5835c9-1a06-4538-91e4-48a9dfc2854d tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Task: {'id': task-1111115, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 757.551972] env[63515]: DEBUG nova.compute.manager [req-dc362ad1-9b5b-4655-884d-fb7499ec646d req-1bef55a5-0c42-46f6-9c34-3a0453620cc7 service nova] [instance: 30b8b872-40ca-4297-b98e-a64c83a75483] Received event network-vif-plugged-b8fe94ee-46b7-489a-819d-0699bc02bbb7 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 757.552274] env[63515]: DEBUG oslo_concurrency.lockutils [req-dc362ad1-9b5b-4655-884d-fb7499ec646d req-1bef55a5-0c42-46f6-9c34-3a0453620cc7 service nova] Acquiring lock "30b8b872-40ca-4297-b98e-a64c83a75483-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 757.552523] env[63515]: DEBUG oslo_concurrency.lockutils [req-dc362ad1-9b5b-4655-884d-fb7499ec646d req-1bef55a5-0c42-46f6-9c34-3a0453620cc7 service nova] Lock "30b8b872-40ca-4297-b98e-a64c83a75483-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 757.552753] env[63515]: DEBUG oslo_concurrency.lockutils [req-dc362ad1-9b5b-4655-884d-fb7499ec646d req-1bef55a5-0c42-46f6-9c34-3a0453620cc7 service nova] Lock "30b8b872-40ca-4297-b98e-a64c83a75483-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 757.553193] env[63515]: DEBUG nova.compute.manager [req-dc362ad1-9b5b-4655-884d-fb7499ec646d req-1bef55a5-0c42-46f6-9c34-3a0453620cc7 service nova] [instance: 30b8b872-40ca-4297-b98e-a64c83a75483] No waiting events found dispatching network-vif-plugged-b8fe94ee-46b7-489a-819d-0699bc02bbb7 {{(pid=63515) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 757.553433] env[63515]: WARNING nova.compute.manager [req-dc362ad1-9b5b-4655-884d-fb7499ec646d req-1bef55a5-0c42-46f6-9c34-3a0453620cc7 service nova] [instance: 30b8b872-40ca-4297-b98e-a64c83a75483] Received unexpected event network-vif-plugged-b8fe94ee-46b7-489a-819d-0699bc02bbb7 for instance with vm_state building and task_state spawning. [ 757.559510] env[63515]: DEBUG oslo_vmware.api [None req-dfa187f2-adf0-4e24-9cf4-7df6d0016e12 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Task: {'id': task-1111112, 'name': PowerOnVM_Task, 'duration_secs': 0.905824} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 757.559882] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-dfa187f2-adf0-4e24-9cf4-7df6d0016e12 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] [instance: 38d6d030-06b0-4185-904d-44a038b3a752] Powered on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 757.560130] env[63515]: INFO nova.compute.manager [None req-dfa187f2-adf0-4e24-9cf4-7df6d0016e12 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] [instance: 38d6d030-06b0-4185-904d-44a038b3a752] Took 8.01 seconds to spawn the instance on the hypervisor. [ 757.560359] env[63515]: DEBUG nova.compute.manager [None req-dfa187f2-adf0-4e24-9cf4-7df6d0016e12 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] [instance: 38d6d030-06b0-4185-904d-44a038b3a752] Checking state {{(pid=63515) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 757.561183] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7385849c-ce3b-40b1-9b34-aac7e4907630 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.596464] env[63515]: INFO nova.compute.manager [-] [instance: 452f3ca2-6141-43b2-a77a-c9ab5754192d] Took 1.26 seconds to deallocate network for instance. [ 757.706894] env[63515]: DEBUG nova.network.neutron [None req-b8b2a7f6-808d-40d4-8de9-c8863e5b55c5 tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] [instance: 30b8b872-40ca-4297-b98e-a64c83a75483] Successfully updated port: b8fe94ee-46b7-489a-819d-0699bc02bbb7 {{(pid=63515) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 757.730520] env[63515]: DEBUG oslo_concurrency.lockutils [None req-8f7a1df0-5f0d-424a-ae8f-76bee2ee3dbb tempest-ServersAaction247Test-2019176257 tempest-ServersAaction247Test-2019176257-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.586s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 757.731107] env[63515]: DEBUG nova.compute.manager [None req-8f7a1df0-5f0d-424a-ae8f-76bee2ee3dbb tempest-ServersAaction247Test-2019176257 tempest-ServersAaction247Test-2019176257-project-member] [instance: 5d50fbd6-a7fb-422c-9dd7-df7140fd0aae] Start building networks asynchronously for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 757.734517] env[63515]: DEBUG oslo_concurrency.lockutils [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 21.522s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 757.737521] env[63515]: DEBUG nova.compute.manager [req-f2d3f820-d001-41c6-a6b6-62b46c806215 req-7d04764a-e284-473e-ae34-4a8d9efecdee service nova] [instance: 30b8b872-40ca-4297-b98e-a64c83a75483] Received event network-changed-b8fe94ee-46b7-489a-819d-0699bc02bbb7 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 757.737521] env[63515]: DEBUG nova.compute.manager [req-f2d3f820-d001-41c6-a6b6-62b46c806215 req-7d04764a-e284-473e-ae34-4a8d9efecdee service nova] [instance: 30b8b872-40ca-4297-b98e-a64c83a75483] Refreshing instance network info cache due to event network-changed-b8fe94ee-46b7-489a-819d-0699bc02bbb7. {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 757.737521] env[63515]: DEBUG oslo_concurrency.lockutils [req-f2d3f820-d001-41c6-a6b6-62b46c806215 req-7d04764a-e284-473e-ae34-4a8d9efecdee service nova] Acquiring lock "refresh_cache-30b8b872-40ca-4297-b98e-a64c83a75483" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 757.737819] env[63515]: DEBUG oslo_concurrency.lockutils [req-f2d3f820-d001-41c6-a6b6-62b46c806215 req-7d04764a-e284-473e-ae34-4a8d9efecdee service nova] Acquired lock "refresh_cache-30b8b872-40ca-4297-b98e-a64c83a75483" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 757.739450] env[63515]: DEBUG nova.network.neutron [req-f2d3f820-d001-41c6-a6b6-62b46c806215 req-7d04764a-e284-473e-ae34-4a8d9efecdee service nova] [instance: 30b8b872-40ca-4297-b98e-a64c83a75483] Refreshing network info cache for port b8fe94ee-46b7-489a-819d-0699bc02bbb7 {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 757.758608] env[63515]: DEBUG oslo_vmware.api [None req-78f15907-fa94-44dd-a2b3-ccdc7dec0a0f tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Task: {'id': task-1111113, 'name': CloneVM_Task} progress is 94%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 757.848089] env[63515]: DEBUG oslo_vmware.api [None req-750d4038-2a95-4b14-b9d7-c9c6803c64b6 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] Task: {'id': task-1111114, 'name': ReconfigVM_Task, 'duration_secs': 0.759522} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 757.848477] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-750d4038-2a95-4b14-b9d7-c9c6803c64b6 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] [instance: 8eb17506-ff93-4d25-b9af-ec5886569e65] Reconfigured VM instance instance-00000034 to attach disk [datastore2] 8eb17506-ff93-4d25-b9af-ec5886569e65/8eb17506-ff93-4d25-b9af-ec5886569e65.vmdk or device None with type sparse {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 757.851944] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-5f3e33f7-4930-42e9-b6f8-0b137500d6ec {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.858656] env[63515]: DEBUG oslo_vmware.api [None req-0e5835c9-1a06-4538-91e4-48a9dfc2854d tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Task: {'id': task-1111115, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.072588} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 757.862063] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-0e5835c9-1a06-4538-91e4-48a9dfc2854d tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: 001a48e1-b237-41f8-a6f4-2fe98d173512] Extended root virtual disk {{(pid=63515) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 757.862285] env[63515]: DEBUG nova.network.neutron [-] [instance: 37e6c27e-317b-45d2-bd55-2fd78ccf009f] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 757.863363] env[63515]: DEBUG oslo_vmware.api [None req-750d4038-2a95-4b14-b9d7-c9c6803c64b6 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] Waiting for the task: (returnval){ [ 757.863363] env[63515]: value = "task-1111116" [ 757.863363] env[63515]: _type = "Task" [ 757.863363] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 757.866271] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-846420fe-00cd-43b4-9e16-235858ec0590 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.895561] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-0e5835c9-1a06-4538-91e4-48a9dfc2854d tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: 001a48e1-b237-41f8-a6f4-2fe98d173512] Reconfiguring VM instance instance-00000030 to attach disk [datastore2] 001a48e1-b237-41f8-a6f4-2fe98d173512/001a48e1-b237-41f8-a6f4-2fe98d173512.vmdk or device None with type sparse {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 757.898082] env[63515]: DEBUG oslo_concurrency.lockutils [None req-4b5baf79-d4c3-467e-88d4-a131ea6620d7 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 757.900767] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5b561af2-f907-4c2d-99f0-8c6185a42718 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.925134] env[63515]: DEBUG oslo_vmware.api [None req-750d4038-2a95-4b14-b9d7-c9c6803c64b6 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] Task: {'id': task-1111116, 'name': Rename_Task} progress is 14%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 757.932217] env[63515]: DEBUG oslo_vmware.api [None req-0e5835c9-1a06-4538-91e4-48a9dfc2854d tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Waiting for the task: (returnval){ [ 757.932217] env[63515]: value = "task-1111117" [ 757.932217] env[63515]: _type = "Task" [ 757.932217] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 757.949079] env[63515]: DEBUG oslo_vmware.api [None req-0e5835c9-1a06-4538-91e4-48a9dfc2854d tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Task: {'id': task-1111117, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 758.092445] env[63515]: INFO nova.compute.manager [None req-dfa187f2-adf0-4e24-9cf4-7df6d0016e12 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] [instance: 38d6d030-06b0-4185-904d-44a038b3a752] Took 33.85 seconds to build instance. [ 758.103143] env[63515]: DEBUG oslo_concurrency.lockutils [None req-64ffe526-19dd-467f-8550-ae608ae5d75b tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 758.209721] env[63515]: DEBUG oslo_concurrency.lockutils [None req-b8b2a7f6-808d-40d4-8de9-c8863e5b55c5 tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] Acquiring lock "refresh_cache-30b8b872-40ca-4297-b98e-a64c83a75483" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 758.240860] env[63515]: DEBUG nova.compute.utils [None req-8f7a1df0-5f0d-424a-ae8f-76bee2ee3dbb tempest-ServersAaction247Test-2019176257 tempest-ServersAaction247Test-2019176257-project-member] Using /dev/sd instead of None {{(pid=63515) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 758.253733] env[63515]: DEBUG nova.compute.manager [None req-8f7a1df0-5f0d-424a-ae8f-76bee2ee3dbb tempest-ServersAaction247Test-2019176257 tempest-ServersAaction247Test-2019176257-project-member] [instance: 5d50fbd6-a7fb-422c-9dd7-df7140fd0aae] Not allocating networking since 'none' was specified. {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 758.265052] env[63515]: DEBUG oslo_vmware.api [None req-78f15907-fa94-44dd-a2b3-ccdc7dec0a0f tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Task: {'id': task-1111113, 'name': CloneVM_Task} progress is 95%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 758.301246] env[63515]: DEBUG nova.network.neutron [req-f2d3f820-d001-41c6-a6b6-62b46c806215 req-7d04764a-e284-473e-ae34-4a8d9efecdee service nova] [instance: 30b8b872-40ca-4297-b98e-a64c83a75483] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 758.366625] env[63515]: INFO nova.compute.manager [-] [instance: 37e6c27e-317b-45d2-bd55-2fd78ccf009f] Took 1.31 seconds to deallocate network for instance. [ 758.381181] env[63515]: DEBUG oslo_vmware.api [None req-750d4038-2a95-4b14-b9d7-c9c6803c64b6 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] Task: {'id': task-1111116, 'name': Rename_Task, 'duration_secs': 0.14786} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 758.381470] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-750d4038-2a95-4b14-b9d7-c9c6803c64b6 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] [instance: 8eb17506-ff93-4d25-b9af-ec5886569e65] Powering on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 758.381796] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a1c97a1a-8e3a-400e-baf4-cbe16b2f029e {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.387953] env[63515]: DEBUG oslo_vmware.api [None req-750d4038-2a95-4b14-b9d7-c9c6803c64b6 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] Waiting for the task: (returnval){ [ 758.387953] env[63515]: value = "task-1111118" [ 758.387953] env[63515]: _type = "Task" [ 758.387953] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 758.396581] env[63515]: DEBUG oslo_vmware.api [None req-750d4038-2a95-4b14-b9d7-c9c6803c64b6 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] Task: {'id': task-1111118, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 758.426484] env[63515]: DEBUG nova.network.neutron [req-f2d3f820-d001-41c6-a6b6-62b46c806215 req-7d04764a-e284-473e-ae34-4a8d9efecdee service nova] [instance: 30b8b872-40ca-4297-b98e-a64c83a75483] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 758.443584] env[63515]: DEBUG oslo_vmware.api [None req-0e5835c9-1a06-4538-91e4-48a9dfc2854d tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Task: {'id': task-1111117, 'name': ReconfigVM_Task} progress is 99%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 758.594961] env[63515]: DEBUG oslo_concurrency.lockutils [None req-dfa187f2-adf0-4e24-9cf4-7df6d0016e12 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Lock "38d6d030-06b0-4185-904d-44a038b3a752" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 118.759s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 758.646388] env[63515]: DEBUG nova.compute.manager [req-bffd337f-99ae-46bf-a9cd-915152e4e9ea req-4bf3656e-eef9-4540-8e4a-77a0f4ade648 service nova] [instance: 37e6c27e-317b-45d2-bd55-2fd78ccf009f] Received event network-vif-deleted-21c8f7a2-4403-41cd-93f8-3091de7cdc40 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 758.752295] env[63515]: DEBUG nova.compute.manager [None req-8f7a1df0-5f0d-424a-ae8f-76bee2ee3dbb tempest-ServersAaction247Test-2019176257 tempest-ServersAaction247Test-2019176257-project-member] [instance: 5d50fbd6-a7fb-422c-9dd7-df7140fd0aae] Start building block device mappings for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 758.755715] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Applying migration context for instance f1d01b75-ac9d-458d-8cc2-ae64cffca4e8 as it has an incoming, in-progress migration 802930d6-330b-4516-9996-d196743b8ab8. Migration status is reverting {{(pid=63515) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1016}} [ 758.757837] env[63515]: INFO nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] [instance: f1d01b75-ac9d-458d-8cc2-ae64cffca4e8] Updating resource usage from migration 802930d6-330b-4516-9996-d196743b8ab8 [ 758.778715] env[63515]: DEBUG oslo_vmware.api [None req-78f15907-fa94-44dd-a2b3-ccdc7dec0a0f tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Task: {'id': task-1111113, 'name': CloneVM_Task, 'duration_secs': 1.843656} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 758.778715] env[63515]: INFO nova.virt.vmwareapi.vmops [None req-78f15907-fa94-44dd-a2b3-ccdc7dec0a0f tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] [instance: 87c468d9-9594-4804-b461-527f01f6118f] Created linked-clone VM from snapshot [ 758.779463] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f3d38ff-f7e3-47ee-99b8-79e3138a1d26 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.788894] env[63515]: DEBUG nova.virt.vmwareapi.images [None req-78f15907-fa94-44dd-a2b3-ccdc7dec0a0f tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] [instance: 87c468d9-9594-4804-b461-527f01f6118f] Uploading image 499b6029-07e9-499c-80d5-a53d8c43ffb7 {{(pid=63515) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 758.791958] env[63515]: WARNING nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Instance b911a5b5-9617-4fb3-9b5e-fb8c492e4931 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 758.791958] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Instance 5b42f744-fdd6-45b1-8563-896869648c23 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63515) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 758.791958] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Instance 90f4930b-aaa0-4c4b-9ab8-92aed45e200b actively managed on this compute host and has allocations in placement: {'resources': {'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63515) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 758.792205] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Instance 396e49dd-48c0-496b-a1ec-190c33a22c5e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63515) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 758.792205] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Instance 87c468d9-9594-4804-b461-527f01f6118f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63515) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 758.792359] env[63515]: WARNING nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Instance 452f3ca2-6141-43b2-a77a-c9ab5754192d is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 758.792558] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Instance 37e6c27e-317b-45d2-bd55-2fd78ccf009f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63515) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 758.792755] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Instance b4477e66-ae12-4929-90ed-b7b652e0f207 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63515) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 758.792901] env[63515]: WARNING nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Instance b85e9a70-7f5b-4d32-b616-f2a97e3186c8 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 758.793031] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Instance d7e8dcdb-41cb-46fb-8b61-d251e7c2d372 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63515) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 758.793149] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Instance e91aa479-1540-4950-851b-b2409e5f89f1 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63515) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 758.793261] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Instance ed550b10-d58f-45b8-b766-198f431c3788 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63515) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 758.793383] env[63515]: WARNING nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Instance de210780-5c0f-4fba-883c-13707566a2e1 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 758.793494] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Instance 001a48e1-b237-41f8-a6f4-2fe98d173512 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63515) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 758.793604] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Instance d2a4f66a-f1a2-4f0d-826d-0b7c4f04bd23 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63515) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 758.793712] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Instance 6688cdaa-29ba-413a-8131-4f834cdb70e4 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63515) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 758.793820] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Instance 38d6d030-06b0-4185-904d-44a038b3a752 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=63515) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 758.793926] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Instance 8eb17506-ff93-4d25-b9af-ec5886569e65 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63515) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 758.794046] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Instance 30b8b872-40ca-4297-b98e-a64c83a75483 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63515) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 758.794145] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Instance 5d50fbd6-a7fb-422c-9dd7-df7140fd0aae actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63515) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 758.813197] env[63515]: DEBUG oslo_vmware.rw_handles [None req-78f15907-fa94-44dd-a2b3-ccdc7dec0a0f tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 758.813197] env[63515]: value = "vm-243441" [ 758.813197] env[63515]: _type = "VirtualMachine" [ 758.813197] env[63515]: }. {{(pid=63515) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 758.813462] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-0867d001-d1b4-4f81-b382-e257d9ef4e78 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.820363] env[63515]: DEBUG oslo_vmware.rw_handles [None req-78f15907-fa94-44dd-a2b3-ccdc7dec0a0f tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Lease: (returnval){ [ 758.820363] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]5229f9bb-bd3f-2ca0-4da9-7fa6cd04e36b" [ 758.820363] env[63515]: _type = "HttpNfcLease" [ 758.820363] env[63515]: } obtained for exporting VM: (result){ [ 758.820363] env[63515]: value = "vm-243441" [ 758.820363] env[63515]: _type = "VirtualMachine" [ 758.820363] env[63515]: }. {{(pid=63515) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 758.820725] env[63515]: DEBUG oslo_vmware.api [None req-78f15907-fa94-44dd-a2b3-ccdc7dec0a0f tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Waiting for the lease: (returnval){ [ 758.820725] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]5229f9bb-bd3f-2ca0-4da9-7fa6cd04e36b" [ 758.820725] env[63515]: _type = "HttpNfcLease" [ 758.820725] env[63515]: } to be ready. {{(pid=63515) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 758.827693] env[63515]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 758.827693] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]5229f9bb-bd3f-2ca0-4da9-7fa6cd04e36b" [ 758.827693] env[63515]: _type = "HttpNfcLease" [ 758.827693] env[63515]: } is initializing. {{(pid=63515) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 758.877688] env[63515]: DEBUG oslo_concurrency.lockutils [None req-938cccb1-5a46-437d-9e93-3058a8134cd5 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 758.898115] env[63515]: DEBUG oslo_vmware.api [None req-750d4038-2a95-4b14-b9d7-c9c6803c64b6 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] Task: {'id': task-1111118, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 758.931393] env[63515]: DEBUG oslo_concurrency.lockutils [req-f2d3f820-d001-41c6-a6b6-62b46c806215 req-7d04764a-e284-473e-ae34-4a8d9efecdee service nova] Releasing lock "refresh_cache-30b8b872-40ca-4297-b98e-a64c83a75483" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 758.931831] env[63515]: DEBUG oslo_concurrency.lockutils [None req-b8b2a7f6-808d-40d4-8de9-c8863e5b55c5 tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] Acquired lock "refresh_cache-30b8b872-40ca-4297-b98e-a64c83a75483" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 758.931987] env[63515]: DEBUG nova.network.neutron [None req-b8b2a7f6-808d-40d4-8de9-c8863e5b55c5 tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] [instance: 30b8b872-40ca-4297-b98e-a64c83a75483] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 758.944400] env[63515]: DEBUG oslo_vmware.api [None req-0e5835c9-1a06-4538-91e4-48a9dfc2854d tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Task: {'id': task-1111117, 'name': ReconfigVM_Task, 'duration_secs': 0.557565} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 758.945267] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-0e5835c9-1a06-4538-91e4-48a9dfc2854d tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: 001a48e1-b237-41f8-a6f4-2fe98d173512] Reconfigured VM instance instance-00000030 to attach disk [datastore2] 001a48e1-b237-41f8-a6f4-2fe98d173512/001a48e1-b237-41f8-a6f4-2fe98d173512.vmdk or device None with type sparse {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 758.945883] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6ded8c45-c343-40f5-8b84-d98c8fa3235b {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.952819] env[63515]: DEBUG oslo_vmware.api [None req-0e5835c9-1a06-4538-91e4-48a9dfc2854d tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Waiting for the task: (returnval){ [ 758.952819] env[63515]: value = "task-1111120" [ 758.952819] env[63515]: _type = "Task" [ 758.952819] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 758.961109] env[63515]: DEBUG oslo_vmware.api [None req-0e5835c9-1a06-4538-91e4-48a9dfc2854d tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Task: {'id': task-1111120, 'name': Rename_Task} progress is 5%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 759.098069] env[63515]: DEBUG nova.compute.manager [None req-66c3e202-5a3e-433e-b543-4bfb02c8ad30 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775-project-member] [instance: 4e453127-1f3e-40ea-819f-6678479826c8] Starting instance... {{(pid=63515) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 759.298156] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Instance 84dbf321-2c70-4e08-b430-cb5a06fc6829 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63515) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 759.330186] env[63515]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 759.330186] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]5229f9bb-bd3f-2ca0-4da9-7fa6cd04e36b" [ 759.330186] env[63515]: _type = "HttpNfcLease" [ 759.330186] env[63515]: } is ready. {{(pid=63515) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 759.330622] env[63515]: DEBUG oslo_vmware.rw_handles [None req-78f15907-fa94-44dd-a2b3-ccdc7dec0a0f tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 759.330622] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]5229f9bb-bd3f-2ca0-4da9-7fa6cd04e36b" [ 759.330622] env[63515]: _type = "HttpNfcLease" [ 759.330622] env[63515]: }. {{(pid=63515) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 759.331290] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fcc1fe20-4039-4564-8bc1-4f6c5b40f624 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.340128] env[63515]: DEBUG oslo_vmware.rw_handles [None req-78f15907-fa94-44dd-a2b3-ccdc7dec0a0f tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52dac88e-b167-9d0e-e4fa-7c0241d36201/disk-0.vmdk from lease info. {{(pid=63515) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 759.340360] env[63515]: DEBUG oslo_vmware.rw_handles [None req-78f15907-fa94-44dd-a2b3-ccdc7dec0a0f tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52dac88e-b167-9d0e-e4fa-7c0241d36201/disk-0.vmdk for reading. {{(pid=63515) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 759.412265] env[63515]: DEBUG oslo_vmware.api [None req-750d4038-2a95-4b14-b9d7-c9c6803c64b6 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] Task: {'id': task-1111118, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 759.452683] env[63515]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-9f820e8f-56a8-4764-b07f-474468bb1994 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.465725] env[63515]: DEBUG oslo_vmware.api [None req-0e5835c9-1a06-4538-91e4-48a9dfc2854d tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Task: {'id': task-1111120, 'name': Rename_Task, 'duration_secs': 0.356375} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 759.466110] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-0e5835c9-1a06-4538-91e4-48a9dfc2854d tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: 001a48e1-b237-41f8-a6f4-2fe98d173512] Powering on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 759.466347] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-056277cc-14e0-4ada-85ec-f8e4e4469ce7 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.474199] env[63515]: DEBUG oslo_vmware.api [None req-0e5835c9-1a06-4538-91e4-48a9dfc2854d tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Waiting for the task: (returnval){ [ 759.474199] env[63515]: value = "task-1111121" [ 759.474199] env[63515]: _type = "Task" [ 759.474199] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 759.485979] env[63515]: DEBUG oslo_vmware.api [None req-0e5835c9-1a06-4538-91e4-48a9dfc2854d tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Task: {'id': task-1111121, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 759.486863] env[63515]: DEBUG nova.network.neutron [None req-b8b2a7f6-808d-40d4-8de9-c8863e5b55c5 tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] [instance: 30b8b872-40ca-4297-b98e-a64c83a75483] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 759.624573] env[63515]: DEBUG oslo_concurrency.lockutils [None req-66c3e202-5a3e-433e-b543-4bfb02c8ad30 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 759.681324] env[63515]: DEBUG nova.network.neutron [None req-b8b2a7f6-808d-40d4-8de9-c8863e5b55c5 tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] [instance: 30b8b872-40ca-4297-b98e-a64c83a75483] Updating instance_info_cache with network_info: [{"id": "b8fe94ee-46b7-489a-819d-0699bc02bbb7", "address": "fa:16:3e:42:a9:82", "network": {"id": "bd2f642c-518b-478b-8358-7426cebde731", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-1534671780-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "72b65c09ed6d4492864022352180ff49", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "92f3cfd6-c130-4390-8910-865fbc42afd1", "external-id": "nsx-vlan-transportzone-142", "segmentation_id": 142, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb8fe94ee-46", "ovs_interfaceid": "b8fe94ee-46b7-489a-819d-0699bc02bbb7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 759.766393] env[63515]: DEBUG nova.compute.manager [None req-8f7a1df0-5f0d-424a-ae8f-76bee2ee3dbb tempest-ServersAaction247Test-2019176257 tempest-ServersAaction247Test-2019176257-project-member] [instance: 5d50fbd6-a7fb-422c-9dd7-df7140fd0aae] Start spawning the instance on the hypervisor. {{(pid=63515) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 759.794413] env[63515]: DEBUG nova.virt.hardware [None req-8f7a1df0-5f0d-424a-ae8f-76bee2ee3dbb tempest-ServersAaction247Test-2019176257 tempest-ServersAaction247Test-2019176257-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T02:52:14Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T02:51:56Z,direct_url=,disk_format='vmdk',id=8a120570-cb06-4099-b262-554ca0ad15c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b569255cc43e42e7ae3f2b4ad37c6ef4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T02:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 759.794413] env[63515]: DEBUG nova.virt.hardware [None req-8f7a1df0-5f0d-424a-ae8f-76bee2ee3dbb tempest-ServersAaction247Test-2019176257 tempest-ServersAaction247Test-2019176257-project-member] Flavor limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 759.795113] env[63515]: DEBUG nova.virt.hardware [None req-8f7a1df0-5f0d-424a-ae8f-76bee2ee3dbb tempest-ServersAaction247Test-2019176257 tempest-ServersAaction247Test-2019176257-project-member] Image limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 759.795464] env[63515]: DEBUG nova.virt.hardware [None req-8f7a1df0-5f0d-424a-ae8f-76bee2ee3dbb tempest-ServersAaction247Test-2019176257 tempest-ServersAaction247Test-2019176257-project-member] Flavor pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 759.795795] env[63515]: DEBUG nova.virt.hardware [None req-8f7a1df0-5f0d-424a-ae8f-76bee2ee3dbb tempest-ServersAaction247Test-2019176257 tempest-ServersAaction247Test-2019176257-project-member] Image pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 759.796221] env[63515]: DEBUG nova.virt.hardware [None req-8f7a1df0-5f0d-424a-ae8f-76bee2ee3dbb tempest-ServersAaction247Test-2019176257 tempest-ServersAaction247Test-2019176257-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 759.799019] env[63515]: DEBUG nova.virt.hardware [None req-8f7a1df0-5f0d-424a-ae8f-76bee2ee3dbb tempest-ServersAaction247Test-2019176257 tempest-ServersAaction247Test-2019176257-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 759.799019] env[63515]: DEBUG nova.virt.hardware [None req-8f7a1df0-5f0d-424a-ae8f-76bee2ee3dbb tempest-ServersAaction247Test-2019176257 tempest-ServersAaction247Test-2019176257-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 759.799019] env[63515]: DEBUG nova.virt.hardware [None req-8f7a1df0-5f0d-424a-ae8f-76bee2ee3dbb tempest-ServersAaction247Test-2019176257 tempest-ServersAaction247Test-2019176257-project-member] Got 1 possible topologies {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 759.799019] env[63515]: DEBUG nova.virt.hardware [None req-8f7a1df0-5f0d-424a-ae8f-76bee2ee3dbb tempest-ServersAaction247Test-2019176257 tempest-ServersAaction247Test-2019176257-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 759.799019] env[63515]: DEBUG nova.virt.hardware [None req-8f7a1df0-5f0d-424a-ae8f-76bee2ee3dbb tempest-ServersAaction247Test-2019176257 tempest-ServersAaction247Test-2019176257-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 759.799531] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f16905e9-6007-48c5-bb50-7f0c7ecfa2a0 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.802830] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Instance 70ed982f-affd-4dd1-bc90-c64e7c6d49d2 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63515) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 759.816161] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e95895e7-db0e-4b14-9956-5d9a94aacaa5 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.832992] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-8f7a1df0-5f0d-424a-ae8f-76bee2ee3dbb tempest-ServersAaction247Test-2019176257 tempest-ServersAaction247Test-2019176257-project-member] [instance: 5d50fbd6-a7fb-422c-9dd7-df7140fd0aae] Instance VIF info [] {{(pid=63515) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 759.839090] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-8f7a1df0-5f0d-424a-ae8f-76bee2ee3dbb tempest-ServersAaction247Test-2019176257 tempest-ServersAaction247Test-2019176257-project-member] Creating folder: Project (6e76c6d5d24c4b9381de103981bf709a). Parent ref: group-v243370. {{(pid=63515) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 759.839785] env[63515]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-0600a514-eb4c-4c2b-8a14-93f0b908ce90 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.853852] env[63515]: INFO nova.virt.vmwareapi.vm_util [None req-8f7a1df0-5f0d-424a-ae8f-76bee2ee3dbb tempest-ServersAaction247Test-2019176257 tempest-ServersAaction247Test-2019176257-project-member] Created folder: Project (6e76c6d5d24c4b9381de103981bf709a) in parent group-v243370. [ 759.854146] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-8f7a1df0-5f0d-424a-ae8f-76bee2ee3dbb tempest-ServersAaction247Test-2019176257 tempest-ServersAaction247Test-2019176257-project-member] Creating folder: Instances. Parent ref: group-v243442. {{(pid=63515) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 759.854844] env[63515]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-43a7cde0-7bd5-4998-80ac-129b777d32be {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.868069] env[63515]: INFO nova.virt.vmwareapi.vm_util [None req-8f7a1df0-5f0d-424a-ae8f-76bee2ee3dbb tempest-ServersAaction247Test-2019176257 tempest-ServersAaction247Test-2019176257-project-member] Created folder: Instances in parent group-v243442. [ 759.868069] env[63515]: DEBUG oslo.service.loopingcall [None req-8f7a1df0-5f0d-424a-ae8f-76bee2ee3dbb tempest-ServersAaction247Test-2019176257 tempest-ServersAaction247Test-2019176257-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 759.868359] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5d50fbd6-a7fb-422c-9dd7-df7140fd0aae] Creating VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 759.868424] env[63515]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-131b37df-b5f9-45ea-897f-c89eba5246c3 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.895775] env[63515]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 759.895775] env[63515]: value = "task-1111124" [ 759.895775] env[63515]: _type = "Task" [ 759.895775] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 759.909060] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1111124, 'name': CreateVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 759.916117] env[63515]: DEBUG oslo_vmware.api [None req-750d4038-2a95-4b14-b9d7-c9c6803c64b6 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] Task: {'id': task-1111118, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 759.990105] env[63515]: DEBUG oslo_vmware.api [None req-0e5835c9-1a06-4538-91e4-48a9dfc2854d tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Task: {'id': task-1111121, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 760.185249] env[63515]: DEBUG oslo_concurrency.lockutils [None req-b8b2a7f6-808d-40d4-8de9-c8863e5b55c5 tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] Releasing lock "refresh_cache-30b8b872-40ca-4297-b98e-a64c83a75483" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 760.185856] env[63515]: DEBUG nova.compute.manager [None req-b8b2a7f6-808d-40d4-8de9-c8863e5b55c5 tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] [instance: 30b8b872-40ca-4297-b98e-a64c83a75483] Instance network_info: |[{"id": "b8fe94ee-46b7-489a-819d-0699bc02bbb7", "address": "fa:16:3e:42:a9:82", "network": {"id": "bd2f642c-518b-478b-8358-7426cebde731", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-1534671780-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "72b65c09ed6d4492864022352180ff49", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "92f3cfd6-c130-4390-8910-865fbc42afd1", "external-id": "nsx-vlan-transportzone-142", "segmentation_id": 142, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb8fe94ee-46", "ovs_interfaceid": "b8fe94ee-46b7-489a-819d-0699bc02bbb7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 760.186498] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-b8b2a7f6-808d-40d4-8de9-c8863e5b55c5 tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] [instance: 30b8b872-40ca-4297-b98e-a64c83a75483] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:42:a9:82', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '92f3cfd6-c130-4390-8910-865fbc42afd1', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b8fe94ee-46b7-489a-819d-0699bc02bbb7', 'vif_model': 'vmxnet3'}] {{(pid=63515) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 760.199277] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-b8b2a7f6-808d-40d4-8de9-c8863e5b55c5 tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] Creating folder: Project (72b65c09ed6d4492864022352180ff49). Parent ref: group-v243370. {{(pid=63515) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 760.200496] env[63515]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b2fa3a41-c4ab-4699-8598-2f43f8ee89e2 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.219659] env[63515]: INFO nova.virt.vmwareapi.vm_util [None req-b8b2a7f6-808d-40d4-8de9-c8863e5b55c5 tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] Created folder: Project (72b65c09ed6d4492864022352180ff49) in parent group-v243370. [ 760.220327] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-b8b2a7f6-808d-40d4-8de9-c8863e5b55c5 tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] Creating folder: Instances. Parent ref: group-v243445. {{(pid=63515) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 760.221146] env[63515]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-67c7d09c-4f13-4e69-b839-f284ea713650 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.247866] env[63515]: INFO nova.virt.vmwareapi.vm_util [None req-b8b2a7f6-808d-40d4-8de9-c8863e5b55c5 tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] Created folder: Instances in parent group-v243445. [ 760.248771] env[63515]: DEBUG oslo.service.loopingcall [None req-b8b2a7f6-808d-40d4-8de9-c8863e5b55c5 tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 760.249460] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 30b8b872-40ca-4297-b98e-a64c83a75483] Creating VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 760.250129] env[63515]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c4c7f8ec-e216-47f3-a6bd-0004ffd2be12 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.275303] env[63515]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 760.275303] env[63515]: value = "task-1111127" [ 760.275303] env[63515]: _type = "Task" [ 760.275303] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 760.284880] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1111127, 'name': CreateVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 760.306691] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Instance 192137b0-03e5-4bc4-b911-4b4f1a874f74 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63515) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 760.390556] env[63515]: DEBUG oslo_concurrency.lockutils [None req-2e592662-996b-4dfa-8941-00965c1317fc tempest-ServersTestBootFromVolume-2042560423 tempest-ServersTestBootFromVolume-2042560423-project-member] Acquiring lock "90f4930b-aaa0-4c4b-9ab8-92aed45e200b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 760.390905] env[63515]: DEBUG oslo_concurrency.lockutils [None req-2e592662-996b-4dfa-8941-00965c1317fc tempest-ServersTestBootFromVolume-2042560423 tempest-ServersTestBootFromVolume-2042560423-project-member] Lock "90f4930b-aaa0-4c4b-9ab8-92aed45e200b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 760.391082] env[63515]: DEBUG oslo_concurrency.lockutils [None req-2e592662-996b-4dfa-8941-00965c1317fc tempest-ServersTestBootFromVolume-2042560423 tempest-ServersTestBootFromVolume-2042560423-project-member] Acquiring lock "90f4930b-aaa0-4c4b-9ab8-92aed45e200b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 760.391292] env[63515]: DEBUG oslo_concurrency.lockutils [None req-2e592662-996b-4dfa-8941-00965c1317fc tempest-ServersTestBootFromVolume-2042560423 tempest-ServersTestBootFromVolume-2042560423-project-member] Lock "90f4930b-aaa0-4c4b-9ab8-92aed45e200b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 760.391503] env[63515]: DEBUG oslo_concurrency.lockutils [None req-2e592662-996b-4dfa-8941-00965c1317fc tempest-ServersTestBootFromVolume-2042560423 tempest-ServersTestBootFromVolume-2042560423-project-member] Lock "90f4930b-aaa0-4c4b-9ab8-92aed45e200b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 760.396323] env[63515]: INFO nova.compute.manager [None req-2e592662-996b-4dfa-8941-00965c1317fc tempest-ServersTestBootFromVolume-2042560423 tempest-ServersTestBootFromVolume-2042560423-project-member] [instance: 90f4930b-aaa0-4c4b-9ab8-92aed45e200b] Terminating instance [ 760.404085] env[63515]: DEBUG nova.compute.manager [None req-2e592662-996b-4dfa-8941-00965c1317fc tempest-ServersTestBootFromVolume-2042560423 tempest-ServersTestBootFromVolume-2042560423-project-member] [instance: 90f4930b-aaa0-4c4b-9ab8-92aed45e200b] Start destroying the instance on the hypervisor. {{(pid=63515) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 760.404436] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-2e592662-996b-4dfa-8941-00965c1317fc tempest-ServersTestBootFromVolume-2042560423 tempest-ServersTestBootFromVolume-2042560423-project-member] [instance: 90f4930b-aaa0-4c4b-9ab8-92aed45e200b] Powering off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 760.405073] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d4f729f6-74e7-4584-94fd-02044a2a9d3e {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.418966] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1111124, 'name': CreateVM_Task, 'duration_secs': 0.385437} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 760.423974] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5d50fbd6-a7fb-422c-9dd7-df7140fd0aae] Created VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 760.424629] env[63515]: DEBUG oslo_vmware.api [None req-750d4038-2a95-4b14-b9d7-c9c6803c64b6 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] Task: {'id': task-1111118, 'name': PowerOnVM_Task, 'duration_secs': 1.802743} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 760.425152] env[63515]: DEBUG oslo_vmware.api [None req-2e592662-996b-4dfa-8941-00965c1317fc tempest-ServersTestBootFromVolume-2042560423 tempest-ServersTestBootFromVolume-2042560423-project-member] Waiting for the task: (returnval){ [ 760.425152] env[63515]: value = "task-1111128" [ 760.425152] env[63515]: _type = "Task" [ 760.425152] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 760.426492] env[63515]: DEBUG oslo_concurrency.lockutils [None req-8f7a1df0-5f0d-424a-ae8f-76bee2ee3dbb tempest-ServersAaction247Test-2019176257 tempest-ServersAaction247Test-2019176257-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 760.426492] env[63515]: DEBUG oslo_concurrency.lockutils [None req-8f7a1df0-5f0d-424a-ae8f-76bee2ee3dbb tempest-ServersAaction247Test-2019176257 tempest-ServersAaction247Test-2019176257-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 760.426492] env[63515]: DEBUG oslo_concurrency.lockutils [None req-8f7a1df0-5f0d-424a-ae8f-76bee2ee3dbb tempest-ServersAaction247Test-2019176257 tempest-ServersAaction247Test-2019176257-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 760.426724] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-750d4038-2a95-4b14-b9d7-c9c6803c64b6 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] [instance: 8eb17506-ff93-4d25-b9af-ec5886569e65] Powered on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 760.427017] env[63515]: INFO nova.compute.manager [None req-750d4038-2a95-4b14-b9d7-c9c6803c64b6 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] [instance: 8eb17506-ff93-4d25-b9af-ec5886569e65] Took 7.86 seconds to spawn the instance on the hypervisor. [ 760.427648] env[63515]: DEBUG nova.compute.manager [None req-750d4038-2a95-4b14-b9d7-c9c6803c64b6 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] [instance: 8eb17506-ff93-4d25-b9af-ec5886569e65] Checking state {{(pid=63515) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 760.427986] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-49bf08d8-f3e1-41ac-b23a-68960b2a766c {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.430837] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-696c8616-cd09-4beb-a364-c66c618d3200 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.441184] env[63515]: DEBUG oslo_vmware.api [None req-8f7a1df0-5f0d-424a-ae8f-76bee2ee3dbb tempest-ServersAaction247Test-2019176257 tempest-ServersAaction247Test-2019176257-project-member] Waiting for the task: (returnval){ [ 760.441184] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]521a193e-40b8-3686-6928-4de9421d6ce4" [ 760.441184] env[63515]: _type = "Task" [ 760.441184] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 760.446591] env[63515]: DEBUG oslo_vmware.api [None req-2e592662-996b-4dfa-8941-00965c1317fc tempest-ServersTestBootFromVolume-2042560423 tempest-ServersTestBootFromVolume-2042560423-project-member] Task: {'id': task-1111128, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 760.464162] env[63515]: DEBUG oslo_vmware.api [None req-8f7a1df0-5f0d-424a-ae8f-76bee2ee3dbb tempest-ServersAaction247Test-2019176257 tempest-ServersAaction247Test-2019176257-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]521a193e-40b8-3686-6928-4de9421d6ce4, 'name': SearchDatastore_Task, 'duration_secs': 0.013175} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 760.464738] env[63515]: DEBUG oslo_concurrency.lockutils [None req-8f7a1df0-5f0d-424a-ae8f-76bee2ee3dbb tempest-ServersAaction247Test-2019176257 tempest-ServersAaction247Test-2019176257-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 760.465147] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-8f7a1df0-5f0d-424a-ae8f-76bee2ee3dbb tempest-ServersAaction247Test-2019176257 tempest-ServersAaction247Test-2019176257-project-member] [instance: 5d50fbd6-a7fb-422c-9dd7-df7140fd0aae] Processing image 8a120570-cb06-4099-b262-554ca0ad15c5 {{(pid=63515) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 760.465565] env[63515]: DEBUG oslo_concurrency.lockutils [None req-8f7a1df0-5f0d-424a-ae8f-76bee2ee3dbb tempest-ServersAaction247Test-2019176257 tempest-ServersAaction247Test-2019176257-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 760.465918] env[63515]: DEBUG oslo_concurrency.lockutils [None req-8f7a1df0-5f0d-424a-ae8f-76bee2ee3dbb tempest-ServersAaction247Test-2019176257 tempest-ServersAaction247Test-2019176257-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 760.466276] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-8f7a1df0-5f0d-424a-ae8f-76bee2ee3dbb tempest-ServersAaction247Test-2019176257 tempest-ServersAaction247Test-2019176257-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 760.466742] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-73961de5-64c9-4ed8-8625-38f52adf6d40 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.483236] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-8f7a1df0-5f0d-424a-ae8f-76bee2ee3dbb tempest-ServersAaction247Test-2019176257 tempest-ServersAaction247Test-2019176257-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 760.483494] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-8f7a1df0-5f0d-424a-ae8f-76bee2ee3dbb tempest-ServersAaction247Test-2019176257 tempest-ServersAaction247Test-2019176257-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63515) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 760.485111] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3ae3da09-0949-4e68-95f6-19ff4858d64b {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.491588] env[63515]: DEBUG oslo_vmware.api [None req-0e5835c9-1a06-4538-91e4-48a9dfc2854d tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Task: {'id': task-1111121, 'name': PowerOnVM_Task, 'duration_secs': 0.782384} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 760.492227] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-0e5835c9-1a06-4538-91e4-48a9dfc2854d tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: 001a48e1-b237-41f8-a6f4-2fe98d173512] Powered on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 760.492227] env[63515]: DEBUG nova.compute.manager [None req-0e5835c9-1a06-4538-91e4-48a9dfc2854d tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: 001a48e1-b237-41f8-a6f4-2fe98d173512] Checking state {{(pid=63515) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 760.493682] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41ea9e9d-9be7-495d-8d9e-7e7d2e013274 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.498009] env[63515]: DEBUG oslo_vmware.api [None req-8f7a1df0-5f0d-424a-ae8f-76bee2ee3dbb tempest-ServersAaction247Test-2019176257 tempest-ServersAaction247Test-2019176257-project-member] Waiting for the task: (returnval){ [ 760.498009] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]52449585-264e-78a9-fbc9-575f74830769" [ 760.498009] env[63515]: _type = "Task" [ 760.498009] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 760.514711] env[63515]: DEBUG oslo_vmware.api [None req-8f7a1df0-5f0d-424a-ae8f-76bee2ee3dbb tempest-ServersAaction247Test-2019176257 tempest-ServersAaction247Test-2019176257-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52449585-264e-78a9-fbc9-575f74830769, 'name': SearchDatastore_Task, 'duration_secs': 0.013869} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 760.514973] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b904d3d1-b7b5-4732-88a1-3b0fdbe4d528 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.524672] env[63515]: DEBUG oslo_vmware.api [None req-8f7a1df0-5f0d-424a-ae8f-76bee2ee3dbb tempest-ServersAaction247Test-2019176257 tempest-ServersAaction247Test-2019176257-project-member] Waiting for the task: (returnval){ [ 760.524672] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]52f3d976-5b03-5926-59ab-0bdbc8729ee3" [ 760.524672] env[63515]: _type = "Task" [ 760.524672] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 760.535704] env[63515]: DEBUG oslo_vmware.api [None req-8f7a1df0-5f0d-424a-ae8f-76bee2ee3dbb tempest-ServersAaction247Test-2019176257 tempest-ServersAaction247Test-2019176257-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52f3d976-5b03-5926-59ab-0bdbc8729ee3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 760.787323] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1111127, 'name': CreateVM_Task, 'duration_secs': 0.438407} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 760.787578] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 30b8b872-40ca-4297-b98e-a64c83a75483] Created VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 760.788181] env[63515]: DEBUG oslo_concurrency.lockutils [None req-b8b2a7f6-808d-40d4-8de9-c8863e5b55c5 tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 760.788503] env[63515]: DEBUG oslo_concurrency.lockutils [None req-b8b2a7f6-808d-40d4-8de9-c8863e5b55c5 tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 760.789020] env[63515]: DEBUG oslo_concurrency.lockutils [None req-b8b2a7f6-808d-40d4-8de9-c8863e5b55c5 tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 760.789467] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-33826a2d-c012-4388-a0ee-ac238fc44d5a {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.796523] env[63515]: DEBUG oslo_vmware.api [None req-b8b2a7f6-808d-40d4-8de9-c8863e5b55c5 tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] Waiting for the task: (returnval){ [ 760.796523] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]52504369-eb1f-19b0-0774-a2444b8f2db2" [ 760.796523] env[63515]: _type = "Task" [ 760.796523] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 760.806264] env[63515]: DEBUG oslo_vmware.api [None req-b8b2a7f6-808d-40d4-8de9-c8863e5b55c5 tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52504369-eb1f-19b0-0774-a2444b8f2db2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 760.809657] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Instance 5cbce760-0163-4b27-8ae3-e46c926c8916 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63515) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 760.832885] env[63515]: DEBUG oslo_concurrency.lockutils [None req-7c74c7c8-c879-4fe1-a77d-f24b849e5c36 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Acquiring lock "d2a4f66a-f1a2-4f0d-826d-0b7c4f04bd23" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 760.833275] env[63515]: DEBUG oslo_concurrency.lockutils [None req-7c74c7c8-c879-4fe1-a77d-f24b849e5c36 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Lock "d2a4f66a-f1a2-4f0d-826d-0b7c4f04bd23" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 760.833470] env[63515]: DEBUG nova.compute.manager [None req-7c74c7c8-c879-4fe1-a77d-f24b849e5c36 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] [instance: d2a4f66a-f1a2-4f0d-826d-0b7c4f04bd23] Checking state {{(pid=63515) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 760.834362] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32a73060-4495-4ea3-b762-0e4512ce0b75 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.842641] env[63515]: DEBUG nova.compute.manager [None req-7c74c7c8-c879-4fe1-a77d-f24b849e5c36 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] [instance: d2a4f66a-f1a2-4f0d-826d-0b7c4f04bd23] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=63515) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 760.843668] env[63515]: DEBUG nova.objects.instance [None req-7c74c7c8-c879-4fe1-a77d-f24b849e5c36 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Lazy-loading 'flavor' on Instance uuid d2a4f66a-f1a2-4f0d-826d-0b7c4f04bd23 {{(pid=63515) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 760.942213] env[63515]: DEBUG oslo_vmware.api [None req-2e592662-996b-4dfa-8941-00965c1317fc tempest-ServersTestBootFromVolume-2042560423 tempest-ServersTestBootFromVolume-2042560423-project-member] Task: {'id': task-1111128, 'name': PowerOffVM_Task, 'duration_secs': 0.261984} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 760.942213] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-2e592662-996b-4dfa-8941-00965c1317fc tempest-ServersTestBootFromVolume-2042560423 tempest-ServersTestBootFromVolume-2042560423-project-member] [instance: 90f4930b-aaa0-4c4b-9ab8-92aed45e200b] Powered off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 760.942525] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-2e592662-996b-4dfa-8941-00965c1317fc tempest-ServersTestBootFromVolume-2042560423 tempest-ServersTestBootFromVolume-2042560423-project-member] [instance: 90f4930b-aaa0-4c4b-9ab8-92aed45e200b] Volume detach. Driver type: vmdk {{(pid=63515) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 760.942702] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-2e592662-996b-4dfa-8941-00965c1317fc tempest-ServersTestBootFromVolume-2042560423 tempest-ServersTestBootFromVolume-2042560423-project-member] [instance: 90f4930b-aaa0-4c4b-9ab8-92aed45e200b] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-243378', 'volume_id': '921350b3-bd4a-4fa1-848a-057b53f51f5c', 'name': 'volume-921350b3-bd4a-4fa1-848a-057b53f51f5c', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '90f4930b-aaa0-4c4b-9ab8-92aed45e200b', 'attached_at': '', 'detached_at': '', 'volume_id': '921350b3-bd4a-4fa1-848a-057b53f51f5c', 'serial': '921350b3-bd4a-4fa1-848a-057b53f51f5c'} {{(pid=63515) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 760.943570] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8050f6c5-b2f9-44f2-87af-e8b19b0e0cdb {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.967905] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72b63b57-221e-43c2-989d-589606304e54 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.973846] env[63515]: INFO nova.compute.manager [None req-750d4038-2a95-4b14-b9d7-c9c6803c64b6 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] [instance: 8eb17506-ff93-4d25-b9af-ec5886569e65] Took 35.75 seconds to build instance. [ 760.979053] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04ac5759-4525-40f9-bd09-0097c1174952 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.001719] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10a4f3e8-c2d2-43b5-8e1f-9d7e214d7b71 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.022966] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-2e592662-996b-4dfa-8941-00965c1317fc tempest-ServersTestBootFromVolume-2042560423 tempest-ServersTestBootFromVolume-2042560423-project-member] The volume has not been displaced from its original location: [datastore2] volume-921350b3-bd4a-4fa1-848a-057b53f51f5c/volume-921350b3-bd4a-4fa1-848a-057b53f51f5c.vmdk. No consolidation needed. {{(pid=63515) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 761.028740] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-2e592662-996b-4dfa-8941-00965c1317fc tempest-ServersTestBootFromVolume-2042560423 tempest-ServersTestBootFromVolume-2042560423-project-member] [instance: 90f4930b-aaa0-4c4b-9ab8-92aed45e200b] Reconfiguring VM instance instance-00000023 to detach disk 2000 {{(pid=63515) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 761.030966] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c9baa4f7-1892-4a85-b930-ad6195c54ebc {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.044615] env[63515]: DEBUG oslo_concurrency.lockutils [None req-0e5835c9-1a06-4538-91e4-48a9dfc2854d tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 761.054861] env[63515]: DEBUG oslo_vmware.api [None req-8f7a1df0-5f0d-424a-ae8f-76bee2ee3dbb tempest-ServersAaction247Test-2019176257 tempest-ServersAaction247Test-2019176257-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52f3d976-5b03-5926-59ab-0bdbc8729ee3, 'name': SearchDatastore_Task, 'duration_secs': 0.014243} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 761.056192] env[63515]: DEBUG oslo_concurrency.lockutils [None req-8f7a1df0-5f0d-424a-ae8f-76bee2ee3dbb tempest-ServersAaction247Test-2019176257 tempest-ServersAaction247Test-2019176257-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 761.056317] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-8f7a1df0-5f0d-424a-ae8f-76bee2ee3dbb tempest-ServersAaction247Test-2019176257 tempest-ServersAaction247Test-2019176257-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk to [datastore1] 5d50fbd6-a7fb-422c-9dd7-df7140fd0aae/5d50fbd6-a7fb-422c-9dd7-df7140fd0aae.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 761.057357] env[63515]: DEBUG oslo_vmware.api [None req-2e592662-996b-4dfa-8941-00965c1317fc tempest-ServersTestBootFromVolume-2042560423 tempest-ServersTestBootFromVolume-2042560423-project-member] Waiting for the task: (returnval){ [ 761.057357] env[63515]: value = "task-1111129" [ 761.057357] env[63515]: _type = "Task" [ 761.057357] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 761.057357] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-dfa0dd02-9203-4004-b79c-ba2c0832dfbd {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.067141] env[63515]: DEBUG oslo_vmware.api [None req-2e592662-996b-4dfa-8941-00965c1317fc tempest-ServersTestBootFromVolume-2042560423 tempest-ServersTestBootFromVolume-2042560423-project-member] Task: {'id': task-1111129, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 761.068411] env[63515]: DEBUG oslo_vmware.api [None req-8f7a1df0-5f0d-424a-ae8f-76bee2ee3dbb tempest-ServersAaction247Test-2019176257 tempest-ServersAaction247Test-2019176257-project-member] Waiting for the task: (returnval){ [ 761.068411] env[63515]: value = "task-1111130" [ 761.068411] env[63515]: _type = "Task" [ 761.068411] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 761.077218] env[63515]: DEBUG oslo_vmware.api [None req-8f7a1df0-5f0d-424a-ae8f-76bee2ee3dbb tempest-ServersAaction247Test-2019176257 tempest-ServersAaction247Test-2019176257-project-member] Task: {'id': task-1111130, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 761.308460] env[63515]: DEBUG oslo_vmware.api [None req-b8b2a7f6-808d-40d4-8de9-c8863e5b55c5 tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52504369-eb1f-19b0-0774-a2444b8f2db2, 'name': SearchDatastore_Task, 'duration_secs': 0.02769} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 761.309595] env[63515]: DEBUG oslo_concurrency.lockutils [None req-b8b2a7f6-808d-40d4-8de9-c8863e5b55c5 tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 761.309595] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-b8b2a7f6-808d-40d4-8de9-c8863e5b55c5 tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] [instance: 30b8b872-40ca-4297-b98e-a64c83a75483] Processing image 8a120570-cb06-4099-b262-554ca0ad15c5 {{(pid=63515) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 761.309595] env[63515]: DEBUG oslo_concurrency.lockutils [None req-b8b2a7f6-808d-40d4-8de9-c8863e5b55c5 tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 761.309880] env[63515]: DEBUG oslo_concurrency.lockutils [None req-b8b2a7f6-808d-40d4-8de9-c8863e5b55c5 tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 761.310086] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-b8b2a7f6-808d-40d4-8de9-c8863e5b55c5 tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 761.312924] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Instance 63ce797d-7180-4209-ac2c-81978bf7607a has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63515) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 761.314419] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1f7bc55e-e3db-4594-9302-c24828c3413e {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.327059] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-b8b2a7f6-808d-40d4-8de9-c8863e5b55c5 tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 761.327352] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-b8b2a7f6-808d-40d4-8de9-c8863e5b55c5 tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63515) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 761.329018] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e249fbf1-8124-458f-aca6-bdd443da9d6a {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.338595] env[63515]: DEBUG oslo_vmware.api [None req-b8b2a7f6-808d-40d4-8de9-c8863e5b55c5 tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] Waiting for the task: (returnval){ [ 761.338595] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]52a055b2-c1c8-4463-45ac-06aed493739e" [ 761.338595] env[63515]: _type = "Task" [ 761.338595] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 761.348104] env[63515]: DEBUG oslo_vmware.api [None req-b8b2a7f6-808d-40d4-8de9-c8863e5b55c5 tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52a055b2-c1c8-4463-45ac-06aed493739e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 761.348661] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-7c74c7c8-c879-4fe1-a77d-f24b849e5c36 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] [instance: d2a4f66a-f1a2-4f0d-826d-0b7c4f04bd23] Powering off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 761.349448] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e276ab84-c8bf-4ec1-8fe3-8353bfe0f822 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.358346] env[63515]: DEBUG oslo_vmware.api [None req-7c74c7c8-c879-4fe1-a77d-f24b849e5c36 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Waiting for the task: (returnval){ [ 761.358346] env[63515]: value = "task-1111131" [ 761.358346] env[63515]: _type = "Task" [ 761.358346] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 761.367996] env[63515]: DEBUG oslo_vmware.api [None req-7c74c7c8-c879-4fe1-a77d-f24b849e5c36 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Task: {'id': task-1111131, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 761.476795] env[63515]: DEBUG oslo_concurrency.lockutils [None req-750d4038-2a95-4b14-b9d7-c9c6803c64b6 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] Lock "8eb17506-ff93-4d25-b9af-ec5886569e65" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 120.123s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 761.577623] env[63515]: DEBUG oslo_vmware.api [None req-2e592662-996b-4dfa-8941-00965c1317fc tempest-ServersTestBootFromVolume-2042560423 tempest-ServersTestBootFromVolume-2042560423-project-member] Task: {'id': task-1111129, 'name': ReconfigVM_Task, 'duration_secs': 0.276123} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 761.578642] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-2e592662-996b-4dfa-8941-00965c1317fc tempest-ServersTestBootFromVolume-2042560423 tempest-ServersTestBootFromVolume-2042560423-project-member] [instance: 90f4930b-aaa0-4c4b-9ab8-92aed45e200b] Reconfigured VM instance instance-00000023 to detach disk 2000 {{(pid=63515) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 761.588486] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4cab8b69-d410-48dd-a84b-88c4099861ed {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.599009] env[63515]: DEBUG oslo_vmware.api [None req-8f7a1df0-5f0d-424a-ae8f-76bee2ee3dbb tempest-ServersAaction247Test-2019176257 tempest-ServersAaction247Test-2019176257-project-member] Task: {'id': task-1111130, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 761.607654] env[63515]: DEBUG oslo_vmware.api [None req-2e592662-996b-4dfa-8941-00965c1317fc tempest-ServersTestBootFromVolume-2042560423 tempest-ServersTestBootFromVolume-2042560423-project-member] Waiting for the task: (returnval){ [ 761.607654] env[63515]: value = "task-1111132" [ 761.607654] env[63515]: _type = "Task" [ 761.607654] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 761.619760] env[63515]: DEBUG oslo_vmware.api [None req-2e592662-996b-4dfa-8941-00965c1317fc tempest-ServersTestBootFromVolume-2042560423 tempest-ServersTestBootFromVolume-2042560423-project-member] Task: {'id': task-1111132, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 761.818511] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Instance e666825c-ff4e-4a0e-93c0-43c00f167bbb has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63515) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 761.818935] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Migration 802930d6-330b-4516-9996-d196743b8ab8 is active on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63515) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1712}} [ 761.818935] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Instance f1d01b75-ac9d-458d-8cc2-ae64cffca4e8 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=63515) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 761.850227] env[63515]: DEBUG oslo_vmware.api [None req-b8b2a7f6-808d-40d4-8de9-c8863e5b55c5 tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52a055b2-c1c8-4463-45ac-06aed493739e, 'name': SearchDatastore_Task, 'duration_secs': 0.018947} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 761.851066] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d3efea6a-bf76-4742-a8fc-11e7da9f0539 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.858168] env[63515]: DEBUG oslo_vmware.api [None req-b8b2a7f6-808d-40d4-8de9-c8863e5b55c5 tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] Waiting for the task: (returnval){ [ 761.858168] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]526e7055-694e-ba0c-4241-9457152cdacc" [ 761.858168] env[63515]: _type = "Task" [ 761.858168] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 761.872124] env[63515]: DEBUG oslo_vmware.api [None req-b8b2a7f6-808d-40d4-8de9-c8863e5b55c5 tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]526e7055-694e-ba0c-4241-9457152cdacc, 'name': SearchDatastore_Task} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 761.875363] env[63515]: DEBUG oslo_concurrency.lockutils [None req-b8b2a7f6-808d-40d4-8de9-c8863e5b55c5 tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 761.875522] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-b8b2a7f6-808d-40d4-8de9-c8863e5b55c5 tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk to [datastore1] 30b8b872-40ca-4297-b98e-a64c83a75483/30b8b872-40ca-4297-b98e-a64c83a75483.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 761.876631] env[63515]: DEBUG oslo_vmware.api [None req-7c74c7c8-c879-4fe1-a77d-f24b849e5c36 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Task: {'id': task-1111131, 'name': PowerOffVM_Task} progress is 100%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 761.876631] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-84c14283-10fc-4dc6-bd65-61182178e4a6 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.884629] env[63515]: DEBUG oslo_vmware.api [None req-b8b2a7f6-808d-40d4-8de9-c8863e5b55c5 tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] Waiting for the task: (returnval){ [ 761.884629] env[63515]: value = "task-1111133" [ 761.884629] env[63515]: _type = "Task" [ 761.884629] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 761.898148] env[63515]: DEBUG oslo_vmware.api [None req-b8b2a7f6-808d-40d4-8de9-c8863e5b55c5 tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] Task: {'id': task-1111133, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 761.935749] env[63515]: DEBUG oslo_concurrency.lockutils [None req-aef604e2-0f8c-4f11-a420-cb96f46c502e tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Acquiring lock "001a48e1-b237-41f8-a6f4-2fe98d173512" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 761.936127] env[63515]: DEBUG oslo_concurrency.lockutils [None req-aef604e2-0f8c-4f11-a420-cb96f46c502e tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Lock "001a48e1-b237-41f8-a6f4-2fe98d173512" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 761.936382] env[63515]: DEBUG oslo_concurrency.lockutils [None req-aef604e2-0f8c-4f11-a420-cb96f46c502e tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Acquiring lock "001a48e1-b237-41f8-a6f4-2fe98d173512-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 761.936606] env[63515]: DEBUG oslo_concurrency.lockutils [None req-aef604e2-0f8c-4f11-a420-cb96f46c502e tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Lock "001a48e1-b237-41f8-a6f4-2fe98d173512-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 761.936845] env[63515]: DEBUG oslo_concurrency.lockutils [None req-aef604e2-0f8c-4f11-a420-cb96f46c502e tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Lock "001a48e1-b237-41f8-a6f4-2fe98d173512-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 761.939473] env[63515]: INFO nova.compute.manager [None req-aef604e2-0f8c-4f11-a420-cb96f46c502e tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: 001a48e1-b237-41f8-a6f4-2fe98d173512] Terminating instance [ 761.941503] env[63515]: DEBUG nova.compute.manager [None req-aef604e2-0f8c-4f11-a420-cb96f46c502e tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: 001a48e1-b237-41f8-a6f4-2fe98d173512] Start destroying the instance on the hypervisor. {{(pid=63515) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 761.941737] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-aef604e2-0f8c-4f11-a420-cb96f46c502e tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: 001a48e1-b237-41f8-a6f4-2fe98d173512] Destroying instance {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 761.942779] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b297df78-cf2b-46a4-9d59-d4e18e5e6406 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.954406] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-aef604e2-0f8c-4f11-a420-cb96f46c502e tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: 001a48e1-b237-41f8-a6f4-2fe98d173512] Powering off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 761.954878] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b7497948-6a91-4379-8844-8172c115efdd {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.965554] env[63515]: DEBUG oslo_vmware.api [None req-aef604e2-0f8c-4f11-a420-cb96f46c502e tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Waiting for the task: (returnval){ [ 761.965554] env[63515]: value = "task-1111134" [ 761.965554] env[63515]: _type = "Task" [ 761.965554] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 761.977853] env[63515]: DEBUG oslo_vmware.api [None req-aef604e2-0f8c-4f11-a420-cb96f46c502e tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Task: {'id': task-1111134, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 761.979593] env[63515]: DEBUG nova.compute.manager [None req-95950a02-20c9-4a3f-9ea8-1f7f19725cc1 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: 4ef8066a-b68c-457b-9964-b1c34bab0fc3] Starting instance... {{(pid=63515) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 762.081837] env[63515]: DEBUG oslo_vmware.api [None req-8f7a1df0-5f0d-424a-ae8f-76bee2ee3dbb tempest-ServersAaction247Test-2019176257 tempest-ServersAaction247Test-2019176257-project-member] Task: {'id': task-1111130, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.678518} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 762.082268] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-8f7a1df0-5f0d-424a-ae8f-76bee2ee3dbb tempest-ServersAaction247Test-2019176257 tempest-ServersAaction247Test-2019176257-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk to [datastore1] 5d50fbd6-a7fb-422c-9dd7-df7140fd0aae/5d50fbd6-a7fb-422c-9dd7-df7140fd0aae.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 762.082530] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-8f7a1df0-5f0d-424a-ae8f-76bee2ee3dbb tempest-ServersAaction247Test-2019176257 tempest-ServersAaction247Test-2019176257-project-member] [instance: 5d50fbd6-a7fb-422c-9dd7-df7140fd0aae] Extending root virtual disk to 1048576 {{(pid=63515) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 762.083113] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-5cf196cc-9444-4cc0-8333-f3b9a9c7b5c8 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.092971] env[63515]: DEBUG oslo_vmware.api [None req-8f7a1df0-5f0d-424a-ae8f-76bee2ee3dbb tempest-ServersAaction247Test-2019176257 tempest-ServersAaction247Test-2019176257-project-member] Waiting for the task: (returnval){ [ 762.092971] env[63515]: value = "task-1111135" [ 762.092971] env[63515]: _type = "Task" [ 762.092971] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 762.107047] env[63515]: DEBUG oslo_vmware.api [None req-8f7a1df0-5f0d-424a-ae8f-76bee2ee3dbb tempest-ServersAaction247Test-2019176257 tempest-ServersAaction247Test-2019176257-project-member] Task: {'id': task-1111135, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 762.119887] env[63515]: DEBUG oslo_vmware.api [None req-2e592662-996b-4dfa-8941-00965c1317fc tempest-ServersTestBootFromVolume-2042560423 tempest-ServersTestBootFromVolume-2042560423-project-member] Task: {'id': task-1111132, 'name': ReconfigVM_Task, 'duration_secs': 0.19866} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 762.120244] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-2e592662-996b-4dfa-8941-00965c1317fc tempest-ServersTestBootFromVolume-2042560423 tempest-ServersTestBootFromVolume-2042560423-project-member] [instance: 90f4930b-aaa0-4c4b-9ab8-92aed45e200b] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-243378', 'volume_id': '921350b3-bd4a-4fa1-848a-057b53f51f5c', 'name': 'volume-921350b3-bd4a-4fa1-848a-057b53f51f5c', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '90f4930b-aaa0-4c4b-9ab8-92aed45e200b', 'attached_at': '', 'detached_at': '', 'volume_id': '921350b3-bd4a-4fa1-848a-057b53f51f5c', 'serial': '921350b3-bd4a-4fa1-848a-057b53f51f5c'} {{(pid=63515) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 762.120538] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-2e592662-996b-4dfa-8941-00965c1317fc tempest-ServersTestBootFromVolume-2042560423 tempest-ServersTestBootFromVolume-2042560423-project-member] [instance: 90f4930b-aaa0-4c4b-9ab8-92aed45e200b] Destroying instance {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 762.121458] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-defc5ed7-abbe-4065-80cb-fddc76be60e1 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.130958] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-2e592662-996b-4dfa-8941-00965c1317fc tempest-ServersTestBootFromVolume-2042560423 tempest-ServersTestBootFromVolume-2042560423-project-member] [instance: 90f4930b-aaa0-4c4b-9ab8-92aed45e200b] Unregistering the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 762.131292] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-565a20a1-5ba5-47b1-a06e-299c0f127eaf {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.174985] env[63515]: INFO nova.compute.manager [None req-db31b1fd-34e1-40f4-b7c8-92b451eb1400 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] [instance: 8eb17506-ff93-4d25-b9af-ec5886569e65] Rebuilding instance [ 762.211978] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-2e592662-996b-4dfa-8941-00965c1317fc tempest-ServersTestBootFromVolume-2042560423 tempest-ServersTestBootFromVolume-2042560423-project-member] [instance: 90f4930b-aaa0-4c4b-9ab8-92aed45e200b] Unregistered the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 762.212284] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-2e592662-996b-4dfa-8941-00965c1317fc tempest-ServersTestBootFromVolume-2042560423 tempest-ServersTestBootFromVolume-2042560423-project-member] [instance: 90f4930b-aaa0-4c4b-9ab8-92aed45e200b] Deleting contents of the VM from datastore datastore2 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 762.212849] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-2e592662-996b-4dfa-8941-00965c1317fc tempest-ServersTestBootFromVolume-2042560423 tempest-ServersTestBootFromVolume-2042560423-project-member] Deleting the datastore file [datastore2] 90f4930b-aaa0-4c4b-9ab8-92aed45e200b {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 762.213747] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-97e2cc82-5a6c-425e-a69c-8f09f34a43e6 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.227532] env[63515]: DEBUG oslo_vmware.api [None req-2e592662-996b-4dfa-8941-00965c1317fc tempest-ServersTestBootFromVolume-2042560423 tempest-ServersTestBootFromVolume-2042560423-project-member] Waiting for the task: (returnval){ [ 762.227532] env[63515]: value = "task-1111137" [ 762.227532] env[63515]: _type = "Task" [ 762.227532] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 762.230264] env[63515]: DEBUG nova.compute.manager [None req-db31b1fd-34e1-40f4-b7c8-92b451eb1400 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] [instance: 8eb17506-ff93-4d25-b9af-ec5886569e65] Checking state {{(pid=63515) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 762.230908] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-681bd5d3-1ac6-4768-958a-ad5e5597b6f4 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.247168] env[63515]: DEBUG oslo_vmware.api [None req-2e592662-996b-4dfa-8941-00965c1317fc tempest-ServersTestBootFromVolume-2042560423 tempest-ServersTestBootFromVolume-2042560423-project-member] Task: {'id': task-1111137, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 762.322199] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Instance 4e453127-1f3e-40ea-819f-6678479826c8 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63515) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 762.377757] env[63515]: DEBUG oslo_vmware.api [None req-7c74c7c8-c879-4fe1-a77d-f24b849e5c36 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Task: {'id': task-1111131, 'name': PowerOffVM_Task, 'duration_secs': 0.53074} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 762.378113] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-7c74c7c8-c879-4fe1-a77d-f24b849e5c36 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] [instance: d2a4f66a-f1a2-4f0d-826d-0b7c4f04bd23] Powered off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 762.378306] env[63515]: DEBUG nova.compute.manager [None req-7c74c7c8-c879-4fe1-a77d-f24b849e5c36 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] [instance: d2a4f66a-f1a2-4f0d-826d-0b7c4f04bd23] Checking state {{(pid=63515) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 762.379357] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48feb5d0-f4e7-4464-92ee-c948676bdc3b {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.401621] env[63515]: DEBUG oslo_vmware.api [None req-b8b2a7f6-808d-40d4-8de9-c8863e5b55c5 tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] Task: {'id': task-1111133, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 762.486761] env[63515]: DEBUG oslo_vmware.api [None req-aef604e2-0f8c-4f11-a420-cb96f46c502e tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Task: {'id': task-1111134, 'name': PowerOffVM_Task, 'duration_secs': 0.242813} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 762.489327] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-aef604e2-0f8c-4f11-a420-cb96f46c502e tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: 001a48e1-b237-41f8-a6f4-2fe98d173512] Powered off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 762.489563] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-aef604e2-0f8c-4f11-a420-cb96f46c502e tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: 001a48e1-b237-41f8-a6f4-2fe98d173512] Unregistering the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 762.490092] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-13b49645-b052-4edf-9365-2ef8f914d813 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.504450] env[63515]: DEBUG oslo_concurrency.lockutils [None req-95950a02-20c9-4a3f-9ea8-1f7f19725cc1 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 762.556427] env[63515]: DEBUG oslo_concurrency.lockutils [None req-41b4a543-33b4-43d9-863c-4344bf9d0a46 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Acquiring lock "d926c699-a64a-4942-9ef4-f0166414661d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 762.556749] env[63515]: DEBUG oslo_concurrency.lockutils [None req-41b4a543-33b4-43d9-863c-4344bf9d0a46 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Lock "d926c699-a64a-4942-9ef4-f0166414661d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 762.572290] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-aef604e2-0f8c-4f11-a420-cb96f46c502e tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: 001a48e1-b237-41f8-a6f4-2fe98d173512] Unregistered the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 762.572555] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-aef604e2-0f8c-4f11-a420-cb96f46c502e tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: 001a48e1-b237-41f8-a6f4-2fe98d173512] Deleting contents of the VM from datastore datastore2 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 762.572712] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-aef604e2-0f8c-4f11-a420-cb96f46c502e tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Deleting the datastore file [datastore2] 001a48e1-b237-41f8-a6f4-2fe98d173512 {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 762.573335] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-957de698-3db9-4416-a54e-8977a50ded14 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.582409] env[63515]: DEBUG oslo_vmware.api [None req-aef604e2-0f8c-4f11-a420-cb96f46c502e tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Waiting for the task: (returnval){ [ 762.582409] env[63515]: value = "task-1111139" [ 762.582409] env[63515]: _type = "Task" [ 762.582409] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 762.591607] env[63515]: DEBUG oslo_vmware.api [None req-aef604e2-0f8c-4f11-a420-cb96f46c502e tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Task: {'id': task-1111139, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 762.602190] env[63515]: DEBUG oslo_vmware.api [None req-8f7a1df0-5f0d-424a-ae8f-76bee2ee3dbb tempest-ServersAaction247Test-2019176257 tempest-ServersAaction247Test-2019176257-project-member] Task: {'id': task-1111135, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.107801} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 762.602479] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-8f7a1df0-5f0d-424a-ae8f-76bee2ee3dbb tempest-ServersAaction247Test-2019176257 tempest-ServersAaction247Test-2019176257-project-member] [instance: 5d50fbd6-a7fb-422c-9dd7-df7140fd0aae] Extended root virtual disk {{(pid=63515) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 762.603251] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d208ad9-9521-437d-b80a-3aa9b3fd94c8 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.623180] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-8f7a1df0-5f0d-424a-ae8f-76bee2ee3dbb tempest-ServersAaction247Test-2019176257 tempest-ServersAaction247Test-2019176257-project-member] [instance: 5d50fbd6-a7fb-422c-9dd7-df7140fd0aae] Reconfiguring VM instance instance-00000036 to attach disk [datastore1] 5d50fbd6-a7fb-422c-9dd7-df7140fd0aae/5d50fbd6-a7fb-422c-9dd7-df7140fd0aae.vmdk or device None with type sparse {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 762.623932] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-10b33392-37e5-457f-b14f-2e0813c81d86 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.646225] env[63515]: DEBUG oslo_vmware.api [None req-8f7a1df0-5f0d-424a-ae8f-76bee2ee3dbb tempest-ServersAaction247Test-2019176257 tempest-ServersAaction247Test-2019176257-project-member] Waiting for the task: (returnval){ [ 762.646225] env[63515]: value = "task-1111140" [ 762.646225] env[63515]: _type = "Task" [ 762.646225] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 762.655832] env[63515]: DEBUG oslo_vmware.api [None req-8f7a1df0-5f0d-424a-ae8f-76bee2ee3dbb tempest-ServersAaction247Test-2019176257 tempest-ServersAaction247Test-2019176257-project-member] Task: {'id': task-1111140, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 762.742059] env[63515]: DEBUG oslo_vmware.api [None req-2e592662-996b-4dfa-8941-00965c1317fc tempest-ServersTestBootFromVolume-2042560423 tempest-ServersTestBootFromVolume-2042560423-project-member] Task: {'id': task-1111137, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.238447} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 762.742059] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-2e592662-996b-4dfa-8941-00965c1317fc tempest-ServersTestBootFromVolume-2042560423 tempest-ServersTestBootFromVolume-2042560423-project-member] Deleted the datastore file {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 762.742059] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-2e592662-996b-4dfa-8941-00965c1317fc tempest-ServersTestBootFromVolume-2042560423 tempest-ServersTestBootFromVolume-2042560423-project-member] [instance: 90f4930b-aaa0-4c4b-9ab8-92aed45e200b] Deleted contents of the VM from datastore datastore2 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 762.742059] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-2e592662-996b-4dfa-8941-00965c1317fc tempest-ServersTestBootFromVolume-2042560423 tempest-ServersTestBootFromVolume-2042560423-project-member] [instance: 90f4930b-aaa0-4c4b-9ab8-92aed45e200b] Instance destroyed {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 762.742332] env[63515]: INFO nova.compute.manager [None req-2e592662-996b-4dfa-8941-00965c1317fc tempest-ServersTestBootFromVolume-2042560423 tempest-ServersTestBootFromVolume-2042560423-project-member] [instance: 90f4930b-aaa0-4c4b-9ab8-92aed45e200b] Took 2.34 seconds to destroy the instance on the hypervisor. [ 762.742332] env[63515]: DEBUG oslo.service.loopingcall [None req-2e592662-996b-4dfa-8941-00965c1317fc tempest-ServersTestBootFromVolume-2042560423 tempest-ServersTestBootFromVolume-2042560423-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 762.742541] env[63515]: DEBUG nova.compute.manager [-] [instance: 90f4930b-aaa0-4c4b-9ab8-92aed45e200b] Deallocating network for instance {{(pid=63515) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 762.742642] env[63515]: DEBUG nova.network.neutron [-] [instance: 90f4930b-aaa0-4c4b-9ab8-92aed45e200b] deallocate_for_instance() {{(pid=63515) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 762.757039] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-db31b1fd-34e1-40f4-b7c8-92b451eb1400 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] [instance: 8eb17506-ff93-4d25-b9af-ec5886569e65] Powering off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 762.757420] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3b134ab0-2184-4de4-89c1-89d5697f32a9 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.766429] env[63515]: DEBUG oslo_vmware.api [None req-db31b1fd-34e1-40f4-b7c8-92b451eb1400 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] Waiting for the task: (returnval){ [ 762.766429] env[63515]: value = "task-1111141" [ 762.766429] env[63515]: _type = "Task" [ 762.766429] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 762.776717] env[63515]: DEBUG oslo_vmware.api [None req-db31b1fd-34e1-40f4-b7c8-92b451eb1400 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] Task: {'id': task-1111141, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 762.825194] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Instance 4ef8066a-b68c-457b-9964-b1c34bab0fc3 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63515) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 762.825538] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Total usable vcpus: 48, total allocated vcpus: 18 {{(pid=63515) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 762.825697] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=4096MB phys_disk=200GB used_disk=17GB total_vcpus=48 used_vcpus=18 pci_stats=[] {{(pid=63515) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 762.849744] env[63515]: DEBUG oslo_concurrency.lockutils [None req-655858bd-8f63-4e55-ac5e-dbecfbd88a62 tempest-ServersV294TestFqdnHostnames-2112993702 tempest-ServersV294TestFqdnHostnames-2112993702-project-member] Acquiring lock "396e49dd-48c0-496b-a1ec-190c33a22c5e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 762.850068] env[63515]: DEBUG oslo_concurrency.lockutils [None req-655858bd-8f63-4e55-ac5e-dbecfbd88a62 tempest-ServersV294TestFqdnHostnames-2112993702 tempest-ServersV294TestFqdnHostnames-2112993702-project-member] Lock "396e49dd-48c0-496b-a1ec-190c33a22c5e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 762.850351] env[63515]: DEBUG oslo_concurrency.lockutils [None req-655858bd-8f63-4e55-ac5e-dbecfbd88a62 tempest-ServersV294TestFqdnHostnames-2112993702 tempest-ServersV294TestFqdnHostnames-2112993702-project-member] Acquiring lock "396e49dd-48c0-496b-a1ec-190c33a22c5e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 762.850628] env[63515]: DEBUG oslo_concurrency.lockutils [None req-655858bd-8f63-4e55-ac5e-dbecfbd88a62 tempest-ServersV294TestFqdnHostnames-2112993702 tempest-ServersV294TestFqdnHostnames-2112993702-project-member] Lock "396e49dd-48c0-496b-a1ec-190c33a22c5e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 762.850834] env[63515]: DEBUG oslo_concurrency.lockutils [None req-655858bd-8f63-4e55-ac5e-dbecfbd88a62 tempest-ServersV294TestFqdnHostnames-2112993702 tempest-ServersV294TestFqdnHostnames-2112993702-project-member] Lock "396e49dd-48c0-496b-a1ec-190c33a22c5e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 762.853765] env[63515]: INFO nova.compute.manager [None req-655858bd-8f63-4e55-ac5e-dbecfbd88a62 tempest-ServersV294TestFqdnHostnames-2112993702 tempest-ServersV294TestFqdnHostnames-2112993702-project-member] [instance: 396e49dd-48c0-496b-a1ec-190c33a22c5e] Terminating instance [ 762.858716] env[63515]: DEBUG nova.compute.manager [None req-655858bd-8f63-4e55-ac5e-dbecfbd88a62 tempest-ServersV294TestFqdnHostnames-2112993702 tempest-ServersV294TestFqdnHostnames-2112993702-project-member] [instance: 396e49dd-48c0-496b-a1ec-190c33a22c5e] Start destroying the instance on the hypervisor. {{(pid=63515) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 762.858716] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-655858bd-8f63-4e55-ac5e-dbecfbd88a62 tempest-ServersV294TestFqdnHostnames-2112993702 tempest-ServersV294TestFqdnHostnames-2112993702-project-member] [instance: 396e49dd-48c0-496b-a1ec-190c33a22c5e] Destroying instance {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 762.859943] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7da6cee-6a6b-4cb6-851c-deb759463ec0 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.877346] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-655858bd-8f63-4e55-ac5e-dbecfbd88a62 tempest-ServersV294TestFqdnHostnames-2112993702 tempest-ServersV294TestFqdnHostnames-2112993702-project-member] [instance: 396e49dd-48c0-496b-a1ec-190c33a22c5e] Powering off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 762.877757] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1d4aa9f6-08c9-4b07-aa34-3d2d24bdc1d1 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.892921] env[63515]: DEBUG oslo_vmware.api [None req-655858bd-8f63-4e55-ac5e-dbecfbd88a62 tempest-ServersV294TestFqdnHostnames-2112993702 tempest-ServersV294TestFqdnHostnames-2112993702-project-member] Waiting for the task: (returnval){ [ 762.892921] env[63515]: value = "task-1111142" [ 762.892921] env[63515]: _type = "Task" [ 762.892921] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 762.913053] env[63515]: DEBUG oslo_concurrency.lockutils [None req-7c74c7c8-c879-4fe1-a77d-f24b849e5c36 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Lock "d2a4f66a-f1a2-4f0d-826d-0b7c4f04bd23" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 2.080s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 762.915818] env[63515]: DEBUG oslo_vmware.api [None req-b8b2a7f6-808d-40d4-8de9-c8863e5b55c5 tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] Task: {'id': task-1111133, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.631938} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 762.917072] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-b8b2a7f6-808d-40d4-8de9-c8863e5b55c5 tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk to [datastore1] 30b8b872-40ca-4297-b98e-a64c83a75483/30b8b872-40ca-4297-b98e-a64c83a75483.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 762.917429] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-b8b2a7f6-808d-40d4-8de9-c8863e5b55c5 tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] [instance: 30b8b872-40ca-4297-b98e-a64c83a75483] Extending root virtual disk to 1048576 {{(pid=63515) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 762.917793] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-69475bef-21a4-44b4-a938-9821e83d0faf {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.925765] env[63515]: DEBUG oslo_vmware.api [None req-655858bd-8f63-4e55-ac5e-dbecfbd88a62 tempest-ServersV294TestFqdnHostnames-2112993702 tempest-ServersV294TestFqdnHostnames-2112993702-project-member] Task: {'id': task-1111142, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 762.932423] env[63515]: DEBUG oslo_vmware.api [None req-b8b2a7f6-808d-40d4-8de9-c8863e5b55c5 tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] Waiting for the task: (returnval){ [ 762.932423] env[63515]: value = "task-1111143" [ 762.932423] env[63515]: _type = "Task" [ 762.932423] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 762.949746] env[63515]: DEBUG oslo_vmware.api [None req-b8b2a7f6-808d-40d4-8de9-c8863e5b55c5 tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] Task: {'id': task-1111143, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 763.094863] env[63515]: DEBUG oslo_vmware.api [None req-aef604e2-0f8c-4f11-a420-cb96f46c502e tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Task: {'id': task-1111139, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.299279} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 763.095355] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-aef604e2-0f8c-4f11-a420-cb96f46c502e tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Deleted the datastore file {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 763.095572] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-aef604e2-0f8c-4f11-a420-cb96f46c502e tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: 001a48e1-b237-41f8-a6f4-2fe98d173512] Deleted contents of the VM from datastore datastore2 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 763.095801] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-aef604e2-0f8c-4f11-a420-cb96f46c502e tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: 001a48e1-b237-41f8-a6f4-2fe98d173512] Instance destroyed {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 763.096042] env[63515]: INFO nova.compute.manager [None req-aef604e2-0f8c-4f11-a420-cb96f46c502e tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: 001a48e1-b237-41f8-a6f4-2fe98d173512] Took 1.15 seconds to destroy the instance on the hypervisor. [ 763.096719] env[63515]: DEBUG oslo.service.loopingcall [None req-aef604e2-0f8c-4f11-a420-cb96f46c502e tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 763.100417] env[63515]: DEBUG nova.compute.manager [-] [instance: 001a48e1-b237-41f8-a6f4-2fe98d173512] Deallocating network for instance {{(pid=63515) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 763.100605] env[63515]: DEBUG nova.network.neutron [-] [instance: 001a48e1-b237-41f8-a6f4-2fe98d173512] deallocate_for_instance() {{(pid=63515) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 763.159443] env[63515]: DEBUG oslo_vmware.api [None req-8f7a1df0-5f0d-424a-ae8f-76bee2ee3dbb tempest-ServersAaction247Test-2019176257 tempest-ServersAaction247Test-2019176257-project-member] Task: {'id': task-1111140, 'name': ReconfigVM_Task, 'duration_secs': 0.424945} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 763.160285] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-8f7a1df0-5f0d-424a-ae8f-76bee2ee3dbb tempest-ServersAaction247Test-2019176257 tempest-ServersAaction247Test-2019176257-project-member] [instance: 5d50fbd6-a7fb-422c-9dd7-df7140fd0aae] Reconfigured VM instance instance-00000036 to attach disk [datastore1] 5d50fbd6-a7fb-422c-9dd7-df7140fd0aae/5d50fbd6-a7fb-422c-9dd7-df7140fd0aae.vmdk or device None with type sparse {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 763.161848] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-2831ffe3-8734-4008-a254-1afc8c821f40 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.171425] env[63515]: DEBUG oslo_vmware.api [None req-8f7a1df0-5f0d-424a-ae8f-76bee2ee3dbb tempest-ServersAaction247Test-2019176257 tempest-ServersAaction247Test-2019176257-project-member] Waiting for the task: (returnval){ [ 763.171425] env[63515]: value = "task-1111144" [ 763.171425] env[63515]: _type = "Task" [ 763.171425] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 763.180615] env[63515]: DEBUG oslo_vmware.api [None req-8f7a1df0-5f0d-424a-ae8f-76bee2ee3dbb tempest-ServersAaction247Test-2019176257 tempest-ServersAaction247Test-2019176257-project-member] Task: {'id': task-1111144, 'name': Rename_Task} progress is 5%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 763.277769] env[63515]: DEBUG oslo_vmware.api [None req-db31b1fd-34e1-40f4-b7c8-92b451eb1400 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] Task: {'id': task-1111141, 'name': PowerOffVM_Task, 'duration_secs': 0.195983} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 763.277769] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-db31b1fd-34e1-40f4-b7c8-92b451eb1400 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] [instance: 8eb17506-ff93-4d25-b9af-ec5886569e65] Powered off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 763.277769] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-db31b1fd-34e1-40f4-b7c8-92b451eb1400 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] [instance: 8eb17506-ff93-4d25-b9af-ec5886569e65] Destroying instance {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 763.278555] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c06c6448-f1d8-40b5-9d3e-87e8d551d9e0 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.286906] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-db31b1fd-34e1-40f4-b7c8-92b451eb1400 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] [instance: 8eb17506-ff93-4d25-b9af-ec5886569e65] Unregistering the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 763.286906] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-cfdfba09-ae80-4193-8f43-b48a670c265c {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.314852] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-db31b1fd-34e1-40f4-b7c8-92b451eb1400 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] [instance: 8eb17506-ff93-4d25-b9af-ec5886569e65] Unregistered the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 763.315220] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-db31b1fd-34e1-40f4-b7c8-92b451eb1400 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] [instance: 8eb17506-ff93-4d25-b9af-ec5886569e65] Deleting contents of the VM from datastore datastore2 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 763.316270] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-db31b1fd-34e1-40f4-b7c8-92b451eb1400 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] Deleting the datastore file [datastore2] 8eb17506-ff93-4d25-b9af-ec5886569e65 {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 763.316270] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-157ea376-0720-43e9-a801-788a04dccda0 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.324583] env[63515]: DEBUG oslo_vmware.api [None req-db31b1fd-34e1-40f4-b7c8-92b451eb1400 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] Waiting for the task: (returnval){ [ 763.324583] env[63515]: value = "task-1111146" [ 763.324583] env[63515]: _type = "Task" [ 763.324583] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 763.333124] env[63515]: DEBUG oslo_vmware.api [None req-db31b1fd-34e1-40f4-b7c8-92b451eb1400 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] Task: {'id': task-1111146, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 763.359912] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8aac355-4959-4361-99f6-c57eb6ac76b4 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.367680] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70c7b088-21a0-4206-8624-619cbda26be7 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.404503] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad4ed419-0565-45c0-a0cc-503e87455284 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.412884] env[63515]: DEBUG oslo_vmware.api [None req-655858bd-8f63-4e55-ac5e-dbecfbd88a62 tempest-ServersV294TestFqdnHostnames-2112993702 tempest-ServersV294TestFqdnHostnames-2112993702-project-member] Task: {'id': task-1111142, 'name': PowerOffVM_Task, 'duration_secs': 0.230227} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 763.415421] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-655858bd-8f63-4e55-ac5e-dbecfbd88a62 tempest-ServersV294TestFqdnHostnames-2112993702 tempest-ServersV294TestFqdnHostnames-2112993702-project-member] [instance: 396e49dd-48c0-496b-a1ec-190c33a22c5e] Powered off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 763.415658] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-655858bd-8f63-4e55-ac5e-dbecfbd88a62 tempest-ServersV294TestFqdnHostnames-2112993702 tempest-ServersV294TestFqdnHostnames-2112993702-project-member] [instance: 396e49dd-48c0-496b-a1ec-190c33a22c5e] Unregistering the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 763.416840] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-71a64301-c7d2-4518-b4f7-07c4fa47673f {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.419510] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-882273a6-70fb-4e58-891d-0d5387516ebf {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.435270] env[63515]: DEBUG nova.compute.provider_tree [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Updating inventory in ProviderTree for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 763.445199] env[63515]: DEBUG oslo_vmware.api [None req-b8b2a7f6-808d-40d4-8de9-c8863e5b55c5 tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] Task: {'id': task-1111143, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.121237} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 763.446711] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-b8b2a7f6-808d-40d4-8de9-c8863e5b55c5 tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] [instance: 30b8b872-40ca-4297-b98e-a64c83a75483] Extended root virtual disk {{(pid=63515) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 763.447716] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f361562d-6f74-4bb3-ad67-a51affa7626f {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.474726] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-b8b2a7f6-808d-40d4-8de9-c8863e5b55c5 tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] [instance: 30b8b872-40ca-4297-b98e-a64c83a75483] Reconfiguring VM instance instance-00000035 to attach disk [datastore1] 30b8b872-40ca-4297-b98e-a64c83a75483/30b8b872-40ca-4297-b98e-a64c83a75483.vmdk or device None with type sparse {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 763.474726] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ea9c0a0c-08f3-478f-932a-83aec0934937 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.494383] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-655858bd-8f63-4e55-ac5e-dbecfbd88a62 tempest-ServersV294TestFqdnHostnames-2112993702 tempest-ServersV294TestFqdnHostnames-2112993702-project-member] [instance: 396e49dd-48c0-496b-a1ec-190c33a22c5e] Unregistered the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 763.494450] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-655858bd-8f63-4e55-ac5e-dbecfbd88a62 tempest-ServersV294TestFqdnHostnames-2112993702 tempest-ServersV294TestFqdnHostnames-2112993702-project-member] [instance: 396e49dd-48c0-496b-a1ec-190c33a22c5e] Deleting contents of the VM from datastore datastore2 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 763.495015] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-655858bd-8f63-4e55-ac5e-dbecfbd88a62 tempest-ServersV294TestFqdnHostnames-2112993702 tempest-ServersV294TestFqdnHostnames-2112993702-project-member] Deleting the datastore file [datastore2] 396e49dd-48c0-496b-a1ec-190c33a22c5e {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 763.496560] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-010ae520-caf3-4fd0-a6ab-3f8c95818a39 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.498692] env[63515]: DEBUG oslo_vmware.api [None req-b8b2a7f6-808d-40d4-8de9-c8863e5b55c5 tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] Waiting for the task: (returnval){ [ 763.498692] env[63515]: value = "task-1111148" [ 763.498692] env[63515]: _type = "Task" [ 763.498692] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 763.505527] env[63515]: DEBUG oslo_vmware.api [None req-655858bd-8f63-4e55-ac5e-dbecfbd88a62 tempest-ServersV294TestFqdnHostnames-2112993702 tempest-ServersV294TestFqdnHostnames-2112993702-project-member] Waiting for the task: (returnval){ [ 763.505527] env[63515]: value = "task-1111149" [ 763.505527] env[63515]: _type = "Task" [ 763.505527] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 763.510168] env[63515]: DEBUG oslo_vmware.api [None req-b8b2a7f6-808d-40d4-8de9-c8863e5b55c5 tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] Task: {'id': task-1111148, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 763.519709] env[63515]: DEBUG oslo_vmware.api [None req-655858bd-8f63-4e55-ac5e-dbecfbd88a62 tempest-ServersV294TestFqdnHostnames-2112993702 tempest-ServersV294TestFqdnHostnames-2112993702-project-member] Task: {'id': task-1111149, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 763.604861] env[63515]: DEBUG nova.compute.manager [req-274f21b0-84cb-49ec-b7a7-4186c7aaf909 req-79f189b4-749c-46b2-8d39-edf7a1c24496 service nova] [instance: 90f4930b-aaa0-4c4b-9ab8-92aed45e200b] Received event network-vif-deleted-a35de7c4-c1af-482b-99a5-1059368cefbb {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 763.605077] env[63515]: INFO nova.compute.manager [req-274f21b0-84cb-49ec-b7a7-4186c7aaf909 req-79f189b4-749c-46b2-8d39-edf7a1c24496 service nova] [instance: 90f4930b-aaa0-4c4b-9ab8-92aed45e200b] Neutron deleted interface a35de7c4-c1af-482b-99a5-1059368cefbb; detaching it from the instance and deleting it from the info cache [ 763.605288] env[63515]: DEBUG nova.network.neutron [req-274f21b0-84cb-49ec-b7a7-4186c7aaf909 req-79f189b4-749c-46b2-8d39-edf7a1c24496 service nova] [instance: 90f4930b-aaa0-4c4b-9ab8-92aed45e200b] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 763.681109] env[63515]: DEBUG nova.compute.manager [req-eb32f99f-adc6-44f1-b2c2-c43f7eaa52a1 req-e860fdcb-8847-492b-886f-487aa50bb1ef service nova] [instance: 001a48e1-b237-41f8-a6f4-2fe98d173512] Received event network-vif-deleted-ed0b7b69-0679-4b35-af6f-44c60792b411 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 763.681339] env[63515]: INFO nova.compute.manager [req-eb32f99f-adc6-44f1-b2c2-c43f7eaa52a1 req-e860fdcb-8847-492b-886f-487aa50bb1ef service nova] [instance: 001a48e1-b237-41f8-a6f4-2fe98d173512] Neutron deleted interface ed0b7b69-0679-4b35-af6f-44c60792b411; detaching it from the instance and deleting it from the info cache [ 763.682170] env[63515]: DEBUG nova.network.neutron [req-eb32f99f-adc6-44f1-b2c2-c43f7eaa52a1 req-e860fdcb-8847-492b-886f-487aa50bb1ef service nova] [instance: 001a48e1-b237-41f8-a6f4-2fe98d173512] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 763.686337] env[63515]: DEBUG oslo_vmware.api [None req-8f7a1df0-5f0d-424a-ae8f-76bee2ee3dbb tempest-ServersAaction247Test-2019176257 tempest-ServersAaction247Test-2019176257-project-member] Task: {'id': task-1111144, 'name': Rename_Task, 'duration_secs': 0.185244} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 763.687084] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-8f7a1df0-5f0d-424a-ae8f-76bee2ee3dbb tempest-ServersAaction247Test-2019176257 tempest-ServersAaction247Test-2019176257-project-member] [instance: 5d50fbd6-a7fb-422c-9dd7-df7140fd0aae] Powering on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 763.687598] env[63515]: DEBUG nova.objects.instance [None req-e1d868d5-2e2f-4122-bebc-bb1318459f7f tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Lazy-loading 'flavor' on Instance uuid d2a4f66a-f1a2-4f0d-826d-0b7c4f04bd23 {{(pid=63515) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 763.689109] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7877a127-0c04-42f5-ab79-ec0c4bf7d679 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.697227] env[63515]: DEBUG oslo_vmware.api [None req-8f7a1df0-5f0d-424a-ae8f-76bee2ee3dbb tempest-ServersAaction247Test-2019176257 tempest-ServersAaction247Test-2019176257-project-member] Waiting for the task: (returnval){ [ 763.697227] env[63515]: value = "task-1111150" [ 763.697227] env[63515]: _type = "Task" [ 763.697227] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 763.707151] env[63515]: DEBUG oslo_vmware.api [None req-8f7a1df0-5f0d-424a-ae8f-76bee2ee3dbb tempest-ServersAaction247Test-2019176257 tempest-ServersAaction247Test-2019176257-project-member] Task: {'id': task-1111150, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 763.836195] env[63515]: DEBUG oslo_vmware.api [None req-db31b1fd-34e1-40f4-b7c8-92b451eb1400 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] Task: {'id': task-1111146, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.239931} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 763.836608] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-db31b1fd-34e1-40f4-b7c8-92b451eb1400 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] Deleted the datastore file {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 763.836823] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-db31b1fd-34e1-40f4-b7c8-92b451eb1400 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] [instance: 8eb17506-ff93-4d25-b9af-ec5886569e65] Deleted contents of the VM from datastore datastore2 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 763.836965] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-db31b1fd-34e1-40f4-b7c8-92b451eb1400 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] [instance: 8eb17506-ff93-4d25-b9af-ec5886569e65] Instance destroyed {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 763.949040] env[63515]: DEBUG nova.network.neutron [-] [instance: 90f4930b-aaa0-4c4b-9ab8-92aed45e200b] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 763.964385] env[63515]: ERROR nova.scheduler.client.report [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] [req-b02e03ec-8242-4fb0-8342-383da653a985] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 2bd86232-2b6f-44d5-9057-1a3a6b27185a. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-b02e03ec-8242-4fb0-8342-383da653a985"}]} [ 763.986231] env[63515]: DEBUG nova.scheduler.client.report [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Refreshing inventories for resource provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 764.001415] env[63515]: DEBUG nova.scheduler.client.report [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Updating ProviderTree inventory for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 764.001415] env[63515]: DEBUG nova.compute.provider_tree [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Updating inventory in ProviderTree for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 764.011956] env[63515]: DEBUG oslo_vmware.api [None req-b8b2a7f6-808d-40d4-8de9-c8863e5b55c5 tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] Task: {'id': task-1111148, 'name': ReconfigVM_Task, 'duration_secs': 0.448698} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 764.015955] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-b8b2a7f6-808d-40d4-8de9-c8863e5b55c5 tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] [instance: 30b8b872-40ca-4297-b98e-a64c83a75483] Reconfigured VM instance instance-00000035 to attach disk [datastore1] 30b8b872-40ca-4297-b98e-a64c83a75483/30b8b872-40ca-4297-b98e-a64c83a75483.vmdk or device None with type sparse {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 764.017321] env[63515]: DEBUG nova.scheduler.client.report [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Refreshing aggregate associations for resource provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a, aggregates: None {{(pid=63515) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 764.019335] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f233039f-9a3c-4613-9ef7-a4b7dc3faf65 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.021773] env[63515]: DEBUG nova.network.neutron [-] [instance: 001a48e1-b237-41f8-a6f4-2fe98d173512] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 764.026562] env[63515]: DEBUG oslo_vmware.api [None req-655858bd-8f63-4e55-ac5e-dbecfbd88a62 tempest-ServersV294TestFqdnHostnames-2112993702 tempest-ServersV294TestFqdnHostnames-2112993702-project-member] Task: {'id': task-1111149, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.305268} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 764.027944] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-655858bd-8f63-4e55-ac5e-dbecfbd88a62 tempest-ServersV294TestFqdnHostnames-2112993702 tempest-ServersV294TestFqdnHostnames-2112993702-project-member] Deleted the datastore file {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 764.028166] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-655858bd-8f63-4e55-ac5e-dbecfbd88a62 tempest-ServersV294TestFqdnHostnames-2112993702 tempest-ServersV294TestFqdnHostnames-2112993702-project-member] [instance: 396e49dd-48c0-496b-a1ec-190c33a22c5e] Deleted contents of the VM from datastore datastore2 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 764.028319] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-655858bd-8f63-4e55-ac5e-dbecfbd88a62 tempest-ServersV294TestFqdnHostnames-2112993702 tempest-ServersV294TestFqdnHostnames-2112993702-project-member] [instance: 396e49dd-48c0-496b-a1ec-190c33a22c5e] Instance destroyed {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 764.028500] env[63515]: INFO nova.compute.manager [None req-655858bd-8f63-4e55-ac5e-dbecfbd88a62 tempest-ServersV294TestFqdnHostnames-2112993702 tempest-ServersV294TestFqdnHostnames-2112993702-project-member] [instance: 396e49dd-48c0-496b-a1ec-190c33a22c5e] Took 1.17 seconds to destroy the instance on the hypervisor. [ 764.029165] env[63515]: DEBUG oslo.service.loopingcall [None req-655858bd-8f63-4e55-ac5e-dbecfbd88a62 tempest-ServersV294TestFqdnHostnames-2112993702 tempest-ServersV294TestFqdnHostnames-2112993702-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 764.029165] env[63515]: DEBUG oslo_vmware.api [None req-b8b2a7f6-808d-40d4-8de9-c8863e5b55c5 tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] Waiting for the task: (returnval){ [ 764.029165] env[63515]: value = "task-1111151" [ 764.029165] env[63515]: _type = "Task" [ 764.029165] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 764.030022] env[63515]: DEBUG nova.compute.manager [-] [instance: 396e49dd-48c0-496b-a1ec-190c33a22c5e] Deallocating network for instance {{(pid=63515) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 764.030022] env[63515]: DEBUG nova.network.neutron [-] [instance: 396e49dd-48c0-496b-a1ec-190c33a22c5e] deallocate_for_instance() {{(pid=63515) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 764.041981] env[63515]: DEBUG oslo_vmware.api [None req-b8b2a7f6-808d-40d4-8de9-c8863e5b55c5 tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] Task: {'id': task-1111151, 'name': Rename_Task} progress is 6%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 764.043055] env[63515]: DEBUG nova.scheduler.client.report [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Refreshing trait associations for resource provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a, traits: COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=63515) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 764.109028] env[63515]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c8dab1b7-546f-444b-ad4e-076d6f8e692f {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.121273] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df663bf4-677a-4edb-8398-0ebd36f6d5ea {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.159767] env[63515]: DEBUG nova.compute.manager [req-274f21b0-84cb-49ec-b7a7-4186c7aaf909 req-79f189b4-749c-46b2-8d39-edf7a1c24496 service nova] [instance: 90f4930b-aaa0-4c4b-9ab8-92aed45e200b] Detach interface failed, port_id=a35de7c4-c1af-482b-99a5-1059368cefbb, reason: Instance 90f4930b-aaa0-4c4b-9ab8-92aed45e200b could not be found. {{(pid=63515) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 764.189604] env[63515]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-7dbdc5c0-6ab1-456e-9497-817e52272611 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.197089] env[63515]: DEBUG oslo_concurrency.lockutils [None req-e1d868d5-2e2f-4122-bebc-bb1318459f7f tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Acquiring lock "refresh_cache-d2a4f66a-f1a2-4f0d-826d-0b7c4f04bd23" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 764.197311] env[63515]: DEBUG oslo_concurrency.lockutils [None req-e1d868d5-2e2f-4122-bebc-bb1318459f7f tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Acquired lock "refresh_cache-d2a4f66a-f1a2-4f0d-826d-0b7c4f04bd23" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 764.197491] env[63515]: DEBUG nova.network.neutron [None req-e1d868d5-2e2f-4122-bebc-bb1318459f7f tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] [instance: d2a4f66a-f1a2-4f0d-826d-0b7c4f04bd23] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 764.197679] env[63515]: DEBUG nova.objects.instance [None req-e1d868d5-2e2f-4122-bebc-bb1318459f7f tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Lazy-loading 'info_cache' on Instance uuid d2a4f66a-f1a2-4f0d-826d-0b7c4f04bd23 {{(pid=63515) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 764.210020] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6895cb17-4c48-4a02-8636-08b1113fca79 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.229266] env[63515]: DEBUG oslo_vmware.api [None req-8f7a1df0-5f0d-424a-ae8f-76bee2ee3dbb tempest-ServersAaction247Test-2019176257 tempest-ServersAaction247Test-2019176257-project-member] Task: {'id': task-1111150, 'name': PowerOnVM_Task} progress is 88%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 764.243300] env[63515]: DEBUG nova.compute.manager [req-eb32f99f-adc6-44f1-b2c2-c43f7eaa52a1 req-e860fdcb-8847-492b-886f-487aa50bb1ef service nova] [instance: 001a48e1-b237-41f8-a6f4-2fe98d173512] Detach interface failed, port_id=ed0b7b69-0679-4b35-af6f-44c60792b411, reason: Instance 001a48e1-b237-41f8-a6f4-2fe98d173512 could not be found. {{(pid=63515) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 764.451717] env[63515]: INFO nova.compute.manager [-] [instance: 90f4930b-aaa0-4c4b-9ab8-92aed45e200b] Took 1.71 seconds to deallocate network for instance. [ 764.524223] env[63515]: INFO nova.compute.manager [-] [instance: 001a48e1-b237-41f8-a6f4-2fe98d173512] Took 1.42 seconds to deallocate network for instance. [ 764.542284] env[63515]: DEBUG oslo_vmware.api [None req-b8b2a7f6-808d-40d4-8de9-c8863e5b55c5 tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] Task: {'id': task-1111151, 'name': Rename_Task, 'duration_secs': 0.284079} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 764.545921] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-b8b2a7f6-808d-40d4-8de9-c8863e5b55c5 tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] [instance: 30b8b872-40ca-4297-b98e-a64c83a75483] Powering on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 764.545921] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-39817fc3-15ab-4d64-add8-cbf3c0713edb {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.553594] env[63515]: DEBUG oslo_vmware.api [None req-b8b2a7f6-808d-40d4-8de9-c8863e5b55c5 tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] Waiting for the task: (returnval){ [ 764.553594] env[63515]: value = "task-1111152" [ 764.553594] env[63515]: _type = "Task" [ 764.553594] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 764.565733] env[63515]: DEBUG oslo_vmware.api [None req-b8b2a7f6-808d-40d4-8de9-c8863e5b55c5 tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] Task: {'id': task-1111152, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 764.576026] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-135d2bb9-df56-415e-9df2-794343c415a3 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.583647] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4622316c-0e37-4398-a37d-5749a6416bf7 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.617423] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2dae45be-e1c2-49b8-ad21-c2d159e4addf {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.625528] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-723d4334-7850-45ee-a86d-7bb638a6f42a {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.645537] env[63515]: DEBUG nova.compute.provider_tree [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Updating inventory in ProviderTree for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 764.704711] env[63515]: DEBUG nova.objects.base [None req-e1d868d5-2e2f-4122-bebc-bb1318459f7f tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Object Instance lazy-loaded attributes: flavor,info_cache {{(pid=63515) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 764.716715] env[63515]: DEBUG oslo_vmware.api [None req-8f7a1df0-5f0d-424a-ae8f-76bee2ee3dbb tempest-ServersAaction247Test-2019176257 tempest-ServersAaction247Test-2019176257-project-member] Task: {'id': task-1111150, 'name': PowerOnVM_Task, 'duration_secs': 0.727883} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 764.717022] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-8f7a1df0-5f0d-424a-ae8f-76bee2ee3dbb tempest-ServersAaction247Test-2019176257 tempest-ServersAaction247Test-2019176257-project-member] [instance: 5d50fbd6-a7fb-422c-9dd7-df7140fd0aae] Powered on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 764.717218] env[63515]: INFO nova.compute.manager [None req-8f7a1df0-5f0d-424a-ae8f-76bee2ee3dbb tempest-ServersAaction247Test-2019176257 tempest-ServersAaction247Test-2019176257-project-member] [instance: 5d50fbd6-a7fb-422c-9dd7-df7140fd0aae] Took 4.95 seconds to spawn the instance on the hypervisor. [ 764.717440] env[63515]: DEBUG nova.compute.manager [None req-8f7a1df0-5f0d-424a-ae8f-76bee2ee3dbb tempest-ServersAaction247Test-2019176257 tempest-ServersAaction247Test-2019176257-project-member] [instance: 5d50fbd6-a7fb-422c-9dd7-df7140fd0aae] Checking state {{(pid=63515) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 764.718418] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8b7d14e-f89f-45b8-9838-857840f7884e {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.895293] env[63515]: DEBUG nova.virt.hardware [None req-db31b1fd-34e1-40f4-b7c8-92b451eb1400 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T02:52:14Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T02:51:56Z,direct_url=,disk_format='vmdk',id=8a120570-cb06-4099-b262-554ca0ad15c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b569255cc43e42e7ae3f2b4ad37c6ef4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T02:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 764.895592] env[63515]: DEBUG nova.virt.hardware [None req-db31b1fd-34e1-40f4-b7c8-92b451eb1400 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] Flavor limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 764.895746] env[63515]: DEBUG nova.virt.hardware [None req-db31b1fd-34e1-40f4-b7c8-92b451eb1400 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] Image limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 764.895937] env[63515]: DEBUG nova.virt.hardware [None req-db31b1fd-34e1-40f4-b7c8-92b451eb1400 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] Flavor pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 764.896142] env[63515]: DEBUG nova.virt.hardware [None req-db31b1fd-34e1-40f4-b7c8-92b451eb1400 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] Image pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 764.896312] env[63515]: DEBUG nova.virt.hardware [None req-db31b1fd-34e1-40f4-b7c8-92b451eb1400 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 764.896533] env[63515]: DEBUG nova.virt.hardware [None req-db31b1fd-34e1-40f4-b7c8-92b451eb1400 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 764.896697] env[63515]: DEBUG nova.virt.hardware [None req-db31b1fd-34e1-40f4-b7c8-92b451eb1400 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 764.896929] env[63515]: DEBUG nova.virt.hardware [None req-db31b1fd-34e1-40f4-b7c8-92b451eb1400 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] Got 1 possible topologies {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 764.897038] env[63515]: DEBUG nova.virt.hardware [None req-db31b1fd-34e1-40f4-b7c8-92b451eb1400 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 764.897226] env[63515]: DEBUG nova.virt.hardware [None req-db31b1fd-34e1-40f4-b7c8-92b451eb1400 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 764.899637] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c2fabad-39b9-49dc-b25d-3d839e99db94 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.906232] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7bcb3377-a129-43fd-a053-4de657f3c480 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.920308] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-db31b1fd-34e1-40f4-b7c8-92b451eb1400 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] [instance: 8eb17506-ff93-4d25-b9af-ec5886569e65] Instance VIF info [] {{(pid=63515) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 764.926129] env[63515]: DEBUG oslo.service.loopingcall [None req-db31b1fd-34e1-40f4-b7c8-92b451eb1400 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 764.926427] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8eb17506-ff93-4d25-b9af-ec5886569e65] Creating VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 764.926880] env[63515]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-83046b01-637a-44fc-8372-4f059f8c118a {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.943144] env[63515]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 764.943144] env[63515]: value = "task-1111153" [ 764.943144] env[63515]: _type = "Task" [ 764.943144] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 764.951167] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1111153, 'name': CreateVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 765.018958] env[63515]: INFO nova.compute.manager [None req-2e592662-996b-4dfa-8941-00965c1317fc tempest-ServersTestBootFromVolume-2042560423 tempest-ServersTestBootFromVolume-2042560423-project-member] [instance: 90f4930b-aaa0-4c4b-9ab8-92aed45e200b] Took 0.57 seconds to detach 1 volumes for instance. [ 765.022524] env[63515]: DEBUG nova.compute.manager [None req-2e592662-996b-4dfa-8941-00965c1317fc tempest-ServersTestBootFromVolume-2042560423 tempest-ServersTestBootFromVolume-2042560423-project-member] [instance: 90f4930b-aaa0-4c4b-9ab8-92aed45e200b] Deleting volume: 921350b3-bd4a-4fa1-848a-057b53f51f5c {{(pid=63515) _cleanup_volumes /opt/stack/nova/nova/compute/manager.py:3247}} [ 765.033685] env[63515]: DEBUG oslo_concurrency.lockutils [None req-aef604e2-0f8c-4f11-a420-cb96f46c502e tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 765.065267] env[63515]: DEBUG oslo_vmware.api [None req-b8b2a7f6-808d-40d4-8de9-c8863e5b55c5 tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] Task: {'id': task-1111152, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 765.160196] env[63515]: DEBUG nova.network.neutron [-] [instance: 396e49dd-48c0-496b-a1ec-190c33a22c5e] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 765.193562] env[63515]: DEBUG nova.scheduler.client.report [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Updated inventory for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a with generation 76 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 765.193720] env[63515]: DEBUG nova.compute.provider_tree [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Updating resource provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a generation from 76 to 77 during operation: update_inventory {{(pid=63515) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 765.193857] env[63515]: DEBUG nova.compute.provider_tree [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Updating inventory in ProviderTree for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 765.242907] env[63515]: INFO nova.compute.manager [None req-8f7a1df0-5f0d-424a-ae8f-76bee2ee3dbb tempest-ServersAaction247Test-2019176257 tempest-ServersAaction247Test-2019176257-project-member] [instance: 5d50fbd6-a7fb-422c-9dd7-df7140fd0aae] Took 30.24 seconds to build instance. [ 765.460973] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1111153, 'name': CreateVM_Task, 'duration_secs': 0.37337} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 765.461222] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8eb17506-ff93-4d25-b9af-ec5886569e65] Created VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 765.461732] env[63515]: DEBUG oslo_concurrency.lockutils [None req-db31b1fd-34e1-40f4-b7c8-92b451eb1400 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 765.461961] env[63515]: DEBUG oslo_concurrency.lockutils [None req-db31b1fd-34e1-40f4-b7c8-92b451eb1400 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 765.462371] env[63515]: DEBUG oslo_concurrency.lockutils [None req-db31b1fd-34e1-40f4-b7c8-92b451eb1400 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 765.462708] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7c4243f5-4b80-4e1e-b011-0a9a70463c57 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.467779] env[63515]: DEBUG oslo_vmware.api [None req-db31b1fd-34e1-40f4-b7c8-92b451eb1400 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] Waiting for the task: (returnval){ [ 765.467779] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]528adac8-92df-8f4a-d9c2-b9d3c5727139" [ 765.467779] env[63515]: _type = "Task" [ 765.467779] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 765.481086] env[63515]: DEBUG oslo_vmware.api [None req-db31b1fd-34e1-40f4-b7c8-92b451eb1400 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]528adac8-92df-8f4a-d9c2-b9d3c5727139, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 765.568377] env[63515]: DEBUG oslo_vmware.api [None req-b8b2a7f6-808d-40d4-8de9-c8863e5b55c5 tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] Task: {'id': task-1111152, 'name': PowerOnVM_Task, 'duration_secs': 0.959436} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 765.568377] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-b8b2a7f6-808d-40d4-8de9-c8863e5b55c5 tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] [instance: 30b8b872-40ca-4297-b98e-a64c83a75483] Powered on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 765.568377] env[63515]: INFO nova.compute.manager [None req-b8b2a7f6-808d-40d4-8de9-c8863e5b55c5 tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] [instance: 30b8b872-40ca-4297-b98e-a64c83a75483] Took 8.39 seconds to spawn the instance on the hypervisor. [ 765.568377] env[63515]: DEBUG nova.compute.manager [None req-b8b2a7f6-808d-40d4-8de9-c8863e5b55c5 tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] [instance: 30b8b872-40ca-4297-b98e-a64c83a75483] Checking state {{(pid=63515) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 765.571783] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6cf2c510-9e35-46a1-9224-768910f4f4d3 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.575442] env[63515]: DEBUG oslo_concurrency.lockutils [None req-2e592662-996b-4dfa-8941-00965c1317fc tempest-ServersTestBootFromVolume-2042560423 tempest-ServersTestBootFromVolume-2042560423-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 765.636767] env[63515]: DEBUG nova.network.neutron [None req-e1d868d5-2e2f-4122-bebc-bb1318459f7f tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] [instance: d2a4f66a-f1a2-4f0d-826d-0b7c4f04bd23] Updating instance_info_cache with network_info: [{"id": "3bca52da-5931-4d22-bdf1-33072a3aca76", "address": "fa:16:3e:4e:8c:2d", "network": {"id": "d2aef558-9770-46ad-b51c-61bccb0de7ec", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-2087172546-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6f501ed813754f759464e338884993c1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6d4ef133-b6f3-41d1-add4-92a1482195cf", "external-id": "nsx-vlan-transportzone-446", "segmentation_id": 446, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3bca52da-59", "ovs_interfaceid": "3bca52da-5931-4d22-bdf1-33072a3aca76", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 765.663578] env[63515]: INFO nova.compute.manager [-] [instance: 396e49dd-48c0-496b-a1ec-190c33a22c5e] Took 1.63 seconds to deallocate network for instance. [ 765.699819] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=63515) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 765.699819] env[63515]: DEBUG oslo_concurrency.lockutils [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 7.965s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 765.699992] env[63515]: DEBUG oslo_concurrency.lockutils [None req-549cc3f8-330f-4c6a-979d-35eedc6f8146 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 28.530s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 765.701492] env[63515]: INFO nova.compute.claims [None req-549cc3f8-330f-4c6a-979d-35eedc6f8146 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: 84dbf321-2c70-4e08-b430-cb5a06fc6829] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 765.746546] env[63515]: DEBUG oslo_concurrency.lockutils [None req-8f7a1df0-5f0d-424a-ae8f-76bee2ee3dbb tempest-ServersAaction247Test-2019176257 tempest-ServersAaction247Test-2019176257-project-member] Lock "5d50fbd6-a7fb-422c-9dd7-df7140fd0aae" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 92.170s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 765.858801] env[63515]: DEBUG nova.compute.manager [req-ae1221eb-d5d8-4b28-8d0a-54d0937b82f4 req-88c18348-b242-4954-ae17-3a5a404a4684 service nova] [instance: 396e49dd-48c0-496b-a1ec-190c33a22c5e] Received event network-vif-deleted-77fbf94a-db5a-4569-b4e7-2d9edbe411a1 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 765.978856] env[63515]: DEBUG oslo_vmware.api [None req-db31b1fd-34e1-40f4-b7c8-92b451eb1400 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]528adac8-92df-8f4a-d9c2-b9d3c5727139, 'name': SearchDatastore_Task, 'duration_secs': 0.015099} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 765.979189] env[63515]: DEBUG oslo_concurrency.lockutils [None req-db31b1fd-34e1-40f4-b7c8-92b451eb1400 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 765.979416] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-db31b1fd-34e1-40f4-b7c8-92b451eb1400 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] [instance: 8eb17506-ff93-4d25-b9af-ec5886569e65] Processing image 8a120570-cb06-4099-b262-554ca0ad15c5 {{(pid=63515) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 765.979650] env[63515]: DEBUG oslo_concurrency.lockutils [None req-db31b1fd-34e1-40f4-b7c8-92b451eb1400 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 765.979797] env[63515]: DEBUG oslo_concurrency.lockutils [None req-db31b1fd-34e1-40f4-b7c8-92b451eb1400 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 765.979978] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-db31b1fd-34e1-40f4-b7c8-92b451eb1400 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 765.980280] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e9128973-3656-4c37-bd0e-eb58fad3c3d4 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.990029] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-db31b1fd-34e1-40f4-b7c8-92b451eb1400 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 765.990029] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-db31b1fd-34e1-40f4-b7c8-92b451eb1400 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63515) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 765.990335] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e921fcde-bc18-416d-8c09-54254bccb7de {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.996090] env[63515]: DEBUG oslo_vmware.api [None req-db31b1fd-34e1-40f4-b7c8-92b451eb1400 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] Waiting for the task: (returnval){ [ 765.996090] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]52c61ff2-2e7f-9cbc-02c0-ca8471bf647c" [ 765.996090] env[63515]: _type = "Task" [ 765.996090] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 766.003857] env[63515]: DEBUG oslo_vmware.api [None req-db31b1fd-34e1-40f4-b7c8-92b451eb1400 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52c61ff2-2e7f-9cbc-02c0-ca8471bf647c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 766.094206] env[63515]: INFO nova.compute.manager [None req-b8b2a7f6-808d-40d4-8de9-c8863e5b55c5 tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] [instance: 30b8b872-40ca-4297-b98e-a64c83a75483] Took 36.25 seconds to build instance. [ 766.139327] env[63515]: DEBUG oslo_concurrency.lockutils [None req-e1d868d5-2e2f-4122-bebc-bb1318459f7f tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Releasing lock "refresh_cache-d2a4f66a-f1a2-4f0d-826d-0b7c4f04bd23" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 766.170466] env[63515]: DEBUG oslo_concurrency.lockutils [None req-655858bd-8f63-4e55-ac5e-dbecfbd88a62 tempest-ServersV294TestFqdnHostnames-2112993702 tempest-ServersV294TestFqdnHostnames-2112993702-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 766.250036] env[63515]: DEBUG nova.compute.manager [None req-41b4a543-33b4-43d9-863c-4344bf9d0a46 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: d926c699-a64a-4942-9ef4-f0166414661d] Starting instance... {{(pid=63515) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 766.343033] env[63515]: DEBUG nova.compute.manager [None req-faa3e5a0-454d-4232-b593-1cbff3aacd1a tempest-ServersAaction247Test-2019176257 tempest-ServersAaction247Test-2019176257-project-member] [instance: 5d50fbd6-a7fb-422c-9dd7-df7140fd0aae] Checking state {{(pid=63515) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 766.343716] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef2d7909-2bcf-419b-984e-637e341abb70 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.509524] env[63515]: DEBUG oslo_vmware.api [None req-db31b1fd-34e1-40f4-b7c8-92b451eb1400 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52c61ff2-2e7f-9cbc-02c0-ca8471bf647c, 'name': SearchDatastore_Task, 'duration_secs': 0.008845} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 766.510504] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a6ccbafd-f98e-42e3-aada-86b326aeaeda {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.515742] env[63515]: DEBUG oslo_vmware.api [None req-db31b1fd-34e1-40f4-b7c8-92b451eb1400 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] Waiting for the task: (returnval){ [ 766.515742] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]523733e4-85b5-0ce6-1261-781b673eafc0" [ 766.515742] env[63515]: _type = "Task" [ 766.515742] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 766.523883] env[63515]: DEBUG oslo_vmware.api [None req-db31b1fd-34e1-40f4-b7c8-92b451eb1400 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]523733e4-85b5-0ce6-1261-781b673eafc0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 766.546505] env[63515]: DEBUG oslo_concurrency.lockutils [None req-65f1d514-d43e-4354-a5f1-c564a5380800 tempest-ServersAaction247Test-2019176257 tempest-ServersAaction247Test-2019176257-project-member] Acquiring lock "5d50fbd6-a7fb-422c-9dd7-df7140fd0aae" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 766.546779] env[63515]: DEBUG oslo_concurrency.lockutils [None req-65f1d514-d43e-4354-a5f1-c564a5380800 tempest-ServersAaction247Test-2019176257 tempest-ServersAaction247Test-2019176257-project-member] Lock "5d50fbd6-a7fb-422c-9dd7-df7140fd0aae" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 766.546966] env[63515]: DEBUG oslo_concurrency.lockutils [None req-65f1d514-d43e-4354-a5f1-c564a5380800 tempest-ServersAaction247Test-2019176257 tempest-ServersAaction247Test-2019176257-project-member] Acquiring lock "5d50fbd6-a7fb-422c-9dd7-df7140fd0aae-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 766.547171] env[63515]: DEBUG oslo_concurrency.lockutils [None req-65f1d514-d43e-4354-a5f1-c564a5380800 tempest-ServersAaction247Test-2019176257 tempest-ServersAaction247Test-2019176257-project-member] Lock "5d50fbd6-a7fb-422c-9dd7-df7140fd0aae-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 766.547357] env[63515]: DEBUG oslo_concurrency.lockutils [None req-65f1d514-d43e-4354-a5f1-c564a5380800 tempest-ServersAaction247Test-2019176257 tempest-ServersAaction247Test-2019176257-project-member] Lock "5d50fbd6-a7fb-422c-9dd7-df7140fd0aae-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 766.551185] env[63515]: INFO nova.compute.manager [None req-65f1d514-d43e-4354-a5f1-c564a5380800 tempest-ServersAaction247Test-2019176257 tempest-ServersAaction247Test-2019176257-project-member] [instance: 5d50fbd6-a7fb-422c-9dd7-df7140fd0aae] Terminating instance [ 766.553857] env[63515]: DEBUG oslo_concurrency.lockutils [None req-65f1d514-d43e-4354-a5f1-c564a5380800 tempest-ServersAaction247Test-2019176257 tempest-ServersAaction247Test-2019176257-project-member] Acquiring lock "refresh_cache-5d50fbd6-a7fb-422c-9dd7-df7140fd0aae" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 766.554064] env[63515]: DEBUG oslo_concurrency.lockutils [None req-65f1d514-d43e-4354-a5f1-c564a5380800 tempest-ServersAaction247Test-2019176257 tempest-ServersAaction247Test-2019176257-project-member] Acquired lock "refresh_cache-5d50fbd6-a7fb-422c-9dd7-df7140fd0aae" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 766.554201] env[63515]: DEBUG nova.network.neutron [None req-65f1d514-d43e-4354-a5f1-c564a5380800 tempest-ServersAaction247Test-2019176257 tempest-ServersAaction247Test-2019176257-project-member] [instance: 5d50fbd6-a7fb-422c-9dd7-df7140fd0aae] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 766.572958] env[63515]: DEBUG nova.compute.manager [req-d675111a-4edc-47a6-9392-551814bc026b req-e6b1df56-3d3d-4e1e-97a8-2a50ff5f99d4 service nova] [instance: 30b8b872-40ca-4297-b98e-a64c83a75483] Received event network-changed-b8fe94ee-46b7-489a-819d-0699bc02bbb7 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 766.572958] env[63515]: DEBUG nova.compute.manager [req-d675111a-4edc-47a6-9392-551814bc026b req-e6b1df56-3d3d-4e1e-97a8-2a50ff5f99d4 service nova] [instance: 30b8b872-40ca-4297-b98e-a64c83a75483] Refreshing instance network info cache due to event network-changed-b8fe94ee-46b7-489a-819d-0699bc02bbb7. {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 766.572958] env[63515]: DEBUG oslo_concurrency.lockutils [req-d675111a-4edc-47a6-9392-551814bc026b req-e6b1df56-3d3d-4e1e-97a8-2a50ff5f99d4 service nova] Acquiring lock "refresh_cache-30b8b872-40ca-4297-b98e-a64c83a75483" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 766.572958] env[63515]: DEBUG oslo_concurrency.lockutils [req-d675111a-4edc-47a6-9392-551814bc026b req-e6b1df56-3d3d-4e1e-97a8-2a50ff5f99d4 service nova] Acquired lock "refresh_cache-30b8b872-40ca-4297-b98e-a64c83a75483" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 766.572958] env[63515]: DEBUG nova.network.neutron [req-d675111a-4edc-47a6-9392-551814bc026b req-e6b1df56-3d3d-4e1e-97a8-2a50ff5f99d4 service nova] [instance: 30b8b872-40ca-4297-b98e-a64c83a75483] Refreshing network info cache for port b8fe94ee-46b7-489a-819d-0699bc02bbb7 {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 766.597018] env[63515]: DEBUG oslo_concurrency.lockutils [None req-b8b2a7f6-808d-40d4-8de9-c8863e5b55c5 tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] Lock "30b8b872-40ca-4297-b98e-a64c83a75483" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 123.197s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 766.644135] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-e1d868d5-2e2f-4122-bebc-bb1318459f7f tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] [instance: d2a4f66a-f1a2-4f0d-826d-0b7c4f04bd23] Powering on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 766.644135] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-cfb2cdcd-0469-4ded-b6c2-e9eb02b69e24 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.649885] env[63515]: DEBUG oslo_vmware.api [None req-e1d868d5-2e2f-4122-bebc-bb1318459f7f tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Waiting for the task: (returnval){ [ 766.649885] env[63515]: value = "task-1111155" [ 766.649885] env[63515]: _type = "Task" [ 766.649885] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 766.661670] env[63515]: DEBUG oslo_vmware.api [None req-e1d868d5-2e2f-4122-bebc-bb1318459f7f tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Task: {'id': task-1111155, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 766.767882] env[63515]: DEBUG oslo_concurrency.lockutils [None req-41b4a543-33b4-43d9-863c-4344bf9d0a46 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 766.854815] env[63515]: INFO nova.compute.manager [None req-faa3e5a0-454d-4232-b593-1cbff3aacd1a tempest-ServersAaction247Test-2019176257 tempest-ServersAaction247Test-2019176257-project-member] [instance: 5d50fbd6-a7fb-422c-9dd7-df7140fd0aae] instance snapshotting [ 766.855502] env[63515]: DEBUG nova.objects.instance [None req-faa3e5a0-454d-4232-b593-1cbff3aacd1a tempest-ServersAaction247Test-2019176257 tempest-ServersAaction247Test-2019176257-project-member] Lazy-loading 'flavor' on Instance uuid 5d50fbd6-a7fb-422c-9dd7-df7140fd0aae {{(pid=63515) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 767.029460] env[63515]: DEBUG oslo_vmware.api [None req-db31b1fd-34e1-40f4-b7c8-92b451eb1400 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]523733e4-85b5-0ce6-1261-781b673eafc0, 'name': SearchDatastore_Task, 'duration_secs': 0.011799} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 767.032051] env[63515]: DEBUG oslo_concurrency.lockutils [None req-db31b1fd-34e1-40f4-b7c8-92b451eb1400 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 767.032342] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-db31b1fd-34e1-40f4-b7c8-92b451eb1400 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk to [datastore2] 8eb17506-ff93-4d25-b9af-ec5886569e65/8eb17506-ff93-4d25-b9af-ec5886569e65.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 767.032827] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-71cfb51b-9d39-45f8-aeeb-b51dde9a8df9 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.040076] env[63515]: DEBUG oslo_vmware.api [None req-db31b1fd-34e1-40f4-b7c8-92b451eb1400 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] Waiting for the task: (returnval){ [ 767.040076] env[63515]: value = "task-1111156" [ 767.040076] env[63515]: _type = "Task" [ 767.040076] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 767.053784] env[63515]: DEBUG oslo_vmware.api [None req-db31b1fd-34e1-40f4-b7c8-92b451eb1400 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] Task: {'id': task-1111156, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 767.084493] env[63515]: DEBUG nova.network.neutron [None req-65f1d514-d43e-4354-a5f1-c564a5380800 tempest-ServersAaction247Test-2019176257 tempest-ServersAaction247Test-2019176257-project-member] [instance: 5d50fbd6-a7fb-422c-9dd7-df7140fd0aae] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 767.159202] env[63515]: DEBUG nova.network.neutron [None req-65f1d514-d43e-4354-a5f1-c564a5380800 tempest-ServersAaction247Test-2019176257 tempest-ServersAaction247Test-2019176257-project-member] [instance: 5d50fbd6-a7fb-422c-9dd7-df7140fd0aae] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 767.164166] env[63515]: DEBUG oslo_vmware.api [None req-e1d868d5-2e2f-4122-bebc-bb1318459f7f tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Task: {'id': task-1111155, 'name': PowerOnVM_Task, 'duration_secs': 0.458837} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 767.164877] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-e1d868d5-2e2f-4122-bebc-bb1318459f7f tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] [instance: d2a4f66a-f1a2-4f0d-826d-0b7c4f04bd23] Powered on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 767.165257] env[63515]: DEBUG nova.compute.manager [None req-e1d868d5-2e2f-4122-bebc-bb1318459f7f tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] [instance: d2a4f66a-f1a2-4f0d-826d-0b7c4f04bd23] Checking state {{(pid=63515) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 767.166423] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-006dbfef-d517-4b7d-a252-6e710ad67fb5 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.171852] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-237787e4-6c34-49f2-9f40-302869a6b524 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.189263] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c363cfcd-a347-4a85-915e-e8fe471099b7 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.196112] env[63515]: DEBUG oslo_concurrency.lockutils [None req-a6c8eb46-cd36-421b-bb36-d08f2b34cdc9 tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] Acquiring lock "30b8b872-40ca-4297-b98e-a64c83a75483" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 767.196112] env[63515]: DEBUG oslo_concurrency.lockutils [None req-a6c8eb46-cd36-421b-bb36-d08f2b34cdc9 tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] Lock "30b8b872-40ca-4297-b98e-a64c83a75483" acquired by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 767.196112] env[63515]: INFO nova.compute.manager [None req-a6c8eb46-cd36-421b-bb36-d08f2b34cdc9 tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] [instance: 30b8b872-40ca-4297-b98e-a64c83a75483] Rebooting instance [ 767.235413] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34551363-e22d-4411-b605-9c2857378461 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.245066] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8caca0d0-5f0e-4952-ac12-374377a92418 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.263921] env[63515]: DEBUG nova.compute.provider_tree [None req-549cc3f8-330f-4c6a-979d-35eedc6f8146 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 767.266304] env[63515]: DEBUG oslo_concurrency.lockutils [None req-a6c8eb46-cd36-421b-bb36-d08f2b34cdc9 tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] Acquiring lock "refresh_cache-30b8b872-40ca-4297-b98e-a64c83a75483" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 767.361667] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28a6404c-0f4a-4142-8a95-e8dac37b1afa {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.384638] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9b22ffe-2ea2-4208-b8e2-3ce61c620735 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.428941] env[63515]: DEBUG nova.network.neutron [req-d675111a-4edc-47a6-9392-551814bc026b req-e6b1df56-3d3d-4e1e-97a8-2a50ff5f99d4 service nova] [instance: 30b8b872-40ca-4297-b98e-a64c83a75483] Updated VIF entry in instance network info cache for port b8fe94ee-46b7-489a-819d-0699bc02bbb7. {{(pid=63515) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 767.429421] env[63515]: DEBUG nova.network.neutron [req-d675111a-4edc-47a6-9392-551814bc026b req-e6b1df56-3d3d-4e1e-97a8-2a50ff5f99d4 service nova] [instance: 30b8b872-40ca-4297-b98e-a64c83a75483] Updating instance_info_cache with network_info: [{"id": "b8fe94ee-46b7-489a-819d-0699bc02bbb7", "address": "fa:16:3e:42:a9:82", "network": {"id": "bd2f642c-518b-478b-8358-7426cebde731", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-1534671780-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "72b65c09ed6d4492864022352180ff49", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "92f3cfd6-c130-4390-8910-865fbc42afd1", "external-id": "nsx-vlan-transportzone-142", "segmentation_id": 142, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb8fe94ee-46", "ovs_interfaceid": "b8fe94ee-46b7-489a-819d-0699bc02bbb7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 767.550465] env[63515]: DEBUG oslo_vmware.api [None req-db31b1fd-34e1-40f4-b7c8-92b451eb1400 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] Task: {'id': task-1111156, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 767.666018] env[63515]: DEBUG oslo_concurrency.lockutils [None req-65f1d514-d43e-4354-a5f1-c564a5380800 tempest-ServersAaction247Test-2019176257 tempest-ServersAaction247Test-2019176257-project-member] Releasing lock "refresh_cache-5d50fbd6-a7fb-422c-9dd7-df7140fd0aae" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 767.666499] env[63515]: DEBUG nova.compute.manager [None req-65f1d514-d43e-4354-a5f1-c564a5380800 tempest-ServersAaction247Test-2019176257 tempest-ServersAaction247Test-2019176257-project-member] [instance: 5d50fbd6-a7fb-422c-9dd7-df7140fd0aae] Start destroying the instance on the hypervisor. {{(pid=63515) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 767.666709] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-65f1d514-d43e-4354-a5f1-c564a5380800 tempest-ServersAaction247Test-2019176257 tempest-ServersAaction247Test-2019176257-project-member] [instance: 5d50fbd6-a7fb-422c-9dd7-df7140fd0aae] Destroying instance {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 767.667627] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b97e5f5-e2bb-4bf0-8dfc-9f6e2bfb2076 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.675209] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-65f1d514-d43e-4354-a5f1-c564a5380800 tempest-ServersAaction247Test-2019176257 tempest-ServersAaction247Test-2019176257-project-member] [instance: 5d50fbd6-a7fb-422c-9dd7-df7140fd0aae] Powering off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 767.675513] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e39d1ba2-52fd-4312-a276-79ea515e42f8 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.681691] env[63515]: DEBUG oslo_vmware.api [None req-65f1d514-d43e-4354-a5f1-c564a5380800 tempest-ServersAaction247Test-2019176257 tempest-ServersAaction247Test-2019176257-project-member] Waiting for the task: (returnval){ [ 767.681691] env[63515]: value = "task-1111157" [ 767.681691] env[63515]: _type = "Task" [ 767.681691] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 767.692880] env[63515]: DEBUG oslo_vmware.api [None req-65f1d514-d43e-4354-a5f1-c564a5380800 tempest-ServersAaction247Test-2019176257 tempest-ServersAaction247Test-2019176257-project-member] Task: {'id': task-1111157, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 767.768089] env[63515]: DEBUG nova.scheduler.client.report [None req-549cc3f8-330f-4c6a-979d-35eedc6f8146 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 767.896465] env[63515]: DEBUG nova.compute.manager [None req-faa3e5a0-454d-4232-b593-1cbff3aacd1a tempest-ServersAaction247Test-2019176257 tempest-ServersAaction247Test-2019176257-project-member] [instance: 5d50fbd6-a7fb-422c-9dd7-df7140fd0aae] Instance disappeared during snapshot {{(pid=63515) _snapshot_instance /opt/stack/nova/nova/compute/manager.py:4494}} [ 767.932733] env[63515]: DEBUG oslo_concurrency.lockutils [req-d675111a-4edc-47a6-9392-551814bc026b req-e6b1df56-3d3d-4e1e-97a8-2a50ff5f99d4 service nova] Releasing lock "refresh_cache-30b8b872-40ca-4297-b98e-a64c83a75483" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 767.933257] env[63515]: DEBUG oslo_concurrency.lockutils [None req-a6c8eb46-cd36-421b-bb36-d08f2b34cdc9 tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] Acquired lock "refresh_cache-30b8b872-40ca-4297-b98e-a64c83a75483" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 767.933546] env[63515]: DEBUG nova.network.neutron [None req-a6c8eb46-cd36-421b-bb36-d08f2b34cdc9 tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] [instance: 30b8b872-40ca-4297-b98e-a64c83a75483] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 768.047256] env[63515]: DEBUG nova.compute.manager [None req-faa3e5a0-454d-4232-b593-1cbff3aacd1a tempest-ServersAaction247Test-2019176257 tempest-ServersAaction247Test-2019176257-project-member] [instance: 5d50fbd6-a7fb-422c-9dd7-df7140fd0aae] Found 0 images (rotation: 2) {{(pid=63515) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4554}} [ 768.054198] env[63515]: DEBUG oslo_vmware.api [None req-db31b1fd-34e1-40f4-b7c8-92b451eb1400 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] Task: {'id': task-1111156, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.530261} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 768.054535] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-db31b1fd-34e1-40f4-b7c8-92b451eb1400 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk to [datastore2] 8eb17506-ff93-4d25-b9af-ec5886569e65/8eb17506-ff93-4d25-b9af-ec5886569e65.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 768.054762] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-db31b1fd-34e1-40f4-b7c8-92b451eb1400 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] [instance: 8eb17506-ff93-4d25-b9af-ec5886569e65] Extending root virtual disk to 1048576 {{(pid=63515) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 768.055139] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-4c11d6e1-6396-48d7-83e7-ea895034fb99 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.062670] env[63515]: DEBUG oslo_vmware.api [None req-db31b1fd-34e1-40f4-b7c8-92b451eb1400 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] Waiting for the task: (returnval){ [ 768.062670] env[63515]: value = "task-1111158" [ 768.062670] env[63515]: _type = "Task" [ 768.062670] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 768.070836] env[63515]: DEBUG oslo_vmware.api [None req-db31b1fd-34e1-40f4-b7c8-92b451eb1400 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] Task: {'id': task-1111158, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 768.192630] env[63515]: DEBUG oslo_vmware.api [None req-65f1d514-d43e-4354-a5f1-c564a5380800 tempest-ServersAaction247Test-2019176257 tempest-ServersAaction247Test-2019176257-project-member] Task: {'id': task-1111157, 'name': PowerOffVM_Task, 'duration_secs': 0.263245} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 768.193367] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-65f1d514-d43e-4354-a5f1-c564a5380800 tempest-ServersAaction247Test-2019176257 tempest-ServersAaction247Test-2019176257-project-member] [instance: 5d50fbd6-a7fb-422c-9dd7-df7140fd0aae] Powered off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 768.193559] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-65f1d514-d43e-4354-a5f1-c564a5380800 tempest-ServersAaction247Test-2019176257 tempest-ServersAaction247Test-2019176257-project-member] [instance: 5d50fbd6-a7fb-422c-9dd7-df7140fd0aae] Unregistering the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 768.193894] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-6b4b9c26-c7a5-4392-9846-15a3b655f7e4 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.219036] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-65f1d514-d43e-4354-a5f1-c564a5380800 tempest-ServersAaction247Test-2019176257 tempest-ServersAaction247Test-2019176257-project-member] [instance: 5d50fbd6-a7fb-422c-9dd7-df7140fd0aae] Unregistered the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 768.219036] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-65f1d514-d43e-4354-a5f1-c564a5380800 tempest-ServersAaction247Test-2019176257 tempest-ServersAaction247Test-2019176257-project-member] [instance: 5d50fbd6-a7fb-422c-9dd7-df7140fd0aae] Deleting contents of the VM from datastore datastore1 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 768.219178] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-65f1d514-d43e-4354-a5f1-c564a5380800 tempest-ServersAaction247Test-2019176257 tempest-ServersAaction247Test-2019176257-project-member] Deleting the datastore file [datastore1] 5d50fbd6-a7fb-422c-9dd7-df7140fd0aae {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 768.219377] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-db4302b5-2794-4b5d-b1fb-22045e9261cc {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.226028] env[63515]: DEBUG oslo_vmware.api [None req-65f1d514-d43e-4354-a5f1-c564a5380800 tempest-ServersAaction247Test-2019176257 tempest-ServersAaction247Test-2019176257-project-member] Waiting for the task: (returnval){ [ 768.226028] env[63515]: value = "task-1111160" [ 768.226028] env[63515]: _type = "Task" [ 768.226028] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 768.234428] env[63515]: DEBUG oslo_vmware.api [None req-65f1d514-d43e-4354-a5f1-c564a5380800 tempest-ServersAaction247Test-2019176257 tempest-ServersAaction247Test-2019176257-project-member] Task: {'id': task-1111160, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 768.273734] env[63515]: DEBUG oslo_concurrency.lockutils [None req-549cc3f8-330f-4c6a-979d-35eedc6f8146 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.574s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 768.274358] env[63515]: DEBUG nova.compute.manager [None req-549cc3f8-330f-4c6a-979d-35eedc6f8146 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: 84dbf321-2c70-4e08-b430-cb5a06fc6829] Start building networks asynchronously for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 768.277348] env[63515]: DEBUG oslo_concurrency.lockutils [None req-996c1e42-eed8-414b-a716-3f08d65ec493 tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 30.851s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 768.277551] env[63515]: DEBUG nova.objects.instance [None req-996c1e42-eed8-414b-a716-3f08d65ec493 tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] [instance: b85e9a70-7f5b-4d32-b616-f2a97e3186c8] Trying to apply a migration context that does not seem to be set for this instance {{(pid=63515) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 768.576145] env[63515]: DEBUG oslo_vmware.api [None req-db31b1fd-34e1-40f4-b7c8-92b451eb1400 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] Task: {'id': task-1111158, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.08302} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 768.576499] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-db31b1fd-34e1-40f4-b7c8-92b451eb1400 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] [instance: 8eb17506-ff93-4d25-b9af-ec5886569e65] Extended root virtual disk {{(pid=63515) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 768.577303] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e60aad2-1200-4588-9d21-6c79b05bd1bf {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.601447] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-db31b1fd-34e1-40f4-b7c8-92b451eb1400 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] [instance: 8eb17506-ff93-4d25-b9af-ec5886569e65] Reconfiguring VM instance instance-00000034 to attach disk [datastore2] 8eb17506-ff93-4d25-b9af-ec5886569e65/8eb17506-ff93-4d25-b9af-ec5886569e65.vmdk or device None with type sparse {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 768.601883] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-591b3d3c-4d9d-418f-99b5-247dca1dc77f {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.625336] env[63515]: DEBUG oslo_vmware.api [None req-db31b1fd-34e1-40f4-b7c8-92b451eb1400 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] Waiting for the task: (returnval){ [ 768.625336] env[63515]: value = "task-1111161" [ 768.625336] env[63515]: _type = "Task" [ 768.625336] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 768.634551] env[63515]: DEBUG oslo_vmware.api [None req-db31b1fd-34e1-40f4-b7c8-92b451eb1400 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] Task: {'id': task-1111161, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 768.736518] env[63515]: DEBUG oslo_vmware.api [None req-65f1d514-d43e-4354-a5f1-c564a5380800 tempest-ServersAaction247Test-2019176257 tempest-ServersAaction247Test-2019176257-project-member] Task: {'id': task-1111160, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.257232} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 768.736634] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-65f1d514-d43e-4354-a5f1-c564a5380800 tempest-ServersAaction247Test-2019176257 tempest-ServersAaction247Test-2019176257-project-member] Deleted the datastore file {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 768.736824] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-65f1d514-d43e-4354-a5f1-c564a5380800 tempest-ServersAaction247Test-2019176257 tempest-ServersAaction247Test-2019176257-project-member] [instance: 5d50fbd6-a7fb-422c-9dd7-df7140fd0aae] Deleted contents of the VM from datastore datastore1 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 768.737018] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-65f1d514-d43e-4354-a5f1-c564a5380800 tempest-ServersAaction247Test-2019176257 tempest-ServersAaction247Test-2019176257-project-member] [instance: 5d50fbd6-a7fb-422c-9dd7-df7140fd0aae] Instance destroyed {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 768.737212] env[63515]: INFO nova.compute.manager [None req-65f1d514-d43e-4354-a5f1-c564a5380800 tempest-ServersAaction247Test-2019176257 tempest-ServersAaction247Test-2019176257-project-member] [instance: 5d50fbd6-a7fb-422c-9dd7-df7140fd0aae] Took 1.07 seconds to destroy the instance on the hypervisor. [ 768.737467] env[63515]: DEBUG oslo.service.loopingcall [None req-65f1d514-d43e-4354-a5f1-c564a5380800 tempest-ServersAaction247Test-2019176257 tempest-ServersAaction247Test-2019176257-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 768.737667] env[63515]: DEBUG nova.compute.manager [-] [instance: 5d50fbd6-a7fb-422c-9dd7-df7140fd0aae] Deallocating network for instance {{(pid=63515) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 768.737765] env[63515]: DEBUG nova.network.neutron [-] [instance: 5d50fbd6-a7fb-422c-9dd7-df7140fd0aae] deallocate_for_instance() {{(pid=63515) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 768.753033] env[63515]: DEBUG nova.network.neutron [-] [instance: 5d50fbd6-a7fb-422c-9dd7-df7140fd0aae] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 768.781616] env[63515]: DEBUG nova.network.neutron [None req-a6c8eb46-cd36-421b-bb36-d08f2b34cdc9 tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] [instance: 30b8b872-40ca-4297-b98e-a64c83a75483] Updating instance_info_cache with network_info: [{"id": "b8fe94ee-46b7-489a-819d-0699bc02bbb7", "address": "fa:16:3e:42:a9:82", "network": {"id": "bd2f642c-518b-478b-8358-7426cebde731", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-1534671780-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "72b65c09ed6d4492864022352180ff49", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "92f3cfd6-c130-4390-8910-865fbc42afd1", "external-id": "nsx-vlan-transportzone-142", "segmentation_id": 142, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb8fe94ee-46", "ovs_interfaceid": "b8fe94ee-46b7-489a-819d-0699bc02bbb7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 768.787532] env[63515]: DEBUG nova.compute.utils [None req-549cc3f8-330f-4c6a-979d-35eedc6f8146 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Using /dev/sd instead of None {{(pid=63515) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 768.788933] env[63515]: DEBUG nova.compute.manager [None req-549cc3f8-330f-4c6a-979d-35eedc6f8146 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: 84dbf321-2c70-4e08-b430-cb5a06fc6829] Allocating IP information in the background. {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 768.789351] env[63515]: DEBUG nova.network.neutron [None req-549cc3f8-330f-4c6a-979d-35eedc6f8146 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: 84dbf321-2c70-4e08-b430-cb5a06fc6829] allocate_for_instance() {{(pid=63515) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 768.793298] env[63515]: DEBUG oslo_service.periodic_task [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=63515) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 768.794546] env[63515]: DEBUG oslo_service.periodic_task [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=63515) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 768.845999] env[63515]: DEBUG nova.policy [None req-549cc3f8-330f-4c6a-979d-35eedc6f8146 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '002c1a9364184b82b649fc18d017d97f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2259667ddc0c42639ea1af3b6e6fa0e8', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63515) authorize /opt/stack/nova/nova/policy.py:201}} [ 769.138969] env[63515]: DEBUG oslo_vmware.api [None req-db31b1fd-34e1-40f4-b7c8-92b451eb1400 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] Task: {'id': task-1111161, 'name': ReconfigVM_Task, 'duration_secs': 0.364174} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 769.139499] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-db31b1fd-34e1-40f4-b7c8-92b451eb1400 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] [instance: 8eb17506-ff93-4d25-b9af-ec5886569e65] Reconfigured VM instance instance-00000034 to attach disk [datastore2] 8eb17506-ff93-4d25-b9af-ec5886569e65/8eb17506-ff93-4d25-b9af-ec5886569e65.vmdk or device None with type sparse {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 769.140225] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-3a0e633a-4fde-4370-b202-389709d4fe33 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.147174] env[63515]: DEBUG oslo_vmware.api [None req-db31b1fd-34e1-40f4-b7c8-92b451eb1400 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] Waiting for the task: (returnval){ [ 769.147174] env[63515]: value = "task-1111162" [ 769.147174] env[63515]: _type = "Task" [ 769.147174] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 769.156605] env[63515]: DEBUG oslo_vmware.api [None req-db31b1fd-34e1-40f4-b7c8-92b451eb1400 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] Task: {'id': task-1111162, 'name': Rename_Task} progress is 5%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 769.256432] env[63515]: DEBUG nova.network.neutron [-] [instance: 5d50fbd6-a7fb-422c-9dd7-df7140fd0aae] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 769.290248] env[63515]: DEBUG oslo_concurrency.lockutils [None req-996c1e42-eed8-414b-a716-3f08d65ec493 tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.013s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 769.291512] env[63515]: DEBUG oslo_concurrency.lockutils [None req-a6c8eb46-cd36-421b-bb36-d08f2b34cdc9 tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] Releasing lock "refresh_cache-30b8b872-40ca-4297-b98e-a64c83a75483" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 769.296947] env[63515]: DEBUG oslo_concurrency.lockutils [None req-2786351a-303c-45a8-b63b-65e218eb6ef6 tempest-ServersTestJSON-324454119 tempest-ServersTestJSON-324454119-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 29.179s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 769.298525] env[63515]: INFO nova.compute.claims [None req-2786351a-303c-45a8-b63b-65e218eb6ef6 tempest-ServersTestJSON-324454119 tempest-ServersTestJSON-324454119-project-member] [instance: 70ed982f-affd-4dd1-bc90-c64e7c6d49d2] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 769.305055] env[63515]: DEBUG nova.compute.manager [None req-549cc3f8-330f-4c6a-979d-35eedc6f8146 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: 84dbf321-2c70-4e08-b430-cb5a06fc6829] Start building block device mappings for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 769.315710] env[63515]: DEBUG nova.compute.manager [None req-a6c8eb46-cd36-421b-bb36-d08f2b34cdc9 tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] [instance: 30b8b872-40ca-4297-b98e-a64c83a75483] Checking state {{(pid=63515) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 769.322185] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28daf85a-e308-4883-9013-ed2397676d6d {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.325440] env[63515]: DEBUG oslo_service.periodic_task [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=63515) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 769.325685] env[63515]: DEBUG nova.compute.manager [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Starting heal instance info cache {{(pid=63515) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9926}} [ 769.434497] env[63515]: DEBUG nova.network.neutron [None req-549cc3f8-330f-4c6a-979d-35eedc6f8146 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: 84dbf321-2c70-4e08-b430-cb5a06fc6829] Successfully created port: ad6685ff-3a87-4cc5-bdfa-5eb3bb2f0590 {{(pid=63515) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 769.660338] env[63515]: DEBUG oslo_vmware.api [None req-db31b1fd-34e1-40f4-b7c8-92b451eb1400 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] Task: {'id': task-1111162, 'name': Rename_Task, 'duration_secs': 0.302199} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 769.660338] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-db31b1fd-34e1-40f4-b7c8-92b451eb1400 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] [instance: 8eb17506-ff93-4d25-b9af-ec5886569e65] Powering on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 769.660338] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-41ea6b9a-f9f0-4a11-ab5c-78f14be58cad {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.667282] env[63515]: DEBUG oslo_vmware.api [None req-db31b1fd-34e1-40f4-b7c8-92b451eb1400 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] Waiting for the task: (returnval){ [ 769.667282] env[63515]: value = "task-1111163" [ 769.667282] env[63515]: _type = "Task" [ 769.667282] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 769.679718] env[63515]: DEBUG oslo_vmware.api [None req-db31b1fd-34e1-40f4-b7c8-92b451eb1400 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] Task: {'id': task-1111163, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 769.759804] env[63515]: INFO nova.compute.manager [-] [instance: 5d50fbd6-a7fb-422c-9dd7-df7140fd0aae] Took 1.02 seconds to deallocate network for instance. [ 770.178644] env[63515]: DEBUG oslo_vmware.api [None req-db31b1fd-34e1-40f4-b7c8-92b451eb1400 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] Task: {'id': task-1111163, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 770.270210] env[63515]: DEBUG oslo_concurrency.lockutils [None req-65f1d514-d43e-4354-a5f1-c564a5380800 tempest-ServersAaction247Test-2019176257 tempest-ServersAaction247Test-2019176257-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 770.324742] env[63515]: DEBUG nova.compute.manager [None req-549cc3f8-330f-4c6a-979d-35eedc6f8146 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: 84dbf321-2c70-4e08-b430-cb5a06fc6829] Start spawning the instance on the hypervisor. {{(pid=63515) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 770.341593] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5090eb2-c8ee-4fa3-a218-baee99e8bb34 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.349673] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-a6c8eb46-cd36-421b-bb36-d08f2b34cdc9 tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] [instance: 30b8b872-40ca-4297-b98e-a64c83a75483] Doing hard reboot of VM {{(pid=63515) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1063}} [ 770.349936] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ResetVM_Task with opID=oslo.vmware-b5e2c429-72b6-492b-b970-1178de95736e {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.359892] env[63515]: DEBUG oslo_vmware.api [None req-a6c8eb46-cd36-421b-bb36-d08f2b34cdc9 tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] Waiting for the task: (returnval){ [ 770.359892] env[63515]: value = "task-1111164" [ 770.359892] env[63515]: _type = "Task" [ 770.359892] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 770.363715] env[63515]: DEBUG nova.virt.hardware [None req-549cc3f8-330f-4c6a-979d-35eedc6f8146 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T02:52:14Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T02:51:56Z,direct_url=,disk_format='vmdk',id=8a120570-cb06-4099-b262-554ca0ad15c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b569255cc43e42e7ae3f2b4ad37c6ef4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T02:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 770.364082] env[63515]: DEBUG nova.virt.hardware [None req-549cc3f8-330f-4c6a-979d-35eedc6f8146 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Flavor limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 770.364208] env[63515]: DEBUG nova.virt.hardware [None req-549cc3f8-330f-4c6a-979d-35eedc6f8146 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Image limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 770.364407] env[63515]: DEBUG nova.virt.hardware [None req-549cc3f8-330f-4c6a-979d-35eedc6f8146 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Flavor pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 770.364572] env[63515]: DEBUG nova.virt.hardware [None req-549cc3f8-330f-4c6a-979d-35eedc6f8146 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Image pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 770.364730] env[63515]: DEBUG nova.virt.hardware [None req-549cc3f8-330f-4c6a-979d-35eedc6f8146 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 770.364958] env[63515]: DEBUG nova.virt.hardware [None req-549cc3f8-330f-4c6a-979d-35eedc6f8146 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 770.365527] env[63515]: DEBUG nova.virt.hardware [None req-549cc3f8-330f-4c6a-979d-35eedc6f8146 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 770.365527] env[63515]: DEBUG nova.virt.hardware [None req-549cc3f8-330f-4c6a-979d-35eedc6f8146 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Got 1 possible topologies {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 770.365674] env[63515]: DEBUG nova.virt.hardware [None req-549cc3f8-330f-4c6a-979d-35eedc6f8146 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 770.365788] env[63515]: DEBUG nova.virt.hardware [None req-549cc3f8-330f-4c6a-979d-35eedc6f8146 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 770.366744] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67e78423-16d1-410c-8bf3-697165b3bdd9 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.370390] env[63515]: DEBUG oslo_concurrency.lockutils [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Acquiring lock "refresh_cache-5b42f744-fdd6-45b1-8563-896869648c23" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 770.370390] env[63515]: DEBUG oslo_concurrency.lockutils [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Acquired lock "refresh_cache-5b42f744-fdd6-45b1-8563-896869648c23" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 770.370478] env[63515]: DEBUG nova.network.neutron [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] [instance: 5b42f744-fdd6-45b1-8563-896869648c23] Forcefully refreshing network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2004}} [ 770.390729] env[63515]: DEBUG oslo_vmware.api [None req-a6c8eb46-cd36-421b-bb36-d08f2b34cdc9 tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] Task: {'id': task-1111164, 'name': ResetVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 770.392881] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7dcfebc2-08bb-4140-9ec4-02bad5d1ae0d {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.681472] env[63515]: DEBUG oslo_vmware.api [None req-db31b1fd-34e1-40f4-b7c8-92b451eb1400 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] Task: {'id': task-1111163, 'name': PowerOnVM_Task, 'duration_secs': 0.711349} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 770.681675] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-db31b1fd-34e1-40f4-b7c8-92b451eb1400 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] [instance: 8eb17506-ff93-4d25-b9af-ec5886569e65] Powered on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 770.682096] env[63515]: DEBUG nova.compute.manager [None req-db31b1fd-34e1-40f4-b7c8-92b451eb1400 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] [instance: 8eb17506-ff93-4d25-b9af-ec5886569e65] Checking state {{(pid=63515) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 770.682622] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d929661-60f6-4800-b219-a79dc83f3fc8 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.839332] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90845638-ebc8-4817-809c-e320ec3d6417 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.848250] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a614933d-db48-4c11-ba53-d921947a6bad {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.883024] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35fc6d73-0e1f-47ae-a92d-871ea8234915 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.894597] env[63515]: DEBUG oslo_vmware.api [None req-a6c8eb46-cd36-421b-bb36-d08f2b34cdc9 tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] Task: {'id': task-1111164, 'name': ResetVM_Task, 'duration_secs': 0.129128} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 770.895074] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-a6c8eb46-cd36-421b-bb36-d08f2b34cdc9 tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] [instance: 30b8b872-40ca-4297-b98e-a64c83a75483] Did hard reboot of VM {{(pid=63515) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1067}} [ 770.895251] env[63515]: DEBUG nova.compute.manager [None req-a6c8eb46-cd36-421b-bb36-d08f2b34cdc9 tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] [instance: 30b8b872-40ca-4297-b98e-a64c83a75483] Checking state {{(pid=63515) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 770.897464] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-300acb18-c495-482c-a840-047ae84a7808 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.904513] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02572248-c03b-4cf8-a730-ac6143d60c9a {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.925266] env[63515]: DEBUG nova.compute.provider_tree [None req-2786351a-303c-45a8-b63b-65e218eb6ef6 tempest-ServersTestJSON-324454119 tempest-ServersTestJSON-324454119-project-member] Updating inventory in ProviderTree for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 771.039570] env[63515]: DEBUG oslo_vmware.rw_handles [None req-78f15907-fa94-44dd-a2b3-ccdc7dec0a0f tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52dac88e-b167-9d0e-e4fa-7c0241d36201/disk-0.vmdk. {{(pid=63515) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 771.041351] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1bcc209-cc15-40d7-8915-b4297c747f8f {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.047247] env[63515]: DEBUG oslo_vmware.rw_handles [None req-78f15907-fa94-44dd-a2b3-ccdc7dec0a0f tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52dac88e-b167-9d0e-e4fa-7c0241d36201/disk-0.vmdk is in state: ready. {{(pid=63515) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 771.047437] env[63515]: ERROR oslo_vmware.rw_handles [None req-78f15907-fa94-44dd-a2b3-ccdc7dec0a0f tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52dac88e-b167-9d0e-e4fa-7c0241d36201/disk-0.vmdk due to incomplete transfer. [ 771.047665] env[63515]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-f921cbec-72ef-44b3-acf9-07d805f30c66 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.054146] env[63515]: DEBUG oslo_vmware.rw_handles [None req-78f15907-fa94-44dd-a2b3-ccdc7dec0a0f tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52dac88e-b167-9d0e-e4fa-7c0241d36201/disk-0.vmdk. {{(pid=63515) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 771.054350] env[63515]: DEBUG nova.virt.vmwareapi.images [None req-78f15907-fa94-44dd-a2b3-ccdc7dec0a0f tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] [instance: 87c468d9-9594-4804-b461-527f01f6118f] Uploaded image 499b6029-07e9-499c-80d5-a53d8c43ffb7 to the Glance image server {{(pid=63515) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 771.056421] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-78f15907-fa94-44dd-a2b3-ccdc7dec0a0f tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] [instance: 87c468d9-9594-4804-b461-527f01f6118f] Destroying the VM {{(pid=63515) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 771.056662] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-afba7e71-cb65-4ac6-889f-52720596c62b {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.062163] env[63515]: DEBUG oslo_vmware.api [None req-78f15907-fa94-44dd-a2b3-ccdc7dec0a0f tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Waiting for the task: (returnval){ [ 771.062163] env[63515]: value = "task-1111165" [ 771.062163] env[63515]: _type = "Task" [ 771.062163] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 771.072156] env[63515]: DEBUG oslo_vmware.api [None req-78f15907-fa94-44dd-a2b3-ccdc7dec0a0f tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Task: {'id': task-1111165, 'name': Destroy_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 771.201856] env[63515]: DEBUG oslo_concurrency.lockutils [None req-db31b1fd-34e1-40f4-b7c8-92b451eb1400 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 771.208245] env[63515]: DEBUG nova.network.neutron [None req-549cc3f8-330f-4c6a-979d-35eedc6f8146 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: 84dbf321-2c70-4e08-b430-cb5a06fc6829] Successfully updated port: ad6685ff-3a87-4cc5-bdfa-5eb3bb2f0590 {{(pid=63515) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 771.223049] env[63515]: DEBUG nova.compute.manager [req-85f9d1ed-6f38-444f-a592-6905abaca16d req-74ba4191-5b5c-4a55-8442-4b5b8dc00150 service nova] [instance: 84dbf321-2c70-4e08-b430-cb5a06fc6829] Received event network-vif-plugged-ad6685ff-3a87-4cc5-bdfa-5eb3bb2f0590 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 771.223049] env[63515]: DEBUG oslo_concurrency.lockutils [req-85f9d1ed-6f38-444f-a592-6905abaca16d req-74ba4191-5b5c-4a55-8442-4b5b8dc00150 service nova] Acquiring lock "84dbf321-2c70-4e08-b430-cb5a06fc6829-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 771.223203] env[63515]: DEBUG oslo_concurrency.lockutils [req-85f9d1ed-6f38-444f-a592-6905abaca16d req-74ba4191-5b5c-4a55-8442-4b5b8dc00150 service nova] Lock "84dbf321-2c70-4e08-b430-cb5a06fc6829-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 771.223334] env[63515]: DEBUG oslo_concurrency.lockutils [req-85f9d1ed-6f38-444f-a592-6905abaca16d req-74ba4191-5b5c-4a55-8442-4b5b8dc00150 service nova] Lock "84dbf321-2c70-4e08-b430-cb5a06fc6829-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 771.223580] env[63515]: DEBUG nova.compute.manager [req-85f9d1ed-6f38-444f-a592-6905abaca16d req-74ba4191-5b5c-4a55-8442-4b5b8dc00150 service nova] [instance: 84dbf321-2c70-4e08-b430-cb5a06fc6829] No waiting events found dispatching network-vif-plugged-ad6685ff-3a87-4cc5-bdfa-5eb3bb2f0590 {{(pid=63515) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 771.223806] env[63515]: WARNING nova.compute.manager [req-85f9d1ed-6f38-444f-a592-6905abaca16d req-74ba4191-5b5c-4a55-8442-4b5b8dc00150 service nova] [instance: 84dbf321-2c70-4e08-b430-cb5a06fc6829] Received unexpected event network-vif-plugged-ad6685ff-3a87-4cc5-bdfa-5eb3bb2f0590 for instance with vm_state building and task_state spawning. [ 771.436559] env[63515]: DEBUG oslo_concurrency.lockutils [None req-a6c8eb46-cd36-421b-bb36-d08f2b34cdc9 tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] Lock "30b8b872-40ca-4297-b98e-a64c83a75483" "released" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: held 4.241s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 771.470388] env[63515]: DEBUG nova.scheduler.client.report [None req-2786351a-303c-45a8-b63b-65e218eb6ef6 tempest-ServersTestJSON-324454119 tempest-ServersTestJSON-324454119-project-member] Updated inventory for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a with generation 77 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 771.470634] env[63515]: DEBUG nova.compute.provider_tree [None req-2786351a-303c-45a8-b63b-65e218eb6ef6 tempest-ServersTestJSON-324454119 tempest-ServersTestJSON-324454119-project-member] Updating resource provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a generation from 77 to 78 during operation: update_inventory {{(pid=63515) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 771.470958] env[63515]: DEBUG nova.compute.provider_tree [None req-2786351a-303c-45a8-b63b-65e218eb6ef6 tempest-ServersTestJSON-324454119 tempest-ServersTestJSON-324454119-project-member] Updating inventory in ProviderTree for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 771.572281] env[63515]: DEBUG oslo_vmware.api [None req-78f15907-fa94-44dd-a2b3-ccdc7dec0a0f tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Task: {'id': task-1111165, 'name': Destroy_Task} progress is 33%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 771.712192] env[63515]: DEBUG oslo_concurrency.lockutils [None req-549cc3f8-330f-4c6a-979d-35eedc6f8146 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Acquiring lock "refresh_cache-84dbf321-2c70-4e08-b430-cb5a06fc6829" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 771.712192] env[63515]: DEBUG oslo_concurrency.lockutils [None req-549cc3f8-330f-4c6a-979d-35eedc6f8146 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Acquired lock "refresh_cache-84dbf321-2c70-4e08-b430-cb5a06fc6829" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 771.712192] env[63515]: DEBUG nova.network.neutron [None req-549cc3f8-330f-4c6a-979d-35eedc6f8146 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: 84dbf321-2c70-4e08-b430-cb5a06fc6829] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 771.759217] env[63515]: DEBUG nova.network.neutron [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] [instance: 5b42f744-fdd6-45b1-8563-896869648c23] Updating instance_info_cache with network_info: [{"id": "3a630639-5423-4e62-9beb-ee1e7054f47d", "address": "fa:16:3e:b6:20:83", "network": {"id": "b4abcbd0-0eac-4bde-ad11-7f93afcedcef", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.28", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "b569255cc43e42e7ae3f2b4ad37c6ef4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0cd5d325-3053-407e-a4ee-f627e82a23f9", "external-id": "nsx-vlan-transportzone-809", "segmentation_id": 809, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3a630639-54", "ovs_interfaceid": "3a630639-5423-4e62-9beb-ee1e7054f47d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 771.793963] env[63515]: DEBUG oslo_concurrency.lockutils [None req-6d0510cf-97fb-4582-ac56-e14cb4885964 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Acquiring lock "38d6d030-06b0-4185-904d-44a038b3a752" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 771.794284] env[63515]: DEBUG oslo_concurrency.lockutils [None req-6d0510cf-97fb-4582-ac56-e14cb4885964 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Lock "38d6d030-06b0-4185-904d-44a038b3a752" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 771.794491] env[63515]: DEBUG oslo_concurrency.lockutils [None req-6d0510cf-97fb-4582-ac56-e14cb4885964 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Acquiring lock "38d6d030-06b0-4185-904d-44a038b3a752-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 771.794673] env[63515]: DEBUG oslo_concurrency.lockutils [None req-6d0510cf-97fb-4582-ac56-e14cb4885964 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Lock "38d6d030-06b0-4185-904d-44a038b3a752-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 771.794849] env[63515]: DEBUG oslo_concurrency.lockutils [None req-6d0510cf-97fb-4582-ac56-e14cb4885964 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Lock "38d6d030-06b0-4185-904d-44a038b3a752-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 771.797787] env[63515]: INFO nova.compute.manager [None req-6d0510cf-97fb-4582-ac56-e14cb4885964 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] [instance: 38d6d030-06b0-4185-904d-44a038b3a752] Terminating instance [ 771.799776] env[63515]: DEBUG nova.compute.manager [None req-6d0510cf-97fb-4582-ac56-e14cb4885964 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] [instance: 38d6d030-06b0-4185-904d-44a038b3a752] Start destroying the instance on the hypervisor. {{(pid=63515) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 771.800068] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-6d0510cf-97fb-4582-ac56-e14cb4885964 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] [instance: 38d6d030-06b0-4185-904d-44a038b3a752] Destroying instance {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 771.801316] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe367b4b-f49f-4d8b-b13e-3750b1429ad9 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.810403] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-6d0510cf-97fb-4582-ac56-e14cb4885964 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] [instance: 38d6d030-06b0-4185-904d-44a038b3a752] Powering off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 771.810846] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0725d048-f544-4a08-8dd9-ed75497e7155 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.817145] env[63515]: DEBUG oslo_vmware.api [None req-6d0510cf-97fb-4582-ac56-e14cb4885964 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Waiting for the task: (returnval){ [ 771.817145] env[63515]: value = "task-1111166" [ 771.817145] env[63515]: _type = "Task" [ 771.817145] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 771.825726] env[63515]: DEBUG oslo_vmware.api [None req-6d0510cf-97fb-4582-ac56-e14cb4885964 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Task: {'id': task-1111166, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 771.975875] env[63515]: DEBUG oslo_concurrency.lockutils [None req-9044ca9f-2eb8-445e-a5e5-e15900a98ac3 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] Acquiring lock "8eb17506-ff93-4d25-b9af-ec5886569e65" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 771.976249] env[63515]: DEBUG oslo_concurrency.lockutils [None req-9044ca9f-2eb8-445e-a5e5-e15900a98ac3 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] Lock "8eb17506-ff93-4d25-b9af-ec5886569e65" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 771.976526] env[63515]: DEBUG oslo_concurrency.lockutils [None req-9044ca9f-2eb8-445e-a5e5-e15900a98ac3 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] Acquiring lock "8eb17506-ff93-4d25-b9af-ec5886569e65-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 771.976769] env[63515]: DEBUG oslo_concurrency.lockutils [None req-9044ca9f-2eb8-445e-a5e5-e15900a98ac3 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] Lock "8eb17506-ff93-4d25-b9af-ec5886569e65-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 771.977016] env[63515]: DEBUG oslo_concurrency.lockutils [None req-9044ca9f-2eb8-445e-a5e5-e15900a98ac3 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] Lock "8eb17506-ff93-4d25-b9af-ec5886569e65-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 771.980651] env[63515]: DEBUG oslo_concurrency.lockutils [None req-2786351a-303c-45a8-b63b-65e218eb6ef6 tempest-ServersTestJSON-324454119 tempest-ServersTestJSON-324454119-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.682s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 771.980651] env[63515]: DEBUG nova.compute.manager [None req-2786351a-303c-45a8-b63b-65e218eb6ef6 tempest-ServersTestJSON-324454119 tempest-ServersTestJSON-324454119-project-member] [instance: 70ed982f-affd-4dd1-bc90-c64e7c6d49d2] Start building networks asynchronously for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 771.983922] env[63515]: INFO nova.compute.manager [None req-9044ca9f-2eb8-445e-a5e5-e15900a98ac3 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] [instance: 8eb17506-ff93-4d25-b9af-ec5886569e65] Terminating instance [ 771.986072] env[63515]: DEBUG oslo_concurrency.lockutils [None req-d4ac6a62-1537-4962-916c-41ec7cc28737 tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 29.875s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 771.986072] env[63515]: DEBUG oslo_concurrency.lockutils [None req-d4ac6a62-1537-4962-916c-41ec7cc28737 tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 771.987863] env[63515]: DEBUG oslo_concurrency.lockutils [None req-8407c401-88c7-47ad-b91e-e7693f87e0f1 tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 26.366s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 771.989611] env[63515]: INFO nova.compute.claims [None req-8407c401-88c7-47ad-b91e-e7693f87e0f1 tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] [instance: 192137b0-03e5-4bc4-b911-4b4f1a874f74] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 771.994897] env[63515]: DEBUG oslo_concurrency.lockutils [None req-9044ca9f-2eb8-445e-a5e5-e15900a98ac3 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] Acquiring lock "refresh_cache-8eb17506-ff93-4d25-b9af-ec5886569e65" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 771.995167] env[63515]: DEBUG oslo_concurrency.lockutils [None req-9044ca9f-2eb8-445e-a5e5-e15900a98ac3 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] Acquired lock "refresh_cache-8eb17506-ff93-4d25-b9af-ec5886569e65" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 771.995504] env[63515]: DEBUG nova.network.neutron [None req-9044ca9f-2eb8-445e-a5e5-e15900a98ac3 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] [instance: 8eb17506-ff93-4d25-b9af-ec5886569e65] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 772.017860] env[63515]: INFO nova.scheduler.client.report [None req-d4ac6a62-1537-4962-916c-41ec7cc28737 tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Deleted allocations for instance b85e9a70-7f5b-4d32-b616-f2a97e3186c8 [ 772.024955] env[63515]: DEBUG oslo_concurrency.lockutils [None req-efd46494-ea54-4600-906e-1e7f1371fb6c tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Acquiring lock "d7e8dcdb-41cb-46fb-8b61-d251e7c2d372" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 772.025267] env[63515]: DEBUG oslo_concurrency.lockutils [None req-efd46494-ea54-4600-906e-1e7f1371fb6c tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Lock "d7e8dcdb-41cb-46fb-8b61-d251e7c2d372" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 772.073907] env[63515]: DEBUG oslo_vmware.api [None req-78f15907-fa94-44dd-a2b3-ccdc7dec0a0f tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Task: {'id': task-1111165, 'name': Destroy_Task} progress is 33%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 772.248075] env[63515]: DEBUG nova.network.neutron [None req-549cc3f8-330f-4c6a-979d-35eedc6f8146 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: 84dbf321-2c70-4e08-b430-cb5a06fc6829] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 772.261350] env[63515]: DEBUG oslo_concurrency.lockutils [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Releasing lock "refresh_cache-5b42f744-fdd6-45b1-8563-896869648c23" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 772.261586] env[63515]: DEBUG nova.compute.manager [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] [instance: 5b42f744-fdd6-45b1-8563-896869648c23] Updated the network info_cache for instance {{(pid=63515) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9997}} [ 772.261808] env[63515]: DEBUG oslo_service.periodic_task [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=63515) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 772.326624] env[63515]: DEBUG oslo_vmware.api [None req-6d0510cf-97fb-4582-ac56-e14cb4885964 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Task: {'id': task-1111166, 'name': PowerOffVM_Task, 'duration_secs': 0.229778} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 772.326624] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-6d0510cf-97fb-4582-ac56-e14cb4885964 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] [instance: 38d6d030-06b0-4185-904d-44a038b3a752] Powered off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 772.326624] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-6d0510cf-97fb-4582-ac56-e14cb4885964 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] [instance: 38d6d030-06b0-4185-904d-44a038b3a752] Unregistering the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 772.326986] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9767c852-3b32-493b-b0e4-bd7ad73ca3fa {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.382612] env[63515]: DEBUG nova.network.neutron [None req-549cc3f8-330f-4c6a-979d-35eedc6f8146 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: 84dbf321-2c70-4e08-b430-cb5a06fc6829] Updating instance_info_cache with network_info: [{"id": "ad6685ff-3a87-4cc5-bdfa-5eb3bb2f0590", "address": "fa:16:3e:5d:a9:21", "network": {"id": "320921bd-a6c9-4726-a801-19009d70d16e", "bridge": "br-int", "label": "tempest-ImagesTestJSON-304158429-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2259667ddc0c42639ea1af3b6e6fa0e8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2d94740a-bce8-4103-8ecf-230d02ec0a44", "external-id": "nsx-vlan-transportzone-149", "segmentation_id": 149, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapad6685ff-3a", "ovs_interfaceid": "ad6685ff-3a87-4cc5-bdfa-5eb3bb2f0590", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 772.387614] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-6d0510cf-97fb-4582-ac56-e14cb4885964 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] [instance: 38d6d030-06b0-4185-904d-44a038b3a752] Unregistered the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 772.387799] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-6d0510cf-97fb-4582-ac56-e14cb4885964 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] [instance: 38d6d030-06b0-4185-904d-44a038b3a752] Deleting contents of the VM from datastore datastore2 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 772.387976] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-6d0510cf-97fb-4582-ac56-e14cb4885964 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Deleting the datastore file [datastore2] 38d6d030-06b0-4185-904d-44a038b3a752 {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 772.388289] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e38ba3a0-9595-47db-b75b-4c6f140954f7 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.394410] env[63515]: DEBUG oslo_vmware.api [None req-6d0510cf-97fb-4582-ac56-e14cb4885964 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Waiting for the task: (returnval){ [ 772.394410] env[63515]: value = "task-1111168" [ 772.394410] env[63515]: _type = "Task" [ 772.394410] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 772.404297] env[63515]: DEBUG oslo_vmware.api [None req-6d0510cf-97fb-4582-ac56-e14cb4885964 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Task: {'id': task-1111168, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 772.499124] env[63515]: DEBUG nova.compute.utils [None req-2786351a-303c-45a8-b63b-65e218eb6ef6 tempest-ServersTestJSON-324454119 tempest-ServersTestJSON-324454119-project-member] Using /dev/sd instead of None {{(pid=63515) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 772.504470] env[63515]: DEBUG nova.compute.manager [None req-2786351a-303c-45a8-b63b-65e218eb6ef6 tempest-ServersTestJSON-324454119 tempest-ServersTestJSON-324454119-project-member] [instance: 70ed982f-affd-4dd1-bc90-c64e7c6d49d2] Allocating IP information in the background. {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 772.504646] env[63515]: DEBUG nova.network.neutron [None req-2786351a-303c-45a8-b63b-65e218eb6ef6 tempest-ServersTestJSON-324454119 tempest-ServersTestJSON-324454119-project-member] [instance: 70ed982f-affd-4dd1-bc90-c64e7c6d49d2] allocate_for_instance() {{(pid=63515) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 772.520219] env[63515]: DEBUG nova.network.neutron [None req-9044ca9f-2eb8-445e-a5e5-e15900a98ac3 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] [instance: 8eb17506-ff93-4d25-b9af-ec5886569e65] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 772.526021] env[63515]: DEBUG oslo_concurrency.lockutils [None req-d4ac6a62-1537-4962-916c-41ec7cc28737 tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Lock "b85e9a70-7f5b-4d32-b616-f2a97e3186c8" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 34.153s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 772.528948] env[63515]: DEBUG nova.compute.utils [None req-efd46494-ea54-4600-906e-1e7f1371fb6c tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Using /dev/sd instead of None {{(pid=63515) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 772.574263] env[63515]: DEBUG oslo_vmware.api [None req-78f15907-fa94-44dd-a2b3-ccdc7dec0a0f tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Task: {'id': task-1111165, 'name': Destroy_Task, 'duration_secs': 1.040917} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 772.575283] env[63515]: DEBUG nova.network.neutron [None req-9044ca9f-2eb8-445e-a5e5-e15900a98ac3 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] [instance: 8eb17506-ff93-4d25-b9af-ec5886569e65] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 772.576502] env[63515]: INFO nova.virt.vmwareapi.vm_util [None req-78f15907-fa94-44dd-a2b3-ccdc7dec0a0f tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] [instance: 87c468d9-9594-4804-b461-527f01f6118f] Destroyed the VM [ 772.576902] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-78f15907-fa94-44dd-a2b3-ccdc7dec0a0f tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] [instance: 87c468d9-9594-4804-b461-527f01f6118f] Deleting Snapshot of the VM instance {{(pid=63515) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 772.577346] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-d0f5ed9f-8fc3-40f2-ad44-c49f3796b005 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.580813] env[63515]: DEBUG oslo_concurrency.lockutils [None req-67d44811-1ff4-4a5b-a720-2ef182e4f0da tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] Acquiring lock "30b8b872-40ca-4297-b98e-a64c83a75483" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 772.581048] env[63515]: DEBUG oslo_concurrency.lockutils [None req-67d44811-1ff4-4a5b-a720-2ef182e4f0da tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] Lock "30b8b872-40ca-4297-b98e-a64c83a75483" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 772.581268] env[63515]: DEBUG oslo_concurrency.lockutils [None req-67d44811-1ff4-4a5b-a720-2ef182e4f0da tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] Acquiring lock "30b8b872-40ca-4297-b98e-a64c83a75483-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 772.581483] env[63515]: DEBUG oslo_concurrency.lockutils [None req-67d44811-1ff4-4a5b-a720-2ef182e4f0da tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] Lock "30b8b872-40ca-4297-b98e-a64c83a75483-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 772.581691] env[63515]: DEBUG oslo_concurrency.lockutils [None req-67d44811-1ff4-4a5b-a720-2ef182e4f0da tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] Lock "30b8b872-40ca-4297-b98e-a64c83a75483-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 772.585639] env[63515]: DEBUG nova.policy [None req-2786351a-303c-45a8-b63b-65e218eb6ef6 tempest-ServersTestJSON-324454119 tempest-ServersTestJSON-324454119-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'edab0912753c4d009433de19039a7e2b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f8662644a3354b46a19edbd1f8ce2c28', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63515) authorize /opt/stack/nova/nova/policy.py:201}} [ 772.587333] env[63515]: INFO nova.compute.manager [None req-67d44811-1ff4-4a5b-a720-2ef182e4f0da tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] [instance: 30b8b872-40ca-4297-b98e-a64c83a75483] Terminating instance [ 772.588969] env[63515]: DEBUG oslo_vmware.api [None req-78f15907-fa94-44dd-a2b3-ccdc7dec0a0f tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Waiting for the task: (returnval){ [ 772.588969] env[63515]: value = "task-1111169" [ 772.588969] env[63515]: _type = "Task" [ 772.588969] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 772.589955] env[63515]: DEBUG nova.compute.manager [None req-67d44811-1ff4-4a5b-a720-2ef182e4f0da tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] [instance: 30b8b872-40ca-4297-b98e-a64c83a75483] Start destroying the instance on the hypervisor. {{(pid=63515) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 772.590173] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-67d44811-1ff4-4a5b-a720-2ef182e4f0da tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] [instance: 30b8b872-40ca-4297-b98e-a64c83a75483] Destroying instance {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 772.591319] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9764b002-2679-4360-b582-3fa462709b9e {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.604645] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-67d44811-1ff4-4a5b-a720-2ef182e4f0da tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] [instance: 30b8b872-40ca-4297-b98e-a64c83a75483] Powering off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 772.605237] env[63515]: DEBUG oslo_vmware.api [None req-78f15907-fa94-44dd-a2b3-ccdc7dec0a0f tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Task: {'id': task-1111169, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 772.605473] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-32dd60cb-2ffd-4a0e-97c4-69ceb133df1a {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.614822] env[63515]: DEBUG oslo_vmware.api [None req-67d44811-1ff4-4a5b-a720-2ef182e4f0da tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] Waiting for the task: (returnval){ [ 772.614822] env[63515]: value = "task-1111170" [ 772.614822] env[63515]: _type = "Task" [ 772.614822] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 772.627421] env[63515]: DEBUG oslo_vmware.api [None req-67d44811-1ff4-4a5b-a720-2ef182e4f0da tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] Task: {'id': task-1111170, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 772.888029] env[63515]: DEBUG oslo_concurrency.lockutils [None req-549cc3f8-330f-4c6a-979d-35eedc6f8146 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Releasing lock "refresh_cache-84dbf321-2c70-4e08-b430-cb5a06fc6829" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 772.888029] env[63515]: DEBUG nova.compute.manager [None req-549cc3f8-330f-4c6a-979d-35eedc6f8146 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: 84dbf321-2c70-4e08-b430-cb5a06fc6829] Instance network_info: |[{"id": "ad6685ff-3a87-4cc5-bdfa-5eb3bb2f0590", "address": "fa:16:3e:5d:a9:21", "network": {"id": "320921bd-a6c9-4726-a801-19009d70d16e", "bridge": "br-int", "label": "tempest-ImagesTestJSON-304158429-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2259667ddc0c42639ea1af3b6e6fa0e8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2d94740a-bce8-4103-8ecf-230d02ec0a44", "external-id": "nsx-vlan-transportzone-149", "segmentation_id": 149, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapad6685ff-3a", "ovs_interfaceid": "ad6685ff-3a87-4cc5-bdfa-5eb3bb2f0590", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 772.888401] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-549cc3f8-330f-4c6a-979d-35eedc6f8146 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: 84dbf321-2c70-4e08-b430-cb5a06fc6829] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:5d:a9:21', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '2d94740a-bce8-4103-8ecf-230d02ec0a44', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ad6685ff-3a87-4cc5-bdfa-5eb3bb2f0590', 'vif_model': 'vmxnet3'}] {{(pid=63515) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 772.896699] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-549cc3f8-330f-4c6a-979d-35eedc6f8146 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Creating folder: Project (2259667ddc0c42639ea1af3b6e6fa0e8). Parent ref: group-v243370. {{(pid=63515) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 772.897197] env[63515]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b6caad33-f958-4207-8685-531473214343 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.909124] env[63515]: DEBUG oslo_vmware.api [None req-6d0510cf-97fb-4582-ac56-e14cb4885964 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Task: {'id': task-1111168, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.297001} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 772.909458] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-6d0510cf-97fb-4582-ac56-e14cb4885964 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Deleted the datastore file {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 772.909683] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-6d0510cf-97fb-4582-ac56-e14cb4885964 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] [instance: 38d6d030-06b0-4185-904d-44a038b3a752] Deleted contents of the VM from datastore datastore2 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 772.909905] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-6d0510cf-97fb-4582-ac56-e14cb4885964 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] [instance: 38d6d030-06b0-4185-904d-44a038b3a752] Instance destroyed {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 772.910135] env[63515]: INFO nova.compute.manager [None req-6d0510cf-97fb-4582-ac56-e14cb4885964 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] [instance: 38d6d030-06b0-4185-904d-44a038b3a752] Took 1.11 seconds to destroy the instance on the hypervisor. [ 772.910497] env[63515]: DEBUG oslo.service.loopingcall [None req-6d0510cf-97fb-4582-ac56-e14cb4885964 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 772.911714] env[63515]: DEBUG nova.compute.manager [-] [instance: 38d6d030-06b0-4185-904d-44a038b3a752] Deallocating network for instance {{(pid=63515) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 772.911849] env[63515]: DEBUG nova.network.neutron [-] [instance: 38d6d030-06b0-4185-904d-44a038b3a752] deallocate_for_instance() {{(pid=63515) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 772.913542] env[63515]: INFO nova.virt.vmwareapi.vm_util [None req-549cc3f8-330f-4c6a-979d-35eedc6f8146 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Created folder: Project (2259667ddc0c42639ea1af3b6e6fa0e8) in parent group-v243370. [ 772.913767] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-549cc3f8-330f-4c6a-979d-35eedc6f8146 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Creating folder: Instances. Parent ref: group-v243449. {{(pid=63515) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 772.914044] env[63515]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b94cfc49-053c-446e-ba77-7c6095fc1351 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.923216] env[63515]: INFO nova.virt.vmwareapi.vm_util [None req-549cc3f8-330f-4c6a-979d-35eedc6f8146 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Created folder: Instances in parent group-v243449. [ 772.923559] env[63515]: DEBUG oslo.service.loopingcall [None req-549cc3f8-330f-4c6a-979d-35eedc6f8146 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 772.923829] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 84dbf321-2c70-4e08-b430-cb5a06fc6829] Creating VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 772.924138] env[63515]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-fe9abc7c-4d5d-43cc-af57-67f64827c271 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.948397] env[63515]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 772.948397] env[63515]: value = "task-1111173" [ 772.948397] env[63515]: _type = "Task" [ 772.948397] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 772.959603] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1111173, 'name': CreateVM_Task} progress is 5%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 773.006150] env[63515]: DEBUG nova.compute.manager [None req-2786351a-303c-45a8-b63b-65e218eb6ef6 tempest-ServersTestJSON-324454119 tempest-ServersTestJSON-324454119-project-member] [instance: 70ed982f-affd-4dd1-bc90-c64e7c6d49d2] Start building block device mappings for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 773.011159] env[63515]: DEBUG nova.network.neutron [None req-2786351a-303c-45a8-b63b-65e218eb6ef6 tempest-ServersTestJSON-324454119 tempest-ServersTestJSON-324454119-project-member] [instance: 70ed982f-affd-4dd1-bc90-c64e7c6d49d2] Successfully created port: fe7c09da-d6bd-4f72-afa2-1a70bc663407 {{(pid=63515) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 773.033420] env[63515]: DEBUG oslo_concurrency.lockutils [None req-efd46494-ea54-4600-906e-1e7f1371fb6c tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Lock "d7e8dcdb-41cb-46fb-8b61-d251e7c2d372" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.006s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 773.078103] env[63515]: DEBUG oslo_concurrency.lockutils [None req-9044ca9f-2eb8-445e-a5e5-e15900a98ac3 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] Releasing lock "refresh_cache-8eb17506-ff93-4d25-b9af-ec5886569e65" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 773.078665] env[63515]: DEBUG nova.compute.manager [None req-9044ca9f-2eb8-445e-a5e5-e15900a98ac3 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] [instance: 8eb17506-ff93-4d25-b9af-ec5886569e65] Start destroying the instance on the hypervisor. {{(pid=63515) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 773.078946] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-9044ca9f-2eb8-445e-a5e5-e15900a98ac3 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] [instance: 8eb17506-ff93-4d25-b9af-ec5886569e65] Destroying instance {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 773.080215] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24e3449f-5f09-4e93-b5a7-3e06cabd0575 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.089589] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-9044ca9f-2eb8-445e-a5e5-e15900a98ac3 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] [instance: 8eb17506-ff93-4d25-b9af-ec5886569e65] Powering off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 773.089885] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ca00d700-4618-4e8a-bce8-1e7d1a208913 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.102485] env[63515]: DEBUG oslo_vmware.api [None req-9044ca9f-2eb8-445e-a5e5-e15900a98ac3 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] Waiting for the task: (returnval){ [ 773.102485] env[63515]: value = "task-1111174" [ 773.102485] env[63515]: _type = "Task" [ 773.102485] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 773.111770] env[63515]: DEBUG oslo_vmware.api [None req-78f15907-fa94-44dd-a2b3-ccdc7dec0a0f tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Task: {'id': task-1111169, 'name': RemoveSnapshot_Task, 'duration_secs': 0.39981} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 773.112891] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-78f15907-fa94-44dd-a2b3-ccdc7dec0a0f tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] [instance: 87c468d9-9594-4804-b461-527f01f6118f] Deleted Snapshot of the VM instance {{(pid=63515) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 773.113418] env[63515]: DEBUG nova.compute.manager [None req-78f15907-fa94-44dd-a2b3-ccdc7dec0a0f tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] [instance: 87c468d9-9594-4804-b461-527f01f6118f] Checking state {{(pid=63515) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 773.114372] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd010ddf-a2f8-4f79-95fa-3e087d2999c6 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.125040] env[63515]: DEBUG oslo_vmware.api [None req-9044ca9f-2eb8-445e-a5e5-e15900a98ac3 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] Task: {'id': task-1111174, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 773.137212] env[63515]: DEBUG oslo_vmware.api [None req-67d44811-1ff4-4a5b-a720-2ef182e4f0da tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] Task: {'id': task-1111170, 'name': PowerOffVM_Task, 'duration_secs': 0.187891} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 773.142453] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-67d44811-1ff4-4a5b-a720-2ef182e4f0da tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] [instance: 30b8b872-40ca-4297-b98e-a64c83a75483] Powered off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 773.143932] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-67d44811-1ff4-4a5b-a720-2ef182e4f0da tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] [instance: 30b8b872-40ca-4297-b98e-a64c83a75483] Unregistering the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 773.143932] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-4c1e2a53-2f8d-4033-80aa-45e301353a28 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.220766] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-67d44811-1ff4-4a5b-a720-2ef182e4f0da tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] [instance: 30b8b872-40ca-4297-b98e-a64c83a75483] Unregistered the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 773.221042] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-67d44811-1ff4-4a5b-a720-2ef182e4f0da tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] [instance: 30b8b872-40ca-4297-b98e-a64c83a75483] Deleting contents of the VM from datastore datastore1 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 773.221373] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-67d44811-1ff4-4a5b-a720-2ef182e4f0da tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] Deleting the datastore file [datastore1] 30b8b872-40ca-4297-b98e-a64c83a75483 {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 773.221704] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e68ed062-17ed-40da-b142-bff4bb8e189b {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.230586] env[63515]: DEBUG oslo_vmware.api [None req-67d44811-1ff4-4a5b-a720-2ef182e4f0da tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] Waiting for the task: (returnval){ [ 773.230586] env[63515]: value = "task-1111176" [ 773.230586] env[63515]: _type = "Task" [ 773.230586] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 773.244068] env[63515]: DEBUG oslo_vmware.api [None req-67d44811-1ff4-4a5b-a720-2ef182e4f0da tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] Task: {'id': task-1111176, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 773.462395] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1111173, 'name': CreateVM_Task, 'duration_secs': 0.390402} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 773.462395] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 84dbf321-2c70-4e08-b430-cb5a06fc6829] Created VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 773.463565] env[63515]: DEBUG oslo_concurrency.lockutils [None req-549cc3f8-330f-4c6a-979d-35eedc6f8146 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 773.463565] env[63515]: DEBUG oslo_concurrency.lockutils [None req-549cc3f8-330f-4c6a-979d-35eedc6f8146 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 773.463565] env[63515]: DEBUG oslo_concurrency.lockutils [None req-549cc3f8-330f-4c6a-979d-35eedc6f8146 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 773.463565] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1126f17c-4342-42eb-a27a-fb03af4e0b35 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.470400] env[63515]: DEBUG oslo_vmware.api [None req-549cc3f8-330f-4c6a-979d-35eedc6f8146 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Waiting for the task: (returnval){ [ 773.470400] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]520d924b-94f9-099f-ca2c-304c78decc4c" [ 773.470400] env[63515]: _type = "Task" [ 773.470400] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 773.478338] env[63515]: DEBUG oslo_vmware.api [None req-549cc3f8-330f-4c6a-979d-35eedc6f8146 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]520d924b-94f9-099f-ca2c-304c78decc4c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 773.479741] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b149dabb-1de0-4aa0-bae3-0cc663d9e468 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.485923] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3bc55cc2-b03e-4b8b-b6f6-2b3a94ef2484 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.515311] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d733f9a4-181d-4a8c-973b-cb77fa52dc8f {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.527968] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47cc3fc2-db59-490d-bee6-756b88c538e4 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.534416] env[63515]: DEBUG oslo_service.periodic_task [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=63515) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 773.534716] env[63515]: DEBUG oslo_service.periodic_task [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Running periodic task ComputeManager._run_pending_deletes {{(pid=63515) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 773.537595] env[63515]: DEBUG nova.compute.manager [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Cleaning up deleted instances {{(pid=63515) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11226}} [ 773.545389] env[63515]: DEBUG nova.compute.provider_tree [None req-8407c401-88c7-47ad-b91e-e7693f87e0f1 tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 773.616530] env[63515]: DEBUG oslo_vmware.api [None req-9044ca9f-2eb8-445e-a5e5-e15900a98ac3 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] Task: {'id': task-1111174, 'name': PowerOffVM_Task, 'duration_secs': 0.29344} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 773.616530] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-9044ca9f-2eb8-445e-a5e5-e15900a98ac3 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] [instance: 8eb17506-ff93-4d25-b9af-ec5886569e65] Powered off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 773.616667] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-9044ca9f-2eb8-445e-a5e5-e15900a98ac3 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] [instance: 8eb17506-ff93-4d25-b9af-ec5886569e65] Unregistering the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 773.616798] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f48f7877-dcef-42ab-a2f4-12b3bac0bf97 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.642942] env[63515]: INFO nova.compute.manager [None req-78f15907-fa94-44dd-a2b3-ccdc7dec0a0f tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] [instance: 87c468d9-9594-4804-b461-527f01f6118f] Shelve offloading [ 773.644050] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-9044ca9f-2eb8-445e-a5e5-e15900a98ac3 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] [instance: 8eb17506-ff93-4d25-b9af-ec5886569e65] Unregistered the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 773.644299] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-9044ca9f-2eb8-445e-a5e5-e15900a98ac3 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] [instance: 8eb17506-ff93-4d25-b9af-ec5886569e65] Deleting contents of the VM from datastore datastore2 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 773.644423] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-9044ca9f-2eb8-445e-a5e5-e15900a98ac3 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] Deleting the datastore file [datastore2] 8eb17506-ff93-4d25-b9af-ec5886569e65 {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 773.645036] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-25906d5a-8eed-4669-8af7-e5367e4a575d {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.647143] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-78f15907-fa94-44dd-a2b3-ccdc7dec0a0f tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] [instance: 87c468d9-9594-4804-b461-527f01f6118f] Powering off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 773.647376] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7f72f8eb-e144-43e1-b467-f0dae5277b84 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.656345] env[63515]: DEBUG oslo_vmware.api [None req-78f15907-fa94-44dd-a2b3-ccdc7dec0a0f tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Waiting for the task: (returnval){ [ 773.656345] env[63515]: value = "task-1111179" [ 773.656345] env[63515]: _type = "Task" [ 773.656345] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 773.658356] env[63515]: DEBUG oslo_vmware.api [None req-9044ca9f-2eb8-445e-a5e5-e15900a98ac3 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] Waiting for the task: (returnval){ [ 773.658356] env[63515]: value = "task-1111178" [ 773.658356] env[63515]: _type = "Task" [ 773.658356] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 773.665412] env[63515]: DEBUG nova.network.neutron [-] [instance: 38d6d030-06b0-4185-904d-44a038b3a752] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 773.672912] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-78f15907-fa94-44dd-a2b3-ccdc7dec0a0f tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] [instance: 87c468d9-9594-4804-b461-527f01f6118f] VM already powered off {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 773.673135] env[63515]: DEBUG nova.compute.manager [None req-78f15907-fa94-44dd-a2b3-ccdc7dec0a0f tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] [instance: 87c468d9-9594-4804-b461-527f01f6118f] Checking state {{(pid=63515) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 773.673555] env[63515]: DEBUG oslo_vmware.api [None req-9044ca9f-2eb8-445e-a5e5-e15900a98ac3 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] Task: {'id': task-1111178, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 773.674280] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18b64804-a1d0-429b-ad54-1ad51a329469 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.680972] env[63515]: DEBUG oslo_concurrency.lockutils [None req-78f15907-fa94-44dd-a2b3-ccdc7dec0a0f tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Acquiring lock "refresh_cache-87c468d9-9594-4804-b461-527f01f6118f" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 773.681172] env[63515]: DEBUG oslo_concurrency.lockutils [None req-78f15907-fa94-44dd-a2b3-ccdc7dec0a0f tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Acquired lock "refresh_cache-87c468d9-9594-4804-b461-527f01f6118f" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 773.681364] env[63515]: DEBUG nova.network.neutron [None req-78f15907-fa94-44dd-a2b3-ccdc7dec0a0f tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] [instance: 87c468d9-9594-4804-b461-527f01f6118f] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 773.692931] env[63515]: DEBUG nova.compute.manager [req-1dafc83d-d43c-4ab3-8a9f-9b6108b30720 req-7242ceb2-9072-4d32-87a5-91f79482ac88 service nova] [instance: 84dbf321-2c70-4e08-b430-cb5a06fc6829] Received event network-changed-ad6685ff-3a87-4cc5-bdfa-5eb3bb2f0590 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 773.693137] env[63515]: DEBUG nova.compute.manager [req-1dafc83d-d43c-4ab3-8a9f-9b6108b30720 req-7242ceb2-9072-4d32-87a5-91f79482ac88 service nova] [instance: 84dbf321-2c70-4e08-b430-cb5a06fc6829] Refreshing instance network info cache due to event network-changed-ad6685ff-3a87-4cc5-bdfa-5eb3bb2f0590. {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 773.693403] env[63515]: DEBUG oslo_concurrency.lockutils [req-1dafc83d-d43c-4ab3-8a9f-9b6108b30720 req-7242ceb2-9072-4d32-87a5-91f79482ac88 service nova] Acquiring lock "refresh_cache-84dbf321-2c70-4e08-b430-cb5a06fc6829" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 773.693541] env[63515]: DEBUG oslo_concurrency.lockutils [req-1dafc83d-d43c-4ab3-8a9f-9b6108b30720 req-7242ceb2-9072-4d32-87a5-91f79482ac88 service nova] Acquired lock "refresh_cache-84dbf321-2c70-4e08-b430-cb5a06fc6829" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 773.693714] env[63515]: DEBUG nova.network.neutron [req-1dafc83d-d43c-4ab3-8a9f-9b6108b30720 req-7242ceb2-9072-4d32-87a5-91f79482ac88 service nova] [instance: 84dbf321-2c70-4e08-b430-cb5a06fc6829] Refreshing network info cache for port ad6685ff-3a87-4cc5-bdfa-5eb3bb2f0590 {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 773.740089] env[63515]: DEBUG oslo_vmware.api [None req-67d44811-1ff4-4a5b-a720-2ef182e4f0da tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] Task: {'id': task-1111176, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.498533} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 773.740370] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-67d44811-1ff4-4a5b-a720-2ef182e4f0da tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] Deleted the datastore file {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 773.740563] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-67d44811-1ff4-4a5b-a720-2ef182e4f0da tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] [instance: 30b8b872-40ca-4297-b98e-a64c83a75483] Deleted contents of the VM from datastore datastore1 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 773.740744] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-67d44811-1ff4-4a5b-a720-2ef182e4f0da tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] [instance: 30b8b872-40ca-4297-b98e-a64c83a75483] Instance destroyed {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 773.740987] env[63515]: INFO nova.compute.manager [None req-67d44811-1ff4-4a5b-a720-2ef182e4f0da tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] [instance: 30b8b872-40ca-4297-b98e-a64c83a75483] Took 1.15 seconds to destroy the instance on the hypervisor. [ 773.741204] env[63515]: DEBUG oslo.service.loopingcall [None req-67d44811-1ff4-4a5b-a720-2ef182e4f0da tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 773.741399] env[63515]: DEBUG nova.compute.manager [-] [instance: 30b8b872-40ca-4297-b98e-a64c83a75483] Deallocating network for instance {{(pid=63515) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 773.741579] env[63515]: DEBUG nova.network.neutron [-] [instance: 30b8b872-40ca-4297-b98e-a64c83a75483] deallocate_for_instance() {{(pid=63515) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 773.981285] env[63515]: DEBUG oslo_vmware.api [None req-549cc3f8-330f-4c6a-979d-35eedc6f8146 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]520d924b-94f9-099f-ca2c-304c78decc4c, 'name': SearchDatastore_Task, 'duration_secs': 0.028093} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 773.981680] env[63515]: DEBUG oslo_concurrency.lockutils [None req-549cc3f8-330f-4c6a-979d-35eedc6f8146 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 773.981949] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-549cc3f8-330f-4c6a-979d-35eedc6f8146 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: 84dbf321-2c70-4e08-b430-cb5a06fc6829] Processing image 8a120570-cb06-4099-b262-554ca0ad15c5 {{(pid=63515) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 773.982216] env[63515]: DEBUG oslo_concurrency.lockutils [None req-549cc3f8-330f-4c6a-979d-35eedc6f8146 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 773.982374] env[63515]: DEBUG oslo_concurrency.lockutils [None req-549cc3f8-330f-4c6a-979d-35eedc6f8146 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 773.982564] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-549cc3f8-330f-4c6a-979d-35eedc6f8146 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 773.982877] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-fe675d88-30cd-472b-b810-776a2c5c4863 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.997363] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-549cc3f8-330f-4c6a-979d-35eedc6f8146 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 773.997613] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-549cc3f8-330f-4c6a-979d-35eedc6f8146 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63515) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 773.998382] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-940aa7d8-b9d9-48d1-94ab-ab7485a49594 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.003559] env[63515]: DEBUG oslo_vmware.api [None req-549cc3f8-330f-4c6a-979d-35eedc6f8146 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Waiting for the task: (returnval){ [ 774.003559] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]5273ee6e-0735-253f-5bb8-e05247682a46" [ 774.003559] env[63515]: _type = "Task" [ 774.003559] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 774.011476] env[63515]: DEBUG oslo_vmware.api [None req-549cc3f8-330f-4c6a-979d-35eedc6f8146 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]5273ee6e-0735-253f-5bb8-e05247682a46, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 774.024690] env[63515]: DEBUG nova.compute.manager [None req-2786351a-303c-45a8-b63b-65e218eb6ef6 tempest-ServersTestJSON-324454119 tempest-ServersTestJSON-324454119-project-member] [instance: 70ed982f-affd-4dd1-bc90-c64e7c6d49d2] Start spawning the instance on the hypervisor. {{(pid=63515) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 774.039034] env[63515]: DEBUG nova.compute.manager [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] There are 6 instances to clean {{(pid=63515) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11235}} [ 774.039444] env[63515]: DEBUG nova.compute.manager [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] [instance: b85e9a70-7f5b-4d32-b616-f2a97e3186c8] Instance has had 0 of 5 cleanup attempts {{(pid=63515) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 774.041576] env[63515]: DEBUG oslo_concurrency.lockutils [None req-ca1a9b4c-30ff-4beb-9aec-3fa3a8a9ff12 tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Acquiring lock "b4477e66-ae12-4929-90ed-b7b652e0f207" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 774.041849] env[63515]: DEBUG oslo_concurrency.lockutils [None req-ca1a9b4c-30ff-4beb-9aec-3fa3a8a9ff12 tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Lock "b4477e66-ae12-4929-90ed-b7b652e0f207" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 774.042075] env[63515]: DEBUG oslo_concurrency.lockutils [None req-ca1a9b4c-30ff-4beb-9aec-3fa3a8a9ff12 tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Acquiring lock "b4477e66-ae12-4929-90ed-b7b652e0f207-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 774.042269] env[63515]: DEBUG oslo_concurrency.lockutils [None req-ca1a9b4c-30ff-4beb-9aec-3fa3a8a9ff12 tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Lock "b4477e66-ae12-4929-90ed-b7b652e0f207-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 774.042766] env[63515]: DEBUG oslo_concurrency.lockutils [None req-ca1a9b4c-30ff-4beb-9aec-3fa3a8a9ff12 tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Lock "b4477e66-ae12-4929-90ed-b7b652e0f207-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 774.045214] env[63515]: INFO nova.compute.manager [None req-ca1a9b4c-30ff-4beb-9aec-3fa3a8a9ff12 tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] [instance: b4477e66-ae12-4929-90ed-b7b652e0f207] Terminating instance [ 774.046809] env[63515]: DEBUG oslo_concurrency.lockutils [None req-ca1a9b4c-30ff-4beb-9aec-3fa3a8a9ff12 tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Acquiring lock "refresh_cache-b4477e66-ae12-4929-90ed-b7b652e0f207" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 774.046950] env[63515]: DEBUG oslo_concurrency.lockutils [None req-ca1a9b4c-30ff-4beb-9aec-3fa3a8a9ff12 tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Acquired lock "refresh_cache-b4477e66-ae12-4929-90ed-b7b652e0f207" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 774.047506] env[63515]: DEBUG nova.network.neutron [None req-ca1a9b4c-30ff-4beb-9aec-3fa3a8a9ff12 tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] [instance: b4477e66-ae12-4929-90ed-b7b652e0f207] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 774.048489] env[63515]: DEBUG nova.scheduler.client.report [None req-8407c401-88c7-47ad-b91e-e7693f87e0f1 tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 774.063010] env[63515]: DEBUG nova.virt.hardware [None req-2786351a-303c-45a8-b63b-65e218eb6ef6 tempest-ServersTestJSON-324454119 tempest-ServersTestJSON-324454119-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T02:52:14Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T02:51:56Z,direct_url=,disk_format='vmdk',id=8a120570-cb06-4099-b262-554ca0ad15c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b569255cc43e42e7ae3f2b4ad37c6ef4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T02:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 774.063280] env[63515]: DEBUG nova.virt.hardware [None req-2786351a-303c-45a8-b63b-65e218eb6ef6 tempest-ServersTestJSON-324454119 tempest-ServersTestJSON-324454119-project-member] Flavor limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 774.063445] env[63515]: DEBUG nova.virt.hardware [None req-2786351a-303c-45a8-b63b-65e218eb6ef6 tempest-ServersTestJSON-324454119 tempest-ServersTestJSON-324454119-project-member] Image limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 774.063633] env[63515]: DEBUG nova.virt.hardware [None req-2786351a-303c-45a8-b63b-65e218eb6ef6 tempest-ServersTestJSON-324454119 tempest-ServersTestJSON-324454119-project-member] Flavor pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 774.063783] env[63515]: DEBUG nova.virt.hardware [None req-2786351a-303c-45a8-b63b-65e218eb6ef6 tempest-ServersTestJSON-324454119 tempest-ServersTestJSON-324454119-project-member] Image pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 774.063934] env[63515]: DEBUG nova.virt.hardware [None req-2786351a-303c-45a8-b63b-65e218eb6ef6 tempest-ServersTestJSON-324454119 tempest-ServersTestJSON-324454119-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 774.064166] env[63515]: DEBUG nova.virt.hardware [None req-2786351a-303c-45a8-b63b-65e218eb6ef6 tempest-ServersTestJSON-324454119 tempest-ServersTestJSON-324454119-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 774.064357] env[63515]: DEBUG nova.virt.hardware [None req-2786351a-303c-45a8-b63b-65e218eb6ef6 tempest-ServersTestJSON-324454119 tempest-ServersTestJSON-324454119-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 774.064506] env[63515]: DEBUG nova.virt.hardware [None req-2786351a-303c-45a8-b63b-65e218eb6ef6 tempest-ServersTestJSON-324454119 tempest-ServersTestJSON-324454119-project-member] Got 1 possible topologies {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 774.064673] env[63515]: DEBUG nova.virt.hardware [None req-2786351a-303c-45a8-b63b-65e218eb6ef6 tempest-ServersTestJSON-324454119 tempest-ServersTestJSON-324454119-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 774.064856] env[63515]: DEBUG nova.virt.hardware [None req-2786351a-303c-45a8-b63b-65e218eb6ef6 tempest-ServersTestJSON-324454119 tempest-ServersTestJSON-324454119-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 774.066052] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0cd6785e-d127-41a1-8c81-74321aa74e8c {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.074904] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92c06f3b-af7c-4e1f-a855-af7f1e9073d3 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.117317] env[63515]: DEBUG oslo_concurrency.lockutils [None req-efd46494-ea54-4600-906e-1e7f1371fb6c tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Acquiring lock "d7e8dcdb-41cb-46fb-8b61-d251e7c2d372" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 774.117676] env[63515]: DEBUG oslo_concurrency.lockutils [None req-efd46494-ea54-4600-906e-1e7f1371fb6c tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Lock "d7e8dcdb-41cb-46fb-8b61-d251e7c2d372" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 774.117993] env[63515]: INFO nova.compute.manager [None req-efd46494-ea54-4600-906e-1e7f1371fb6c tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] [instance: d7e8dcdb-41cb-46fb-8b61-d251e7c2d372] Attaching volume b1a5444a-1e2f-4443-8931-4a04ec5d9d68 to /dev/sdb [ 774.154446] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df0287c3-9e0d-4b56-9a28-d886dc5761f7 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.161934] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6674c585-dcdf-4370-bdcf-203a1bc0454d {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.171955] env[63515]: DEBUG oslo_vmware.api [None req-9044ca9f-2eb8-445e-a5e5-e15900a98ac3 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] Task: {'id': task-1111178, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.282824} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 774.172227] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-9044ca9f-2eb8-445e-a5e5-e15900a98ac3 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] Deleted the datastore file {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 774.172417] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-9044ca9f-2eb8-445e-a5e5-e15900a98ac3 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] [instance: 8eb17506-ff93-4d25-b9af-ec5886569e65] Deleted contents of the VM from datastore datastore2 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 774.172594] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-9044ca9f-2eb8-445e-a5e5-e15900a98ac3 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] [instance: 8eb17506-ff93-4d25-b9af-ec5886569e65] Instance destroyed {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 774.172803] env[63515]: INFO nova.compute.manager [None req-9044ca9f-2eb8-445e-a5e5-e15900a98ac3 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] [instance: 8eb17506-ff93-4d25-b9af-ec5886569e65] Took 1.09 seconds to destroy the instance on the hypervisor. [ 774.173059] env[63515]: DEBUG oslo.service.loopingcall [None req-9044ca9f-2eb8-445e-a5e5-e15900a98ac3 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 774.173421] env[63515]: INFO nova.compute.manager [-] [instance: 38d6d030-06b0-4185-904d-44a038b3a752] Took 1.26 seconds to deallocate network for instance. [ 774.173633] env[63515]: DEBUG nova.compute.manager [-] [instance: 8eb17506-ff93-4d25-b9af-ec5886569e65] Deallocating network for instance {{(pid=63515) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 774.173789] env[63515]: DEBUG nova.network.neutron [-] [instance: 8eb17506-ff93-4d25-b9af-ec5886569e65] deallocate_for_instance() {{(pid=63515) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 774.179841] env[63515]: DEBUG nova.virt.block_device [None req-efd46494-ea54-4600-906e-1e7f1371fb6c tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] [instance: d7e8dcdb-41cb-46fb-8b61-d251e7c2d372] Updating existing volume attachment record: 97466c82-53b3-4508-aa61-f1f29e00e69d {{(pid=63515) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 774.203601] env[63515]: DEBUG nova.network.neutron [-] [instance: 8eb17506-ff93-4d25-b9af-ec5886569e65] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 774.517290] env[63515]: DEBUG oslo_vmware.api [None req-549cc3f8-330f-4c6a-979d-35eedc6f8146 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]5273ee6e-0735-253f-5bb8-e05247682a46, 'name': SearchDatastore_Task, 'duration_secs': 0.010301} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 774.518264] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-337849de-a6d2-4482-b574-478a37132296 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.524042] env[63515]: DEBUG oslo_vmware.api [None req-549cc3f8-330f-4c6a-979d-35eedc6f8146 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Waiting for the task: (returnval){ [ 774.524042] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]52bd8702-c01f-4fa1-a33f-05c5b3cd805a" [ 774.524042] env[63515]: _type = "Task" [ 774.524042] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 774.535467] env[63515]: DEBUG nova.network.neutron [-] [instance: 30b8b872-40ca-4297-b98e-a64c83a75483] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 774.540384] env[63515]: DEBUG oslo_vmware.api [None req-549cc3f8-330f-4c6a-979d-35eedc6f8146 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52bd8702-c01f-4fa1-a33f-05c5b3cd805a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 774.547390] env[63515]: DEBUG nova.compute.manager [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] [instance: 48668736-df27-4f2a-94d9-132f5b49701b] Instance has had 0 of 5 cleanup attempts {{(pid=63515) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 774.555974] env[63515]: DEBUG oslo_concurrency.lockutils [None req-8407c401-88c7-47ad-b91e-e7693f87e0f1 tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.567s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 774.555974] env[63515]: DEBUG nova.compute.manager [None req-8407c401-88c7-47ad-b91e-e7693f87e0f1 tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] [instance: 192137b0-03e5-4bc4-b911-4b4f1a874f74] Start building networks asynchronously for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 774.561028] env[63515]: DEBUG oslo_concurrency.lockutils [None req-dac2c5cc-03a7-4df4-a3f6-d707671aaebe tempest-AttachInterfacesUnderV243Test-1943594058 tempest-AttachInterfacesUnderV243Test-1943594058-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 27.706s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 774.561028] env[63515]: DEBUG oslo_concurrency.lockutils [None req-dac2c5cc-03a7-4df4-a3f6-d707671aaebe tempest-AttachInterfacesUnderV243Test-1943594058 tempest-AttachInterfacesUnderV243Test-1943594058-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 774.562288] env[63515]: DEBUG oslo_concurrency.lockutils [None req-ffbb9c80-abfa-457c-b756-9497c4635a40 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: waited 25.080s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 774.589284] env[63515]: INFO nova.scheduler.client.report [None req-dac2c5cc-03a7-4df4-a3f6-d707671aaebe tempest-AttachInterfacesUnderV243Test-1943594058 tempest-AttachInterfacesUnderV243Test-1943594058-project-member] Deleted allocations for instance b911a5b5-9617-4fb3-9b5e-fb8c492e4931 [ 774.591122] env[63515]: DEBUG nova.network.neutron [None req-ca1a9b4c-30ff-4beb-9aec-3fa3a8a9ff12 tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] [instance: b4477e66-ae12-4929-90ed-b7b652e0f207] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 774.661364] env[63515]: DEBUG nova.network.neutron [req-1dafc83d-d43c-4ab3-8a9f-9b6108b30720 req-7242ceb2-9072-4d32-87a5-91f79482ac88 service nova] [instance: 84dbf321-2c70-4e08-b430-cb5a06fc6829] Updated VIF entry in instance network info cache for port ad6685ff-3a87-4cc5-bdfa-5eb3bb2f0590. {{(pid=63515) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 774.661364] env[63515]: DEBUG nova.network.neutron [req-1dafc83d-d43c-4ab3-8a9f-9b6108b30720 req-7242ceb2-9072-4d32-87a5-91f79482ac88 service nova] [instance: 84dbf321-2c70-4e08-b430-cb5a06fc6829] Updating instance_info_cache with network_info: [{"id": "ad6685ff-3a87-4cc5-bdfa-5eb3bb2f0590", "address": "fa:16:3e:5d:a9:21", "network": {"id": "320921bd-a6c9-4726-a801-19009d70d16e", "bridge": "br-int", "label": "tempest-ImagesTestJSON-304158429-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2259667ddc0c42639ea1af3b6e6fa0e8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2d94740a-bce8-4103-8ecf-230d02ec0a44", "external-id": "nsx-vlan-transportzone-149", "segmentation_id": 149, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapad6685ff-3a", "ovs_interfaceid": "ad6685ff-3a87-4cc5-bdfa-5eb3bb2f0590", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 774.691018] env[63515]: DEBUG oslo_concurrency.lockutils [None req-6d0510cf-97fb-4582-ac56-e14cb4885964 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 774.694813] env[63515]: DEBUG nova.network.neutron [None req-78f15907-fa94-44dd-a2b3-ccdc7dec0a0f tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] [instance: 87c468d9-9594-4804-b461-527f01f6118f] Updating instance_info_cache with network_info: [{"id": "d2098868-0475-4b06-bae3-785c608a861e", "address": "fa:16:3e:af:f9:f0", "network": {"id": "09fb20ca-2b17-4d46-b915-6aff85b862d8", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1815529427-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.212", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "19ce3b05eec64475b685faad1fd801f6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ded18042-834c-4792-b3e8-b1c377446432", "external-id": "nsx-vlan-transportzone-293", "segmentation_id": 293, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd2098868-04", "ovs_interfaceid": "d2098868-0475-4b06-bae3-785c608a861e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 774.706252] env[63515]: DEBUG nova.network.neutron [-] [instance: 8eb17506-ff93-4d25-b9af-ec5886569e65] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 774.712310] env[63515]: DEBUG nova.network.neutron [None req-ca1a9b4c-30ff-4beb-9aec-3fa3a8a9ff12 tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] [instance: b4477e66-ae12-4929-90ed-b7b652e0f207] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 774.733649] env[63515]: DEBUG nova.compute.manager [req-18f42671-0b02-421c-85c7-584dbf4743c1 req-5c40e90b-adb3-44aa-8805-8608c06d40b3 service nova] [instance: 70ed982f-affd-4dd1-bc90-c64e7c6d49d2] Received event network-vif-plugged-fe7c09da-d6bd-4f72-afa2-1a70bc663407 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 774.733907] env[63515]: DEBUG oslo_concurrency.lockutils [req-18f42671-0b02-421c-85c7-584dbf4743c1 req-5c40e90b-adb3-44aa-8805-8608c06d40b3 service nova] Acquiring lock "70ed982f-affd-4dd1-bc90-c64e7c6d49d2-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 774.734260] env[63515]: DEBUG oslo_concurrency.lockutils [req-18f42671-0b02-421c-85c7-584dbf4743c1 req-5c40e90b-adb3-44aa-8805-8608c06d40b3 service nova] Lock "70ed982f-affd-4dd1-bc90-c64e7c6d49d2-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 774.734502] env[63515]: DEBUG oslo_concurrency.lockutils [req-18f42671-0b02-421c-85c7-584dbf4743c1 req-5c40e90b-adb3-44aa-8805-8608c06d40b3 service nova] Lock "70ed982f-affd-4dd1-bc90-c64e7c6d49d2-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 774.734700] env[63515]: DEBUG nova.compute.manager [req-18f42671-0b02-421c-85c7-584dbf4743c1 req-5c40e90b-adb3-44aa-8805-8608c06d40b3 service nova] [instance: 70ed982f-affd-4dd1-bc90-c64e7c6d49d2] No waiting events found dispatching network-vif-plugged-fe7c09da-d6bd-4f72-afa2-1a70bc663407 {{(pid=63515) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 774.734884] env[63515]: WARNING nova.compute.manager [req-18f42671-0b02-421c-85c7-584dbf4743c1 req-5c40e90b-adb3-44aa-8805-8608c06d40b3 service nova] [instance: 70ed982f-affd-4dd1-bc90-c64e7c6d49d2] Received unexpected event network-vif-plugged-fe7c09da-d6bd-4f72-afa2-1a70bc663407 for instance with vm_state building and task_state spawning. [ 774.774519] env[63515]: DEBUG nova.network.neutron [None req-2786351a-303c-45a8-b63b-65e218eb6ef6 tempest-ServersTestJSON-324454119 tempest-ServersTestJSON-324454119-project-member] [instance: 70ed982f-affd-4dd1-bc90-c64e7c6d49d2] Successfully updated port: fe7c09da-d6bd-4f72-afa2-1a70bc663407 {{(pid=63515) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 774.894949] env[63515]: WARNING oslo_messaging._drivers.amqpdriver [None req-f8bace5a-3f38-4490-8a6d-e315321b7f63 tempest-ServersTestManualDisk-1114050206 tempest-ServersTestManualDisk-1114050206-project-member] Number of call queues is 11, greater than warning threshold: 10. There could be a leak. Increasing threshold to: 20 [ 775.036043] env[63515]: DEBUG oslo_vmware.api [None req-549cc3f8-330f-4c6a-979d-35eedc6f8146 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52bd8702-c01f-4fa1-a33f-05c5b3cd805a, 'name': SearchDatastore_Task, 'duration_secs': 0.009781} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 775.036043] env[63515]: DEBUG oslo_concurrency.lockutils [None req-549cc3f8-330f-4c6a-979d-35eedc6f8146 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 775.036267] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-549cc3f8-330f-4c6a-979d-35eedc6f8146 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk to [datastore2] 84dbf321-2c70-4e08-b430-cb5a06fc6829/84dbf321-2c70-4e08-b430-cb5a06fc6829.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 775.036401] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7a0693e2-a658-4d76-9cc8-9ff3239a16c8 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.038982] env[63515]: INFO nova.compute.manager [-] [instance: 30b8b872-40ca-4297-b98e-a64c83a75483] Took 1.30 seconds to deallocate network for instance. [ 775.048062] env[63515]: DEBUG oslo_vmware.api [None req-549cc3f8-330f-4c6a-979d-35eedc6f8146 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Waiting for the task: (returnval){ [ 775.048062] env[63515]: value = "task-1111183" [ 775.048062] env[63515]: _type = "Task" [ 775.048062] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 775.050241] env[63515]: DEBUG nova.compute.manager [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] [instance: 4087b2c4-6ed1-4b68-8b78-a36e34d935b1] Instance has had 0 of 5 cleanup attempts {{(pid=63515) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 775.054793] env[63515]: DEBUG oslo_vmware.api [None req-549cc3f8-330f-4c6a-979d-35eedc6f8146 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Task: {'id': task-1111183, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 775.066559] env[63515]: DEBUG nova.compute.utils [None req-8407c401-88c7-47ad-b91e-e7693f87e0f1 tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] Using /dev/sd instead of None {{(pid=63515) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 775.068014] env[63515]: DEBUG nova.objects.instance [None req-ffbb9c80-abfa-457c-b756-9497c4635a40 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Lazy-loading 'migration_context' on Instance uuid f1d01b75-ac9d-458d-8cc2-ae64cffca4e8 {{(pid=63515) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 775.069334] env[63515]: DEBUG nova.compute.manager [None req-8407c401-88c7-47ad-b91e-e7693f87e0f1 tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] [instance: 192137b0-03e5-4bc4-b911-4b4f1a874f74] Not allocating networking since 'none' was specified. {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 775.103826] env[63515]: DEBUG oslo_concurrency.lockutils [None req-dac2c5cc-03a7-4df4-a3f6-d707671aaebe tempest-AttachInterfacesUnderV243Test-1943594058 tempest-AttachInterfacesUnderV243Test-1943594058-project-member] Lock "b911a5b5-9617-4fb3-9b5e-fb8c492e4931" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 31.783s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 775.163449] env[63515]: DEBUG oslo_concurrency.lockutils [req-1dafc83d-d43c-4ab3-8a9f-9b6108b30720 req-7242ceb2-9072-4d32-87a5-91f79482ac88 service nova] Releasing lock "refresh_cache-84dbf321-2c70-4e08-b430-cb5a06fc6829" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 775.163794] env[63515]: DEBUG nova.compute.manager [req-1dafc83d-d43c-4ab3-8a9f-9b6108b30720 req-7242ceb2-9072-4d32-87a5-91f79482ac88 service nova] [instance: 30b8b872-40ca-4297-b98e-a64c83a75483] Received event network-changed-b8fe94ee-46b7-489a-819d-0699bc02bbb7 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 775.164071] env[63515]: DEBUG nova.compute.manager [req-1dafc83d-d43c-4ab3-8a9f-9b6108b30720 req-7242ceb2-9072-4d32-87a5-91f79482ac88 service nova] [instance: 30b8b872-40ca-4297-b98e-a64c83a75483] Refreshing instance network info cache due to event network-changed-b8fe94ee-46b7-489a-819d-0699bc02bbb7. {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 775.164349] env[63515]: DEBUG oslo_concurrency.lockutils [req-1dafc83d-d43c-4ab3-8a9f-9b6108b30720 req-7242ceb2-9072-4d32-87a5-91f79482ac88 service nova] Acquiring lock "refresh_cache-30b8b872-40ca-4297-b98e-a64c83a75483" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 775.164554] env[63515]: DEBUG oslo_concurrency.lockutils [req-1dafc83d-d43c-4ab3-8a9f-9b6108b30720 req-7242ceb2-9072-4d32-87a5-91f79482ac88 service nova] Acquired lock "refresh_cache-30b8b872-40ca-4297-b98e-a64c83a75483" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 775.164783] env[63515]: DEBUG nova.network.neutron [req-1dafc83d-d43c-4ab3-8a9f-9b6108b30720 req-7242ceb2-9072-4d32-87a5-91f79482ac88 service nova] [instance: 30b8b872-40ca-4297-b98e-a64c83a75483] Refreshing network info cache for port b8fe94ee-46b7-489a-819d-0699bc02bbb7 {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 775.198240] env[63515]: DEBUG oslo_concurrency.lockutils [None req-78f15907-fa94-44dd-a2b3-ccdc7dec0a0f tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Releasing lock "refresh_cache-87c468d9-9594-4804-b461-527f01f6118f" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 775.215380] env[63515]: DEBUG oslo_concurrency.lockutils [None req-ca1a9b4c-30ff-4beb-9aec-3fa3a8a9ff12 tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Releasing lock "refresh_cache-b4477e66-ae12-4929-90ed-b7b652e0f207" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 775.215906] env[63515]: DEBUG nova.compute.manager [None req-ca1a9b4c-30ff-4beb-9aec-3fa3a8a9ff12 tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] [instance: b4477e66-ae12-4929-90ed-b7b652e0f207] Start destroying the instance on the hypervisor. {{(pid=63515) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 775.216135] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-ca1a9b4c-30ff-4beb-9aec-3fa3a8a9ff12 tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] [instance: b4477e66-ae12-4929-90ed-b7b652e0f207] Destroying instance {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 775.216614] env[63515]: INFO nova.compute.manager [-] [instance: 8eb17506-ff93-4d25-b9af-ec5886569e65] Took 1.04 seconds to deallocate network for instance. [ 775.217578] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-edf2afcd-eaaf-4ff6-a737-5fda7e0f298b {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.229692] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-ca1a9b4c-30ff-4beb-9aec-3fa3a8a9ff12 tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] [instance: b4477e66-ae12-4929-90ed-b7b652e0f207] Powering off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 775.230067] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-eba24202-bb16-4a9a-b88b-d44904aa79bd {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.240405] env[63515]: DEBUG oslo_vmware.api [None req-ca1a9b4c-30ff-4beb-9aec-3fa3a8a9ff12 tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Waiting for the task: (returnval){ [ 775.240405] env[63515]: value = "task-1111184" [ 775.240405] env[63515]: _type = "Task" [ 775.240405] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 775.252149] env[63515]: DEBUG oslo_vmware.api [None req-ca1a9b4c-30ff-4beb-9aec-3fa3a8a9ff12 tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Task: {'id': task-1111184, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 775.277090] env[63515]: DEBUG oslo_concurrency.lockutils [None req-2786351a-303c-45a8-b63b-65e218eb6ef6 tempest-ServersTestJSON-324454119 tempest-ServersTestJSON-324454119-project-member] Acquiring lock "refresh_cache-70ed982f-affd-4dd1-bc90-c64e7c6d49d2" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 775.277270] env[63515]: DEBUG oslo_concurrency.lockutils [None req-2786351a-303c-45a8-b63b-65e218eb6ef6 tempest-ServersTestJSON-324454119 tempest-ServersTestJSON-324454119-project-member] Acquired lock "refresh_cache-70ed982f-affd-4dd1-bc90-c64e7c6d49d2" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 775.277475] env[63515]: DEBUG nova.network.neutron [None req-2786351a-303c-45a8-b63b-65e218eb6ef6 tempest-ServersTestJSON-324454119 tempest-ServersTestJSON-324454119-project-member] [instance: 70ed982f-affd-4dd1-bc90-c64e7c6d49d2] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 775.397964] env[63515]: DEBUG oslo_concurrency.lockutils [None req-f8bace5a-3f38-4490-8a6d-e315321b7f63 tempest-ServersTestManualDisk-1114050206 tempest-ServersTestManualDisk-1114050206-project-member] Acquiring lock "e91aa479-1540-4950-851b-b2409e5f89f1" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 775.398270] env[63515]: DEBUG oslo_concurrency.lockutils [None req-f8bace5a-3f38-4490-8a6d-e315321b7f63 tempest-ServersTestManualDisk-1114050206 tempest-ServersTestManualDisk-1114050206-project-member] Lock "e91aa479-1540-4950-851b-b2409e5f89f1" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 775.398509] env[63515]: DEBUG oslo_concurrency.lockutils [None req-f8bace5a-3f38-4490-8a6d-e315321b7f63 tempest-ServersTestManualDisk-1114050206 tempest-ServersTestManualDisk-1114050206-project-member] Acquiring lock "e91aa479-1540-4950-851b-b2409e5f89f1-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 775.398704] env[63515]: DEBUG oslo_concurrency.lockutils [None req-f8bace5a-3f38-4490-8a6d-e315321b7f63 tempest-ServersTestManualDisk-1114050206 tempest-ServersTestManualDisk-1114050206-project-member] Lock "e91aa479-1540-4950-851b-b2409e5f89f1-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 775.398885] env[63515]: DEBUG oslo_concurrency.lockutils [None req-f8bace5a-3f38-4490-8a6d-e315321b7f63 tempest-ServersTestManualDisk-1114050206 tempest-ServersTestManualDisk-1114050206-project-member] Lock "e91aa479-1540-4950-851b-b2409e5f89f1-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 775.401181] env[63515]: INFO nova.compute.manager [None req-f8bace5a-3f38-4490-8a6d-e315321b7f63 tempest-ServersTestManualDisk-1114050206 tempest-ServersTestManualDisk-1114050206-project-member] [instance: e91aa479-1540-4950-851b-b2409e5f89f1] Terminating instance [ 775.403140] env[63515]: DEBUG nova.compute.manager [None req-f8bace5a-3f38-4490-8a6d-e315321b7f63 tempest-ServersTestManualDisk-1114050206 tempest-ServersTestManualDisk-1114050206-project-member] [instance: e91aa479-1540-4950-851b-b2409e5f89f1] Start destroying the instance on the hypervisor. {{(pid=63515) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 775.403353] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-f8bace5a-3f38-4490-8a6d-e315321b7f63 tempest-ServersTestManualDisk-1114050206 tempest-ServersTestManualDisk-1114050206-project-member] [instance: e91aa479-1540-4950-851b-b2409e5f89f1] Destroying instance {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 775.404216] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c7ecdce-f6d5-478e-afc3-b5acd7beac0b {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.414519] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-f8bace5a-3f38-4490-8a6d-e315321b7f63 tempest-ServersTestManualDisk-1114050206 tempest-ServersTestManualDisk-1114050206-project-member] [instance: e91aa479-1540-4950-851b-b2409e5f89f1] Powering off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 775.414814] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-fc8c5b24-9aae-4b3c-a16c-9c2507dcc564 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.424583] env[63515]: DEBUG oslo_vmware.api [None req-f8bace5a-3f38-4490-8a6d-e315321b7f63 tempest-ServersTestManualDisk-1114050206 tempest-ServersTestManualDisk-1114050206-project-member] Waiting for the task: (returnval){ [ 775.424583] env[63515]: value = "task-1111185" [ 775.424583] env[63515]: _type = "Task" [ 775.424583] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 775.436071] env[63515]: DEBUG oslo_vmware.api [None req-f8bace5a-3f38-4490-8a6d-e315321b7f63 tempest-ServersTestManualDisk-1114050206 tempest-ServersTestManualDisk-1114050206-project-member] Task: {'id': task-1111185, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 775.546559] env[63515]: DEBUG oslo_concurrency.lockutils [None req-67d44811-1ff4-4a5b-a720-2ef182e4f0da tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 775.558100] env[63515]: DEBUG nova.compute.manager [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] [instance: 3ad04a9f-62cf-4b1c-aab2-7dc4a219edc6] Instance has had 0 of 5 cleanup attempts {{(pid=63515) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 775.559038] env[63515]: DEBUG oslo_vmware.api [None req-549cc3f8-330f-4c6a-979d-35eedc6f8146 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Task: {'id': task-1111183, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.472375} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 775.559388] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-549cc3f8-330f-4c6a-979d-35eedc6f8146 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk to [datastore2] 84dbf321-2c70-4e08-b430-cb5a06fc6829/84dbf321-2c70-4e08-b430-cb5a06fc6829.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 775.559658] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-549cc3f8-330f-4c6a-979d-35eedc6f8146 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: 84dbf321-2c70-4e08-b430-cb5a06fc6829] Extending root virtual disk to 1048576 {{(pid=63515) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 775.560211] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-cf2a09ac-f421-4879-a2d6-983deae2facd {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.567770] env[63515]: DEBUG oslo_vmware.api [None req-549cc3f8-330f-4c6a-979d-35eedc6f8146 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Waiting for the task: (returnval){ [ 775.567770] env[63515]: value = "task-1111186" [ 775.567770] env[63515]: _type = "Task" [ 775.567770] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 775.571520] env[63515]: DEBUG nova.compute.manager [None req-8407c401-88c7-47ad-b91e-e7693f87e0f1 tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] [instance: 192137b0-03e5-4bc4-b911-4b4f1a874f74] Start building block device mappings for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 775.589199] env[63515]: DEBUG oslo_vmware.api [None req-549cc3f8-330f-4c6a-979d-35eedc6f8146 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Task: {'id': task-1111186, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 775.629305] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-78f15907-fa94-44dd-a2b3-ccdc7dec0a0f tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] [instance: 87c468d9-9594-4804-b461-527f01f6118f] Destroying instance {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 775.630277] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5dc6c6f5-f3d1-455d-ab74-8577abe21d39 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.639235] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-78f15907-fa94-44dd-a2b3-ccdc7dec0a0f tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] [instance: 87c468d9-9594-4804-b461-527f01f6118f] Unregistering the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 775.639512] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-795c3982-8346-4bce-8649-db1f7fe27d0c {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.689264] env[63515]: DEBUG nova.network.neutron [req-1dafc83d-d43c-4ab3-8a9f-9b6108b30720 req-7242ceb2-9072-4d32-87a5-91f79482ac88 service nova] [instance: 30b8b872-40ca-4297-b98e-a64c83a75483] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 775.719974] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-78f15907-fa94-44dd-a2b3-ccdc7dec0a0f tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] [instance: 87c468d9-9594-4804-b461-527f01f6118f] Unregistered the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 775.720283] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-78f15907-fa94-44dd-a2b3-ccdc7dec0a0f tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] [instance: 87c468d9-9594-4804-b461-527f01f6118f] Deleting contents of the VM from datastore datastore1 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 775.720766] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-78f15907-fa94-44dd-a2b3-ccdc7dec0a0f tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Deleting the datastore file [datastore1] 87c468d9-9594-4804-b461-527f01f6118f {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 775.722182] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-6d870ccb-a89b-4d32-8462-40af664c952d {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.725103] env[63515]: DEBUG nova.compute.manager [req-a7138479-124f-4a67-bac5-b2c595fe2da1 req-e14b77f3-8fdd-4262-a55a-1713edeca627 service nova] [instance: 30b8b872-40ca-4297-b98e-a64c83a75483] Received event network-vif-deleted-b8fe94ee-46b7-489a-819d-0699bc02bbb7 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 775.728574] env[63515]: DEBUG oslo_concurrency.lockutils [None req-9044ca9f-2eb8-445e-a5e5-e15900a98ac3 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 775.734427] env[63515]: DEBUG oslo_vmware.api [None req-78f15907-fa94-44dd-a2b3-ccdc7dec0a0f tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Waiting for the task: (returnval){ [ 775.734427] env[63515]: value = "task-1111188" [ 775.734427] env[63515]: _type = "Task" [ 775.734427] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 775.751082] env[63515]: DEBUG oslo_vmware.api [None req-78f15907-fa94-44dd-a2b3-ccdc7dec0a0f tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Task: {'id': task-1111188, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 775.756366] env[63515]: DEBUG oslo_vmware.api [None req-ca1a9b4c-30ff-4beb-9aec-3fa3a8a9ff12 tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Task: {'id': task-1111184, 'name': PowerOffVM_Task, 'duration_secs': 0.328088} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 775.758954] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-ca1a9b4c-30ff-4beb-9aec-3fa3a8a9ff12 tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] [instance: b4477e66-ae12-4929-90ed-b7b652e0f207] Powered off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 775.759152] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-ca1a9b4c-30ff-4beb-9aec-3fa3a8a9ff12 tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] [instance: b4477e66-ae12-4929-90ed-b7b652e0f207] Unregistering the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 775.759605] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-387e6999-2de8-4485-9691-fa52ad24fc7f {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.789567] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-ca1a9b4c-30ff-4beb-9aec-3fa3a8a9ff12 tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] [instance: b4477e66-ae12-4929-90ed-b7b652e0f207] Unregistered the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 775.789825] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-ca1a9b4c-30ff-4beb-9aec-3fa3a8a9ff12 tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] [instance: b4477e66-ae12-4929-90ed-b7b652e0f207] Deleting contents of the VM from datastore datastore2 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 775.790019] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-ca1a9b4c-30ff-4beb-9aec-3fa3a8a9ff12 tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Deleting the datastore file [datastore2] b4477e66-ae12-4929-90ed-b7b652e0f207 {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 775.790802] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-600f97eb-de6a-4ede-a0cb-4e43b1d3d534 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.801517] env[63515]: DEBUG oslo_vmware.api [None req-ca1a9b4c-30ff-4beb-9aec-3fa3a8a9ff12 tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Waiting for the task: (returnval){ [ 775.801517] env[63515]: value = "task-1111190" [ 775.801517] env[63515]: _type = "Task" [ 775.801517] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 775.813392] env[63515]: DEBUG oslo_vmware.api [None req-ca1a9b4c-30ff-4beb-9aec-3fa3a8a9ff12 tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Task: {'id': task-1111190, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 775.820971] env[63515]: DEBUG nova.network.neutron [req-1dafc83d-d43c-4ab3-8a9f-9b6108b30720 req-7242ceb2-9072-4d32-87a5-91f79482ac88 service nova] [instance: 30b8b872-40ca-4297-b98e-a64c83a75483] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 775.822887] env[63515]: DEBUG nova.network.neutron [None req-2786351a-303c-45a8-b63b-65e218eb6ef6 tempest-ServersTestJSON-324454119 tempest-ServersTestJSON-324454119-project-member] [instance: 70ed982f-affd-4dd1-bc90-c64e7c6d49d2] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 775.935853] env[63515]: DEBUG oslo_vmware.api [None req-f8bace5a-3f38-4490-8a6d-e315321b7f63 tempest-ServersTestManualDisk-1114050206 tempest-ServersTestManualDisk-1114050206-project-member] Task: {'id': task-1111185, 'name': PowerOffVM_Task, 'duration_secs': 0.215464} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 775.936187] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-f8bace5a-3f38-4490-8a6d-e315321b7f63 tempest-ServersTestManualDisk-1114050206 tempest-ServersTestManualDisk-1114050206-project-member] [instance: e91aa479-1540-4950-851b-b2409e5f89f1] Powered off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 775.936369] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-f8bace5a-3f38-4490-8a6d-e315321b7f63 tempest-ServersTestManualDisk-1114050206 tempest-ServersTestManualDisk-1114050206-project-member] [instance: e91aa479-1540-4950-851b-b2409e5f89f1] Unregistering the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 775.936636] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-0693d4bc-1cf4-42ea-b029-015ed25bd16f {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.006918] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-f8bace5a-3f38-4490-8a6d-e315321b7f63 tempest-ServersTestManualDisk-1114050206 tempest-ServersTestManualDisk-1114050206-project-member] [instance: e91aa479-1540-4950-851b-b2409e5f89f1] Unregistered the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 776.007193] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-f8bace5a-3f38-4490-8a6d-e315321b7f63 tempest-ServersTestManualDisk-1114050206 tempest-ServersTestManualDisk-1114050206-project-member] [instance: e91aa479-1540-4950-851b-b2409e5f89f1] Deleting contents of the VM from datastore datastore1 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 776.007413] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-f8bace5a-3f38-4490-8a6d-e315321b7f63 tempest-ServersTestManualDisk-1114050206 tempest-ServersTestManualDisk-1114050206-project-member] Deleting the datastore file [datastore1] e91aa479-1540-4950-851b-b2409e5f89f1 {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 776.007691] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-40c1e89a-b2ec-4e6c-be78-6b3039cbce43 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.016381] env[63515]: DEBUG oslo_vmware.api [None req-f8bace5a-3f38-4490-8a6d-e315321b7f63 tempest-ServersTestManualDisk-1114050206 tempest-ServersTestManualDisk-1114050206-project-member] Waiting for the task: (returnval){ [ 776.016381] env[63515]: value = "task-1111192" [ 776.016381] env[63515]: _type = "Task" [ 776.016381] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 776.029128] env[63515]: DEBUG oslo_vmware.api [None req-f8bace5a-3f38-4490-8a6d-e315321b7f63 tempest-ServersTestManualDisk-1114050206 tempest-ServersTestManualDisk-1114050206-project-member] Task: {'id': task-1111192, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 776.031354] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c20d0cb6-1bae-421a-a4f0-7f4249aa0514 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.038931] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4962d471-104a-4b51-8ca0-0fac82467855 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.074603] env[63515]: DEBUG nova.compute.manager [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] [instance: 4ad5440f-cd18-44c5-8836-0aa39824cf03] Instance has had 0 of 5 cleanup attempts {{(pid=63515) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 776.077930] env[63515]: DEBUG nova.network.neutron [None req-2786351a-303c-45a8-b63b-65e218eb6ef6 tempest-ServersTestJSON-324454119 tempest-ServersTestJSON-324454119-project-member] [instance: 70ed982f-affd-4dd1-bc90-c64e7c6d49d2] Updating instance_info_cache with network_info: [{"id": "fe7c09da-d6bd-4f72-afa2-1a70bc663407", "address": "fa:16:3e:06:ad:d9", "network": {"id": "61e1e43b-1c9a-4769-8de9-6fce1435e861", "bridge": "br-int", "label": "tempest-ServersTestJSON-481843141-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f8662644a3354b46a19edbd1f8ce2c28", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4105fae7-3de1-4a6c-a44b-39097c174699", "external-id": "nsx-vlan-transportzone-773", "segmentation_id": 773, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfe7c09da-d6", "ovs_interfaceid": "fe7c09da-d6bd-4f72-afa2-1a70bc663407", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 776.087929] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-696da7c1-b905-42b7-97a6-3a760b40c8ff {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.099925] env[63515]: DEBUG oslo_vmware.api [None req-549cc3f8-330f-4c6a-979d-35eedc6f8146 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Task: {'id': task-1111186, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.065754} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 776.103187] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-549cc3f8-330f-4c6a-979d-35eedc6f8146 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: 84dbf321-2c70-4e08-b430-cb5a06fc6829] Extended root virtual disk {{(pid=63515) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 776.105178] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3788cb3e-163d-4a62-b1bc-96f18c78f233 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.110010] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81fbd14d-ff76-4564-837e-fc13b0afa08f {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.128187] env[63515]: DEBUG nova.compute.provider_tree [None req-ffbb9c80-abfa-457c-b756-9497c4635a40 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 776.148234] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-549cc3f8-330f-4c6a-979d-35eedc6f8146 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: 84dbf321-2c70-4e08-b430-cb5a06fc6829] Reconfiguring VM instance instance-00000037 to attach disk [datastore2] 84dbf321-2c70-4e08-b430-cb5a06fc6829/84dbf321-2c70-4e08-b430-cb5a06fc6829.vmdk or device None with type sparse {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 776.150045] env[63515]: DEBUG nova.scheduler.client.report [None req-ffbb9c80-abfa-457c-b756-9497c4635a40 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 776.152750] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-321cc752-449b-4c72-ac3c-eec2e1c0dbe4 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.175240] env[63515]: DEBUG oslo_vmware.api [None req-549cc3f8-330f-4c6a-979d-35eedc6f8146 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Waiting for the task: (returnval){ [ 776.175240] env[63515]: value = "task-1111193" [ 776.175240] env[63515]: _type = "Task" [ 776.175240] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 776.184124] env[63515]: DEBUG oslo_vmware.api [None req-549cc3f8-330f-4c6a-979d-35eedc6f8146 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Task: {'id': task-1111193, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 776.248018] env[63515]: DEBUG oslo_vmware.api [None req-78f15907-fa94-44dd-a2b3-ccdc7dec0a0f tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Task: {'id': task-1111188, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.270053} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 776.248018] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-78f15907-fa94-44dd-a2b3-ccdc7dec0a0f tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Deleted the datastore file {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 776.248018] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-78f15907-fa94-44dd-a2b3-ccdc7dec0a0f tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] [instance: 87c468d9-9594-4804-b461-527f01f6118f] Deleted contents of the VM from datastore datastore1 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 776.248018] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-78f15907-fa94-44dd-a2b3-ccdc7dec0a0f tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] [instance: 87c468d9-9594-4804-b461-527f01f6118f] Instance destroyed {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 776.269543] env[63515]: INFO nova.scheduler.client.report [None req-78f15907-fa94-44dd-a2b3-ccdc7dec0a0f tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Deleted allocations for instance 87c468d9-9594-4804-b461-527f01f6118f [ 776.315512] env[63515]: DEBUG oslo_vmware.api [None req-ca1a9b4c-30ff-4beb-9aec-3fa3a8a9ff12 tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Task: {'id': task-1111190, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.131038} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 776.315810] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-ca1a9b4c-30ff-4beb-9aec-3fa3a8a9ff12 tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Deleted the datastore file {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 776.316015] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-ca1a9b4c-30ff-4beb-9aec-3fa3a8a9ff12 tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] [instance: b4477e66-ae12-4929-90ed-b7b652e0f207] Deleted contents of the VM from datastore datastore2 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 776.316262] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-ca1a9b4c-30ff-4beb-9aec-3fa3a8a9ff12 tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] [instance: b4477e66-ae12-4929-90ed-b7b652e0f207] Instance destroyed {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 776.316463] env[63515]: INFO nova.compute.manager [None req-ca1a9b4c-30ff-4beb-9aec-3fa3a8a9ff12 tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] [instance: b4477e66-ae12-4929-90ed-b7b652e0f207] Took 1.10 seconds to destroy the instance on the hypervisor. [ 776.316716] env[63515]: DEBUG oslo.service.loopingcall [None req-ca1a9b4c-30ff-4beb-9aec-3fa3a8a9ff12 tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 776.316913] env[63515]: DEBUG nova.compute.manager [-] [instance: b4477e66-ae12-4929-90ed-b7b652e0f207] Deallocating network for instance {{(pid=63515) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 776.317016] env[63515]: DEBUG nova.network.neutron [-] [instance: b4477e66-ae12-4929-90ed-b7b652e0f207] deallocate_for_instance() {{(pid=63515) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 776.328797] env[63515]: DEBUG oslo_concurrency.lockutils [req-1dafc83d-d43c-4ab3-8a9f-9b6108b30720 req-7242ceb2-9072-4d32-87a5-91f79482ac88 service nova] Releasing lock "refresh_cache-30b8b872-40ca-4297-b98e-a64c83a75483" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 776.329064] env[63515]: DEBUG nova.compute.manager [req-1dafc83d-d43c-4ab3-8a9f-9b6108b30720 req-7242ceb2-9072-4d32-87a5-91f79482ac88 service nova] [instance: 38d6d030-06b0-4185-904d-44a038b3a752] Received event network-vif-deleted-2242e364-37c0-4750-8755-d807d47e899d {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 776.338153] env[63515]: DEBUG nova.network.neutron [-] [instance: b4477e66-ae12-4929-90ed-b7b652e0f207] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 776.528815] env[63515]: DEBUG oslo_vmware.api [None req-f8bace5a-3f38-4490-8a6d-e315321b7f63 tempest-ServersTestManualDisk-1114050206 tempest-ServersTestManualDisk-1114050206-project-member] Task: {'id': task-1111192, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.379725} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 776.530310] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-f8bace5a-3f38-4490-8a6d-e315321b7f63 tempest-ServersTestManualDisk-1114050206 tempest-ServersTestManualDisk-1114050206-project-member] Deleted the datastore file {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 776.530310] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-f8bace5a-3f38-4490-8a6d-e315321b7f63 tempest-ServersTestManualDisk-1114050206 tempest-ServersTestManualDisk-1114050206-project-member] [instance: e91aa479-1540-4950-851b-b2409e5f89f1] Deleted contents of the VM from datastore datastore1 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 776.530310] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-f8bace5a-3f38-4490-8a6d-e315321b7f63 tempest-ServersTestManualDisk-1114050206 tempest-ServersTestManualDisk-1114050206-project-member] [instance: e91aa479-1540-4950-851b-b2409e5f89f1] Instance destroyed {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 776.530310] env[63515]: INFO nova.compute.manager [None req-f8bace5a-3f38-4490-8a6d-e315321b7f63 tempest-ServersTestManualDisk-1114050206 tempest-ServersTestManualDisk-1114050206-project-member] [instance: e91aa479-1540-4950-851b-b2409e5f89f1] Took 1.13 seconds to destroy the instance on the hypervisor. [ 776.530310] env[63515]: DEBUG oslo.service.loopingcall [None req-f8bace5a-3f38-4490-8a6d-e315321b7f63 tempest-ServersTestManualDisk-1114050206 tempest-ServersTestManualDisk-1114050206-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 776.530679] env[63515]: DEBUG nova.compute.manager [-] [instance: e91aa479-1540-4950-851b-b2409e5f89f1] Deallocating network for instance {{(pid=63515) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 776.530679] env[63515]: DEBUG nova.network.neutron [-] [instance: e91aa479-1540-4950-851b-b2409e5f89f1] deallocate_for_instance() {{(pid=63515) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 776.577764] env[63515]: DEBUG nova.compute.manager [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] [instance: 93b28142-8454-43c0-b0a7-d61aa95c8fc8] Instance has had 0 of 5 cleanup attempts {{(pid=63515) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 776.595291] env[63515]: DEBUG oslo_concurrency.lockutils [None req-2786351a-303c-45a8-b63b-65e218eb6ef6 tempest-ServersTestJSON-324454119 tempest-ServersTestJSON-324454119-project-member] Releasing lock "refresh_cache-70ed982f-affd-4dd1-bc90-c64e7c6d49d2" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 776.595875] env[63515]: DEBUG nova.compute.manager [None req-2786351a-303c-45a8-b63b-65e218eb6ef6 tempest-ServersTestJSON-324454119 tempest-ServersTestJSON-324454119-project-member] [instance: 70ed982f-affd-4dd1-bc90-c64e7c6d49d2] Instance network_info: |[{"id": "fe7c09da-d6bd-4f72-afa2-1a70bc663407", "address": "fa:16:3e:06:ad:d9", "network": {"id": "61e1e43b-1c9a-4769-8de9-6fce1435e861", "bridge": "br-int", "label": "tempest-ServersTestJSON-481843141-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f8662644a3354b46a19edbd1f8ce2c28", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4105fae7-3de1-4a6c-a44b-39097c174699", "external-id": "nsx-vlan-transportzone-773", "segmentation_id": 773, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfe7c09da-d6", "ovs_interfaceid": "fe7c09da-d6bd-4f72-afa2-1a70bc663407", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 776.598022] env[63515]: DEBUG nova.compute.manager [None req-8407c401-88c7-47ad-b91e-e7693f87e0f1 tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] [instance: 192137b0-03e5-4bc4-b911-4b4f1a874f74] Start spawning the instance on the hypervisor. {{(pid=63515) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 776.599863] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-2786351a-303c-45a8-b63b-65e218eb6ef6 tempest-ServersTestJSON-324454119 tempest-ServersTestJSON-324454119-project-member] [instance: 70ed982f-affd-4dd1-bc90-c64e7c6d49d2] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:06:ad:d9', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '4105fae7-3de1-4a6c-a44b-39097c174699', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'fe7c09da-d6bd-4f72-afa2-1a70bc663407', 'vif_model': 'vmxnet3'}] {{(pid=63515) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 776.609854] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-2786351a-303c-45a8-b63b-65e218eb6ef6 tempest-ServersTestJSON-324454119 tempest-ServersTestJSON-324454119-project-member] Creating folder: Project (f8662644a3354b46a19edbd1f8ce2c28). Parent ref: group-v243370. {{(pid=63515) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 776.611719] env[63515]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a5c673c8-2066-4296-9292-82724205bff6 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.624384] env[63515]: INFO nova.virt.vmwareapi.vm_util [None req-2786351a-303c-45a8-b63b-65e218eb6ef6 tempest-ServersTestJSON-324454119 tempest-ServersTestJSON-324454119-project-member] Created folder: Project (f8662644a3354b46a19edbd1f8ce2c28) in parent group-v243370. [ 776.624384] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-2786351a-303c-45a8-b63b-65e218eb6ef6 tempest-ServersTestJSON-324454119 tempest-ServersTestJSON-324454119-project-member] Creating folder: Instances. Parent ref: group-v243454. {{(pid=63515) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 776.624384] env[63515]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-865311f4-9346-441e-8ba2-62422277120b {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.632857] env[63515]: INFO nova.virt.vmwareapi.vm_util [None req-2786351a-303c-45a8-b63b-65e218eb6ef6 tempest-ServersTestJSON-324454119 tempest-ServersTestJSON-324454119-project-member] Created folder: Instances in parent group-v243454. [ 776.633182] env[63515]: DEBUG oslo.service.loopingcall [None req-2786351a-303c-45a8-b63b-65e218eb6ef6 tempest-ServersTestJSON-324454119 tempest-ServersTestJSON-324454119-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 776.635298] env[63515]: DEBUG nova.virt.hardware [None req-8407c401-88c7-47ad-b91e-e7693f87e0f1 tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T02:52:14Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T02:51:56Z,direct_url=,disk_format='vmdk',id=8a120570-cb06-4099-b262-554ca0ad15c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b569255cc43e42e7ae3f2b4ad37c6ef4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T02:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 776.635533] env[63515]: DEBUG nova.virt.hardware [None req-8407c401-88c7-47ad-b91e-e7693f87e0f1 tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] Flavor limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 776.635723] env[63515]: DEBUG nova.virt.hardware [None req-8407c401-88c7-47ad-b91e-e7693f87e0f1 tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] Image limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 776.635879] env[63515]: DEBUG nova.virt.hardware [None req-8407c401-88c7-47ad-b91e-e7693f87e0f1 tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] Flavor pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 776.636036] env[63515]: DEBUG nova.virt.hardware [None req-8407c401-88c7-47ad-b91e-e7693f87e0f1 tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] Image pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 776.636216] env[63515]: DEBUG nova.virt.hardware [None req-8407c401-88c7-47ad-b91e-e7693f87e0f1 tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 776.636508] env[63515]: DEBUG nova.virt.hardware [None req-8407c401-88c7-47ad-b91e-e7693f87e0f1 tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 776.636705] env[63515]: DEBUG nova.virt.hardware [None req-8407c401-88c7-47ad-b91e-e7693f87e0f1 tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 776.636883] env[63515]: DEBUG nova.virt.hardware [None req-8407c401-88c7-47ad-b91e-e7693f87e0f1 tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] Got 1 possible topologies {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 776.637062] env[63515]: DEBUG nova.virt.hardware [None req-8407c401-88c7-47ad-b91e-e7693f87e0f1 tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 776.637282] env[63515]: DEBUG nova.virt.hardware [None req-8407c401-88c7-47ad-b91e-e7693f87e0f1 tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 776.637504] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 70ed982f-affd-4dd1-bc90-c64e7c6d49d2] Creating VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 776.638236] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3d8d97e-0f46-4ae9-8e24-f752560f0d61 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.641078] env[63515]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-746e046b-3c7f-4f28-bfc9-48465fcf914a {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.663898] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6a85a49-123c-46b5-8d55-78eb99c38980 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.667767] env[63515]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 776.667767] env[63515]: value = "task-1111197" [ 776.667767] env[63515]: _type = "Task" [ 776.667767] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 776.686378] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-8407c401-88c7-47ad-b91e-e7693f87e0f1 tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] [instance: 192137b0-03e5-4bc4-b911-4b4f1a874f74] Instance VIF info [] {{(pid=63515) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 776.692067] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-8407c401-88c7-47ad-b91e-e7693f87e0f1 tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] Creating folder: Project (c9351fd59ec14074ac54920a537862a1). Parent ref: group-v243370. {{(pid=63515) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 776.696377] env[63515]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d4f361a9-9283-4e90-8a3c-b3fcc2fa33e2 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.701950] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1111197, 'name': CreateVM_Task} progress is 15%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 776.706039] env[63515]: DEBUG oslo_vmware.api [None req-549cc3f8-330f-4c6a-979d-35eedc6f8146 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Task: {'id': task-1111193, 'name': ReconfigVM_Task, 'duration_secs': 0.297444} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 776.706347] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-549cc3f8-330f-4c6a-979d-35eedc6f8146 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: 84dbf321-2c70-4e08-b430-cb5a06fc6829] Reconfigured VM instance instance-00000037 to attach disk [datastore2] 84dbf321-2c70-4e08-b430-cb5a06fc6829/84dbf321-2c70-4e08-b430-cb5a06fc6829.vmdk or device None with type sparse {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 776.706968] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-16b20710-3685-4bbf-be8f-baaaf1468981 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.710164] env[63515]: INFO nova.virt.vmwareapi.vm_util [None req-8407c401-88c7-47ad-b91e-e7693f87e0f1 tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] Created folder: Project (c9351fd59ec14074ac54920a537862a1) in parent group-v243370. [ 776.710396] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-8407c401-88c7-47ad-b91e-e7693f87e0f1 tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] Creating folder: Instances. Parent ref: group-v243457. {{(pid=63515) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 776.710995] env[63515]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-bfb21832-6609-420e-b6bd-5fca44075060 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.715152] env[63515]: DEBUG oslo_vmware.api [None req-549cc3f8-330f-4c6a-979d-35eedc6f8146 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Waiting for the task: (returnval){ [ 776.715152] env[63515]: value = "task-1111199" [ 776.715152] env[63515]: _type = "Task" [ 776.715152] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 776.719834] env[63515]: INFO nova.virt.vmwareapi.vm_util [None req-8407c401-88c7-47ad-b91e-e7693f87e0f1 tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] Created folder: Instances in parent group-v243457. [ 776.720123] env[63515]: DEBUG oslo.service.loopingcall [None req-8407c401-88c7-47ad-b91e-e7693f87e0f1 tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 776.720665] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 192137b0-03e5-4bc4-b911-4b4f1a874f74] Creating VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 776.720882] env[63515]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-104d88e1-9fde-485f-a8b0-37bd5c41b787 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.736632] env[63515]: DEBUG oslo_vmware.api [None req-549cc3f8-330f-4c6a-979d-35eedc6f8146 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Task: {'id': task-1111199, 'name': Rename_Task} progress is 6%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 776.740999] env[63515]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 776.740999] env[63515]: value = "task-1111201" [ 776.740999] env[63515]: _type = "Task" [ 776.740999] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 776.748997] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1111201, 'name': CreateVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 776.773467] env[63515]: DEBUG oslo_concurrency.lockutils [None req-78f15907-fa94-44dd-a2b3-ccdc7dec0a0f tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 776.827196] env[63515]: DEBUG nova.compute.manager [req-2c5b2027-abb2-44c4-bd99-2ad564ff7b2f req-43fa149b-9f98-47b6-893a-6613a60250f1 service nova] [instance: 70ed982f-affd-4dd1-bc90-c64e7c6d49d2] Received event network-changed-fe7c09da-d6bd-4f72-afa2-1a70bc663407 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 776.827682] env[63515]: DEBUG nova.compute.manager [req-2c5b2027-abb2-44c4-bd99-2ad564ff7b2f req-43fa149b-9f98-47b6-893a-6613a60250f1 service nova] [instance: 70ed982f-affd-4dd1-bc90-c64e7c6d49d2] Refreshing instance network info cache due to event network-changed-fe7c09da-d6bd-4f72-afa2-1a70bc663407. {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 776.828010] env[63515]: DEBUG oslo_concurrency.lockutils [req-2c5b2027-abb2-44c4-bd99-2ad564ff7b2f req-43fa149b-9f98-47b6-893a-6613a60250f1 service nova] Acquiring lock "refresh_cache-70ed982f-affd-4dd1-bc90-c64e7c6d49d2" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 776.828181] env[63515]: DEBUG oslo_concurrency.lockutils [req-2c5b2027-abb2-44c4-bd99-2ad564ff7b2f req-43fa149b-9f98-47b6-893a-6613a60250f1 service nova] Acquired lock "refresh_cache-70ed982f-affd-4dd1-bc90-c64e7c6d49d2" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 776.828358] env[63515]: DEBUG nova.network.neutron [req-2c5b2027-abb2-44c4-bd99-2ad564ff7b2f req-43fa149b-9f98-47b6-893a-6613a60250f1 service nova] [instance: 70ed982f-affd-4dd1-bc90-c64e7c6d49d2] Refreshing network info cache for port fe7c09da-d6bd-4f72-afa2-1a70bc663407 {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 776.840534] env[63515]: DEBUG nova.network.neutron [-] [instance: b4477e66-ae12-4929-90ed-b7b652e0f207] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 777.083324] env[63515]: DEBUG oslo_service.periodic_task [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Running periodic task ComputeManager._cleanup_incomplete_migrations {{(pid=63515) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 777.083512] env[63515]: DEBUG nova.compute.manager [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Cleaning up deleted instances with incomplete migration {{(pid=63515) _cleanup_incomplete_migrations /opt/stack/nova/nova/compute/manager.py:11264}} [ 777.180050] env[63515]: DEBUG oslo_concurrency.lockutils [None req-ffbb9c80-abfa-457c-b756-9497c4635a40 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: held 2.617s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 777.189556] env[63515]: DEBUG oslo_concurrency.lockutils [None req-83fbd51b-fc37-49eb-8423-cdaa019d3b0c tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 27.577s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 777.191136] env[63515]: INFO nova.compute.claims [None req-83fbd51b-fc37-49eb-8423-cdaa019d3b0c tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] [instance: 5cbce760-0163-4b27-8ae3-e46c926c8916] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 777.194262] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1111197, 'name': CreateVM_Task, 'duration_secs': 0.413056} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 777.196138] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 70ed982f-affd-4dd1-bc90-c64e7c6d49d2] Created VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 777.199815] env[63515]: DEBUG oslo_concurrency.lockutils [None req-2786351a-303c-45a8-b63b-65e218eb6ef6 tempest-ServersTestJSON-324454119 tempest-ServersTestJSON-324454119-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 777.200058] env[63515]: DEBUG oslo_concurrency.lockutils [None req-2786351a-303c-45a8-b63b-65e218eb6ef6 tempest-ServersTestJSON-324454119 tempest-ServersTestJSON-324454119-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 777.201818] env[63515]: DEBUG oslo_concurrency.lockutils [None req-2786351a-303c-45a8-b63b-65e218eb6ef6 tempest-ServersTestJSON-324454119 tempest-ServersTestJSON-324454119-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 777.201818] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-848afd21-72e8-424a-893f-682a81a59c42 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.210591] env[63515]: DEBUG oslo_vmware.api [None req-2786351a-303c-45a8-b63b-65e218eb6ef6 tempest-ServersTestJSON-324454119 tempest-ServersTestJSON-324454119-project-member] Waiting for the task: (returnval){ [ 777.210591] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]525fcb26-2c11-d580-b963-dec93ed9d62f" [ 777.210591] env[63515]: _type = "Task" [ 777.210591] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 777.231065] env[63515]: DEBUG oslo_vmware.api [None req-2786351a-303c-45a8-b63b-65e218eb6ef6 tempest-ServersTestJSON-324454119 tempest-ServersTestJSON-324454119-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]525fcb26-2c11-d580-b963-dec93ed9d62f, 'name': SearchDatastore_Task, 'duration_secs': 0.008759} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 777.231547] env[63515]: DEBUG oslo_concurrency.lockutils [None req-2786351a-303c-45a8-b63b-65e218eb6ef6 tempest-ServersTestJSON-324454119 tempest-ServersTestJSON-324454119-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 777.231905] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-2786351a-303c-45a8-b63b-65e218eb6ef6 tempest-ServersTestJSON-324454119 tempest-ServersTestJSON-324454119-project-member] [instance: 70ed982f-affd-4dd1-bc90-c64e7c6d49d2] Processing image 8a120570-cb06-4099-b262-554ca0ad15c5 {{(pid=63515) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 777.232525] env[63515]: DEBUG oslo_concurrency.lockutils [None req-2786351a-303c-45a8-b63b-65e218eb6ef6 tempest-ServersTestJSON-324454119 tempest-ServersTestJSON-324454119-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 777.232959] env[63515]: DEBUG oslo_concurrency.lockutils [None req-2786351a-303c-45a8-b63b-65e218eb6ef6 tempest-ServersTestJSON-324454119 tempest-ServersTestJSON-324454119-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 777.233426] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-2786351a-303c-45a8-b63b-65e218eb6ef6 tempest-ServersTestJSON-324454119 tempest-ServersTestJSON-324454119-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 777.238038] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ca14923b-6e48-4aed-9229-33edc362e3cf {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.240051] env[63515]: DEBUG oslo_vmware.api [None req-549cc3f8-330f-4c6a-979d-35eedc6f8146 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Task: {'id': task-1111199, 'name': Rename_Task} progress is 14%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 777.247658] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-2786351a-303c-45a8-b63b-65e218eb6ef6 tempest-ServersTestJSON-324454119 tempest-ServersTestJSON-324454119-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 777.248022] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-2786351a-303c-45a8-b63b-65e218eb6ef6 tempest-ServersTestJSON-324454119 tempest-ServersTestJSON-324454119-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63515) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 777.251997] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d073556b-8544-4f73-9745-c100e1aa75b5 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.254262] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1111201, 'name': CreateVM_Task} progress is 99%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 777.258050] env[63515]: DEBUG oslo_vmware.api [None req-2786351a-303c-45a8-b63b-65e218eb6ef6 tempest-ServersTestJSON-324454119 tempest-ServersTestJSON-324454119-project-member] Waiting for the task: (returnval){ [ 777.258050] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]5232a22e-3a0e-3b5d-0d4c-a43c8e0cb97e" [ 777.258050] env[63515]: _type = "Task" [ 777.258050] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 777.266468] env[63515]: DEBUG oslo_vmware.api [None req-2786351a-303c-45a8-b63b-65e218eb6ef6 tempest-ServersTestJSON-324454119 tempest-ServersTestJSON-324454119-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]5232a22e-3a0e-3b5d-0d4c-a43c8e0cb97e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 777.344032] env[63515]: INFO nova.compute.manager [-] [instance: b4477e66-ae12-4929-90ed-b7b652e0f207] Took 1.03 seconds to deallocate network for instance. [ 777.585896] env[63515]: DEBUG oslo_service.periodic_task [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Running periodic task ComputeManager._cleanup_expired_console_auth_tokens {{(pid=63515) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 777.647360] env[63515]: DEBUG nova.network.neutron [req-2c5b2027-abb2-44c4-bd99-2ad564ff7b2f req-43fa149b-9f98-47b6-893a-6613a60250f1 service nova] [instance: 70ed982f-affd-4dd1-bc90-c64e7c6d49d2] Updated VIF entry in instance network info cache for port fe7c09da-d6bd-4f72-afa2-1a70bc663407. {{(pid=63515) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 777.647723] env[63515]: DEBUG nova.network.neutron [req-2c5b2027-abb2-44c4-bd99-2ad564ff7b2f req-43fa149b-9f98-47b6-893a-6613a60250f1 service nova] [instance: 70ed982f-affd-4dd1-bc90-c64e7c6d49d2] Updating instance_info_cache with network_info: [{"id": "fe7c09da-d6bd-4f72-afa2-1a70bc663407", "address": "fa:16:3e:06:ad:d9", "network": {"id": "61e1e43b-1c9a-4769-8de9-6fce1435e861", "bridge": "br-int", "label": "tempest-ServersTestJSON-481843141-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f8662644a3354b46a19edbd1f8ce2c28", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4105fae7-3de1-4a6c-a44b-39097c174699", "external-id": "nsx-vlan-transportzone-773", "segmentation_id": 773, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfe7c09da-d6", "ovs_interfaceid": "fe7c09da-d6bd-4f72-afa2-1a70bc663407", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 777.727805] env[63515]: DEBUG oslo_vmware.api [None req-549cc3f8-330f-4c6a-979d-35eedc6f8146 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Task: {'id': task-1111199, 'name': Rename_Task} progress is 14%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 777.754026] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1111201, 'name': CreateVM_Task} progress is 99%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 777.768714] env[63515]: DEBUG oslo_vmware.api [None req-2786351a-303c-45a8-b63b-65e218eb6ef6 tempest-ServersTestJSON-324454119 tempest-ServersTestJSON-324454119-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]5232a22e-3a0e-3b5d-0d4c-a43c8e0cb97e, 'name': SearchDatastore_Task, 'duration_secs': 0.007913} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 777.769549] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cca04143-c627-4d01-9491-3d2f5ffc6256 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.775453] env[63515]: DEBUG oslo_vmware.api [None req-2786351a-303c-45a8-b63b-65e218eb6ef6 tempest-ServersTestJSON-324454119 tempest-ServersTestJSON-324454119-project-member] Waiting for the task: (returnval){ [ 777.775453] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]522dc56f-bf75-239a-cd2b-5f5e71c9ca5b" [ 777.775453] env[63515]: _type = "Task" [ 777.775453] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 777.783710] env[63515]: DEBUG oslo_vmware.api [None req-2786351a-303c-45a8-b63b-65e218eb6ef6 tempest-ServersTestJSON-324454119 tempest-ServersTestJSON-324454119-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]522dc56f-bf75-239a-cd2b-5f5e71c9ca5b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 777.854296] env[63515]: DEBUG oslo_concurrency.lockutils [None req-ca1a9b4c-30ff-4beb-9aec-3fa3a8a9ff12 tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 778.031661] env[63515]: DEBUG nova.network.neutron [-] [instance: e91aa479-1540-4950-851b-b2409e5f89f1] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 778.153314] env[63515]: DEBUG oslo_concurrency.lockutils [req-2c5b2027-abb2-44c4-bd99-2ad564ff7b2f req-43fa149b-9f98-47b6-893a-6613a60250f1 service nova] Releasing lock "refresh_cache-70ed982f-affd-4dd1-bc90-c64e7c6d49d2" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 778.153314] env[63515]: DEBUG nova.compute.manager [req-2c5b2027-abb2-44c4-bd99-2ad564ff7b2f req-43fa149b-9f98-47b6-893a-6613a60250f1 service nova] [instance: 87c468d9-9594-4804-b461-527f01f6118f] Received event network-vif-unplugged-d2098868-0475-4b06-bae3-785c608a861e {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 778.153314] env[63515]: DEBUG oslo_concurrency.lockutils [req-2c5b2027-abb2-44c4-bd99-2ad564ff7b2f req-43fa149b-9f98-47b6-893a-6613a60250f1 service nova] Acquiring lock "87c468d9-9594-4804-b461-527f01f6118f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 778.153314] env[63515]: DEBUG oslo_concurrency.lockutils [req-2c5b2027-abb2-44c4-bd99-2ad564ff7b2f req-43fa149b-9f98-47b6-893a-6613a60250f1 service nova] Lock "87c468d9-9594-4804-b461-527f01f6118f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 778.153314] env[63515]: DEBUG oslo_concurrency.lockutils [req-2c5b2027-abb2-44c4-bd99-2ad564ff7b2f req-43fa149b-9f98-47b6-893a-6613a60250f1 service nova] Lock "87c468d9-9594-4804-b461-527f01f6118f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 778.153517] env[63515]: DEBUG nova.compute.manager [req-2c5b2027-abb2-44c4-bd99-2ad564ff7b2f req-43fa149b-9f98-47b6-893a-6613a60250f1 service nova] [instance: 87c468d9-9594-4804-b461-527f01f6118f] No waiting events found dispatching network-vif-unplugged-d2098868-0475-4b06-bae3-785c608a861e {{(pid=63515) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 778.153517] env[63515]: WARNING nova.compute.manager [req-2c5b2027-abb2-44c4-bd99-2ad564ff7b2f req-43fa149b-9f98-47b6-893a-6613a60250f1 service nova] [instance: 87c468d9-9594-4804-b461-527f01f6118f] Received unexpected event network-vif-unplugged-d2098868-0475-4b06-bae3-785c608a861e for instance with vm_state shelved_offloaded and task_state None. [ 778.153517] env[63515]: DEBUG nova.compute.manager [req-2c5b2027-abb2-44c4-bd99-2ad564ff7b2f req-43fa149b-9f98-47b6-893a-6613a60250f1 service nova] [instance: 87c468d9-9594-4804-b461-527f01f6118f] Received event network-changed-d2098868-0475-4b06-bae3-785c608a861e {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 778.153517] env[63515]: DEBUG nova.compute.manager [req-2c5b2027-abb2-44c4-bd99-2ad564ff7b2f req-43fa149b-9f98-47b6-893a-6613a60250f1 service nova] [instance: 87c468d9-9594-4804-b461-527f01f6118f] Refreshing instance network info cache due to event network-changed-d2098868-0475-4b06-bae3-785c608a861e. {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 778.153695] env[63515]: DEBUG oslo_concurrency.lockutils [req-2c5b2027-abb2-44c4-bd99-2ad564ff7b2f req-43fa149b-9f98-47b6-893a-6613a60250f1 service nova] Acquiring lock "refresh_cache-87c468d9-9594-4804-b461-527f01f6118f" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 778.153729] env[63515]: DEBUG oslo_concurrency.lockutils [req-2c5b2027-abb2-44c4-bd99-2ad564ff7b2f req-43fa149b-9f98-47b6-893a-6613a60250f1 service nova] Acquired lock "refresh_cache-87c468d9-9594-4804-b461-527f01f6118f" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 778.156652] env[63515]: DEBUG nova.network.neutron [req-2c5b2027-abb2-44c4-bd99-2ad564ff7b2f req-43fa149b-9f98-47b6-893a-6613a60250f1 service nova] [instance: 87c468d9-9594-4804-b461-527f01f6118f] Refreshing network info cache for port d2098868-0475-4b06-bae3-785c608a861e {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 778.230162] env[63515]: DEBUG oslo_vmware.api [None req-549cc3f8-330f-4c6a-979d-35eedc6f8146 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Task: {'id': task-1111199, 'name': Rename_Task, 'duration_secs': 1.082305} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 778.230499] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-549cc3f8-330f-4c6a-979d-35eedc6f8146 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: 84dbf321-2c70-4e08-b430-cb5a06fc6829] Powering on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 778.230672] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-cba1fe67-fbac-4ce3-bb7d-c455d0577c0d {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.238036] env[63515]: DEBUG oslo_vmware.api [None req-549cc3f8-330f-4c6a-979d-35eedc6f8146 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Waiting for the task: (returnval){ [ 778.238036] env[63515]: value = "task-1111202" [ 778.238036] env[63515]: _type = "Task" [ 778.238036] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 778.263964] env[63515]: DEBUG nova.compute.manager [req-b6751c82-024a-4034-83b1-4c261ebbb9a6 req-69cf7b96-b2d4-4c95-bd3c-0018271c3768 service nova] [instance: e91aa479-1540-4950-851b-b2409e5f89f1] Received event network-vif-deleted-fcab3239-8739-4ac7-9f35-4a121630aa6f {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 778.264639] env[63515]: DEBUG oslo_vmware.api [None req-549cc3f8-330f-4c6a-979d-35eedc6f8146 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Task: {'id': task-1111202, 'name': PowerOnVM_Task} progress is 33%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 778.264830] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1111201, 'name': CreateVM_Task} progress is 99%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 778.287048] env[63515]: DEBUG oslo_vmware.api [None req-2786351a-303c-45a8-b63b-65e218eb6ef6 tempest-ServersTestJSON-324454119 tempest-ServersTestJSON-324454119-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]522dc56f-bf75-239a-cd2b-5f5e71c9ca5b, 'name': SearchDatastore_Task, 'duration_secs': 0.009183} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 778.289612] env[63515]: DEBUG oslo_concurrency.lockutils [None req-2786351a-303c-45a8-b63b-65e218eb6ef6 tempest-ServersTestJSON-324454119 tempest-ServersTestJSON-324454119-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 778.290066] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-2786351a-303c-45a8-b63b-65e218eb6ef6 tempest-ServersTestJSON-324454119 tempest-ServersTestJSON-324454119-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk to [datastore2] 70ed982f-affd-4dd1-bc90-c64e7c6d49d2/70ed982f-affd-4dd1-bc90-c64e7c6d49d2.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 778.290869] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-df777478-089e-4cd1-be1f-2ba7d22ba7c3 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.299309] env[63515]: DEBUG oslo_vmware.api [None req-2786351a-303c-45a8-b63b-65e218eb6ef6 tempest-ServersTestJSON-324454119 tempest-ServersTestJSON-324454119-project-member] Waiting for the task: (returnval){ [ 778.299309] env[63515]: value = "task-1111203" [ 778.299309] env[63515]: _type = "Task" [ 778.299309] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 778.309636] env[63515]: DEBUG oslo_vmware.api [None req-2786351a-303c-45a8-b63b-65e218eb6ef6 tempest-ServersTestJSON-324454119 tempest-ServersTestJSON-324454119-project-member] Task: {'id': task-1111203, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 778.535206] env[63515]: INFO nova.compute.manager [-] [instance: e91aa479-1540-4950-851b-b2409e5f89f1] Took 2.00 seconds to deallocate network for instance. [ 778.734791] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-efd46494-ea54-4600-906e-1e7f1371fb6c tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] [instance: d7e8dcdb-41cb-46fb-8b61-d251e7c2d372] Volume attach. Driver type: vmdk {{(pid=63515) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 778.735102] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-efd46494-ea54-4600-906e-1e7f1371fb6c tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] [instance: d7e8dcdb-41cb-46fb-8b61-d251e7c2d372] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-243453', 'volume_id': 'b1a5444a-1e2f-4443-8931-4a04ec5d9d68', 'name': 'volume-b1a5444a-1e2f-4443-8931-4a04ec5d9d68', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'd7e8dcdb-41cb-46fb-8b61-d251e7c2d372', 'attached_at': '', 'detached_at': '', 'volume_id': 'b1a5444a-1e2f-4443-8931-4a04ec5d9d68', 'serial': 'b1a5444a-1e2f-4443-8931-4a04ec5d9d68'} {{(pid=63515) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 778.736743] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0ff0145-2bed-4ad5-b79f-874561b9d409 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.743943] env[63515]: INFO nova.compute.manager [None req-ffbb9c80-abfa-457c-b756-9497c4635a40 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] [instance: f1d01b75-ac9d-458d-8cc2-ae64cffca4e8] Swapping old allocation on dict_keys(['2bd86232-2b6f-44d5-9057-1a3a6b27185a']) held by migration 802930d6-330b-4516-9996-d196743b8ab8 for instance [ 778.771838] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e412bd86-1180-4262-9dbd-536274cadedd {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.778947] env[63515]: DEBUG oslo_vmware.api [None req-549cc3f8-330f-4c6a-979d-35eedc6f8146 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Task: {'id': task-1111202, 'name': PowerOnVM_Task} progress is 89%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 778.786567] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1111201, 'name': CreateVM_Task} progress is 99%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 778.802826] env[63515]: DEBUG oslo_concurrency.lockutils [None req-8c8e056d-f20f-4ef4-b199-3c0616617ccf tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Acquiring lock "interface-ed550b10-d58f-45b8-b766-198f431c3788-None" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 778.803158] env[63515]: DEBUG oslo_concurrency.lockutils [None req-8c8e056d-f20f-4ef4-b199-3c0616617ccf tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Lock "interface-ed550b10-d58f-45b8-b766-198f431c3788-None" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 778.803513] env[63515]: DEBUG nova.objects.instance [None req-8c8e056d-f20f-4ef4-b199-3c0616617ccf tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Lazy-loading 'flavor' on Instance uuid ed550b10-d58f-45b8-b766-198f431c3788 {{(pid=63515) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 778.812406] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-efd46494-ea54-4600-906e-1e7f1371fb6c tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] [instance: d7e8dcdb-41cb-46fb-8b61-d251e7c2d372] Reconfiguring VM instance instance-0000002c to attach disk [datastore2] volume-b1a5444a-1e2f-4443-8931-4a04ec5d9d68/volume-b1a5444a-1e2f-4443-8931-4a04ec5d9d68.vmdk or device None with type thin {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 778.817500] env[63515]: DEBUG nova.scheduler.client.report [None req-ffbb9c80-abfa-457c-b756-9497c4635a40 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Overwriting current allocation {'allocations': {'2bd86232-2b6f-44d5-9057-1a3a6b27185a': {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}, 'generation': 79}}, 'project_id': 'd3d0957a488b47559794495a451f5437', 'user_id': '59a67d99419c47e9b4ff92c6d647656f', 'consumer_generation': 1} on consumer f1d01b75-ac9d-458d-8cc2-ae64cffca4e8 {{(pid=63515) move_allocations /opt/stack/nova/nova/scheduler/client/report.py:2032}} [ 778.821666] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a5b5c722-4410-41f7-a90f-2c851b46a3d8 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.836756] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4cc7579-ff84-4961-a776-eb5d798e7844 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.846475] env[63515]: DEBUG oslo_vmware.api [None req-2786351a-303c-45a8-b63b-65e218eb6ef6 tempest-ServersTestJSON-324454119 tempest-ServersTestJSON-324454119-project-member] Task: {'id': task-1111203, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.46683} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 778.849383] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-2786351a-303c-45a8-b63b-65e218eb6ef6 tempest-ServersTestJSON-324454119 tempest-ServersTestJSON-324454119-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk to [datastore2] 70ed982f-affd-4dd1-bc90-c64e7c6d49d2/70ed982f-affd-4dd1-bc90-c64e7c6d49d2.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 778.849618] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-2786351a-303c-45a8-b63b-65e218eb6ef6 tempest-ServersTestJSON-324454119 tempest-ServersTestJSON-324454119-project-member] [instance: 70ed982f-affd-4dd1-bc90-c64e7c6d49d2] Extending root virtual disk to 1048576 {{(pid=63515) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 778.849980] env[63515]: DEBUG oslo_vmware.api [None req-efd46494-ea54-4600-906e-1e7f1371fb6c tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Waiting for the task: (returnval){ [ 778.849980] env[63515]: value = "task-1111204" [ 778.849980] env[63515]: _type = "Task" [ 778.849980] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 778.850181] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-0f4cf6b4-b414-4414-b074-ab648fa89bbf {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.853136] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c455d4f7-6a44-4995-b2b2-3593f9a5b1f4 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.866015] env[63515]: DEBUG oslo_vmware.api [None req-efd46494-ea54-4600-906e-1e7f1371fb6c tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Task: {'id': task-1111204, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 778.892084] env[63515]: DEBUG oslo_vmware.api [None req-2786351a-303c-45a8-b63b-65e218eb6ef6 tempest-ServersTestJSON-324454119 tempest-ServersTestJSON-324454119-project-member] Waiting for the task: (returnval){ [ 778.892084] env[63515]: value = "task-1111205" [ 778.892084] env[63515]: _type = "Task" [ 778.892084] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 778.892904] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b904c65a-6f75-4455-83f8-b72d481fcbb8 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.908580] env[63515]: DEBUG oslo_concurrency.lockutils [None req-ce131312-5bf5-4677-8923-2d91405ade5d tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Acquiring lock "87c468d9-9594-4804-b461-527f01f6118f" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 778.912583] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ce1f9a6-08a1-41f4-b3e7-7ae6044e7b75 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.916511] env[63515]: DEBUG oslo_vmware.api [None req-2786351a-303c-45a8-b63b-65e218eb6ef6 tempest-ServersTestJSON-324454119 tempest-ServersTestJSON-324454119-project-member] Task: {'id': task-1111205, 'name': ExtendVirtualDisk_Task} progress is 50%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 778.928292] env[63515]: DEBUG nova.compute.provider_tree [None req-83fbd51b-fc37-49eb-8423-cdaa019d3b0c tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 778.937918] env[63515]: DEBUG oslo_concurrency.lockutils [None req-ffbb9c80-abfa-457c-b756-9497c4635a40 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Acquiring lock "refresh_cache-f1d01b75-ac9d-458d-8cc2-ae64cffca4e8" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 778.937918] env[63515]: DEBUG oslo_concurrency.lockutils [None req-ffbb9c80-abfa-457c-b756-9497c4635a40 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Acquired lock "refresh_cache-f1d01b75-ac9d-458d-8cc2-ae64cffca4e8" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 778.937918] env[63515]: DEBUG nova.network.neutron [None req-ffbb9c80-abfa-457c-b756-9497c4635a40 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] [instance: f1d01b75-ac9d-458d-8cc2-ae64cffca4e8] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 779.043557] env[63515]: DEBUG oslo_concurrency.lockutils [None req-f8bace5a-3f38-4490-8a6d-e315321b7f63 tempest-ServersTestManualDisk-1114050206 tempest-ServersTestManualDisk-1114050206-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 779.071641] env[63515]: DEBUG nova.network.neutron [req-2c5b2027-abb2-44c4-bd99-2ad564ff7b2f req-43fa149b-9f98-47b6-893a-6613a60250f1 service nova] [instance: 87c468d9-9594-4804-b461-527f01f6118f] Updated VIF entry in instance network info cache for port d2098868-0475-4b06-bae3-785c608a861e. {{(pid=63515) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 779.071902] env[63515]: DEBUG nova.network.neutron [req-2c5b2027-abb2-44c4-bd99-2ad564ff7b2f req-43fa149b-9f98-47b6-893a-6613a60250f1 service nova] [instance: 87c468d9-9594-4804-b461-527f01f6118f] Updating instance_info_cache with network_info: [{"id": "d2098868-0475-4b06-bae3-785c608a861e", "address": "fa:16:3e:af:f9:f0", "network": {"id": "09fb20ca-2b17-4d46-b915-6aff85b862d8", "bridge": null, "label": "tempest-AttachVolumeShelveTestJSON-1815529427-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.212", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "19ce3b05eec64475b685faad1fd801f6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tapd2098868-04", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 779.092995] env[63515]: DEBUG oslo_service.periodic_task [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=63515) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 779.093258] env[63515]: DEBUG oslo_service.periodic_task [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=63515) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 779.093412] env[63515]: DEBUG nova.compute.manager [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Starting heal instance info cache {{(pid=63515) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9926}} [ 779.253538] env[63515]: DEBUG oslo_vmware.api [None req-549cc3f8-330f-4c6a-979d-35eedc6f8146 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Task: {'id': task-1111202, 'name': PowerOnVM_Task, 'duration_secs': 0.633511} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 779.253839] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-549cc3f8-330f-4c6a-979d-35eedc6f8146 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: 84dbf321-2c70-4e08-b430-cb5a06fc6829] Powered on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 779.254076] env[63515]: INFO nova.compute.manager [None req-549cc3f8-330f-4c6a-979d-35eedc6f8146 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: 84dbf321-2c70-4e08-b430-cb5a06fc6829] Took 8.93 seconds to spawn the instance on the hypervisor. [ 779.254274] env[63515]: DEBUG nova.compute.manager [None req-549cc3f8-330f-4c6a-979d-35eedc6f8146 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: 84dbf321-2c70-4e08-b430-cb5a06fc6829] Checking state {{(pid=63515) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 779.255424] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f45d73ed-1aa5-4a8d-b3f8-d0ea7bae640a {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.272706] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1111201, 'name': CreateVM_Task} progress is 99%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 779.366119] env[63515]: DEBUG oslo_vmware.api [None req-efd46494-ea54-4600-906e-1e7f1371fb6c tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Task: {'id': task-1111204, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 779.402296] env[63515]: DEBUG nova.objects.instance [None req-8c8e056d-f20f-4ef4-b199-3c0616617ccf tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Lazy-loading 'pci_requests' on Instance uuid ed550b10-d58f-45b8-b766-198f431c3788 {{(pid=63515) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 779.406364] env[63515]: DEBUG oslo_vmware.api [None req-2786351a-303c-45a8-b63b-65e218eb6ef6 tempest-ServersTestJSON-324454119 tempest-ServersTestJSON-324454119-project-member] Task: {'id': task-1111205, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.065619} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 779.406889] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-2786351a-303c-45a8-b63b-65e218eb6ef6 tempest-ServersTestJSON-324454119 tempest-ServersTestJSON-324454119-project-member] [instance: 70ed982f-affd-4dd1-bc90-c64e7c6d49d2] Extended root virtual disk {{(pid=63515) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 779.407717] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26c2f0a9-5345-4c79-ae97-99cc06074101 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.429970] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-2786351a-303c-45a8-b63b-65e218eb6ef6 tempest-ServersTestJSON-324454119 tempest-ServersTestJSON-324454119-project-member] [instance: 70ed982f-affd-4dd1-bc90-c64e7c6d49d2] Reconfiguring VM instance instance-00000038 to attach disk [datastore2] 70ed982f-affd-4dd1-bc90-c64e7c6d49d2/70ed982f-affd-4dd1-bc90-c64e7c6d49d2.vmdk or device None with type sparse {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 779.431099] env[63515]: DEBUG nova.scheduler.client.report [None req-83fbd51b-fc37-49eb-8423-cdaa019d3b0c tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 779.434124] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-373b3fea-29d7-4d4c-aac2-381ceb5f9dce {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.457443] env[63515]: DEBUG oslo_vmware.api [None req-2786351a-303c-45a8-b63b-65e218eb6ef6 tempest-ServersTestJSON-324454119 tempest-ServersTestJSON-324454119-project-member] Waiting for the task: (returnval){ [ 779.457443] env[63515]: value = "task-1111206" [ 779.457443] env[63515]: _type = "Task" [ 779.457443] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 779.468602] env[63515]: DEBUG oslo_vmware.api [None req-2786351a-303c-45a8-b63b-65e218eb6ef6 tempest-ServersTestJSON-324454119 tempest-ServersTestJSON-324454119-project-member] Task: {'id': task-1111206, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 779.574402] env[63515]: DEBUG oslo_concurrency.lockutils [req-2c5b2027-abb2-44c4-bd99-2ad564ff7b2f req-43fa149b-9f98-47b6-893a-6613a60250f1 service nova] Releasing lock "refresh_cache-87c468d9-9594-4804-b461-527f01f6118f" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 779.597435] env[63515]: DEBUG oslo_concurrency.lockutils [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Acquiring lock "refresh_cache-f1d01b75-ac9d-458d-8cc2-ae64cffca4e8" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 779.654553] env[63515]: DEBUG nova.network.neutron [None req-ffbb9c80-abfa-457c-b756-9497c4635a40 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] [instance: f1d01b75-ac9d-458d-8cc2-ae64cffca4e8] Updating instance_info_cache with network_info: [{"id": "48f0868f-94db-42f7-8153-d1a27fa02707", "address": "fa:16:3e:7a:74:da", "network": {"id": "b4abcbd0-0eac-4bde-ad11-7f93afcedcef", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.21", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "b569255cc43e42e7ae3f2b4ad37c6ef4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0cd5d325-3053-407e-a4ee-f627e82a23f9", "external-id": "nsx-vlan-transportzone-809", "segmentation_id": 809, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap48f0868f-94", "ovs_interfaceid": "48f0868f-94db-42f7-8153-d1a27fa02707", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 779.784728] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1111201, 'name': CreateVM_Task, 'duration_secs': 3.031214} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 779.785344] env[63515]: INFO nova.compute.manager [None req-549cc3f8-330f-4c6a-979d-35eedc6f8146 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: 84dbf321-2c70-4e08-b430-cb5a06fc6829] Took 42.64 seconds to build instance. [ 779.786677] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 192137b0-03e5-4bc4-b911-4b4f1a874f74] Created VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 779.788240] env[63515]: DEBUG oslo_concurrency.lockutils [None req-8407c401-88c7-47ad-b91e-e7693f87e0f1 tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 779.788597] env[63515]: DEBUG oslo_concurrency.lockutils [None req-8407c401-88c7-47ad-b91e-e7693f87e0f1 tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 779.789891] env[63515]: DEBUG oslo_concurrency.lockutils [None req-8407c401-88c7-47ad-b91e-e7693f87e0f1 tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 779.790672] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-32ed6879-bf54-4ab2-b44d-ff3169940c3b {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.796569] env[63515]: DEBUG oslo_vmware.api [None req-8407c401-88c7-47ad-b91e-e7693f87e0f1 tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] Waiting for the task: (returnval){ [ 779.796569] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]5255e726-3aaf-3ff8-a826-abf2445b510c" [ 779.796569] env[63515]: _type = "Task" [ 779.796569] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 779.805156] env[63515]: DEBUG oslo_vmware.api [None req-8407c401-88c7-47ad-b91e-e7693f87e0f1 tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]5255e726-3aaf-3ff8-a826-abf2445b510c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 779.865689] env[63515]: DEBUG oslo_vmware.api [None req-efd46494-ea54-4600-906e-1e7f1371fb6c tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Task: {'id': task-1111204, 'name': ReconfigVM_Task, 'duration_secs': 0.69018} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 779.866196] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-efd46494-ea54-4600-906e-1e7f1371fb6c tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] [instance: d7e8dcdb-41cb-46fb-8b61-d251e7c2d372] Reconfigured VM instance instance-0000002c to attach disk [datastore2] volume-b1a5444a-1e2f-4443-8931-4a04ec5d9d68/volume-b1a5444a-1e2f-4443-8931-4a04ec5d9d68.vmdk or device None with type thin {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 779.871229] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7da2cfde-7459-4a9c-9d92-9844cf26ce49 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.886412] env[63515]: DEBUG oslo_vmware.api [None req-efd46494-ea54-4600-906e-1e7f1371fb6c tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Waiting for the task: (returnval){ [ 779.886412] env[63515]: value = "task-1111207" [ 779.886412] env[63515]: _type = "Task" [ 779.886412] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 779.895934] env[63515]: DEBUG oslo_vmware.api [None req-efd46494-ea54-4600-906e-1e7f1371fb6c tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Task: {'id': task-1111207, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 779.907757] env[63515]: DEBUG nova.objects.base [None req-8c8e056d-f20f-4ef4-b199-3c0616617ccf tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Object Instance lazy-loaded attributes: flavor,pci_requests {{(pid=63515) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 779.908055] env[63515]: DEBUG nova.network.neutron [None req-8c8e056d-f20f-4ef4-b199-3c0616617ccf tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: ed550b10-d58f-45b8-b766-198f431c3788] allocate_for_instance() {{(pid=63515) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 779.950750] env[63515]: DEBUG oslo_concurrency.lockutils [None req-83fbd51b-fc37-49eb-8423-cdaa019d3b0c tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.761s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 779.951393] env[63515]: DEBUG nova.compute.manager [None req-83fbd51b-fc37-49eb-8423-cdaa019d3b0c tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] [instance: 5cbce760-0163-4b27-8ae3-e46c926c8916] Start building networks asynchronously for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 779.954670] env[63515]: DEBUG oslo_concurrency.lockutils [None req-711301e3-5f69-4e98-bb19-e93c4ab2a0db tempest-ServerTagsTestJSON-1170819896 tempest-ServerTagsTestJSON-1170819896-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 29.427s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 779.955711] env[63515]: DEBUG oslo_concurrency.lockutils [None req-711301e3-5f69-4e98-bb19-e93c4ab2a0db tempest-ServerTagsTestJSON-1170819896 tempest-ServerTagsTestJSON-1170819896-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 779.957553] env[63515]: DEBUG oslo_concurrency.lockutils [None req-fa0cfe6d-3452-46d4-9987-2cb422956d62 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 24.700s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 779.958987] env[63515]: INFO nova.compute.claims [None req-fa0cfe6d-3452-46d4-9987-2cb422956d62 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] [instance: 63ce797d-7180-4209-ac2c-81978bf7607a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 779.975398] env[63515]: DEBUG oslo_vmware.api [None req-2786351a-303c-45a8-b63b-65e218eb6ef6 tempest-ServersTestJSON-324454119 tempest-ServersTestJSON-324454119-project-member] Task: {'id': task-1111206, 'name': ReconfigVM_Task, 'duration_secs': 0.25555} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 779.977340] env[63515]: DEBUG nova.policy [None req-8c8e056d-f20f-4ef4-b199-3c0616617ccf tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '16880abf1c5b4341800c94ada2c756c7', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7d4ada9ec35f42b19c6480a9101d21a4', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63515) authorize /opt/stack/nova/nova/policy.py:201}} [ 779.978556] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-2786351a-303c-45a8-b63b-65e218eb6ef6 tempest-ServersTestJSON-324454119 tempest-ServersTestJSON-324454119-project-member] [instance: 70ed982f-affd-4dd1-bc90-c64e7c6d49d2] Reconfigured VM instance instance-00000038 to attach disk [datastore2] 70ed982f-affd-4dd1-bc90-c64e7c6d49d2/70ed982f-affd-4dd1-bc90-c64e7c6d49d2.vmdk or device None with type sparse {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 779.979272] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-0cb9349b-52e4-43e7-ba82-11c9fe3fd41a {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.986090] env[63515]: DEBUG oslo_vmware.api [None req-2786351a-303c-45a8-b63b-65e218eb6ef6 tempest-ServersTestJSON-324454119 tempest-ServersTestJSON-324454119-project-member] Waiting for the task: (returnval){ [ 779.986090] env[63515]: value = "task-1111208" [ 779.986090] env[63515]: _type = "Task" [ 779.986090] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 779.996564] env[63515]: DEBUG oslo_vmware.api [None req-2786351a-303c-45a8-b63b-65e218eb6ef6 tempest-ServersTestJSON-324454119 tempest-ServersTestJSON-324454119-project-member] Task: {'id': task-1111208, 'name': Rename_Task} progress is 5%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 779.998934] env[63515]: INFO nova.scheduler.client.report [None req-711301e3-5f69-4e98-bb19-e93c4ab2a0db tempest-ServerTagsTestJSON-1170819896 tempest-ServerTagsTestJSON-1170819896-project-member] Deleted allocations for instance de210780-5c0f-4fba-883c-13707566a2e1 [ 780.158442] env[63515]: DEBUG oslo_concurrency.lockutils [None req-ffbb9c80-abfa-457c-b756-9497c4635a40 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Releasing lock "refresh_cache-f1d01b75-ac9d-458d-8cc2-ae64cffca4e8" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 780.158544] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-ffbb9c80-abfa-457c-b756-9497c4635a40 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] [instance: f1d01b75-ac9d-458d-8cc2-ae64cffca4e8] Powering off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 780.158795] env[63515]: DEBUG oslo_concurrency.lockutils [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Acquired lock "refresh_cache-f1d01b75-ac9d-458d-8cc2-ae64cffca4e8" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 780.158943] env[63515]: DEBUG nova.network.neutron [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] [instance: f1d01b75-ac9d-458d-8cc2-ae64cffca4e8] Forcefully refreshing network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2004}} [ 780.160109] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-780504c1-801e-4234-8954-d400bb291d79 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.167896] env[63515]: DEBUG oslo_vmware.api [None req-ffbb9c80-abfa-457c-b756-9497c4635a40 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Waiting for the task: (returnval){ [ 780.167896] env[63515]: value = "task-1111209" [ 780.167896] env[63515]: _type = "Task" [ 780.167896] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 780.177780] env[63515]: DEBUG oslo_vmware.api [None req-ffbb9c80-abfa-457c-b756-9497c4635a40 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Task: {'id': task-1111209, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 780.289752] env[63515]: DEBUG oslo_concurrency.lockutils [None req-549cc3f8-330f-4c6a-979d-35eedc6f8146 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Lock "84dbf321-2c70-4e08-b430-cb5a06fc6829" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 97.504s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 780.310107] env[63515]: DEBUG oslo_vmware.api [None req-8407c401-88c7-47ad-b91e-e7693f87e0f1 tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]5255e726-3aaf-3ff8-a826-abf2445b510c, 'name': SearchDatastore_Task, 'duration_secs': 0.00935} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 780.311551] env[63515]: DEBUG oslo_concurrency.lockutils [None req-8407c401-88c7-47ad-b91e-e7693f87e0f1 tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 780.311551] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-8407c401-88c7-47ad-b91e-e7693f87e0f1 tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] [instance: 192137b0-03e5-4bc4-b911-4b4f1a874f74] Processing image 8a120570-cb06-4099-b262-554ca0ad15c5 {{(pid=63515) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 780.311551] env[63515]: DEBUG oslo_concurrency.lockutils [None req-8407c401-88c7-47ad-b91e-e7693f87e0f1 tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 780.311551] env[63515]: DEBUG oslo_concurrency.lockutils [None req-8407c401-88c7-47ad-b91e-e7693f87e0f1 tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 780.311750] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-8407c401-88c7-47ad-b91e-e7693f87e0f1 tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 780.311789] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2407255c-6197-4774-b818-56414fb40e0a {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.316622] env[63515]: DEBUG nova.network.neutron [None req-8c8e056d-f20f-4ef4-b199-3c0616617ccf tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: ed550b10-d58f-45b8-b766-198f431c3788] Successfully created port: e24932f9-9063-4a16-8498-950bdfeeaa84 {{(pid=63515) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 780.329299] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-8407c401-88c7-47ad-b91e-e7693f87e0f1 tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 780.329600] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-8407c401-88c7-47ad-b91e-e7693f87e0f1 tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63515) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 780.330435] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-03e9e56d-304e-4870-9130-0b96bb023610 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.336208] env[63515]: DEBUG oslo_vmware.api [None req-8407c401-88c7-47ad-b91e-e7693f87e0f1 tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] Waiting for the task: (returnval){ [ 780.336208] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]52bf8522-a405-b7b2-33dd-24be5d0ad609" [ 780.336208] env[63515]: _type = "Task" [ 780.336208] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 780.345309] env[63515]: DEBUG oslo_vmware.api [None req-8407c401-88c7-47ad-b91e-e7693f87e0f1 tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52bf8522-a405-b7b2-33dd-24be5d0ad609, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 780.396617] env[63515]: DEBUG oslo_vmware.api [None req-efd46494-ea54-4600-906e-1e7f1371fb6c tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Task: {'id': task-1111207, 'name': ReconfigVM_Task, 'duration_secs': 0.239014} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 780.397058] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-efd46494-ea54-4600-906e-1e7f1371fb6c tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] [instance: d7e8dcdb-41cb-46fb-8b61-d251e7c2d372] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-243453', 'volume_id': 'b1a5444a-1e2f-4443-8931-4a04ec5d9d68', 'name': 'volume-b1a5444a-1e2f-4443-8931-4a04ec5d9d68', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'd7e8dcdb-41cb-46fb-8b61-d251e7c2d372', 'attached_at': '', 'detached_at': '', 'volume_id': 'b1a5444a-1e2f-4443-8931-4a04ec5d9d68', 'serial': 'b1a5444a-1e2f-4443-8931-4a04ec5d9d68'} {{(pid=63515) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 780.463924] env[63515]: DEBUG nova.compute.utils [None req-83fbd51b-fc37-49eb-8423-cdaa019d3b0c tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Using /dev/sd instead of None {{(pid=63515) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 780.463924] env[63515]: DEBUG nova.compute.manager [None req-83fbd51b-fc37-49eb-8423-cdaa019d3b0c tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] [instance: 5cbce760-0163-4b27-8ae3-e46c926c8916] Allocating IP information in the background. {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 780.463924] env[63515]: DEBUG nova.network.neutron [None req-83fbd51b-fc37-49eb-8423-cdaa019d3b0c tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] [instance: 5cbce760-0163-4b27-8ae3-e46c926c8916] allocate_for_instance() {{(pid=63515) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 780.496017] env[63515]: DEBUG oslo_vmware.api [None req-2786351a-303c-45a8-b63b-65e218eb6ef6 tempest-ServersTestJSON-324454119 tempest-ServersTestJSON-324454119-project-member] Task: {'id': task-1111208, 'name': Rename_Task, 'duration_secs': 0.144368} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 780.496373] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-2786351a-303c-45a8-b63b-65e218eb6ef6 tempest-ServersTestJSON-324454119 tempest-ServersTestJSON-324454119-project-member] [instance: 70ed982f-affd-4dd1-bc90-c64e7c6d49d2] Powering on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 780.496626] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-02571dbc-5851-4207-bebb-44ac8e2c06bf {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.503046] env[63515]: DEBUG oslo_vmware.api [None req-2786351a-303c-45a8-b63b-65e218eb6ef6 tempest-ServersTestJSON-324454119 tempest-ServersTestJSON-324454119-project-member] Waiting for the task: (returnval){ [ 780.503046] env[63515]: value = "task-1111210" [ 780.503046] env[63515]: _type = "Task" [ 780.503046] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 780.510058] env[63515]: DEBUG oslo_concurrency.lockutils [None req-711301e3-5f69-4e98-bb19-e93c4ab2a0db tempest-ServerTagsTestJSON-1170819896 tempest-ServerTagsTestJSON-1170819896-project-member] Lock "de210780-5c0f-4fba-883c-13707566a2e1" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 33.477s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 780.516817] env[63515]: DEBUG oslo_vmware.api [None req-2786351a-303c-45a8-b63b-65e218eb6ef6 tempest-ServersTestJSON-324454119 tempest-ServersTestJSON-324454119-project-member] Task: {'id': task-1111210, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 780.545306] env[63515]: DEBUG nova.policy [None req-83fbd51b-fc37-49eb-8423-cdaa019d3b0c tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2b86a44d6b814d2baeb7e7eb87f12361', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '08c9aa51b1864410809f8877c7cd49c4', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63515) authorize /opt/stack/nova/nova/policy.py:201}} [ 780.680492] env[63515]: DEBUG oslo_vmware.api [None req-ffbb9c80-abfa-457c-b756-9497c4635a40 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Task: {'id': task-1111209, 'name': PowerOffVM_Task, 'duration_secs': 0.180883} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 780.680791] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-ffbb9c80-abfa-457c-b756-9497c4635a40 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] [instance: f1d01b75-ac9d-458d-8cc2-ae64cffca4e8] Powered off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 780.681489] env[63515]: DEBUG nova.virt.hardware [None req-ffbb9c80-abfa-457c-b756-9497c4635a40 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T02:54:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='1fae9a09-456d-4e70-9655-fc9d700f1547',id=38,is_public=True,memory_mb=192,name='tempest-test_resize_flavor_-1433437001',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=8a120570-cb06-4099-b262-554ca0ad15c5,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 780.682436] env[63515]: DEBUG nova.virt.hardware [None req-ffbb9c80-abfa-457c-b756-9497c4635a40 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Flavor limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 780.682436] env[63515]: DEBUG nova.virt.hardware [None req-ffbb9c80-abfa-457c-b756-9497c4635a40 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Image limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 780.684936] env[63515]: DEBUG nova.virt.hardware [None req-ffbb9c80-abfa-457c-b756-9497c4635a40 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Flavor pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 780.684936] env[63515]: DEBUG nova.virt.hardware [None req-ffbb9c80-abfa-457c-b756-9497c4635a40 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Image pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 780.684936] env[63515]: DEBUG nova.virt.hardware [None req-ffbb9c80-abfa-457c-b756-9497c4635a40 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 780.684936] env[63515]: DEBUG nova.virt.hardware [None req-ffbb9c80-abfa-457c-b756-9497c4635a40 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 780.684936] env[63515]: DEBUG nova.virt.hardware [None req-ffbb9c80-abfa-457c-b756-9497c4635a40 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 780.685403] env[63515]: DEBUG nova.virt.hardware [None req-ffbb9c80-abfa-457c-b756-9497c4635a40 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Got 1 possible topologies {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 780.685403] env[63515]: DEBUG nova.virt.hardware [None req-ffbb9c80-abfa-457c-b756-9497c4635a40 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 780.685403] env[63515]: DEBUG nova.virt.hardware [None req-ffbb9c80-abfa-457c-b756-9497c4635a40 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 780.690352] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9d5a738b-ff6b-424d-9092-c544fbc5759c {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.710443] env[63515]: DEBUG oslo_vmware.api [None req-ffbb9c80-abfa-457c-b756-9497c4635a40 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Waiting for the task: (returnval){ [ 780.710443] env[63515]: value = "task-1111211" [ 780.710443] env[63515]: _type = "Task" [ 780.710443] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 780.718439] env[63515]: DEBUG oslo_vmware.api [None req-ffbb9c80-abfa-457c-b756-9497c4635a40 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Task: {'id': task-1111211, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 780.848183] env[63515]: DEBUG oslo_vmware.api [None req-8407c401-88c7-47ad-b91e-e7693f87e0f1 tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52bf8522-a405-b7b2-33dd-24be5d0ad609, 'name': SearchDatastore_Task, 'duration_secs': 0.009415} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 780.849534] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4d55687a-7c72-4a6f-9d10-fe8017757df6 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.855524] env[63515]: DEBUG oslo_vmware.api [None req-8407c401-88c7-47ad-b91e-e7693f87e0f1 tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] Waiting for the task: (returnval){ [ 780.855524] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]52554214-cbe7-8d01-f3a9-47c503111bc9" [ 780.855524] env[63515]: _type = "Task" [ 780.855524] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 780.864131] env[63515]: DEBUG oslo_vmware.api [None req-8407c401-88c7-47ad-b91e-e7693f87e0f1 tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52554214-cbe7-8d01-f3a9-47c503111bc9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 780.971313] env[63515]: DEBUG nova.compute.manager [None req-83fbd51b-fc37-49eb-8423-cdaa019d3b0c tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] [instance: 5cbce760-0163-4b27-8ae3-e46c926c8916] Start building block device mappings for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 781.020740] env[63515]: DEBUG oslo_vmware.api [None req-2786351a-303c-45a8-b63b-65e218eb6ef6 tempest-ServersTestJSON-324454119 tempest-ServersTestJSON-324454119-project-member] Task: {'id': task-1111210, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 781.174750] env[63515]: DEBUG nova.network.neutron [None req-83fbd51b-fc37-49eb-8423-cdaa019d3b0c tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] [instance: 5cbce760-0163-4b27-8ae3-e46c926c8916] Successfully created port: 5fa807c4-5ce1-492a-bea3-d8f7a65d898f {{(pid=63515) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 781.223287] env[63515]: DEBUG oslo_vmware.api [None req-ffbb9c80-abfa-457c-b756-9497c4635a40 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Task: {'id': task-1111211, 'name': ReconfigVM_Task, 'duration_secs': 0.127575} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 781.226798] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-739bf770-42d2-45f2-8234-6a18fce95f67 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.245140] env[63515]: DEBUG nova.virt.hardware [None req-ffbb9c80-abfa-457c-b756-9497c4635a40 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T02:54:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='1fae9a09-456d-4e70-9655-fc9d700f1547',id=38,is_public=True,memory_mb=192,name='tempest-test_resize_flavor_-1433437001',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=8a120570-cb06-4099-b262-554ca0ad15c5,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 781.245418] env[63515]: DEBUG nova.virt.hardware [None req-ffbb9c80-abfa-457c-b756-9497c4635a40 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Flavor limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 781.245592] env[63515]: DEBUG nova.virt.hardware [None req-ffbb9c80-abfa-457c-b756-9497c4635a40 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Image limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 781.245785] env[63515]: DEBUG nova.virt.hardware [None req-ffbb9c80-abfa-457c-b756-9497c4635a40 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Flavor pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 781.245936] env[63515]: DEBUG nova.virt.hardware [None req-ffbb9c80-abfa-457c-b756-9497c4635a40 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Image pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 781.246127] env[63515]: DEBUG nova.virt.hardware [None req-ffbb9c80-abfa-457c-b756-9497c4635a40 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 781.246341] env[63515]: DEBUG nova.virt.hardware [None req-ffbb9c80-abfa-457c-b756-9497c4635a40 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 781.246510] env[63515]: DEBUG nova.virt.hardware [None req-ffbb9c80-abfa-457c-b756-9497c4635a40 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 781.246682] env[63515]: DEBUG nova.virt.hardware [None req-ffbb9c80-abfa-457c-b756-9497c4635a40 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Got 1 possible topologies {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 781.246849] env[63515]: DEBUG nova.virt.hardware [None req-ffbb9c80-abfa-457c-b756-9497c4635a40 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 781.247065] env[63515]: DEBUG nova.virt.hardware [None req-ffbb9c80-abfa-457c-b756-9497c4635a40 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 781.251929] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0841e688-6a22-4426-808d-d01285169400 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.258112] env[63515]: DEBUG oslo_vmware.api [None req-ffbb9c80-abfa-457c-b756-9497c4635a40 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Waiting for the task: (returnval){ [ 781.258112] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]520b32fe-08b7-27db-4b8a-dcb9806d7d36" [ 781.258112] env[63515]: _type = "Task" [ 781.258112] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 781.267179] env[63515]: DEBUG oslo_vmware.api [None req-ffbb9c80-abfa-457c-b756-9497c4635a40 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]520b32fe-08b7-27db-4b8a-dcb9806d7d36, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 781.371017] env[63515]: DEBUG oslo_vmware.api [None req-8407c401-88c7-47ad-b91e-e7693f87e0f1 tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52554214-cbe7-8d01-f3a9-47c503111bc9, 'name': SearchDatastore_Task, 'duration_secs': 0.012657} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 781.371306] env[63515]: DEBUG oslo_concurrency.lockutils [None req-8407c401-88c7-47ad-b91e-e7693f87e0f1 tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 781.371572] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-8407c401-88c7-47ad-b91e-e7693f87e0f1 tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk to [datastore1] 192137b0-03e5-4bc4-b911-4b4f1a874f74/192137b0-03e5-4bc4-b911-4b4f1a874f74.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 781.371839] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-084ccaf2-8b5d-4204-a9cb-ce2b3b5fa89c {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.378552] env[63515]: DEBUG oslo_vmware.api [None req-8407c401-88c7-47ad-b91e-e7693f87e0f1 tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] Waiting for the task: (returnval){ [ 781.378552] env[63515]: value = "task-1111212" [ 781.378552] env[63515]: _type = "Task" [ 781.378552] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 781.389357] env[63515]: DEBUG oslo_vmware.api [None req-8407c401-88c7-47ad-b91e-e7693f87e0f1 tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] Task: {'id': task-1111212, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 781.402321] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9789e0a1-77b4-4dbf-82b8-bf205d33dacd {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.409783] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a66626e9-572f-43b2-bcc4-3b81859239f2 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.442906] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47ca9984-cd5f-4966-9c4f-ecba9abd20d8 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.446404] env[63515]: DEBUG nova.objects.instance [None req-efd46494-ea54-4600-906e-1e7f1371fb6c tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Lazy-loading 'flavor' on Instance uuid d7e8dcdb-41cb-46fb-8b61-d251e7c2d372 {{(pid=63515) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 781.454376] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ff842fa-3af3-4957-b759-a07446f655b4 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.468840] env[63515]: DEBUG nova.compute.provider_tree [None req-fa0cfe6d-3452-46d4-9987-2cb422956d62 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 781.518481] env[63515]: DEBUG oslo_vmware.api [None req-2786351a-303c-45a8-b63b-65e218eb6ef6 tempest-ServersTestJSON-324454119 tempest-ServersTestJSON-324454119-project-member] Task: {'id': task-1111210, 'name': PowerOnVM_Task, 'duration_secs': 0.671985} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 781.518481] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-2786351a-303c-45a8-b63b-65e218eb6ef6 tempest-ServersTestJSON-324454119 tempest-ServersTestJSON-324454119-project-member] [instance: 70ed982f-affd-4dd1-bc90-c64e7c6d49d2] Powered on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 781.518481] env[63515]: INFO nova.compute.manager [None req-2786351a-303c-45a8-b63b-65e218eb6ef6 tempest-ServersTestJSON-324454119 tempest-ServersTestJSON-324454119-project-member] [instance: 70ed982f-affd-4dd1-bc90-c64e7c6d49d2] Took 7.49 seconds to spawn the instance on the hypervisor. [ 781.518769] env[63515]: DEBUG nova.compute.manager [None req-2786351a-303c-45a8-b63b-65e218eb6ef6 tempest-ServersTestJSON-324454119 tempest-ServersTestJSON-324454119-project-member] [instance: 70ed982f-affd-4dd1-bc90-c64e7c6d49d2] Checking state {{(pid=63515) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 781.519526] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6b5e96e-ce29-4b87-aabd-8b0678208898 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.543741] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93673176-917f-4d8a-b963-3cfc1550e128 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.551885] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-7de2916d-b6ba-4a07-a593-ca4b473d8ab6 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: 84dbf321-2c70-4e08-b430-cb5a06fc6829] Suspending the VM {{(pid=63515) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1162}} [ 781.552414] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-fff8047f-b9d4-49c3-b958-bc6e911bc5b6 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.566023] env[63515]: DEBUG oslo_vmware.api [None req-7de2916d-b6ba-4a07-a593-ca4b473d8ab6 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Waiting for the task: (returnval){ [ 781.566023] env[63515]: value = "task-1111213" [ 781.566023] env[63515]: _type = "Task" [ 781.566023] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 781.576300] env[63515]: DEBUG oslo_vmware.api [None req-7de2916d-b6ba-4a07-a593-ca4b473d8ab6 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Task: {'id': task-1111213, 'name': SuspendVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 781.775119] env[63515]: DEBUG oslo_vmware.api [None req-ffbb9c80-abfa-457c-b756-9497c4635a40 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]520b32fe-08b7-27db-4b8a-dcb9806d7d36, 'name': SearchDatastore_Task, 'duration_secs': 0.008325} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 781.785243] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-ffbb9c80-abfa-457c-b756-9497c4635a40 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] [instance: f1d01b75-ac9d-458d-8cc2-ae64cffca4e8] Reconfiguring VM instance instance-00000022 to detach disk 2000 {{(pid=63515) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 781.785521] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-063dfcf7-11be-44fd-a0b0-bf877b45b03e {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.807345] env[63515]: DEBUG nova.network.neutron [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] [instance: f1d01b75-ac9d-458d-8cc2-ae64cffca4e8] Updating instance_info_cache with network_info: [{"id": "48f0868f-94db-42f7-8153-d1a27fa02707", "address": "fa:16:3e:7a:74:da", "network": {"id": "b4abcbd0-0eac-4bde-ad11-7f93afcedcef", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.21", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "b569255cc43e42e7ae3f2b4ad37c6ef4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0cd5d325-3053-407e-a4ee-f627e82a23f9", "external-id": "nsx-vlan-transportzone-809", "segmentation_id": 809, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap48f0868f-94", "ovs_interfaceid": "48f0868f-94db-42f7-8153-d1a27fa02707", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 781.811185] env[63515]: DEBUG oslo_vmware.api [None req-ffbb9c80-abfa-457c-b756-9497c4635a40 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Waiting for the task: (returnval){ [ 781.811185] env[63515]: value = "task-1111214" [ 781.811185] env[63515]: _type = "Task" [ 781.811185] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 781.822703] env[63515]: DEBUG oslo_vmware.api [None req-ffbb9c80-abfa-457c-b756-9497c4635a40 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Task: {'id': task-1111214, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 781.895814] env[63515]: DEBUG oslo_vmware.api [None req-8407c401-88c7-47ad-b91e-e7693f87e0f1 tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] Task: {'id': task-1111212, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 781.953089] env[63515]: DEBUG oslo_concurrency.lockutils [None req-efd46494-ea54-4600-906e-1e7f1371fb6c tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Lock "d7e8dcdb-41cb-46fb-8b61-d251e7c2d372" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.835s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 781.978154] env[63515]: DEBUG nova.scheduler.client.report [None req-fa0cfe6d-3452-46d4-9987-2cb422956d62 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 781.985333] env[63515]: DEBUG nova.compute.manager [None req-83fbd51b-fc37-49eb-8423-cdaa019d3b0c tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] [instance: 5cbce760-0163-4b27-8ae3-e46c926c8916] Start spawning the instance on the hypervisor. {{(pid=63515) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 782.028908] env[63515]: DEBUG nova.virt.hardware [None req-83fbd51b-fc37-49eb-8423-cdaa019d3b0c tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T02:52:14Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T02:51:56Z,direct_url=,disk_format='vmdk',id=8a120570-cb06-4099-b262-554ca0ad15c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b569255cc43e42e7ae3f2b4ad37c6ef4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T02:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 782.029189] env[63515]: DEBUG nova.virt.hardware [None req-83fbd51b-fc37-49eb-8423-cdaa019d3b0c tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Flavor limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 782.029352] env[63515]: DEBUG nova.virt.hardware [None req-83fbd51b-fc37-49eb-8423-cdaa019d3b0c tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Image limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 782.029535] env[63515]: DEBUG nova.virt.hardware [None req-83fbd51b-fc37-49eb-8423-cdaa019d3b0c tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Flavor pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 782.030119] env[63515]: DEBUG nova.virt.hardware [None req-83fbd51b-fc37-49eb-8423-cdaa019d3b0c tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Image pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 782.030119] env[63515]: DEBUG nova.virt.hardware [None req-83fbd51b-fc37-49eb-8423-cdaa019d3b0c tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 782.030407] env[63515]: DEBUG nova.virt.hardware [None req-83fbd51b-fc37-49eb-8423-cdaa019d3b0c tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 782.030635] env[63515]: DEBUG nova.virt.hardware [None req-83fbd51b-fc37-49eb-8423-cdaa019d3b0c tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 782.030822] env[63515]: DEBUG nova.virt.hardware [None req-83fbd51b-fc37-49eb-8423-cdaa019d3b0c tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Got 1 possible topologies {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 782.031017] env[63515]: DEBUG nova.virt.hardware [None req-83fbd51b-fc37-49eb-8423-cdaa019d3b0c tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 782.031206] env[63515]: DEBUG nova.virt.hardware [None req-83fbd51b-fc37-49eb-8423-cdaa019d3b0c tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 782.032113] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a79b4174-3014-4cce-ab0b-a21997460ed5 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.044630] env[63515]: INFO nova.compute.manager [None req-2786351a-303c-45a8-b63b-65e218eb6ef6 tempest-ServersTestJSON-324454119 tempest-ServersTestJSON-324454119-project-member] [instance: 70ed982f-affd-4dd1-bc90-c64e7c6d49d2] Took 41.96 seconds to build instance. [ 782.050167] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a13ae986-dab0-4613-b3ad-abe2664860c0 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.072706] env[63515]: DEBUG oslo_vmware.api [None req-7de2916d-b6ba-4a07-a593-ca4b473d8ab6 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Task: {'id': task-1111213, 'name': SuspendVM_Task} progress is 58%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 782.115650] env[63515]: DEBUG nova.network.neutron [None req-8c8e056d-f20f-4ef4-b199-3c0616617ccf tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: ed550b10-d58f-45b8-b766-198f431c3788] Successfully updated port: e24932f9-9063-4a16-8498-950bdfeeaa84 {{(pid=63515) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 782.313604] env[63515]: DEBUG oslo_concurrency.lockutils [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Releasing lock "refresh_cache-f1d01b75-ac9d-458d-8cc2-ae64cffca4e8" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 782.313858] env[63515]: DEBUG nova.compute.manager [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] [instance: f1d01b75-ac9d-458d-8cc2-ae64cffca4e8] Updated the network info_cache for instance {{(pid=63515) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9997}} [ 782.314538] env[63515]: DEBUG oslo_service.periodic_task [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=63515) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 782.320531] env[63515]: DEBUG oslo_service.periodic_task [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=63515) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 782.320802] env[63515]: DEBUG oslo_service.periodic_task [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=63515) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 782.321484] env[63515]: DEBUG oslo_service.periodic_task [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=63515) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 782.321753] env[63515]: DEBUG oslo_service.periodic_task [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=63515) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 782.321926] env[63515]: DEBUG nova.compute.manager [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=63515) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10545}} [ 782.322091] env[63515]: DEBUG oslo_service.periodic_task [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Running periodic task ComputeManager.update_available_resource {{(pid=63515) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 782.326444] env[63515]: DEBUG oslo_vmware.api [None req-ffbb9c80-abfa-457c-b756-9497c4635a40 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Task: {'id': task-1111214, 'name': ReconfigVM_Task, 'duration_secs': 0.25291} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 782.326953] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-ffbb9c80-abfa-457c-b756-9497c4635a40 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] [instance: f1d01b75-ac9d-458d-8cc2-ae64cffca4e8] Reconfigured VM instance instance-00000022 to detach disk 2000 {{(pid=63515) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 782.329354] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23cf0253-fa33-428a-8cc5-01969787b380 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.354605] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-ffbb9c80-abfa-457c-b756-9497c4635a40 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] [instance: f1d01b75-ac9d-458d-8cc2-ae64cffca4e8] Reconfiguring VM instance instance-00000022 to attach disk [datastore2] f1d01b75-ac9d-458d-8cc2-ae64cffca4e8/f1d01b75-ac9d-458d-8cc2-ae64cffca4e8.vmdk or device None with type thin {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 782.355937] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-085967b0-622c-4e36-817e-161fe52cac99 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.375360] env[63515]: DEBUG oslo_vmware.api [None req-ffbb9c80-abfa-457c-b756-9497c4635a40 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Waiting for the task: (returnval){ [ 782.375360] env[63515]: value = "task-1111215" [ 782.375360] env[63515]: _type = "Task" [ 782.375360] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 782.383529] env[63515]: DEBUG oslo_vmware.api [None req-ffbb9c80-abfa-457c-b756-9497c4635a40 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Task: {'id': task-1111215, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 782.393489] env[63515]: DEBUG oslo_vmware.api [None req-8407c401-88c7-47ad-b91e-e7693f87e0f1 tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] Task: {'id': task-1111212, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.529463} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 782.393785] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-8407c401-88c7-47ad-b91e-e7693f87e0f1 tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk to [datastore1] 192137b0-03e5-4bc4-b911-4b4f1a874f74/192137b0-03e5-4bc4-b911-4b4f1a874f74.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 782.394015] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-8407c401-88c7-47ad-b91e-e7693f87e0f1 tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] [instance: 192137b0-03e5-4bc4-b911-4b4f1a874f74] Extending root virtual disk to 1048576 {{(pid=63515) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 782.394297] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-07c77d49-daa2-4b77-9785-c702b97b759d {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.400992] env[63515]: DEBUG oslo_vmware.api [None req-8407c401-88c7-47ad-b91e-e7693f87e0f1 tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] Waiting for the task: (returnval){ [ 782.400992] env[63515]: value = "task-1111216" [ 782.400992] env[63515]: _type = "Task" [ 782.400992] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 782.409810] env[63515]: DEBUG oslo_vmware.api [None req-8407c401-88c7-47ad-b91e-e7693f87e0f1 tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] Task: {'id': task-1111216, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 782.481446] env[63515]: DEBUG oslo_concurrency.lockutils [None req-fa0cfe6d-3452-46d4-9987-2cb422956d62 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.524s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 782.481983] env[63515]: DEBUG nova.compute.manager [None req-fa0cfe6d-3452-46d4-9987-2cb422956d62 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] [instance: 63ce797d-7180-4209-ac2c-81978bf7607a] Start building networks asynchronously for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 782.485261] env[63515]: DEBUG oslo_concurrency.lockutils [None req-4b5baf79-d4c3-467e-88d4-a131ea6620d7 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 24.588s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 782.486490] env[63515]: INFO nova.compute.claims [None req-4b5baf79-d4c3-467e-88d4-a131ea6620d7 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] [instance: e666825c-ff4e-4a0e-93c0-43c00f167bbb] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 782.549045] env[63515]: DEBUG oslo_concurrency.lockutils [None req-2786351a-303c-45a8-b63b-65e218eb6ef6 tempest-ServersTestJSON-324454119 tempest-ServersTestJSON-324454119-project-member] Lock "70ed982f-affd-4dd1-bc90-c64e7c6d49d2" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 98.100s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 782.578128] env[63515]: DEBUG oslo_vmware.api [None req-7de2916d-b6ba-4a07-a593-ca4b473d8ab6 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Task: {'id': task-1111213, 'name': SuspendVM_Task, 'duration_secs': 0.629146} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 782.578500] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-7de2916d-b6ba-4a07-a593-ca4b473d8ab6 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: 84dbf321-2c70-4e08-b430-cb5a06fc6829] Suspended the VM {{(pid=63515) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1166}} [ 782.578993] env[63515]: DEBUG nova.compute.manager [None req-7de2916d-b6ba-4a07-a593-ca4b473d8ab6 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: 84dbf321-2c70-4e08-b430-cb5a06fc6829] Checking state {{(pid=63515) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 782.579555] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc8688d4-c6e3-45ed-8ffa-3bd98f2ff4a7 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.589956] env[63515]: DEBUG oslo_concurrency.lockutils [None req-8c8e056d-f20f-4ef4-b199-3c0616617ccf tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Acquiring lock "refresh_cache-ed550b10-d58f-45b8-b766-198f431c3788" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 782.590153] env[63515]: DEBUG oslo_concurrency.lockutils [None req-8c8e056d-f20f-4ef4-b199-3c0616617ccf tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Acquired lock "refresh_cache-ed550b10-d58f-45b8-b766-198f431c3788" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 782.590330] env[63515]: DEBUG nova.network.neutron [None req-8c8e056d-f20f-4ef4-b199-3c0616617ccf tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: ed550b10-d58f-45b8-b766-198f431c3788] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 782.828962] env[63515]: DEBUG oslo_concurrency.lockutils [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 782.886016] env[63515]: DEBUG oslo_vmware.api [None req-ffbb9c80-abfa-457c-b756-9497c4635a40 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Task: {'id': task-1111215, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 782.912585] env[63515]: DEBUG oslo_vmware.api [None req-8407c401-88c7-47ad-b91e-e7693f87e0f1 tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] Task: {'id': task-1111216, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.065978} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 782.912585] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-8407c401-88c7-47ad-b91e-e7693f87e0f1 tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] [instance: 192137b0-03e5-4bc4-b911-4b4f1a874f74] Extended root virtual disk {{(pid=63515) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 782.912585] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4cbabc0e-7824-44ea-ba6e-4d8b3ddf5786 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.933027] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-8407c401-88c7-47ad-b91e-e7693f87e0f1 tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] [instance: 192137b0-03e5-4bc4-b911-4b4f1a874f74] Reconfiguring VM instance instance-00000039 to attach disk [datastore1] 192137b0-03e5-4bc4-b911-4b4f1a874f74/192137b0-03e5-4bc4-b911-4b4f1a874f74.vmdk or device None with type sparse {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 782.933027] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-168d9c4a-e515-4a79-8e53-4b2a299ccbb7 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.955780] env[63515]: DEBUG oslo_vmware.api [None req-8407c401-88c7-47ad-b91e-e7693f87e0f1 tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] Waiting for the task: (returnval){ [ 782.955780] env[63515]: value = "task-1111217" [ 782.955780] env[63515]: _type = "Task" [ 782.955780] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 782.964040] env[63515]: DEBUG oslo_vmware.api [None req-8407c401-88c7-47ad-b91e-e7693f87e0f1 tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] Task: {'id': task-1111217, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 782.994665] env[63515]: DEBUG nova.compute.utils [None req-fa0cfe6d-3452-46d4-9987-2cb422956d62 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Using /dev/sd instead of None {{(pid=63515) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 782.995069] env[63515]: DEBUG nova.compute.manager [None req-fa0cfe6d-3452-46d4-9987-2cb422956d62 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] [instance: 63ce797d-7180-4209-ac2c-81978bf7607a] Allocating IP information in the background. {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 782.995311] env[63515]: DEBUG nova.network.neutron [None req-fa0cfe6d-3452-46d4-9987-2cb422956d62 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] [instance: 63ce797d-7180-4209-ac2c-81978bf7607a] allocate_for_instance() {{(pid=63515) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 783.038233] env[63515]: DEBUG nova.policy [None req-fa0cfe6d-3452-46d4-9987-2cb422956d62 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2b86a44d6b814d2baeb7e7eb87f12361', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '08c9aa51b1864410809f8877c7cd49c4', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63515) authorize /opt/stack/nova/nova/policy.py:201}} [ 783.157768] env[63515]: WARNING nova.network.neutron [None req-8c8e056d-f20f-4ef4-b199-3c0616617ccf tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: ed550b10-d58f-45b8-b766-198f431c3788] 88dfb29d-41e0-4df7-8f70-e52cda05620c already exists in list: networks containing: ['88dfb29d-41e0-4df7-8f70-e52cda05620c']. ignoring it [ 783.304631] env[63515]: DEBUG nova.compute.manager [req-b59c9cd8-153e-4c34-b4e9-0fc00cc62eea req-ee40584f-b180-4a2a-a110-821c3372eb65 service nova] [instance: ed550b10-d58f-45b8-b766-198f431c3788] Received event network-vif-plugged-e24932f9-9063-4a16-8498-950bdfeeaa84 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 783.304989] env[63515]: DEBUG oslo_concurrency.lockutils [req-b59c9cd8-153e-4c34-b4e9-0fc00cc62eea req-ee40584f-b180-4a2a-a110-821c3372eb65 service nova] Acquiring lock "ed550b10-d58f-45b8-b766-198f431c3788-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 783.305706] env[63515]: DEBUG oslo_concurrency.lockutils [req-b59c9cd8-153e-4c34-b4e9-0fc00cc62eea req-ee40584f-b180-4a2a-a110-821c3372eb65 service nova] Lock "ed550b10-d58f-45b8-b766-198f431c3788-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 783.305999] env[63515]: DEBUG oslo_concurrency.lockutils [req-b59c9cd8-153e-4c34-b4e9-0fc00cc62eea req-ee40584f-b180-4a2a-a110-821c3372eb65 service nova] Lock "ed550b10-d58f-45b8-b766-198f431c3788-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.001s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 783.306312] env[63515]: DEBUG nova.compute.manager [req-b59c9cd8-153e-4c34-b4e9-0fc00cc62eea req-ee40584f-b180-4a2a-a110-821c3372eb65 service nova] [instance: ed550b10-d58f-45b8-b766-198f431c3788] No waiting events found dispatching network-vif-plugged-e24932f9-9063-4a16-8498-950bdfeeaa84 {{(pid=63515) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 783.306600] env[63515]: WARNING nova.compute.manager [req-b59c9cd8-153e-4c34-b4e9-0fc00cc62eea req-ee40584f-b180-4a2a-a110-821c3372eb65 service nova] [instance: ed550b10-d58f-45b8-b766-198f431c3788] Received unexpected event network-vif-plugged-e24932f9-9063-4a16-8498-950bdfeeaa84 for instance with vm_state active and task_state None. [ 783.329622] env[63515]: DEBUG nova.network.neutron [None req-fa0cfe6d-3452-46d4-9987-2cb422956d62 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] [instance: 63ce797d-7180-4209-ac2c-81978bf7607a] Successfully created port: 25ea9d0d-3405-42da-a576-a2a1f37fa75a {{(pid=63515) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 783.389059] env[63515]: DEBUG oslo_vmware.api [None req-ffbb9c80-abfa-457c-b756-9497c4635a40 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Task: {'id': task-1111215, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 783.473943] env[63515]: DEBUG oslo_vmware.api [None req-8407c401-88c7-47ad-b91e-e7693f87e0f1 tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] Task: {'id': task-1111217, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 783.498031] env[63515]: DEBUG nova.compute.manager [None req-fa0cfe6d-3452-46d4-9987-2cb422956d62 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] [instance: 63ce797d-7180-4209-ac2c-81978bf7607a] Start building block device mappings for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 783.712921] env[63515]: DEBUG nova.network.neutron [None req-83fbd51b-fc37-49eb-8423-cdaa019d3b0c tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] [instance: 5cbce760-0163-4b27-8ae3-e46c926c8916] Successfully updated port: 5fa807c4-5ce1-492a-bea3-d8f7a65d898f {{(pid=63515) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 783.821728] env[63515]: DEBUG nova.network.neutron [None req-8c8e056d-f20f-4ef4-b199-3c0616617ccf tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: ed550b10-d58f-45b8-b766-198f431c3788] Updating instance_info_cache with network_info: [{"id": "ad9dda29-9326-472b-8c82-294a33a7faf0", "address": "fa:16:3e:26:e3:72", "network": {"id": "88dfb29d-41e0-4df7-8f70-e52cda05620c", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1355578400-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.251", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7d4ada9ec35f42b19c6480a9101d21a4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "357d2811-e990-4985-9f9e-b158d10d3699", "external-id": "nsx-vlan-transportzone-641", "segmentation_id": 641, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapad9dda29-93", "ovs_interfaceid": "ad9dda29-9326-472b-8c82-294a33a7faf0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "e24932f9-9063-4a16-8498-950bdfeeaa84", "address": "fa:16:3e:06:c4:2b", "network": {"id": "88dfb29d-41e0-4df7-8f70-e52cda05620c", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1355578400-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7d4ada9ec35f42b19c6480a9101d21a4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "357d2811-e990-4985-9f9e-b158d10d3699", "external-id": "nsx-vlan-transportzone-641", "segmentation_id": 641, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape24932f9-90", "ovs_interfaceid": "e24932f9-9063-4a16-8498-950bdfeeaa84", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 783.888384] env[63515]: DEBUG oslo_vmware.api [None req-ffbb9c80-abfa-457c-b756-9497c4635a40 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Task: {'id': task-1111215, 'name': ReconfigVM_Task, 'duration_secs': 1.213581} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 783.891208] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-ffbb9c80-abfa-457c-b756-9497c4635a40 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] [instance: f1d01b75-ac9d-458d-8cc2-ae64cffca4e8] Reconfigured VM instance instance-00000022 to attach disk [datastore2] f1d01b75-ac9d-458d-8cc2-ae64cffca4e8/f1d01b75-ac9d-458d-8cc2-ae64cffca4e8.vmdk or device None with type thin {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 783.892573] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1663ada-f543-459d-9a00-b0ca41901027 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.915189] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81d7d08c-60e4-4128-82ad-610a18d53446 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.937829] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af74ae99-803a-43fc-a837-979b7b32f2d5 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.959310] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a2757ac-2abf-4f0a-b325-022549a61e5e {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.969928] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-ffbb9c80-abfa-457c-b756-9497c4635a40 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] [instance: f1d01b75-ac9d-458d-8cc2-ae64cffca4e8] Powering on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 783.974036] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b2016a40-78dd-4405-b2f2-b43f05e96d9a {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.975323] env[63515]: DEBUG oslo_vmware.api [None req-8407c401-88c7-47ad-b91e-e7693f87e0f1 tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] Task: {'id': task-1111217, 'name': ReconfigVM_Task, 'duration_secs': 0.70307} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 783.976391] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-8407c401-88c7-47ad-b91e-e7693f87e0f1 tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] [instance: 192137b0-03e5-4bc4-b911-4b4f1a874f74] Reconfigured VM instance instance-00000039 to attach disk [datastore1] 192137b0-03e5-4bc4-b911-4b4f1a874f74/192137b0-03e5-4bc4-b911-4b4f1a874f74.vmdk or device None with type sparse {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 783.977531] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b03982e-5528-4954-a00a-f350648ef41b {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.980388] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-472fd758-6013-43f3-8f14-bdbff8035649 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.984808] env[63515]: DEBUG oslo_vmware.api [None req-ffbb9c80-abfa-457c-b756-9497c4635a40 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Waiting for the task: (returnval){ [ 783.984808] env[63515]: value = "task-1111218" [ 783.984808] env[63515]: _type = "Task" [ 783.984808] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 783.992330] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9eab5218-e0f0-44fa-9088-9c7e18277892 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.995566] env[63515]: DEBUG oslo_vmware.api [None req-8407c401-88c7-47ad-b91e-e7693f87e0f1 tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] Waiting for the task: (returnval){ [ 783.995566] env[63515]: value = "task-1111219" [ 783.995566] env[63515]: _type = "Task" [ 783.995566] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 784.036409] env[63515]: DEBUG oslo_vmware.api [None req-ffbb9c80-abfa-457c-b756-9497c4635a40 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Task: {'id': task-1111218, 'name': PowerOnVM_Task} progress is 33%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 784.036772] env[63515]: DEBUG oslo_concurrency.lockutils [None req-891c9525-5ba2-4bc7-b611-5e1ddfcdb81e tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Acquiring lock "d7e8dcdb-41cb-46fb-8b61-d251e7c2d372" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 784.037014] env[63515]: DEBUG oslo_concurrency.lockutils [None req-891c9525-5ba2-4bc7-b611-5e1ddfcdb81e tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Lock "d7e8dcdb-41cb-46fb-8b61-d251e7c2d372" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 784.037204] env[63515]: DEBUG nova.compute.manager [None req-891c9525-5ba2-4bc7-b611-5e1ddfcdb81e tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] [instance: d7e8dcdb-41cb-46fb-8b61-d251e7c2d372] Checking state {{(pid=63515) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 784.039047] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82a455af-9359-4b4e-8e42-d5345fcb4067 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.042381] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5757ae0b-24aa-4565-8daf-b782d0a396aa {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.048655] env[63515]: DEBUG oslo_vmware.api [None req-8407c401-88c7-47ad-b91e-e7693f87e0f1 tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] Task: {'id': task-1111219, 'name': Rename_Task} progress is 14%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 784.055424] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ab033b3-ec22-4019-9505-63fb05ec0dfe {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.059674] env[63515]: DEBUG nova.compute.manager [None req-891c9525-5ba2-4bc7-b611-5e1ddfcdb81e tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] [instance: d7e8dcdb-41cb-46fb-8b61-d251e7c2d372] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=63515) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 784.060907] env[63515]: DEBUG nova.objects.instance [None req-891c9525-5ba2-4bc7-b611-5e1ddfcdb81e tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Lazy-loading 'flavor' on Instance uuid d7e8dcdb-41cb-46fb-8b61-d251e7c2d372 {{(pid=63515) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 784.072813] env[63515]: DEBUG nova.compute.provider_tree [None req-4b5baf79-d4c3-467e-88d4-a131ea6620d7 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 784.217790] env[63515]: DEBUG oslo_concurrency.lockutils [None req-83fbd51b-fc37-49eb-8423-cdaa019d3b0c tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Acquiring lock "refresh_cache-5cbce760-0163-4b27-8ae3-e46c926c8916" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 784.217932] env[63515]: DEBUG oslo_concurrency.lockutils [None req-83fbd51b-fc37-49eb-8423-cdaa019d3b0c tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Acquired lock "refresh_cache-5cbce760-0163-4b27-8ae3-e46c926c8916" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 784.218112] env[63515]: DEBUG nova.network.neutron [None req-83fbd51b-fc37-49eb-8423-cdaa019d3b0c tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] [instance: 5cbce760-0163-4b27-8ae3-e46c926c8916] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 784.325673] env[63515]: DEBUG oslo_concurrency.lockutils [None req-8c8e056d-f20f-4ef4-b199-3c0616617ccf tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Releasing lock "refresh_cache-ed550b10-d58f-45b8-b766-198f431c3788" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 784.326475] env[63515]: DEBUG oslo_concurrency.lockutils [None req-8c8e056d-f20f-4ef4-b199-3c0616617ccf tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Acquiring lock "ed550b10-d58f-45b8-b766-198f431c3788" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 784.326547] env[63515]: DEBUG oslo_concurrency.lockutils [None req-8c8e056d-f20f-4ef4-b199-3c0616617ccf tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Acquired lock "ed550b10-d58f-45b8-b766-198f431c3788" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 784.327413] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c196f3c5-9e18-423e-8953-0ce76002a77d {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.344547] env[63515]: DEBUG nova.virt.hardware [None req-8c8e056d-f20f-4ef4-b199-3c0616617ccf tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T02:52:14Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 784.344822] env[63515]: DEBUG nova.virt.hardware [None req-8c8e056d-f20f-4ef4-b199-3c0616617ccf tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Flavor limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 784.344996] env[63515]: DEBUG nova.virt.hardware [None req-8c8e056d-f20f-4ef4-b199-3c0616617ccf tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Image limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 784.345276] env[63515]: DEBUG nova.virt.hardware [None req-8c8e056d-f20f-4ef4-b199-3c0616617ccf tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Flavor pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 784.345381] env[63515]: DEBUG nova.virt.hardware [None req-8c8e056d-f20f-4ef4-b199-3c0616617ccf tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Image pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 784.345602] env[63515]: DEBUG nova.virt.hardware [None req-8c8e056d-f20f-4ef4-b199-3c0616617ccf tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 784.345792] env[63515]: DEBUG nova.virt.hardware [None req-8c8e056d-f20f-4ef4-b199-3c0616617ccf tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 784.345974] env[63515]: DEBUG nova.virt.hardware [None req-8c8e056d-f20f-4ef4-b199-3c0616617ccf tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 784.346179] env[63515]: DEBUG nova.virt.hardware [None req-8c8e056d-f20f-4ef4-b199-3c0616617ccf tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Got 1 possible topologies {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 784.346370] env[63515]: DEBUG nova.virt.hardware [None req-8c8e056d-f20f-4ef4-b199-3c0616617ccf tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 784.346564] env[63515]: DEBUG nova.virt.hardware [None req-8c8e056d-f20f-4ef4-b199-3c0616617ccf tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 784.353316] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-8c8e056d-f20f-4ef4-b199-3c0616617ccf tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: ed550b10-d58f-45b8-b766-198f431c3788] Reconfiguring VM to attach interface {{(pid=63515) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1923}} [ 784.353669] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-060163b7-18bf-434f-bb14-32aead795553 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.373492] env[63515]: DEBUG oslo_vmware.api [None req-8c8e056d-f20f-4ef4-b199-3c0616617ccf tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Waiting for the task: (returnval){ [ 784.373492] env[63515]: value = "task-1111220" [ 784.373492] env[63515]: _type = "Task" [ 784.373492] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 784.383450] env[63515]: DEBUG oslo_vmware.api [None req-8c8e056d-f20f-4ef4-b199-3c0616617ccf tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Task: {'id': task-1111220, 'name': ReconfigVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 784.498029] env[63515]: DEBUG oslo_vmware.api [None req-ffbb9c80-abfa-457c-b756-9497c4635a40 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Task: {'id': task-1111218, 'name': PowerOnVM_Task} progress is 84%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 784.506904] env[63515]: DEBUG oslo_vmware.api [None req-8407c401-88c7-47ad-b91e-e7693f87e0f1 tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] Task: {'id': task-1111219, 'name': Rename_Task, 'duration_secs': 0.178163} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 784.507695] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-8407c401-88c7-47ad-b91e-e7693f87e0f1 tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] [instance: 192137b0-03e5-4bc4-b911-4b4f1a874f74] Powering on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 784.507695] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5d57ee94-ad2c-40e4-b39e-3d1f11518102 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.514751] env[63515]: DEBUG oslo_vmware.api [None req-8407c401-88c7-47ad-b91e-e7693f87e0f1 tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] Waiting for the task: (returnval){ [ 784.514751] env[63515]: value = "task-1111221" [ 784.514751] env[63515]: _type = "Task" [ 784.514751] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 784.522740] env[63515]: DEBUG oslo_vmware.api [None req-8407c401-88c7-47ad-b91e-e7693f87e0f1 tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] Task: {'id': task-1111221, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 784.542026] env[63515]: DEBUG nova.compute.manager [None req-fa0cfe6d-3452-46d4-9987-2cb422956d62 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] [instance: 63ce797d-7180-4209-ac2c-81978bf7607a] Start spawning the instance on the hypervisor. {{(pid=63515) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 784.567186] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-891c9525-5ba2-4bc7-b611-5e1ddfcdb81e tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] [instance: d7e8dcdb-41cb-46fb-8b61-d251e7c2d372] Powering off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 784.567519] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-470e4019-5219-4f07-af9a-4ae7395987f1 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.573388] env[63515]: DEBUG nova.virt.hardware [None req-fa0cfe6d-3452-46d4-9987-2cb422956d62 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T02:52:14Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T02:51:56Z,direct_url=,disk_format='vmdk',id=8a120570-cb06-4099-b262-554ca0ad15c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b569255cc43e42e7ae3f2b4ad37c6ef4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T02:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 784.573648] env[63515]: DEBUG nova.virt.hardware [None req-fa0cfe6d-3452-46d4-9987-2cb422956d62 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Flavor limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 784.573885] env[63515]: DEBUG nova.virt.hardware [None req-fa0cfe6d-3452-46d4-9987-2cb422956d62 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Image limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 784.573990] env[63515]: DEBUG nova.virt.hardware [None req-fa0cfe6d-3452-46d4-9987-2cb422956d62 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Flavor pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 784.574648] env[63515]: DEBUG nova.virt.hardware [None req-fa0cfe6d-3452-46d4-9987-2cb422956d62 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Image pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 784.574648] env[63515]: DEBUG nova.virt.hardware [None req-fa0cfe6d-3452-46d4-9987-2cb422956d62 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 784.574648] env[63515]: DEBUG nova.virt.hardware [None req-fa0cfe6d-3452-46d4-9987-2cb422956d62 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 784.574792] env[63515]: DEBUG nova.virt.hardware [None req-fa0cfe6d-3452-46d4-9987-2cb422956d62 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 784.574903] env[63515]: DEBUG nova.virt.hardware [None req-fa0cfe6d-3452-46d4-9987-2cb422956d62 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Got 1 possible topologies {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 784.575042] env[63515]: DEBUG nova.virt.hardware [None req-fa0cfe6d-3452-46d4-9987-2cb422956d62 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 784.575279] env[63515]: DEBUG nova.virt.hardware [None req-fa0cfe6d-3452-46d4-9987-2cb422956d62 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 784.576594] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0fee4737-9538-4f81-8204-24e2c3f2e64a {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.581043] env[63515]: DEBUG nova.scheduler.client.report [None req-4b5baf79-d4c3-467e-88d4-a131ea6620d7 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 784.584371] env[63515]: DEBUG oslo_vmware.api [None req-891c9525-5ba2-4bc7-b611-5e1ddfcdb81e tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Waiting for the task: (returnval){ [ 784.584371] env[63515]: value = "task-1111222" [ 784.584371] env[63515]: _type = "Task" [ 784.584371] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 784.593363] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55c65355-8847-4755-bf5c-522ffcdb983c {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.601440] env[63515]: DEBUG oslo_vmware.api [None req-891c9525-5ba2-4bc7-b611-5e1ddfcdb81e tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Task: {'id': task-1111222, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 784.772013] env[63515]: DEBUG nova.network.neutron [None req-83fbd51b-fc37-49eb-8423-cdaa019d3b0c tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] [instance: 5cbce760-0163-4b27-8ae3-e46c926c8916] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 784.890735] env[63515]: DEBUG oslo_vmware.api [None req-8c8e056d-f20f-4ef4-b199-3c0616617ccf tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Task: {'id': task-1111220, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 784.961547] env[63515]: DEBUG nova.network.neutron [None req-fa0cfe6d-3452-46d4-9987-2cb422956d62 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] [instance: 63ce797d-7180-4209-ac2c-81978bf7607a] Successfully updated port: 25ea9d0d-3405-42da-a576-a2a1f37fa75a {{(pid=63515) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 785.000125] env[63515]: DEBUG oslo_vmware.api [None req-ffbb9c80-abfa-457c-b756-9497c4635a40 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Task: {'id': task-1111218, 'name': PowerOnVM_Task, 'duration_secs': 0.708094} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 785.000125] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-ffbb9c80-abfa-457c-b756-9497c4635a40 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] [instance: f1d01b75-ac9d-458d-8cc2-ae64cffca4e8] Powered on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 785.007746] env[63515]: DEBUG nova.network.neutron [None req-83fbd51b-fc37-49eb-8423-cdaa019d3b0c tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] [instance: 5cbce760-0163-4b27-8ae3-e46c926c8916] Updating instance_info_cache with network_info: [{"id": "5fa807c4-5ce1-492a-bea3-d8f7a65d898f", "address": "fa:16:3e:23:bb:83", "network": {"id": "850bfa95-3df6-4019-8260-ade20ec58799", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-832000501-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "08c9aa51b1864410809f8877c7cd49c4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3cf748a8-7ae0-4dca-817d-e727c30d72f4", "external-id": "nsx-vlan-transportzone-853", "segmentation_id": 853, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5fa807c4-5c", "ovs_interfaceid": "5fa807c4-5ce1-492a-bea3-d8f7a65d898f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 785.013843] env[63515]: DEBUG nova.compute.manager [req-b675062a-0577-4a36-8b50-cf4e59ba01e1 req-12414bb2-04d8-4d3f-9c7b-a376c72d34ce service nova] [instance: 5cbce760-0163-4b27-8ae3-e46c926c8916] Received event network-vif-plugged-5fa807c4-5ce1-492a-bea3-d8f7a65d898f {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 785.013843] env[63515]: DEBUG oslo_concurrency.lockutils [req-b675062a-0577-4a36-8b50-cf4e59ba01e1 req-12414bb2-04d8-4d3f-9c7b-a376c72d34ce service nova] Acquiring lock "5cbce760-0163-4b27-8ae3-e46c926c8916-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 785.013843] env[63515]: DEBUG oslo_concurrency.lockutils [req-b675062a-0577-4a36-8b50-cf4e59ba01e1 req-12414bb2-04d8-4d3f-9c7b-a376c72d34ce service nova] Lock "5cbce760-0163-4b27-8ae3-e46c926c8916-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 785.013843] env[63515]: DEBUG oslo_concurrency.lockutils [req-b675062a-0577-4a36-8b50-cf4e59ba01e1 req-12414bb2-04d8-4d3f-9c7b-a376c72d34ce service nova] Lock "5cbce760-0163-4b27-8ae3-e46c926c8916-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 785.013843] env[63515]: DEBUG nova.compute.manager [req-b675062a-0577-4a36-8b50-cf4e59ba01e1 req-12414bb2-04d8-4d3f-9c7b-a376c72d34ce service nova] [instance: 5cbce760-0163-4b27-8ae3-e46c926c8916] No waiting events found dispatching network-vif-plugged-5fa807c4-5ce1-492a-bea3-d8f7a65d898f {{(pid=63515) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 785.013843] env[63515]: WARNING nova.compute.manager [req-b675062a-0577-4a36-8b50-cf4e59ba01e1 req-12414bb2-04d8-4d3f-9c7b-a376c72d34ce service nova] [instance: 5cbce760-0163-4b27-8ae3-e46c926c8916] Received unexpected event network-vif-plugged-5fa807c4-5ce1-492a-bea3-d8f7a65d898f for instance with vm_state building and task_state spawning. [ 785.026560] env[63515]: DEBUG oslo_vmware.api [None req-8407c401-88c7-47ad-b91e-e7693f87e0f1 tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] Task: {'id': task-1111221, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 785.064729] env[63515]: DEBUG oslo_concurrency.lockutils [None req-49d5860e-9bde-4499-902f-a44ae9cea035 tempest-ServerMetadataNegativeTestJSON-502053314 tempest-ServerMetadataNegativeTestJSON-502053314-project-member] Acquiring lock "9d18c7b6-ce8e-4042-9e70-696bb7f57cb3" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 785.065099] env[63515]: DEBUG oslo_concurrency.lockutils [None req-49d5860e-9bde-4499-902f-a44ae9cea035 tempest-ServerMetadataNegativeTestJSON-502053314 tempest-ServerMetadataNegativeTestJSON-502053314-project-member] Lock "9d18c7b6-ce8e-4042-9e70-696bb7f57cb3" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 785.086652] env[63515]: DEBUG oslo_concurrency.lockutils [None req-4b5baf79-d4c3-467e-88d4-a131ea6620d7 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.601s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 785.087426] env[63515]: DEBUG nova.compute.manager [None req-4b5baf79-d4c3-467e-88d4-a131ea6620d7 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] [instance: e666825c-ff4e-4a0e-93c0-43c00f167bbb] Start building networks asynchronously for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 785.091374] env[63515]: DEBUG oslo_concurrency.lockutils [None req-64ffe526-19dd-467f-8550-ae608ae5d75b tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 26.988s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 785.091572] env[63515]: DEBUG oslo_concurrency.lockutils [None req-64ffe526-19dd-467f-8550-ae608ae5d75b tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 785.093790] env[63515]: DEBUG oslo_concurrency.lockutils [None req-938cccb1-5a46-437d-9e93-3058a8134cd5 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 26.216s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 785.094045] env[63515]: DEBUG nova.objects.instance [None req-938cccb1-5a46-437d-9e93-3058a8134cd5 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Lazy-loading 'resources' on Instance uuid 37e6c27e-317b-45d2-bd55-2fd78ccf009f {{(pid=63515) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 785.106786] env[63515]: DEBUG oslo_vmware.api [None req-891c9525-5ba2-4bc7-b611-5e1ddfcdb81e tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Task: {'id': task-1111222, 'name': PowerOffVM_Task, 'duration_secs': 0.223764} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 785.107855] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-891c9525-5ba2-4bc7-b611-5e1ddfcdb81e tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] [instance: d7e8dcdb-41cb-46fb-8b61-d251e7c2d372] Powered off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 785.108126] env[63515]: DEBUG nova.compute.manager [None req-891c9525-5ba2-4bc7-b611-5e1ddfcdb81e tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] [instance: d7e8dcdb-41cb-46fb-8b61-d251e7c2d372] Checking state {{(pid=63515) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 785.109178] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-515a9271-b2a8-4961-a762-9c9b55fd52cc {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.244046] env[63515]: INFO nova.scheduler.client.report [None req-64ffe526-19dd-467f-8550-ae608ae5d75b tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Deleted allocations for instance 452f3ca2-6141-43b2-a77a-c9ab5754192d [ 785.387790] env[63515]: DEBUG oslo_vmware.api [None req-8c8e056d-f20f-4ef4-b199-3c0616617ccf tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Task: {'id': task-1111220, 'name': ReconfigVM_Task, 'duration_secs': 0.623284} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 785.388363] env[63515]: DEBUG oslo_concurrency.lockutils [None req-8c8e056d-f20f-4ef4-b199-3c0616617ccf tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Releasing lock "ed550b10-d58f-45b8-b766-198f431c3788" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 785.388585] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-8c8e056d-f20f-4ef4-b199-3c0616617ccf tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: ed550b10-d58f-45b8-b766-198f431c3788] Reconfigured VM to attach interface {{(pid=63515) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1937}} [ 785.465148] env[63515]: DEBUG oslo_concurrency.lockutils [None req-fa0cfe6d-3452-46d4-9987-2cb422956d62 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Acquiring lock "refresh_cache-63ce797d-7180-4209-ac2c-81978bf7607a" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 785.465338] env[63515]: DEBUG oslo_concurrency.lockutils [None req-fa0cfe6d-3452-46d4-9987-2cb422956d62 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Acquired lock "refresh_cache-63ce797d-7180-4209-ac2c-81978bf7607a" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 785.465493] env[63515]: DEBUG nova.network.neutron [None req-fa0cfe6d-3452-46d4-9987-2cb422956d62 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] [instance: 63ce797d-7180-4209-ac2c-81978bf7607a] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 785.513044] env[63515]: DEBUG oslo_concurrency.lockutils [None req-83fbd51b-fc37-49eb-8423-cdaa019d3b0c tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Releasing lock "refresh_cache-5cbce760-0163-4b27-8ae3-e46c926c8916" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 785.513339] env[63515]: DEBUG nova.compute.manager [None req-83fbd51b-fc37-49eb-8423-cdaa019d3b0c tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] [instance: 5cbce760-0163-4b27-8ae3-e46c926c8916] Instance network_info: |[{"id": "5fa807c4-5ce1-492a-bea3-d8f7a65d898f", "address": "fa:16:3e:23:bb:83", "network": {"id": "850bfa95-3df6-4019-8260-ade20ec58799", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-832000501-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "08c9aa51b1864410809f8877c7cd49c4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3cf748a8-7ae0-4dca-817d-e727c30d72f4", "external-id": "nsx-vlan-transportzone-853", "segmentation_id": 853, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5fa807c4-5c", "ovs_interfaceid": "5fa807c4-5ce1-492a-bea3-d8f7a65d898f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 785.513741] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-83fbd51b-fc37-49eb-8423-cdaa019d3b0c tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] [instance: 5cbce760-0163-4b27-8ae3-e46c926c8916] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:23:bb:83', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '3cf748a8-7ae0-4dca-817d-e727c30d72f4', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '5fa807c4-5ce1-492a-bea3-d8f7a65d898f', 'vif_model': 'vmxnet3'}] {{(pid=63515) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 785.521645] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-83fbd51b-fc37-49eb-8423-cdaa019d3b0c tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Creating folder: Project (08c9aa51b1864410809f8877c7cd49c4). Parent ref: group-v243370. {{(pid=63515) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 785.522191] env[63515]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-600f29de-5106-4ee8-99ad-895139de6491 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.533483] env[63515]: DEBUG oslo_vmware.api [None req-8407c401-88c7-47ad-b91e-e7693f87e0f1 tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] Task: {'id': task-1111221, 'name': PowerOnVM_Task, 'duration_secs': 0.747486} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 785.533742] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-8407c401-88c7-47ad-b91e-e7693f87e0f1 tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] [instance: 192137b0-03e5-4bc4-b911-4b4f1a874f74] Powered on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 785.533983] env[63515]: INFO nova.compute.manager [None req-8407c401-88c7-47ad-b91e-e7693f87e0f1 tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] [instance: 192137b0-03e5-4bc4-b911-4b4f1a874f74] Took 8.94 seconds to spawn the instance on the hypervisor. [ 785.534198] env[63515]: DEBUG nova.compute.manager [None req-8407c401-88c7-47ad-b91e-e7693f87e0f1 tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] [instance: 192137b0-03e5-4bc4-b911-4b4f1a874f74] Checking state {{(pid=63515) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 785.534934] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d00693ce-ed33-4b27-bf68-9ce987fd99ae {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.538657] env[63515]: INFO nova.virt.vmwareapi.vm_util [None req-83fbd51b-fc37-49eb-8423-cdaa019d3b0c tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Created folder: Project (08c9aa51b1864410809f8877c7cd49c4) in parent group-v243370. [ 785.539024] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-83fbd51b-fc37-49eb-8423-cdaa019d3b0c tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Creating folder: Instances. Parent ref: group-v243460. {{(pid=63515) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 785.539363] env[63515]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-cbacc14e-0a52-4dd8-b77b-5aab4910089b {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.551317] env[63515]: INFO nova.virt.vmwareapi.vm_util [None req-83fbd51b-fc37-49eb-8423-cdaa019d3b0c tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Created folder: Instances in parent group-v243460. [ 785.552699] env[63515]: DEBUG oslo.service.loopingcall [None req-83fbd51b-fc37-49eb-8423-cdaa019d3b0c tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 785.552699] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5cbce760-0163-4b27-8ae3-e46c926c8916] Creating VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 785.552699] env[63515]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c9814731-2049-4216-9ec3-2a5d94860c1a {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.567434] env[63515]: DEBUG nova.compute.manager [None req-49d5860e-9bde-4499-902f-a44ae9cea035 tempest-ServerMetadataNegativeTestJSON-502053314 tempest-ServerMetadataNegativeTestJSON-502053314-project-member] [instance: 9d18c7b6-ce8e-4042-9e70-696bb7f57cb3] Starting instance... {{(pid=63515) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 785.574563] env[63515]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 785.574563] env[63515]: value = "task-1111225" [ 785.574563] env[63515]: _type = "Task" [ 785.574563] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 785.582660] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1111225, 'name': CreateVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 785.593030] env[63515]: DEBUG nova.compute.utils [None req-4b5baf79-d4c3-467e-88d4-a131ea6620d7 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] Using /dev/sd instead of None {{(pid=63515) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 785.593961] env[63515]: DEBUG nova.compute.manager [None req-4b5baf79-d4c3-467e-88d4-a131ea6620d7 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] [instance: e666825c-ff4e-4a0e-93c0-43c00f167bbb] Allocating IP information in the background. {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 785.594094] env[63515]: DEBUG nova.network.neutron [None req-4b5baf79-d4c3-467e-88d4-a131ea6620d7 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] [instance: e666825c-ff4e-4a0e-93c0-43c00f167bbb] allocate_for_instance() {{(pid=63515) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 785.622340] env[63515]: DEBUG oslo_concurrency.lockutils [None req-891c9525-5ba2-4bc7-b611-5e1ddfcdb81e tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Lock "d7e8dcdb-41cb-46fb-8b61-d251e7c2d372" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.585s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 785.657586] env[63515]: DEBUG nova.policy [None req-4b5baf79-d4c3-467e-88d4-a131ea6620d7 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8b6f6567b7ce4f44a860c1f5cca16964', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a1f130751200408982bffb31a54bbcea', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63515) authorize /opt/stack/nova/nova/policy.py:201}} [ 785.751187] env[63515]: DEBUG oslo_concurrency.lockutils [None req-64ffe526-19dd-467f-8550-ae608ae5d75b tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Lock "452f3ca2-6141-43b2-a77a-c9ab5754192d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 30.561s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 785.893855] env[63515]: DEBUG oslo_concurrency.lockutils [None req-8c8e056d-f20f-4ef4-b199-3c0616617ccf tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Lock "interface-ed550b10-d58f-45b8-b766-198f431c3788-None" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 7.091s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 785.974845] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13c5f4de-bcc5-4be6-9b06-f844df96f01b {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.982684] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7fb65d65-d85f-4c0b-9b5e-bdaf1048ebdf {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.020224] env[63515]: DEBUG nova.network.neutron [None req-fa0cfe6d-3452-46d4-9987-2cb422956d62 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] [instance: 63ce797d-7180-4209-ac2c-81978bf7607a] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 786.026455] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac2b4e34-0608-4a8f-ab30-1ffb14f9ff1b {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.026455] env[63515]: INFO nova.compute.manager [None req-ffbb9c80-abfa-457c-b756-9497c4635a40 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] [instance: f1d01b75-ac9d-458d-8cc2-ae64cffca4e8] Updating instance to original state: 'active' [ 786.036539] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-082c2145-c301-4731-a97d-efbceb25dc62 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.058103] env[63515]: DEBUG nova.compute.provider_tree [None req-938cccb1-5a46-437d-9e93-3058a8134cd5 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 786.061561] env[63515]: DEBUG nova.network.neutron [None req-4b5baf79-d4c3-467e-88d4-a131ea6620d7 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] [instance: e666825c-ff4e-4a0e-93c0-43c00f167bbb] Successfully created port: 7872b4c3-d54f-4613-85d2-f568ec3cb47f {{(pid=63515) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 786.068892] env[63515]: INFO nova.compute.manager [None req-8407c401-88c7-47ad-b91e-e7693f87e0f1 tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] [instance: 192137b0-03e5-4bc4-b911-4b4f1a874f74] Took 40.47 seconds to build instance. [ 786.089766] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1111225, 'name': CreateVM_Task, 'duration_secs': 0.416583} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 786.089939] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5cbce760-0163-4b27-8ae3-e46c926c8916] Created VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 786.090735] env[63515]: DEBUG oslo_concurrency.lockutils [None req-83fbd51b-fc37-49eb-8423-cdaa019d3b0c tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 786.090924] env[63515]: DEBUG oslo_concurrency.lockutils [None req-83fbd51b-fc37-49eb-8423-cdaa019d3b0c tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 786.091292] env[63515]: DEBUG oslo_concurrency.lockutils [None req-83fbd51b-fc37-49eb-8423-cdaa019d3b0c tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 786.091919] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9f75dc3d-0147-4e51-80ad-9e759c498f89 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.095041] env[63515]: DEBUG oslo_concurrency.lockutils [None req-49d5860e-9bde-4499-902f-a44ae9cea035 tempest-ServerMetadataNegativeTestJSON-502053314 tempest-ServerMetadataNegativeTestJSON-502053314-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 786.096852] env[63515]: DEBUG nova.compute.manager [None req-4b5baf79-d4c3-467e-88d4-a131ea6620d7 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] [instance: e666825c-ff4e-4a0e-93c0-43c00f167bbb] Start building block device mappings for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 786.101273] env[63515]: DEBUG oslo_vmware.api [None req-83fbd51b-fc37-49eb-8423-cdaa019d3b0c tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Waiting for the task: (returnval){ [ 786.101273] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]52f08120-50c9-6ac1-9c91-d780e0c79e11" [ 786.101273] env[63515]: _type = "Task" [ 786.101273] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 786.113950] env[63515]: DEBUG oslo_vmware.api [None req-83fbd51b-fc37-49eb-8423-cdaa019d3b0c tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52f08120-50c9-6ac1-9c91-d780e0c79e11, 'name': SearchDatastore_Task, 'duration_secs': 0.010519} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 786.113950] env[63515]: DEBUG oslo_concurrency.lockutils [None req-83fbd51b-fc37-49eb-8423-cdaa019d3b0c tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 786.113950] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-83fbd51b-fc37-49eb-8423-cdaa019d3b0c tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] [instance: 5cbce760-0163-4b27-8ae3-e46c926c8916] Processing image 8a120570-cb06-4099-b262-554ca0ad15c5 {{(pid=63515) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 786.114235] env[63515]: DEBUG oslo_concurrency.lockutils [None req-83fbd51b-fc37-49eb-8423-cdaa019d3b0c tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 786.114422] env[63515]: DEBUG oslo_concurrency.lockutils [None req-83fbd51b-fc37-49eb-8423-cdaa019d3b0c tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 786.114609] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-83fbd51b-fc37-49eb-8423-cdaa019d3b0c tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 786.114909] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-071b0f4a-7bf9-4762-91e3-dcba7c57448c {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.123054] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-83fbd51b-fc37-49eb-8423-cdaa019d3b0c tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 786.123282] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-83fbd51b-fc37-49eb-8423-cdaa019d3b0c tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63515) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 786.124537] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9c3aa33f-0e8d-47b2-a008-834c5cd9d684 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.130579] env[63515]: DEBUG oslo_vmware.api [None req-83fbd51b-fc37-49eb-8423-cdaa019d3b0c tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Waiting for the task: (returnval){ [ 786.130579] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]520b195c-de7c-1ee2-abfd-27dd91e0746a" [ 786.130579] env[63515]: _type = "Task" [ 786.130579] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 786.142323] env[63515]: DEBUG oslo_vmware.api [None req-83fbd51b-fc37-49eb-8423-cdaa019d3b0c tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]520b195c-de7c-1ee2-abfd-27dd91e0746a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 786.330802] env[63515]: DEBUG nova.network.neutron [None req-fa0cfe6d-3452-46d4-9987-2cb422956d62 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] [instance: 63ce797d-7180-4209-ac2c-81978bf7607a] Updating instance_info_cache with network_info: [{"id": "25ea9d0d-3405-42da-a576-a2a1f37fa75a", "address": "fa:16:3e:35:5c:ab", "network": {"id": "850bfa95-3df6-4019-8260-ade20ec58799", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-832000501-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "08c9aa51b1864410809f8877c7cd49c4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3cf748a8-7ae0-4dca-817d-e727c30d72f4", "external-id": "nsx-vlan-transportzone-853", "segmentation_id": 853, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap25ea9d0d-34", "ovs_interfaceid": "25ea9d0d-3405-42da-a576-a2a1f37fa75a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 786.571703] env[63515]: DEBUG nova.scheduler.client.report [None req-938cccb1-5a46-437d-9e93-3058a8134cd5 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 786.574658] env[63515]: DEBUG oslo_concurrency.lockutils [None req-8407c401-88c7-47ad-b91e-e7693f87e0f1 tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] Lock "192137b0-03e5-4bc4-b911-4b4f1a874f74" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 98.548s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 786.644588] env[63515]: DEBUG oslo_vmware.api [None req-83fbd51b-fc37-49eb-8423-cdaa019d3b0c tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]520b195c-de7c-1ee2-abfd-27dd91e0746a, 'name': SearchDatastore_Task, 'duration_secs': 0.009784} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 786.645445] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-60c2bd45-7847-4d50-9bde-29ebc3a88557 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.651875] env[63515]: DEBUG oslo_vmware.api [None req-83fbd51b-fc37-49eb-8423-cdaa019d3b0c tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Waiting for the task: (returnval){ [ 786.651875] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]52baeeb8-eb56-7df6-cccc-f4328ed31522" [ 786.651875] env[63515]: _type = "Task" [ 786.651875] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 786.659924] env[63515]: DEBUG oslo_vmware.api [None req-83fbd51b-fc37-49eb-8423-cdaa019d3b0c tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52baeeb8-eb56-7df6-cccc-f4328ed31522, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 786.794096] env[63515]: DEBUG nova.compute.manager [req-c733811e-ef67-46a5-a2d8-172f51057db6 req-32296e40-52e9-4881-ad7b-93b3d7915b31 service nova] [instance: ed550b10-d58f-45b8-b766-198f431c3788] Received event network-changed-e24932f9-9063-4a16-8498-950bdfeeaa84 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 786.794336] env[63515]: DEBUG nova.compute.manager [req-c733811e-ef67-46a5-a2d8-172f51057db6 req-32296e40-52e9-4881-ad7b-93b3d7915b31 service nova] [instance: ed550b10-d58f-45b8-b766-198f431c3788] Refreshing instance network info cache due to event network-changed-e24932f9-9063-4a16-8498-950bdfeeaa84. {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 786.794567] env[63515]: DEBUG oslo_concurrency.lockutils [req-c733811e-ef67-46a5-a2d8-172f51057db6 req-32296e40-52e9-4881-ad7b-93b3d7915b31 service nova] Acquiring lock "refresh_cache-ed550b10-d58f-45b8-b766-198f431c3788" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 786.794715] env[63515]: DEBUG oslo_concurrency.lockutils [req-c733811e-ef67-46a5-a2d8-172f51057db6 req-32296e40-52e9-4881-ad7b-93b3d7915b31 service nova] Acquired lock "refresh_cache-ed550b10-d58f-45b8-b766-198f431c3788" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 786.795175] env[63515]: DEBUG nova.network.neutron [req-c733811e-ef67-46a5-a2d8-172f51057db6 req-32296e40-52e9-4881-ad7b-93b3d7915b31 service nova] [instance: ed550b10-d58f-45b8-b766-198f431c3788] Refreshing network info cache for port e24932f9-9063-4a16-8498-950bdfeeaa84 {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 786.834220] env[63515]: DEBUG oslo_concurrency.lockutils [None req-fa0cfe6d-3452-46d4-9987-2cb422956d62 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Releasing lock "refresh_cache-63ce797d-7180-4209-ac2c-81978bf7607a" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 786.834596] env[63515]: DEBUG nova.compute.manager [None req-fa0cfe6d-3452-46d4-9987-2cb422956d62 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] [instance: 63ce797d-7180-4209-ac2c-81978bf7607a] Instance network_info: |[{"id": "25ea9d0d-3405-42da-a576-a2a1f37fa75a", "address": "fa:16:3e:35:5c:ab", "network": {"id": "850bfa95-3df6-4019-8260-ade20ec58799", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-832000501-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "08c9aa51b1864410809f8877c7cd49c4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3cf748a8-7ae0-4dca-817d-e727c30d72f4", "external-id": "nsx-vlan-transportzone-853", "segmentation_id": 853, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap25ea9d0d-34", "ovs_interfaceid": "25ea9d0d-3405-42da-a576-a2a1f37fa75a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 786.835096] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-fa0cfe6d-3452-46d4-9987-2cb422956d62 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] [instance: 63ce797d-7180-4209-ac2c-81978bf7607a] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:35:5c:ab', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '3cf748a8-7ae0-4dca-817d-e727c30d72f4', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '25ea9d0d-3405-42da-a576-a2a1f37fa75a', 'vif_model': 'vmxnet3'}] {{(pid=63515) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 786.843248] env[63515]: DEBUG oslo.service.loopingcall [None req-fa0cfe6d-3452-46d4-9987-2cb422956d62 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 786.843818] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 63ce797d-7180-4209-ac2c-81978bf7607a] Creating VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 786.844661] env[63515]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a3f76f37-e650-4c4a-ae29-a2a7dce1d41b {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.879573] env[63515]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 786.879573] env[63515]: value = "task-1111226" [ 786.879573] env[63515]: _type = "Task" [ 786.879573] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 786.895237] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1111226, 'name': CreateVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 787.076878] env[63515]: DEBUG oslo_concurrency.lockutils [None req-938cccb1-5a46-437d-9e93-3058a8134cd5 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.983s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 787.079992] env[63515]: DEBUG oslo_concurrency.lockutils [None req-66c3e202-5a3e-433e-b543-4bfb02c8ad30 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 27.455s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 787.081468] env[63515]: INFO nova.compute.claims [None req-66c3e202-5a3e-433e-b543-4bfb02c8ad30 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775-project-member] [instance: 4e453127-1f3e-40ea-819f-6678479826c8] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 787.108784] env[63515]: DEBUG nova.compute.manager [None req-4b5baf79-d4c3-467e-88d4-a131ea6620d7 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] [instance: e666825c-ff4e-4a0e-93c0-43c00f167bbb] Start spawning the instance on the hypervisor. {{(pid=63515) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 787.139959] env[63515]: DEBUG nova.virt.hardware [None req-4b5baf79-d4c3-467e-88d4-a131ea6620d7 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T02:52:14Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T02:51:56Z,direct_url=,disk_format='vmdk',id=8a120570-cb06-4099-b262-554ca0ad15c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b569255cc43e42e7ae3f2b4ad37c6ef4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T02:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 787.140282] env[63515]: DEBUG nova.virt.hardware [None req-4b5baf79-d4c3-467e-88d4-a131ea6620d7 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] Flavor limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 787.140493] env[63515]: DEBUG nova.virt.hardware [None req-4b5baf79-d4c3-467e-88d4-a131ea6620d7 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] Image limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 787.140726] env[63515]: DEBUG nova.virt.hardware [None req-4b5baf79-d4c3-467e-88d4-a131ea6620d7 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] Flavor pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 787.140904] env[63515]: DEBUG nova.virt.hardware [None req-4b5baf79-d4c3-467e-88d4-a131ea6620d7 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] Image pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 787.141093] env[63515]: DEBUG nova.virt.hardware [None req-4b5baf79-d4c3-467e-88d4-a131ea6620d7 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 787.141359] env[63515]: DEBUG nova.virt.hardware [None req-4b5baf79-d4c3-467e-88d4-a131ea6620d7 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 787.141540] env[63515]: DEBUG nova.virt.hardware [None req-4b5baf79-d4c3-467e-88d4-a131ea6620d7 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 787.142398] env[63515]: DEBUG nova.virt.hardware [None req-4b5baf79-d4c3-467e-88d4-a131ea6620d7 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] Got 1 possible topologies {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 787.142653] env[63515]: DEBUG nova.virt.hardware [None req-4b5baf79-d4c3-467e-88d4-a131ea6620d7 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 787.142856] env[63515]: DEBUG nova.virt.hardware [None req-4b5baf79-d4c3-467e-88d4-a131ea6620d7 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 787.143842] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1d5e3fe-079b-4db7-8535-fc11b92692bc {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.152504] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d21ac6f-604e-4fee-8905-abc6e3cf1e78 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.169718] env[63515]: DEBUG oslo_vmware.api [None req-83fbd51b-fc37-49eb-8423-cdaa019d3b0c tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52baeeb8-eb56-7df6-cccc-f4328ed31522, 'name': SearchDatastore_Task, 'duration_secs': 0.025251} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 787.179857] env[63515]: DEBUG oslo_concurrency.lockutils [None req-83fbd51b-fc37-49eb-8423-cdaa019d3b0c tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 787.180471] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-83fbd51b-fc37-49eb-8423-cdaa019d3b0c tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk to [datastore1] 5cbce760-0163-4b27-8ae3-e46c926c8916/5cbce760-0163-4b27-8ae3-e46c926c8916.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 787.181334] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-9004fbb6-a349-4f64-bd7b-9dd43a53df2c {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.183793] env[63515]: INFO nova.scheduler.client.report [None req-938cccb1-5a46-437d-9e93-3058a8134cd5 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Deleted allocations for instance 37e6c27e-317b-45d2-bd55-2fd78ccf009f [ 787.189944] env[63515]: DEBUG oslo_vmware.api [None req-83fbd51b-fc37-49eb-8423-cdaa019d3b0c tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Waiting for the task: (returnval){ [ 787.189944] env[63515]: value = "task-1111227" [ 787.189944] env[63515]: _type = "Task" [ 787.189944] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 787.199754] env[63515]: DEBUG oslo_vmware.api [None req-83fbd51b-fc37-49eb-8423-cdaa019d3b0c tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Task: {'id': task-1111227, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 787.390600] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1111226, 'name': CreateVM_Task, 'duration_secs': 0.357972} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 787.391114] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 63ce797d-7180-4209-ac2c-81978bf7607a] Created VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 787.391569] env[63515]: DEBUG oslo_concurrency.lockutils [None req-fa0cfe6d-3452-46d4-9987-2cb422956d62 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 787.391694] env[63515]: DEBUG oslo_concurrency.lockutils [None req-fa0cfe6d-3452-46d4-9987-2cb422956d62 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 787.392069] env[63515]: DEBUG oslo_concurrency.lockutils [None req-fa0cfe6d-3452-46d4-9987-2cb422956d62 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 787.392340] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5880882a-199a-418e-8bf6-494857e355d2 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.397290] env[63515]: DEBUG oslo_vmware.api [None req-fa0cfe6d-3452-46d4-9987-2cb422956d62 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Waiting for the task: (returnval){ [ 787.397290] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]5221cd63-2acf-d2c6-13ca-00731dd52590" [ 787.397290] env[63515]: _type = "Task" [ 787.397290] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 787.409159] env[63515]: DEBUG oslo_vmware.api [None req-fa0cfe6d-3452-46d4-9987-2cb422956d62 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]5221cd63-2acf-d2c6-13ca-00731dd52590, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 787.567939] env[63515]: DEBUG nova.network.neutron [req-c733811e-ef67-46a5-a2d8-172f51057db6 req-32296e40-52e9-4881-ad7b-93b3d7915b31 service nova] [instance: ed550b10-d58f-45b8-b766-198f431c3788] Updated VIF entry in instance network info cache for port e24932f9-9063-4a16-8498-950bdfeeaa84. {{(pid=63515) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 787.568342] env[63515]: DEBUG nova.network.neutron [req-c733811e-ef67-46a5-a2d8-172f51057db6 req-32296e40-52e9-4881-ad7b-93b3d7915b31 service nova] [instance: ed550b10-d58f-45b8-b766-198f431c3788] Updating instance_info_cache with network_info: [{"id": "ad9dda29-9326-472b-8c82-294a33a7faf0", "address": "fa:16:3e:26:e3:72", "network": {"id": "88dfb29d-41e0-4df7-8f70-e52cda05620c", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1355578400-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.251", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7d4ada9ec35f42b19c6480a9101d21a4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "357d2811-e990-4985-9f9e-b158d10d3699", "external-id": "nsx-vlan-transportzone-641", "segmentation_id": 641, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapad9dda29-93", "ovs_interfaceid": "ad9dda29-9326-472b-8c82-294a33a7faf0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "e24932f9-9063-4a16-8498-950bdfeeaa84", "address": "fa:16:3e:06:c4:2b", "network": {"id": "88dfb29d-41e0-4df7-8f70-e52cda05620c", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1355578400-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7d4ada9ec35f42b19c6480a9101d21a4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "357d2811-e990-4985-9f9e-b158d10d3699", "external-id": "nsx-vlan-transportzone-641", "segmentation_id": 641, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape24932f9-90", "ovs_interfaceid": "e24932f9-9063-4a16-8498-950bdfeeaa84", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 787.643360] env[63515]: DEBUG nova.compute.manager [req-33c1995f-9709-4a34-9f8b-31a49adbd4d2 req-1e1de6e5-388b-4bf7-a6cb-d7faece4dca2 service nova] [instance: 5cbce760-0163-4b27-8ae3-e46c926c8916] Received event network-changed-5fa807c4-5ce1-492a-bea3-d8f7a65d898f {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 787.643360] env[63515]: DEBUG nova.compute.manager [req-33c1995f-9709-4a34-9f8b-31a49adbd4d2 req-1e1de6e5-388b-4bf7-a6cb-d7faece4dca2 service nova] [instance: 5cbce760-0163-4b27-8ae3-e46c926c8916] Refreshing instance network info cache due to event network-changed-5fa807c4-5ce1-492a-bea3-d8f7a65d898f. {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 787.643360] env[63515]: DEBUG oslo_concurrency.lockutils [req-33c1995f-9709-4a34-9f8b-31a49adbd4d2 req-1e1de6e5-388b-4bf7-a6cb-d7faece4dca2 service nova] Acquiring lock "refresh_cache-5cbce760-0163-4b27-8ae3-e46c926c8916" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 787.643360] env[63515]: DEBUG oslo_concurrency.lockutils [req-33c1995f-9709-4a34-9f8b-31a49adbd4d2 req-1e1de6e5-388b-4bf7-a6cb-d7faece4dca2 service nova] Acquired lock "refresh_cache-5cbce760-0163-4b27-8ae3-e46c926c8916" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 787.643360] env[63515]: DEBUG nova.network.neutron [req-33c1995f-9709-4a34-9f8b-31a49adbd4d2 req-1e1de6e5-388b-4bf7-a6cb-d7faece4dca2 service nova] [instance: 5cbce760-0163-4b27-8ae3-e46c926c8916] Refreshing network info cache for port 5fa807c4-5ce1-492a-bea3-d8f7a65d898f {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 787.700336] env[63515]: DEBUG oslo_concurrency.lockutils [None req-938cccb1-5a46-437d-9e93-3058a8134cd5 tempest-ListServersNegativeTestJSON-1448699027 tempest-ListServersNegativeTestJSON-1448699027-project-member] Lock "37e6c27e-317b-45d2-bd55-2fd78ccf009f" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 32.324s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 787.704838] env[63515]: DEBUG oslo_vmware.api [None req-83fbd51b-fc37-49eb-8423-cdaa019d3b0c tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Task: {'id': task-1111227, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 787.735460] env[63515]: DEBUG nova.network.neutron [None req-4b5baf79-d4c3-467e-88d4-a131ea6620d7 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] [instance: e666825c-ff4e-4a0e-93c0-43c00f167bbb] Successfully updated port: 7872b4c3-d54f-4613-85d2-f568ec3cb47f {{(pid=63515) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 787.778036] env[63515]: DEBUG nova.compute.manager [None req-67e41316-801d-46e9-947d-28c19f5a3017 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: 84dbf321-2c70-4e08-b430-cb5a06fc6829] Checking state {{(pid=63515) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 787.778695] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06d52138-7f2d-42f6-9a65-9f7bc1f9b23f {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.909741] env[63515]: DEBUG oslo_vmware.api [None req-fa0cfe6d-3452-46d4-9987-2cb422956d62 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]5221cd63-2acf-d2c6-13ca-00731dd52590, 'name': SearchDatastore_Task, 'duration_secs': 0.010164} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 787.910057] env[63515]: DEBUG oslo_concurrency.lockutils [None req-fa0cfe6d-3452-46d4-9987-2cb422956d62 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 787.910312] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-fa0cfe6d-3452-46d4-9987-2cb422956d62 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] [instance: 63ce797d-7180-4209-ac2c-81978bf7607a] Processing image 8a120570-cb06-4099-b262-554ca0ad15c5 {{(pid=63515) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 787.910562] env[63515]: DEBUG oslo_concurrency.lockutils [None req-fa0cfe6d-3452-46d4-9987-2cb422956d62 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 787.910711] env[63515]: DEBUG oslo_concurrency.lockutils [None req-fa0cfe6d-3452-46d4-9987-2cb422956d62 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 787.910896] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-fa0cfe6d-3452-46d4-9987-2cb422956d62 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 787.911199] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a6b5c7f6-1395-4cc6-b8e8-509e55308136 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.923305] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-fa0cfe6d-3452-46d4-9987-2cb422956d62 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 787.923548] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-fa0cfe6d-3452-46d4-9987-2cb422956d62 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63515) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 787.924719] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ca304479-ce34-4faf-90bb-e7ac940ac4eb {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.930305] env[63515]: DEBUG oslo_vmware.api [None req-fa0cfe6d-3452-46d4-9987-2cb422956d62 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Waiting for the task: (returnval){ [ 787.930305] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]5276f206-e874-2f7e-989e-a9ce51e968ee" [ 787.930305] env[63515]: _type = "Task" [ 787.930305] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 787.938319] env[63515]: DEBUG oslo_vmware.api [None req-fa0cfe6d-3452-46d4-9987-2cb422956d62 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]5276f206-e874-2f7e-989e-a9ce51e968ee, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 788.071759] env[63515]: DEBUG oslo_concurrency.lockutils [req-c733811e-ef67-46a5-a2d8-172f51057db6 req-32296e40-52e9-4881-ad7b-93b3d7915b31 service nova] Releasing lock "refresh_cache-ed550b10-d58f-45b8-b766-198f431c3788" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 788.072829] env[63515]: DEBUG nova.compute.manager [req-c733811e-ef67-46a5-a2d8-172f51057db6 req-32296e40-52e9-4881-ad7b-93b3d7915b31 service nova] [instance: 63ce797d-7180-4209-ac2c-81978bf7607a] Received event network-vif-plugged-25ea9d0d-3405-42da-a576-a2a1f37fa75a {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 788.072829] env[63515]: DEBUG oslo_concurrency.lockutils [req-c733811e-ef67-46a5-a2d8-172f51057db6 req-32296e40-52e9-4881-ad7b-93b3d7915b31 service nova] Acquiring lock "63ce797d-7180-4209-ac2c-81978bf7607a-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 788.072829] env[63515]: DEBUG oslo_concurrency.lockutils [req-c733811e-ef67-46a5-a2d8-172f51057db6 req-32296e40-52e9-4881-ad7b-93b3d7915b31 service nova] Lock "63ce797d-7180-4209-ac2c-81978bf7607a-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 788.072829] env[63515]: DEBUG oslo_concurrency.lockutils [req-c733811e-ef67-46a5-a2d8-172f51057db6 req-32296e40-52e9-4881-ad7b-93b3d7915b31 service nova] Lock "63ce797d-7180-4209-ac2c-81978bf7607a-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 788.072829] env[63515]: DEBUG nova.compute.manager [req-c733811e-ef67-46a5-a2d8-172f51057db6 req-32296e40-52e9-4881-ad7b-93b3d7915b31 service nova] [instance: 63ce797d-7180-4209-ac2c-81978bf7607a] No waiting events found dispatching network-vif-plugged-25ea9d0d-3405-42da-a576-a2a1f37fa75a {{(pid=63515) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 788.073085] env[63515]: WARNING nova.compute.manager [req-c733811e-ef67-46a5-a2d8-172f51057db6 req-32296e40-52e9-4881-ad7b-93b3d7915b31 service nova] [instance: 63ce797d-7180-4209-ac2c-81978bf7607a] Received unexpected event network-vif-plugged-25ea9d0d-3405-42da-a576-a2a1f37fa75a for instance with vm_state building and task_state spawning. [ 788.073214] env[63515]: DEBUG nova.compute.manager [req-c733811e-ef67-46a5-a2d8-172f51057db6 req-32296e40-52e9-4881-ad7b-93b3d7915b31 service nova] [instance: 63ce797d-7180-4209-ac2c-81978bf7607a] Received event network-changed-25ea9d0d-3405-42da-a576-a2a1f37fa75a {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 788.073390] env[63515]: DEBUG nova.compute.manager [req-c733811e-ef67-46a5-a2d8-172f51057db6 req-32296e40-52e9-4881-ad7b-93b3d7915b31 service nova] [instance: 63ce797d-7180-4209-ac2c-81978bf7607a] Refreshing instance network info cache due to event network-changed-25ea9d0d-3405-42da-a576-a2a1f37fa75a. {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 788.073598] env[63515]: DEBUG oslo_concurrency.lockutils [req-c733811e-ef67-46a5-a2d8-172f51057db6 req-32296e40-52e9-4881-ad7b-93b3d7915b31 service nova] Acquiring lock "refresh_cache-63ce797d-7180-4209-ac2c-81978bf7607a" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 788.073745] env[63515]: DEBUG oslo_concurrency.lockutils [req-c733811e-ef67-46a5-a2d8-172f51057db6 req-32296e40-52e9-4881-ad7b-93b3d7915b31 service nova] Acquired lock "refresh_cache-63ce797d-7180-4209-ac2c-81978bf7607a" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 788.073905] env[63515]: DEBUG nova.network.neutron [req-c733811e-ef67-46a5-a2d8-172f51057db6 req-32296e40-52e9-4881-ad7b-93b3d7915b31 service nova] [instance: 63ce797d-7180-4209-ac2c-81978bf7607a] Refreshing network info cache for port 25ea9d0d-3405-42da-a576-a2a1f37fa75a {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 788.170570] env[63515]: INFO nova.compute.manager [None req-b53e671a-9cf4-4e34-a2ac-4321dce89b1c tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] [instance: 192137b0-03e5-4bc4-b911-4b4f1a874f74] Rebuilding instance [ 788.189915] env[63515]: DEBUG nova.objects.instance [None req-9afb130a-5980-40d4-acae-35d697720c32 tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Lazy-loading 'flavor' on Instance uuid d7e8dcdb-41cb-46fb-8b61-d251e7c2d372 {{(pid=63515) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 788.210582] env[63515]: DEBUG oslo_vmware.api [None req-83fbd51b-fc37-49eb-8423-cdaa019d3b0c tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Task: {'id': task-1111227, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.837111} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 788.211508] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-83fbd51b-fc37-49eb-8423-cdaa019d3b0c tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk to [datastore1] 5cbce760-0163-4b27-8ae3-e46c926c8916/5cbce760-0163-4b27-8ae3-e46c926c8916.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 788.211508] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-83fbd51b-fc37-49eb-8423-cdaa019d3b0c tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] [instance: 5cbce760-0163-4b27-8ae3-e46c926c8916] Extending root virtual disk to 1048576 {{(pid=63515) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 788.216465] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-5fad7353-dff1-4fc7-8d56-8df45907a29f {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.227756] env[63515]: DEBUG oslo_vmware.api [None req-83fbd51b-fc37-49eb-8423-cdaa019d3b0c tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Waiting for the task: (returnval){ [ 788.227756] env[63515]: value = "task-1111228" [ 788.227756] env[63515]: _type = "Task" [ 788.227756] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 788.240475] env[63515]: DEBUG oslo_concurrency.lockutils [None req-4b5baf79-d4c3-467e-88d4-a131ea6620d7 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] Acquiring lock "refresh_cache-e666825c-ff4e-4a0e-93c0-43c00f167bbb" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 788.240660] env[63515]: DEBUG oslo_concurrency.lockutils [None req-4b5baf79-d4c3-467e-88d4-a131ea6620d7 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] Acquired lock "refresh_cache-e666825c-ff4e-4a0e-93c0-43c00f167bbb" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 788.240857] env[63515]: DEBUG nova.network.neutron [None req-4b5baf79-d4c3-467e-88d4-a131ea6620d7 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] [instance: e666825c-ff4e-4a0e-93c0-43c00f167bbb] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 788.247079] env[63515]: DEBUG oslo_vmware.api [None req-83fbd51b-fc37-49eb-8423-cdaa019d3b0c tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Task: {'id': task-1111228, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 788.247574] env[63515]: DEBUG nova.compute.manager [None req-b53e671a-9cf4-4e34-a2ac-4321dce89b1c tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] [instance: 192137b0-03e5-4bc4-b911-4b4f1a874f74] Checking state {{(pid=63515) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 788.252253] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51ed1c4c-f6b4-40e7-ae5e-941575a31de9 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.295361] env[63515]: INFO nova.compute.manager [None req-67e41316-801d-46e9-947d-28c19f5a3017 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: 84dbf321-2c70-4e08-b430-cb5a06fc6829] instance snapshotting [ 788.295651] env[63515]: WARNING nova.compute.manager [None req-67e41316-801d-46e9-947d-28c19f5a3017 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: 84dbf321-2c70-4e08-b430-cb5a06fc6829] trying to snapshot a non-running instance: (state: 7 expected: 1) [ 788.300317] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9857870b-b16e-400b-b1e1-e2eb7feaf5ab {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.324017] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07b9a4a0-792a-40d4-88d9-5e9770fb47d7 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.441774] env[63515]: DEBUG oslo_vmware.api [None req-fa0cfe6d-3452-46d4-9987-2cb422956d62 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]5276f206-e874-2f7e-989e-a9ce51e968ee, 'name': SearchDatastore_Task, 'duration_secs': 0.054683} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 788.445557] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2d3fa414-6ae9-4304-ab38-3efd7a173702 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.451924] env[63515]: DEBUG oslo_vmware.api [None req-fa0cfe6d-3452-46d4-9987-2cb422956d62 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Waiting for the task: (returnval){ [ 788.451924] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]52887b1e-a8f2-dfca-7985-19709843c450" [ 788.451924] env[63515]: _type = "Task" [ 788.451924] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 788.462573] env[63515]: DEBUG oslo_vmware.api [None req-fa0cfe6d-3452-46d4-9987-2cb422956d62 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52887b1e-a8f2-dfca-7985-19709843c450, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 788.471497] env[63515]: DEBUG nova.network.neutron [req-33c1995f-9709-4a34-9f8b-31a49adbd4d2 req-1e1de6e5-388b-4bf7-a6cb-d7faece4dca2 service nova] [instance: 5cbce760-0163-4b27-8ae3-e46c926c8916] Updated VIF entry in instance network info cache for port 5fa807c4-5ce1-492a-bea3-d8f7a65d898f. {{(pid=63515) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 788.471835] env[63515]: DEBUG nova.network.neutron [req-33c1995f-9709-4a34-9f8b-31a49adbd4d2 req-1e1de6e5-388b-4bf7-a6cb-d7faece4dca2 service nova] [instance: 5cbce760-0163-4b27-8ae3-e46c926c8916] Updating instance_info_cache with network_info: [{"id": "5fa807c4-5ce1-492a-bea3-d8f7a65d898f", "address": "fa:16:3e:23:bb:83", "network": {"id": "850bfa95-3df6-4019-8260-ade20ec58799", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-832000501-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "08c9aa51b1864410809f8877c7cd49c4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3cf748a8-7ae0-4dca-817d-e727c30d72f4", "external-id": "nsx-vlan-transportzone-853", "segmentation_id": 853, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5fa807c4-5c", "ovs_interfaceid": "5fa807c4-5ce1-492a-bea3-d8f7a65d898f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 788.536643] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b174e3d-d23f-4a2c-ac5e-9711422db9ca {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.547741] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-feba6349-d33b-4b8f-8e42-2cc67efb8105 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.581212] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24e001b7-e7f7-4b6c-b655-dd581b24cc45 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.589343] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af9352e4-7e7c-45b6-8c7a-0904fef46ef7 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.603160] env[63515]: DEBUG nova.compute.provider_tree [None req-66c3e202-5a3e-433e-b543-4bfb02c8ad30 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 788.699045] env[63515]: DEBUG oslo_concurrency.lockutils [None req-9afb130a-5980-40d4-acae-35d697720c32 tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Acquiring lock "refresh_cache-d7e8dcdb-41cb-46fb-8b61-d251e7c2d372" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 788.699387] env[63515]: DEBUG oslo_concurrency.lockutils [None req-9afb130a-5980-40d4-acae-35d697720c32 tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Acquired lock "refresh_cache-d7e8dcdb-41cb-46fb-8b61-d251e7c2d372" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 788.699685] env[63515]: DEBUG nova.network.neutron [None req-9afb130a-5980-40d4-acae-35d697720c32 tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] [instance: d7e8dcdb-41cb-46fb-8b61-d251e7c2d372] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 788.700106] env[63515]: DEBUG nova.objects.instance [None req-9afb130a-5980-40d4-acae-35d697720c32 tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Lazy-loading 'info_cache' on Instance uuid d7e8dcdb-41cb-46fb-8b61-d251e7c2d372 {{(pid=63515) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 788.737395] env[63515]: DEBUG oslo_vmware.api [None req-83fbd51b-fc37-49eb-8423-cdaa019d3b0c tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Task: {'id': task-1111228, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.067813} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 788.743059] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-83fbd51b-fc37-49eb-8423-cdaa019d3b0c tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] [instance: 5cbce760-0163-4b27-8ae3-e46c926c8916] Extended root virtual disk {{(pid=63515) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 788.743059] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e01e2a4f-7cbb-48e9-b9a5-05e0911d7b43 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.769919] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-83fbd51b-fc37-49eb-8423-cdaa019d3b0c tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] [instance: 5cbce760-0163-4b27-8ae3-e46c926c8916] Reconfiguring VM instance instance-0000003a to attach disk [datastore1] 5cbce760-0163-4b27-8ae3-e46c926c8916/5cbce760-0163-4b27-8ae3-e46c926c8916.vmdk or device None with type sparse {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 788.774346] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-b53e671a-9cf4-4e34-a2ac-4321dce89b1c tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] [instance: 192137b0-03e5-4bc4-b911-4b4f1a874f74] Powering off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 788.774346] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0258fb03-67f2-4c1e-b033-8aec4bad1ffa {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.789159] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9b9c4e99-b302-408e-90b4-90565839a566 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.796894] env[63515]: DEBUG oslo_vmware.api [None req-b53e671a-9cf4-4e34-a2ac-4321dce89b1c tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] Waiting for the task: (returnval){ [ 788.796894] env[63515]: value = "task-1111229" [ 788.796894] env[63515]: _type = "Task" [ 788.796894] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 788.798288] env[63515]: DEBUG oslo_vmware.api [None req-83fbd51b-fc37-49eb-8423-cdaa019d3b0c tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Waiting for the task: (returnval){ [ 788.798288] env[63515]: value = "task-1111230" [ 788.798288] env[63515]: _type = "Task" [ 788.798288] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 788.814787] env[63515]: DEBUG oslo_vmware.api [None req-83fbd51b-fc37-49eb-8423-cdaa019d3b0c tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Task: {'id': task-1111230, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 788.817845] env[63515]: DEBUG oslo_vmware.api [None req-b53e671a-9cf4-4e34-a2ac-4321dce89b1c tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] Task: {'id': task-1111229, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 788.837558] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-67e41316-801d-46e9-947d-28c19f5a3017 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: 84dbf321-2c70-4e08-b430-cb5a06fc6829] Creating Snapshot of the VM instance {{(pid=63515) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 788.837893] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-74d928a6-0dcc-41b1-8ea8-4e6d64133e0a {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.844942] env[63515]: DEBUG oslo_vmware.api [None req-67e41316-801d-46e9-947d-28c19f5a3017 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Waiting for the task: (returnval){ [ 788.844942] env[63515]: value = "task-1111231" [ 788.844942] env[63515]: _type = "Task" [ 788.844942] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 788.853961] env[63515]: DEBUG oslo_vmware.api [None req-67e41316-801d-46e9-947d-28c19f5a3017 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Task: {'id': task-1111231, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 788.862717] env[63515]: DEBUG nova.network.neutron [None req-4b5baf79-d4c3-467e-88d4-a131ea6620d7 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] [instance: e666825c-ff4e-4a0e-93c0-43c00f167bbb] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 788.877205] env[63515]: DEBUG nova.network.neutron [req-c733811e-ef67-46a5-a2d8-172f51057db6 req-32296e40-52e9-4881-ad7b-93b3d7915b31 service nova] [instance: 63ce797d-7180-4209-ac2c-81978bf7607a] Updated VIF entry in instance network info cache for port 25ea9d0d-3405-42da-a576-a2a1f37fa75a. {{(pid=63515) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 788.877739] env[63515]: DEBUG nova.network.neutron [req-c733811e-ef67-46a5-a2d8-172f51057db6 req-32296e40-52e9-4881-ad7b-93b3d7915b31 service nova] [instance: 63ce797d-7180-4209-ac2c-81978bf7607a] Updating instance_info_cache with network_info: [{"id": "25ea9d0d-3405-42da-a576-a2a1f37fa75a", "address": "fa:16:3e:35:5c:ab", "network": {"id": "850bfa95-3df6-4019-8260-ade20ec58799", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-832000501-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "08c9aa51b1864410809f8877c7cd49c4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3cf748a8-7ae0-4dca-817d-e727c30d72f4", "external-id": "nsx-vlan-transportzone-853", "segmentation_id": 853, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap25ea9d0d-34", "ovs_interfaceid": "25ea9d0d-3405-42da-a576-a2a1f37fa75a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 788.883083] env[63515]: DEBUG nova.compute.manager [req-dcbed5a2-17ec-49e5-bc73-b36635377dd9 req-62d11547-fdd9-4e8a-a99c-79521cf27edb service nova] [instance: 70ed982f-affd-4dd1-bc90-c64e7c6d49d2] Received event network-changed-fe7c09da-d6bd-4f72-afa2-1a70bc663407 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 788.883330] env[63515]: DEBUG nova.compute.manager [req-dcbed5a2-17ec-49e5-bc73-b36635377dd9 req-62d11547-fdd9-4e8a-a99c-79521cf27edb service nova] [instance: 70ed982f-affd-4dd1-bc90-c64e7c6d49d2] Refreshing instance network info cache due to event network-changed-fe7c09da-d6bd-4f72-afa2-1a70bc663407. {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 788.883584] env[63515]: DEBUG oslo_concurrency.lockutils [req-dcbed5a2-17ec-49e5-bc73-b36635377dd9 req-62d11547-fdd9-4e8a-a99c-79521cf27edb service nova] Acquiring lock "refresh_cache-70ed982f-affd-4dd1-bc90-c64e7c6d49d2" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 788.883751] env[63515]: DEBUG oslo_concurrency.lockutils [req-dcbed5a2-17ec-49e5-bc73-b36635377dd9 req-62d11547-fdd9-4e8a-a99c-79521cf27edb service nova] Acquired lock "refresh_cache-70ed982f-affd-4dd1-bc90-c64e7c6d49d2" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 788.883927] env[63515]: DEBUG nova.network.neutron [req-dcbed5a2-17ec-49e5-bc73-b36635377dd9 req-62d11547-fdd9-4e8a-a99c-79521cf27edb service nova] [instance: 70ed982f-affd-4dd1-bc90-c64e7c6d49d2] Refreshing network info cache for port fe7c09da-d6bd-4f72-afa2-1a70bc663407 {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 788.909102] env[63515]: DEBUG oslo_concurrency.lockutils [None req-65badfd4-59e2-4bf6-bac8-39c1a78dd350 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Acquiring lock "f1d01b75-ac9d-458d-8cc2-ae64cffca4e8" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 788.909407] env[63515]: DEBUG oslo_concurrency.lockutils [None req-65badfd4-59e2-4bf6-bac8-39c1a78dd350 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Lock "f1d01b75-ac9d-458d-8cc2-ae64cffca4e8" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 788.909628] env[63515]: DEBUG oslo_concurrency.lockutils [None req-65badfd4-59e2-4bf6-bac8-39c1a78dd350 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Acquiring lock "f1d01b75-ac9d-458d-8cc2-ae64cffca4e8-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 788.909827] env[63515]: DEBUG oslo_concurrency.lockutils [None req-65badfd4-59e2-4bf6-bac8-39c1a78dd350 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Lock "f1d01b75-ac9d-458d-8cc2-ae64cffca4e8-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 788.910024] env[63515]: DEBUG oslo_concurrency.lockutils [None req-65badfd4-59e2-4bf6-bac8-39c1a78dd350 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Lock "f1d01b75-ac9d-458d-8cc2-ae64cffca4e8-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 788.912378] env[63515]: INFO nova.compute.manager [None req-65badfd4-59e2-4bf6-bac8-39c1a78dd350 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] [instance: f1d01b75-ac9d-458d-8cc2-ae64cffca4e8] Terminating instance [ 788.914539] env[63515]: DEBUG nova.compute.manager [None req-65badfd4-59e2-4bf6-bac8-39c1a78dd350 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] [instance: f1d01b75-ac9d-458d-8cc2-ae64cffca4e8] Start destroying the instance on the hypervisor. {{(pid=63515) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 788.914754] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-65badfd4-59e2-4bf6-bac8-39c1a78dd350 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] [instance: f1d01b75-ac9d-458d-8cc2-ae64cffca4e8] Destroying instance {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 788.916611] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8150976-43ec-4b04-ba6f-65cb656d86a3 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.925951] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-65badfd4-59e2-4bf6-bac8-39c1a78dd350 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] [instance: f1d01b75-ac9d-458d-8cc2-ae64cffca4e8] Powering off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 788.925951] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-190abec8-339a-46b0-9517-4f56913b5485 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.933336] env[63515]: DEBUG oslo_vmware.api [None req-65badfd4-59e2-4bf6-bac8-39c1a78dd350 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Waiting for the task: (returnval){ [ 788.933336] env[63515]: value = "task-1111232" [ 788.933336] env[63515]: _type = "Task" [ 788.933336] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 788.943248] env[63515]: DEBUG oslo_vmware.api [None req-65badfd4-59e2-4bf6-bac8-39c1a78dd350 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Task: {'id': task-1111232, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 788.962348] env[63515]: DEBUG oslo_vmware.api [None req-fa0cfe6d-3452-46d4-9987-2cb422956d62 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52887b1e-a8f2-dfca-7985-19709843c450, 'name': SearchDatastore_Task, 'duration_secs': 0.036479} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 788.962661] env[63515]: DEBUG oslo_concurrency.lockutils [None req-fa0cfe6d-3452-46d4-9987-2cb422956d62 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 788.962936] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-fa0cfe6d-3452-46d4-9987-2cb422956d62 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk to [datastore1] 63ce797d-7180-4209-ac2c-81978bf7607a/63ce797d-7180-4209-ac2c-81978bf7607a.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 788.963237] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-0a1ecdc2-ed04-41b2-920c-a8f1b8d2b850 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.969624] env[63515]: DEBUG oslo_vmware.api [None req-fa0cfe6d-3452-46d4-9987-2cb422956d62 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Waiting for the task: (returnval){ [ 788.969624] env[63515]: value = "task-1111233" [ 788.969624] env[63515]: _type = "Task" [ 788.969624] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 788.975992] env[63515]: DEBUG oslo_concurrency.lockutils [req-33c1995f-9709-4a34-9f8b-31a49adbd4d2 req-1e1de6e5-388b-4bf7-a6cb-d7faece4dca2 service nova] Releasing lock "refresh_cache-5cbce760-0163-4b27-8ae3-e46c926c8916" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 788.979093] env[63515]: DEBUG oslo_vmware.api [None req-fa0cfe6d-3452-46d4-9987-2cb422956d62 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Task: {'id': task-1111233, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 789.028379] env[63515]: DEBUG oslo_concurrency.lockutils [None req-8d4c1f89-f73c-481b-8b79-61983aeba07b tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Acquiring lock "interface-ed550b10-d58f-45b8-b766-198f431c3788-e24932f9-9063-4a16-8498-950bdfeeaa84" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 789.028664] env[63515]: DEBUG oslo_concurrency.lockutils [None req-8d4c1f89-f73c-481b-8b79-61983aeba07b tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Lock "interface-ed550b10-d58f-45b8-b766-198f431c3788-e24932f9-9063-4a16-8498-950bdfeeaa84" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 789.106900] env[63515]: DEBUG nova.scheduler.client.report [None req-66c3e202-5a3e-433e-b543-4bfb02c8ad30 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 789.204324] env[63515]: DEBUG nova.objects.base [None req-9afb130a-5980-40d4-acae-35d697720c32 tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Object Instance lazy-loaded attributes: flavor,info_cache {{(pid=63515) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 789.236037] env[63515]: DEBUG nova.network.neutron [None req-4b5baf79-d4c3-467e-88d4-a131ea6620d7 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] [instance: e666825c-ff4e-4a0e-93c0-43c00f167bbb] Updating instance_info_cache with network_info: [{"id": "7872b4c3-d54f-4613-85d2-f568ec3cb47f", "address": "fa:16:3e:a1:a9:1e", "network": {"id": "b4abcbd0-0eac-4bde-ad11-7f93afcedcef", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.123", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "b569255cc43e42e7ae3f2b4ad37c6ef4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0cd5d325-3053-407e-a4ee-f627e82a23f9", "external-id": "nsx-vlan-transportzone-809", "segmentation_id": 809, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7872b4c3-d5", "ovs_interfaceid": "7872b4c3-d54f-4613-85d2-f568ec3cb47f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 789.308197] env[63515]: DEBUG oslo_vmware.api [None req-b53e671a-9cf4-4e34-a2ac-4321dce89b1c tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] Task: {'id': task-1111229, 'name': PowerOffVM_Task, 'duration_secs': 0.390154} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 789.312559] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-b53e671a-9cf4-4e34-a2ac-4321dce89b1c tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] [instance: 192137b0-03e5-4bc4-b911-4b4f1a874f74] Powered off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 789.312903] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-b53e671a-9cf4-4e34-a2ac-4321dce89b1c tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] [instance: 192137b0-03e5-4bc4-b911-4b4f1a874f74] Destroying instance {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 789.313962] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ecab24f-d422-4552-a0a9-bd93cc1b0fc4 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.323242] env[63515]: DEBUG oslo_vmware.api [None req-83fbd51b-fc37-49eb-8423-cdaa019d3b0c tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Task: {'id': task-1111230, 'name': ReconfigVM_Task, 'duration_secs': 0.312937} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 789.326069] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-83fbd51b-fc37-49eb-8423-cdaa019d3b0c tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] [instance: 5cbce760-0163-4b27-8ae3-e46c926c8916] Reconfigured VM instance instance-0000003a to attach disk [datastore1] 5cbce760-0163-4b27-8ae3-e46c926c8916/5cbce760-0163-4b27-8ae3-e46c926c8916.vmdk or device None with type sparse {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 789.326845] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-b53e671a-9cf4-4e34-a2ac-4321dce89b1c tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] [instance: 192137b0-03e5-4bc4-b911-4b4f1a874f74] Unregistering the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 789.327112] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-8974fcf9-5b91-4005-8d15-578005cd08da {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.328874] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-1fff5fcc-3f71-45a3-8b2d-c88a80de38cb {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.336032] env[63515]: DEBUG oslo_vmware.api [None req-83fbd51b-fc37-49eb-8423-cdaa019d3b0c tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Waiting for the task: (returnval){ [ 789.336032] env[63515]: value = "task-1111234" [ 789.336032] env[63515]: _type = "Task" [ 789.336032] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 789.346578] env[63515]: DEBUG oslo_vmware.api [None req-83fbd51b-fc37-49eb-8423-cdaa019d3b0c tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Task: {'id': task-1111234, 'name': Rename_Task} progress is 5%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 789.355764] env[63515]: DEBUG oslo_vmware.api [None req-67e41316-801d-46e9-947d-28c19f5a3017 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Task: {'id': task-1111231, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 789.357147] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-b53e671a-9cf4-4e34-a2ac-4321dce89b1c tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] [instance: 192137b0-03e5-4bc4-b911-4b4f1a874f74] Unregistered the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 789.357393] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-b53e671a-9cf4-4e34-a2ac-4321dce89b1c tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] [instance: 192137b0-03e5-4bc4-b911-4b4f1a874f74] Deleting contents of the VM from datastore datastore1 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 789.359305] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-b53e671a-9cf4-4e34-a2ac-4321dce89b1c tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] Deleting the datastore file [datastore1] 192137b0-03e5-4bc4-b911-4b4f1a874f74 {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 789.359305] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b7f0ca59-9596-4473-89a2-1fa0e203f5aa {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.364302] env[63515]: DEBUG oslo_vmware.api [None req-b53e671a-9cf4-4e34-a2ac-4321dce89b1c tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] Waiting for the task: (returnval){ [ 789.364302] env[63515]: value = "task-1111236" [ 789.364302] env[63515]: _type = "Task" [ 789.364302] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 789.373350] env[63515]: DEBUG oslo_vmware.api [None req-b53e671a-9cf4-4e34-a2ac-4321dce89b1c tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] Task: {'id': task-1111236, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 789.382404] env[63515]: DEBUG oslo_concurrency.lockutils [req-c733811e-ef67-46a5-a2d8-172f51057db6 req-32296e40-52e9-4881-ad7b-93b3d7915b31 service nova] Releasing lock "refresh_cache-63ce797d-7180-4209-ac2c-81978bf7607a" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 789.444158] env[63515]: DEBUG oslo_vmware.api [None req-65badfd4-59e2-4bf6-bac8-39c1a78dd350 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Task: {'id': task-1111232, 'name': PowerOffVM_Task, 'duration_secs': 0.259779} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 789.447034] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-65badfd4-59e2-4bf6-bac8-39c1a78dd350 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] [instance: f1d01b75-ac9d-458d-8cc2-ae64cffca4e8] Powered off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 789.447348] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-65badfd4-59e2-4bf6-bac8-39c1a78dd350 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] [instance: f1d01b75-ac9d-458d-8cc2-ae64cffca4e8] Unregistering the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 789.447732] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-662ab0e8-f832-4df9-9ae2-48302c56c98b {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.479735] env[63515]: DEBUG oslo_vmware.api [None req-fa0cfe6d-3452-46d4-9987-2cb422956d62 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Task: {'id': task-1111233, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 789.516152] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-65badfd4-59e2-4bf6-bac8-39c1a78dd350 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] [instance: f1d01b75-ac9d-458d-8cc2-ae64cffca4e8] Unregistered the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 789.516476] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-65badfd4-59e2-4bf6-bac8-39c1a78dd350 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] [instance: f1d01b75-ac9d-458d-8cc2-ae64cffca4e8] Deleting contents of the VM from datastore datastore2 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 789.516782] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-65badfd4-59e2-4bf6-bac8-39c1a78dd350 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Deleting the datastore file [datastore2] f1d01b75-ac9d-458d-8cc2-ae64cffca4e8 {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 789.517427] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-282c5478-1b60-429a-aa4c-f18ab8932772 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.526052] env[63515]: DEBUG oslo_vmware.api [None req-65badfd4-59e2-4bf6-bac8-39c1a78dd350 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Waiting for the task: (returnval){ [ 789.526052] env[63515]: value = "task-1111238" [ 789.526052] env[63515]: _type = "Task" [ 789.526052] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 789.534143] env[63515]: DEBUG oslo_concurrency.lockutils [None req-8d4c1f89-f73c-481b-8b79-61983aeba07b tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Acquiring lock "ed550b10-d58f-45b8-b766-198f431c3788" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 789.534143] env[63515]: DEBUG oslo_concurrency.lockutils [None req-8d4c1f89-f73c-481b-8b79-61983aeba07b tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Acquired lock "ed550b10-d58f-45b8-b766-198f431c3788" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 789.534143] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d35ae85-d490-4e1c-acc4-a3ec32ebc6d3 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.539711] env[63515]: DEBUG oslo_vmware.api [None req-65badfd4-59e2-4bf6-bac8-39c1a78dd350 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Task: {'id': task-1111238, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 789.552596] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8475805-b1ae-4084-8956-186b71e17e3a {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.586795] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-8d4c1f89-f73c-481b-8b79-61983aeba07b tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: ed550b10-d58f-45b8-b766-198f431c3788] Reconfiguring VM to detach interface {{(pid=63515) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1968}} [ 789.586795] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5aa6eba4-d28a-419e-90f6-9a021d7a648f {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.606720] env[63515]: DEBUG oslo_vmware.api [None req-8d4c1f89-f73c-481b-8b79-61983aeba07b tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Waiting for the task: (returnval){ [ 789.606720] env[63515]: value = "task-1111239" [ 789.606720] env[63515]: _type = "Task" [ 789.606720] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 789.615643] env[63515]: DEBUG oslo_concurrency.lockutils [None req-66c3e202-5a3e-433e-b543-4bfb02c8ad30 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.536s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 789.616193] env[63515]: DEBUG nova.compute.manager [None req-66c3e202-5a3e-433e-b543-4bfb02c8ad30 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775-project-member] [instance: 4e453127-1f3e-40ea-819f-6678479826c8] Start building networks asynchronously for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 789.618717] env[63515]: DEBUG oslo_vmware.api [None req-8d4c1f89-f73c-481b-8b79-61983aeba07b tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Task: {'id': task-1111239, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 789.618999] env[63515]: DEBUG oslo_concurrency.lockutils [None req-0e5835c9-1a06-4538-91e4-48a9dfc2854d tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 28.575s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 789.619195] env[63515]: DEBUG nova.objects.instance [None req-0e5835c9-1a06-4538-91e4-48a9dfc2854d tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: 001a48e1-b237-41f8-a6f4-2fe98d173512] Trying to apply a migration context that does not seem to be set for this instance {{(pid=63515) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 789.742251] env[63515]: DEBUG oslo_concurrency.lockutils [None req-4b5baf79-d4c3-467e-88d4-a131ea6620d7 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] Releasing lock "refresh_cache-e666825c-ff4e-4a0e-93c0-43c00f167bbb" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 789.742251] env[63515]: DEBUG nova.compute.manager [None req-4b5baf79-d4c3-467e-88d4-a131ea6620d7 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] [instance: e666825c-ff4e-4a0e-93c0-43c00f167bbb] Instance network_info: |[{"id": "7872b4c3-d54f-4613-85d2-f568ec3cb47f", "address": "fa:16:3e:a1:a9:1e", "network": {"id": "b4abcbd0-0eac-4bde-ad11-7f93afcedcef", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.123", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "b569255cc43e42e7ae3f2b4ad37c6ef4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0cd5d325-3053-407e-a4ee-f627e82a23f9", "external-id": "nsx-vlan-transportzone-809", "segmentation_id": 809, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7872b4c3-d5", "ovs_interfaceid": "7872b4c3-d54f-4613-85d2-f568ec3cb47f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 789.742251] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-4b5baf79-d4c3-467e-88d4-a131ea6620d7 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] [instance: e666825c-ff4e-4a0e-93c0-43c00f167bbb] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:a1:a9:1e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '0cd5d325-3053-407e-a4ee-f627e82a23f9', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '7872b4c3-d54f-4613-85d2-f568ec3cb47f', 'vif_model': 'vmxnet3'}] {{(pid=63515) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 789.747825] env[63515]: DEBUG oslo.service.loopingcall [None req-4b5baf79-d4c3-467e-88d4-a131ea6620d7 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 789.748530] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e666825c-ff4e-4a0e-93c0-43c00f167bbb] Creating VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 789.748785] env[63515]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-83eac840-a258-4580-8be7-b4cfb3b099d8 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.772315] env[63515]: DEBUG nova.compute.manager [req-10c14eaf-2c3b-4bbd-8f4f-77f8dafd636a req-75f4ef9e-6e4d-463a-a0d0-2fef3a6dcc20 service nova] [instance: e666825c-ff4e-4a0e-93c0-43c00f167bbb] Received event network-vif-plugged-7872b4c3-d54f-4613-85d2-f568ec3cb47f {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 789.772475] env[63515]: DEBUG oslo_concurrency.lockutils [req-10c14eaf-2c3b-4bbd-8f4f-77f8dafd636a req-75f4ef9e-6e4d-463a-a0d0-2fef3a6dcc20 service nova] Acquiring lock "e666825c-ff4e-4a0e-93c0-43c00f167bbb-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 789.772719] env[63515]: DEBUG oslo_concurrency.lockutils [req-10c14eaf-2c3b-4bbd-8f4f-77f8dafd636a req-75f4ef9e-6e4d-463a-a0d0-2fef3a6dcc20 service nova] Lock "e666825c-ff4e-4a0e-93c0-43c00f167bbb-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 789.772888] env[63515]: DEBUG oslo_concurrency.lockutils [req-10c14eaf-2c3b-4bbd-8f4f-77f8dafd636a req-75f4ef9e-6e4d-463a-a0d0-2fef3a6dcc20 service nova] Lock "e666825c-ff4e-4a0e-93c0-43c00f167bbb-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 789.773700] env[63515]: DEBUG nova.compute.manager [req-10c14eaf-2c3b-4bbd-8f4f-77f8dafd636a req-75f4ef9e-6e4d-463a-a0d0-2fef3a6dcc20 service nova] [instance: e666825c-ff4e-4a0e-93c0-43c00f167bbb] No waiting events found dispatching network-vif-plugged-7872b4c3-d54f-4613-85d2-f568ec3cb47f {{(pid=63515) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 789.773700] env[63515]: WARNING nova.compute.manager [req-10c14eaf-2c3b-4bbd-8f4f-77f8dafd636a req-75f4ef9e-6e4d-463a-a0d0-2fef3a6dcc20 service nova] [instance: e666825c-ff4e-4a0e-93c0-43c00f167bbb] Received unexpected event network-vif-plugged-7872b4c3-d54f-4613-85d2-f568ec3cb47f for instance with vm_state building and task_state spawning. [ 789.773859] env[63515]: DEBUG nova.compute.manager [req-10c14eaf-2c3b-4bbd-8f4f-77f8dafd636a req-75f4ef9e-6e4d-463a-a0d0-2fef3a6dcc20 service nova] [instance: e666825c-ff4e-4a0e-93c0-43c00f167bbb] Received event network-changed-7872b4c3-d54f-4613-85d2-f568ec3cb47f {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 789.774072] env[63515]: DEBUG nova.compute.manager [req-10c14eaf-2c3b-4bbd-8f4f-77f8dafd636a req-75f4ef9e-6e4d-463a-a0d0-2fef3a6dcc20 service nova] [instance: e666825c-ff4e-4a0e-93c0-43c00f167bbb] Refreshing instance network info cache due to event network-changed-7872b4c3-d54f-4613-85d2-f568ec3cb47f. {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 789.774277] env[63515]: DEBUG oslo_concurrency.lockutils [req-10c14eaf-2c3b-4bbd-8f4f-77f8dafd636a req-75f4ef9e-6e4d-463a-a0d0-2fef3a6dcc20 service nova] Acquiring lock "refresh_cache-e666825c-ff4e-4a0e-93c0-43c00f167bbb" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 789.774413] env[63515]: DEBUG oslo_concurrency.lockutils [req-10c14eaf-2c3b-4bbd-8f4f-77f8dafd636a req-75f4ef9e-6e4d-463a-a0d0-2fef3a6dcc20 service nova] Acquired lock "refresh_cache-e666825c-ff4e-4a0e-93c0-43c00f167bbb" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 789.775032] env[63515]: DEBUG nova.network.neutron [req-10c14eaf-2c3b-4bbd-8f4f-77f8dafd636a req-75f4ef9e-6e4d-463a-a0d0-2fef3a6dcc20 service nova] [instance: e666825c-ff4e-4a0e-93c0-43c00f167bbb] Refreshing network info cache for port 7872b4c3-d54f-4613-85d2-f568ec3cb47f {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 789.777422] env[63515]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 789.777422] env[63515]: value = "task-1111240" [ 789.777422] env[63515]: _type = "Task" [ 789.777422] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 789.781326] env[63515]: DEBUG nova.network.neutron [req-dcbed5a2-17ec-49e5-bc73-b36635377dd9 req-62d11547-fdd9-4e8a-a99c-79521cf27edb service nova] [instance: 70ed982f-affd-4dd1-bc90-c64e7c6d49d2] Updated VIF entry in instance network info cache for port fe7c09da-d6bd-4f72-afa2-1a70bc663407. {{(pid=63515) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 789.781850] env[63515]: DEBUG nova.network.neutron [req-dcbed5a2-17ec-49e5-bc73-b36635377dd9 req-62d11547-fdd9-4e8a-a99c-79521cf27edb service nova] [instance: 70ed982f-affd-4dd1-bc90-c64e7c6d49d2] Updating instance_info_cache with network_info: [{"id": "fe7c09da-d6bd-4f72-afa2-1a70bc663407", "address": "fa:16:3e:06:ad:d9", "network": {"id": "61e1e43b-1c9a-4769-8de9-6fce1435e861", "bridge": "br-int", "label": "tempest-ServersTestJSON-481843141-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.245", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f8662644a3354b46a19edbd1f8ce2c28", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4105fae7-3de1-4a6c-a44b-39097c174699", "external-id": "nsx-vlan-transportzone-773", "segmentation_id": 773, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfe7c09da-d6", "ovs_interfaceid": "fe7c09da-d6bd-4f72-afa2-1a70bc663407", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 789.789593] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1111240, 'name': CreateVM_Task} progress is 6%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 789.854210] env[63515]: DEBUG oslo_vmware.api [None req-83fbd51b-fc37-49eb-8423-cdaa019d3b0c tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Task: {'id': task-1111234, 'name': Rename_Task} progress is 99%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 789.859663] env[63515]: DEBUG oslo_vmware.api [None req-67e41316-801d-46e9-947d-28c19f5a3017 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Task: {'id': task-1111231, 'name': CreateSnapshot_Task, 'duration_secs': 0.655792} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 789.859883] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-67e41316-801d-46e9-947d-28c19f5a3017 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: 84dbf321-2c70-4e08-b430-cb5a06fc6829] Created Snapshot of the VM instance {{(pid=63515) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 789.861454] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b77b6200-fac6-4e17-afd9-c33d3fa0c066 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.877926] env[63515]: DEBUG oslo_vmware.api [None req-b53e671a-9cf4-4e34-a2ac-4321dce89b1c tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] Task: {'id': task-1111236, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.203767} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 789.878983] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-b53e671a-9cf4-4e34-a2ac-4321dce89b1c tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] Deleted the datastore file {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 789.878983] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-b53e671a-9cf4-4e34-a2ac-4321dce89b1c tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] [instance: 192137b0-03e5-4bc4-b911-4b4f1a874f74] Deleted contents of the VM from datastore datastore1 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 789.878983] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-b53e671a-9cf4-4e34-a2ac-4321dce89b1c tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] [instance: 192137b0-03e5-4bc4-b911-4b4f1a874f74] Instance destroyed {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 789.984967] env[63515]: DEBUG oslo_vmware.api [None req-fa0cfe6d-3452-46d4-9987-2cb422956d62 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Task: {'id': task-1111233, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.52477} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 789.985177] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-fa0cfe6d-3452-46d4-9987-2cb422956d62 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk to [datastore1] 63ce797d-7180-4209-ac2c-81978bf7607a/63ce797d-7180-4209-ac2c-81978bf7607a.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 789.985418] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-fa0cfe6d-3452-46d4-9987-2cb422956d62 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] [instance: 63ce797d-7180-4209-ac2c-81978bf7607a] Extending root virtual disk to 1048576 {{(pid=63515) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 789.985687] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-0ef3d913-7ee4-4a1d-b478-4dac8e959a46 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.992713] env[63515]: DEBUG oslo_vmware.api [None req-fa0cfe6d-3452-46d4-9987-2cb422956d62 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Waiting for the task: (returnval){ [ 789.992713] env[63515]: value = "task-1111241" [ 789.992713] env[63515]: _type = "Task" [ 789.992713] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 790.002675] env[63515]: DEBUG oslo_vmware.api [None req-fa0cfe6d-3452-46d4-9987-2cb422956d62 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Task: {'id': task-1111241, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 790.039222] env[63515]: DEBUG oslo_vmware.api [None req-65badfd4-59e2-4bf6-bac8-39c1a78dd350 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Task: {'id': task-1111238, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.204896} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 790.039222] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-65badfd4-59e2-4bf6-bac8-39c1a78dd350 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Deleted the datastore file {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 790.039222] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-65badfd4-59e2-4bf6-bac8-39c1a78dd350 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] [instance: f1d01b75-ac9d-458d-8cc2-ae64cffca4e8] Deleted contents of the VM from datastore datastore2 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 790.039222] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-65badfd4-59e2-4bf6-bac8-39c1a78dd350 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] [instance: f1d01b75-ac9d-458d-8cc2-ae64cffca4e8] Instance destroyed {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 790.039222] env[63515]: INFO nova.compute.manager [None req-65badfd4-59e2-4bf6-bac8-39c1a78dd350 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] [instance: f1d01b75-ac9d-458d-8cc2-ae64cffca4e8] Took 1.12 seconds to destroy the instance on the hypervisor. [ 790.039222] env[63515]: DEBUG oslo.service.loopingcall [None req-65badfd4-59e2-4bf6-bac8-39c1a78dd350 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 790.040967] env[63515]: DEBUG nova.compute.manager [-] [instance: f1d01b75-ac9d-458d-8cc2-ae64cffca4e8] Deallocating network for instance {{(pid=63515) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 790.041138] env[63515]: DEBUG nova.network.neutron [-] [instance: f1d01b75-ac9d-458d-8cc2-ae64cffca4e8] deallocate_for_instance() {{(pid=63515) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 790.115286] env[63515]: DEBUG nova.network.neutron [None req-9afb130a-5980-40d4-acae-35d697720c32 tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] [instance: d7e8dcdb-41cb-46fb-8b61-d251e7c2d372] Updating instance_info_cache with network_info: [{"id": "9378f59d-f9bc-4c2e-8336-2c5710f987cf", "address": "fa:16:3e:7b:d9:e1", "network": {"id": "57de0b07-78e5-453d-83ba-b12b7dcaf6ce", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-272799484-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.170", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "12a90929c98e40d9a1263929e2b97532", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a1895250-76cc-41f7-b7f8-2e5679494607", "external-id": "nsx-vlan-transportzone-785", "segmentation_id": 785, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9378f59d-f9", "ovs_interfaceid": "9378f59d-f9bc-4c2e-8336-2c5710f987cf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 790.124467] env[63515]: DEBUG oslo_vmware.api [None req-8d4c1f89-f73c-481b-8b79-61983aeba07b tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Task: {'id': task-1111239, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 790.134362] env[63515]: DEBUG nova.compute.utils [None req-66c3e202-5a3e-433e-b543-4bfb02c8ad30 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775-project-member] Using /dev/sd instead of None {{(pid=63515) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 790.142020] env[63515]: DEBUG nova.compute.manager [None req-66c3e202-5a3e-433e-b543-4bfb02c8ad30 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775-project-member] [instance: 4e453127-1f3e-40ea-819f-6678479826c8] Allocating IP information in the background. {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 790.142020] env[63515]: DEBUG nova.network.neutron [None req-66c3e202-5a3e-433e-b543-4bfb02c8ad30 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775-project-member] [instance: 4e453127-1f3e-40ea-819f-6678479826c8] allocate_for_instance() {{(pid=63515) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 790.186997] env[63515]: DEBUG nova.policy [None req-66c3e202-5a3e-433e-b543-4bfb02c8ad30 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '806793c336c14d95899a6794b96bbc1d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '60d9dcd0f30e48e7965ecfc1a72f5375', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63515) authorize /opt/stack/nova/nova/policy.py:201}} [ 790.284240] env[63515]: DEBUG oslo_concurrency.lockutils [req-dcbed5a2-17ec-49e5-bc73-b36635377dd9 req-62d11547-fdd9-4e8a-a99c-79521cf27edb service nova] Releasing lock "refresh_cache-70ed982f-affd-4dd1-bc90-c64e7c6d49d2" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 790.290266] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1111240, 'name': CreateVM_Task} progress is 99%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 790.351841] env[63515]: DEBUG oslo_vmware.api [None req-83fbd51b-fc37-49eb-8423-cdaa019d3b0c tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Task: {'id': task-1111234, 'name': Rename_Task} progress is 99%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 790.380819] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-67e41316-801d-46e9-947d-28c19f5a3017 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: 84dbf321-2c70-4e08-b430-cb5a06fc6829] Creating linked-clone VM from snapshot {{(pid=63515) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 790.381240] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-b5d02c57-6ad1-4d13-9b96-6d03fc8317b8 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.394744] env[63515]: DEBUG oslo_vmware.api [None req-67e41316-801d-46e9-947d-28c19f5a3017 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Waiting for the task: (returnval){ [ 790.394744] env[63515]: value = "task-1111242" [ 790.394744] env[63515]: _type = "Task" [ 790.394744] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 790.409600] env[63515]: DEBUG oslo_vmware.api [None req-67e41316-801d-46e9-947d-28c19f5a3017 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Task: {'id': task-1111242, 'name': CloneVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 790.503330] env[63515]: DEBUG oslo_vmware.api [None req-fa0cfe6d-3452-46d4-9987-2cb422956d62 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Task: {'id': task-1111241, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.075419} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 790.507180] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-fa0cfe6d-3452-46d4-9987-2cb422956d62 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] [instance: 63ce797d-7180-4209-ac2c-81978bf7607a] Extended root virtual disk {{(pid=63515) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 790.507476] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2e59903-e56c-476b-a922-38f2f69f5ca0 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.540197] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-fa0cfe6d-3452-46d4-9987-2cb422956d62 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] [instance: 63ce797d-7180-4209-ac2c-81978bf7607a] Reconfiguring VM instance instance-0000003b to attach disk [datastore1] 63ce797d-7180-4209-ac2c-81978bf7607a/63ce797d-7180-4209-ac2c-81978bf7607a.vmdk or device None with type sparse {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 790.543476] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-45802c5d-9077-4d0d-8588-0ec38807c7a9 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.567680] env[63515]: DEBUG oslo_vmware.api [None req-fa0cfe6d-3452-46d4-9987-2cb422956d62 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Waiting for the task: (returnval){ [ 790.567680] env[63515]: value = "task-1111243" [ 790.567680] env[63515]: _type = "Task" [ 790.567680] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 790.576400] env[63515]: DEBUG oslo_vmware.api [None req-fa0cfe6d-3452-46d4-9987-2cb422956d62 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Task: {'id': task-1111243, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 790.616924] env[63515]: DEBUG oslo_vmware.api [None req-8d4c1f89-f73c-481b-8b79-61983aeba07b tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Task: {'id': task-1111239, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 790.619480] env[63515]: DEBUG oslo_concurrency.lockutils [None req-9afb130a-5980-40d4-acae-35d697720c32 tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Releasing lock "refresh_cache-d7e8dcdb-41cb-46fb-8b61-d251e7c2d372" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 790.642030] env[63515]: DEBUG oslo_concurrency.lockutils [None req-0e5835c9-1a06-4538-91e4-48a9dfc2854d tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.021s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 790.642030] env[63515]: DEBUG oslo_concurrency.lockutils [None req-95950a02-20c9-4a3f-9ea8-1f7f19725cc1 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 28.137s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 790.643019] env[63515]: INFO nova.compute.claims [None req-95950a02-20c9-4a3f-9ea8-1f7f19725cc1 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: 4ef8066a-b68c-457b-9964-b1c34bab0fc3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 790.648086] env[63515]: DEBUG nova.network.neutron [None req-66c3e202-5a3e-433e-b543-4bfb02c8ad30 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775-project-member] [instance: 4e453127-1f3e-40ea-819f-6678479826c8] Successfully created port: 3d5ad33f-8649-42c8-b899-ca20c9a2636e {{(pid=63515) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 790.651166] env[63515]: DEBUG nova.compute.manager [None req-66c3e202-5a3e-433e-b543-4bfb02c8ad30 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775-project-member] [instance: 4e453127-1f3e-40ea-819f-6678479826c8] Start building block device mappings for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 790.780892] env[63515]: DEBUG nova.network.neutron [req-10c14eaf-2c3b-4bbd-8f4f-77f8dafd636a req-75f4ef9e-6e4d-463a-a0d0-2fef3a6dcc20 service nova] [instance: e666825c-ff4e-4a0e-93c0-43c00f167bbb] Updated VIF entry in instance network info cache for port 7872b4c3-d54f-4613-85d2-f568ec3cb47f. {{(pid=63515) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 790.781278] env[63515]: DEBUG nova.network.neutron [req-10c14eaf-2c3b-4bbd-8f4f-77f8dafd636a req-75f4ef9e-6e4d-463a-a0d0-2fef3a6dcc20 service nova] [instance: e666825c-ff4e-4a0e-93c0-43c00f167bbb] Updating instance_info_cache with network_info: [{"id": "7872b4c3-d54f-4613-85d2-f568ec3cb47f", "address": "fa:16:3e:a1:a9:1e", "network": {"id": "b4abcbd0-0eac-4bde-ad11-7f93afcedcef", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.123", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "b569255cc43e42e7ae3f2b4ad37c6ef4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0cd5d325-3053-407e-a4ee-f627e82a23f9", "external-id": "nsx-vlan-transportzone-809", "segmentation_id": 809, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7872b4c3-d5", "ovs_interfaceid": "7872b4c3-d54f-4613-85d2-f568ec3cb47f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 790.796645] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1111240, 'name': CreateVM_Task, 'duration_secs': 0.557993} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 790.796876] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e666825c-ff4e-4a0e-93c0-43c00f167bbb] Created VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 790.798101] env[63515]: DEBUG oslo_concurrency.lockutils [None req-4b5baf79-d4c3-467e-88d4-a131ea6620d7 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 790.798432] env[63515]: DEBUG oslo_concurrency.lockutils [None req-4b5baf79-d4c3-467e-88d4-a131ea6620d7 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 790.798904] env[63515]: DEBUG oslo_concurrency.lockutils [None req-4b5baf79-d4c3-467e-88d4-a131ea6620d7 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 790.799490] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f5130e4a-77b2-4393-a075-588e907d7d45 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.808963] env[63515]: DEBUG oslo_vmware.api [None req-4b5baf79-d4c3-467e-88d4-a131ea6620d7 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] Waiting for the task: (returnval){ [ 790.808963] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]527d5858-7d8c-cea2-f87f-f300e5880833" [ 790.808963] env[63515]: _type = "Task" [ 790.808963] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 790.818589] env[63515]: DEBUG oslo_vmware.api [None req-4b5baf79-d4c3-467e-88d4-a131ea6620d7 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]527d5858-7d8c-cea2-f87f-f300e5880833, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 790.851758] env[63515]: DEBUG oslo_vmware.api [None req-83fbd51b-fc37-49eb-8423-cdaa019d3b0c tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Task: {'id': task-1111234, 'name': Rename_Task, 'duration_secs': 1.211021} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 790.852511] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-83fbd51b-fc37-49eb-8423-cdaa019d3b0c tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] [instance: 5cbce760-0163-4b27-8ae3-e46c926c8916] Powering on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 790.853048] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-665f7d8a-983b-442b-bfc5-e51e29552281 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.862023] env[63515]: DEBUG oslo_vmware.api [None req-83fbd51b-fc37-49eb-8423-cdaa019d3b0c tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Waiting for the task: (returnval){ [ 790.862023] env[63515]: value = "task-1111244" [ 790.862023] env[63515]: _type = "Task" [ 790.862023] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 790.875594] env[63515]: DEBUG oslo_vmware.api [None req-83fbd51b-fc37-49eb-8423-cdaa019d3b0c tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Task: {'id': task-1111244, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 790.910518] env[63515]: DEBUG oslo_vmware.api [None req-67e41316-801d-46e9-947d-28c19f5a3017 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Task: {'id': task-1111242, 'name': CloneVM_Task} progress is 94%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 790.924764] env[63515]: DEBUG nova.virt.hardware [None req-b53e671a-9cf4-4e34-a2ac-4321dce89b1c tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T02:52:14Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T02:51:56Z,direct_url=,disk_format='vmdk',id=8a120570-cb06-4099-b262-554ca0ad15c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b569255cc43e42e7ae3f2b4ad37c6ef4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T02:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 790.927067] env[63515]: DEBUG nova.virt.hardware [None req-b53e671a-9cf4-4e34-a2ac-4321dce89b1c tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] Flavor limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 790.927067] env[63515]: DEBUG nova.virt.hardware [None req-b53e671a-9cf4-4e34-a2ac-4321dce89b1c tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] Image limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 790.927067] env[63515]: DEBUG nova.virt.hardware [None req-b53e671a-9cf4-4e34-a2ac-4321dce89b1c tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] Flavor pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 790.927067] env[63515]: DEBUG nova.virt.hardware [None req-b53e671a-9cf4-4e34-a2ac-4321dce89b1c tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] Image pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 790.927067] env[63515]: DEBUG nova.virt.hardware [None req-b53e671a-9cf4-4e34-a2ac-4321dce89b1c tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 790.927067] env[63515]: DEBUG nova.virt.hardware [None req-b53e671a-9cf4-4e34-a2ac-4321dce89b1c tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 790.927067] env[63515]: DEBUG nova.virt.hardware [None req-b53e671a-9cf4-4e34-a2ac-4321dce89b1c tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 790.927543] env[63515]: DEBUG nova.virt.hardware [None req-b53e671a-9cf4-4e34-a2ac-4321dce89b1c tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] Got 1 possible topologies {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 790.927875] env[63515]: DEBUG nova.virt.hardware [None req-b53e671a-9cf4-4e34-a2ac-4321dce89b1c tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 790.928201] env[63515]: DEBUG nova.virt.hardware [None req-b53e671a-9cf4-4e34-a2ac-4321dce89b1c tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 790.929869] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9893b3f-0f5a-4bae-9af6-81d8522b3645 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.941390] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4f8ffb4-8f6c-4a77-b611-fa203373ce0e {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.958477] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-b53e671a-9cf4-4e34-a2ac-4321dce89b1c tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] [instance: 192137b0-03e5-4bc4-b911-4b4f1a874f74] Instance VIF info [] {{(pid=63515) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 790.967019] env[63515]: DEBUG oslo.service.loopingcall [None req-b53e671a-9cf4-4e34-a2ac-4321dce89b1c tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 790.967019] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 192137b0-03e5-4bc4-b911-4b4f1a874f74] Creating VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 790.967019] env[63515]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-4daf5e41-3d71-4beb-bf20-37bd5a1d2c7b {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.983603] env[63515]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 790.983603] env[63515]: value = "task-1111245" [ 790.983603] env[63515]: _type = "Task" [ 790.983603] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 790.987535] env[63515]: DEBUG nova.network.neutron [-] [instance: f1d01b75-ac9d-458d-8cc2-ae64cffca4e8] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 790.993702] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1111245, 'name': CreateVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 791.016590] env[63515]: DEBUG nova.compute.manager [req-c2624a85-8488-4b92-80ba-29e80f662ac0 req-e4e4c376-2422-4acd-9f81-771a657972ae service nova] [instance: f1d01b75-ac9d-458d-8cc2-ae64cffca4e8] Received event network-vif-deleted-48f0868f-94db-42f7-8153-d1a27fa02707 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 791.016590] env[63515]: INFO nova.compute.manager [req-c2624a85-8488-4b92-80ba-29e80f662ac0 req-e4e4c376-2422-4acd-9f81-771a657972ae service nova] [instance: f1d01b75-ac9d-458d-8cc2-ae64cffca4e8] Neutron deleted interface 48f0868f-94db-42f7-8153-d1a27fa02707; detaching it from the instance and deleting it from the info cache [ 791.016590] env[63515]: DEBUG nova.network.neutron [req-c2624a85-8488-4b92-80ba-29e80f662ac0 req-e4e4c376-2422-4acd-9f81-771a657972ae service nova] [instance: f1d01b75-ac9d-458d-8cc2-ae64cffca4e8] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 791.078583] env[63515]: DEBUG oslo_vmware.api [None req-fa0cfe6d-3452-46d4-9987-2cb422956d62 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Task: {'id': task-1111243, 'name': ReconfigVM_Task, 'duration_secs': 0.276743} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 791.079026] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-fa0cfe6d-3452-46d4-9987-2cb422956d62 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] [instance: 63ce797d-7180-4209-ac2c-81978bf7607a] Reconfigured VM instance instance-0000003b to attach disk [datastore1] 63ce797d-7180-4209-ac2c-81978bf7607a/63ce797d-7180-4209-ac2c-81978bf7607a.vmdk or device None with type sparse {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 791.080089] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-081014b3-4832-4d86-b9a2-67afef378309 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.087414] env[63515]: DEBUG oslo_vmware.api [None req-fa0cfe6d-3452-46d4-9987-2cb422956d62 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Waiting for the task: (returnval){ [ 791.087414] env[63515]: value = "task-1111246" [ 791.087414] env[63515]: _type = "Task" [ 791.087414] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 791.096953] env[63515]: DEBUG oslo_vmware.api [None req-fa0cfe6d-3452-46d4-9987-2cb422956d62 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Task: {'id': task-1111246, 'name': Rename_Task} progress is 5%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 791.120442] env[63515]: DEBUG oslo_vmware.api [None req-8d4c1f89-f73c-481b-8b79-61983aeba07b tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Task: {'id': task-1111239, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 791.126514] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-9afb130a-5980-40d4-acae-35d697720c32 tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] [instance: d7e8dcdb-41cb-46fb-8b61-d251e7c2d372] Powering on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 791.126841] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0473983a-f2ee-46dc-8a94-4dcae34bc88d {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.133742] env[63515]: DEBUG oslo_vmware.api [None req-9afb130a-5980-40d4-acae-35d697720c32 tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Waiting for the task: (returnval){ [ 791.133742] env[63515]: value = "task-1111247" [ 791.133742] env[63515]: _type = "Task" [ 791.133742] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 791.144287] env[63515]: DEBUG oslo_vmware.api [None req-9afb130a-5980-40d4-acae-35d697720c32 tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Task: {'id': task-1111247, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 791.290598] env[63515]: DEBUG oslo_concurrency.lockutils [req-10c14eaf-2c3b-4bbd-8f4f-77f8dafd636a req-75f4ef9e-6e4d-463a-a0d0-2fef3a6dcc20 service nova] Releasing lock "refresh_cache-e666825c-ff4e-4a0e-93c0-43c00f167bbb" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 791.320755] env[63515]: DEBUG oslo_vmware.api [None req-4b5baf79-d4c3-467e-88d4-a131ea6620d7 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]527d5858-7d8c-cea2-f87f-f300e5880833, 'name': SearchDatastore_Task, 'duration_secs': 0.016023} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 791.321067] env[63515]: DEBUG oslo_concurrency.lockutils [None req-4b5baf79-d4c3-467e-88d4-a131ea6620d7 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 791.321320] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-4b5baf79-d4c3-467e-88d4-a131ea6620d7 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] [instance: e666825c-ff4e-4a0e-93c0-43c00f167bbb] Processing image 8a120570-cb06-4099-b262-554ca0ad15c5 {{(pid=63515) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 791.321635] env[63515]: DEBUG oslo_concurrency.lockutils [None req-4b5baf79-d4c3-467e-88d4-a131ea6620d7 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 791.321693] env[63515]: DEBUG oslo_concurrency.lockutils [None req-4b5baf79-d4c3-467e-88d4-a131ea6620d7 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 791.321883] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-4b5baf79-d4c3-467e-88d4-a131ea6620d7 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 791.322162] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-bfa1c3bb-c225-4b79-a314-118c4df2c574 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.330298] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-4b5baf79-d4c3-467e-88d4-a131ea6620d7 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 791.330488] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-4b5baf79-d4c3-467e-88d4-a131ea6620d7 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63515) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 791.331224] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-190a08c7-d563-487f-b8be-b7fc0b922bce {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.336078] env[63515]: DEBUG oslo_vmware.api [None req-4b5baf79-d4c3-467e-88d4-a131ea6620d7 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] Waiting for the task: (returnval){ [ 791.336078] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]5223fd03-1001-f386-18ee-873361ebbaef" [ 791.336078] env[63515]: _type = "Task" [ 791.336078] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 791.344424] env[63515]: DEBUG oslo_vmware.api [None req-4b5baf79-d4c3-467e-88d4-a131ea6620d7 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]5223fd03-1001-f386-18ee-873361ebbaef, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 791.371035] env[63515]: DEBUG oslo_vmware.api [None req-83fbd51b-fc37-49eb-8423-cdaa019d3b0c tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Task: {'id': task-1111244, 'name': PowerOnVM_Task} progress is 33%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 791.405316] env[63515]: DEBUG oslo_vmware.api [None req-67e41316-801d-46e9-947d-28c19f5a3017 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Task: {'id': task-1111242, 'name': CloneVM_Task} progress is 94%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 791.490364] env[63515]: INFO nova.compute.manager [-] [instance: f1d01b75-ac9d-458d-8cc2-ae64cffca4e8] Took 1.45 seconds to deallocate network for instance. [ 791.495059] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1111245, 'name': CreateVM_Task} progress is 99%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 791.519207] env[63515]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-44c74835-6a4a-4ead-a39e-d5158ac3c550 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.529130] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93e389f7-68a2-4ea5-91a5-21e35bae347e {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.560801] env[63515]: DEBUG nova.compute.manager [req-c2624a85-8488-4b92-80ba-29e80f662ac0 req-e4e4c376-2422-4acd-9f81-771a657972ae service nova] [instance: f1d01b75-ac9d-458d-8cc2-ae64cffca4e8] Detach interface failed, port_id=48f0868f-94db-42f7-8153-d1a27fa02707, reason: Instance f1d01b75-ac9d-458d-8cc2-ae64cffca4e8 could not be found. {{(pid=63515) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 791.598982] env[63515]: DEBUG oslo_vmware.api [None req-fa0cfe6d-3452-46d4-9987-2cb422956d62 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Task: {'id': task-1111246, 'name': Rename_Task, 'duration_secs': 0.147373} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 791.599299] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-fa0cfe6d-3452-46d4-9987-2cb422956d62 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] [instance: 63ce797d-7180-4209-ac2c-81978bf7607a] Powering on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 791.599547] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-cedd1e0c-03b9-48e3-821c-05127dbee6ea {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.605571] env[63515]: DEBUG oslo_vmware.api [None req-fa0cfe6d-3452-46d4-9987-2cb422956d62 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Waiting for the task: (returnval){ [ 791.605571] env[63515]: value = "task-1111248" [ 791.605571] env[63515]: _type = "Task" [ 791.605571] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 791.615908] env[63515]: DEBUG oslo_vmware.api [None req-fa0cfe6d-3452-46d4-9987-2cb422956d62 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Task: {'id': task-1111248, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 791.620814] env[63515]: DEBUG oslo_vmware.api [None req-8d4c1f89-f73c-481b-8b79-61983aeba07b tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Task: {'id': task-1111239, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 791.644253] env[63515]: DEBUG oslo_vmware.api [None req-9afb130a-5980-40d4-acae-35d697720c32 tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Task: {'id': task-1111247, 'name': PowerOnVM_Task, 'duration_secs': 0.425226} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 791.644253] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-9afb130a-5980-40d4-acae-35d697720c32 tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] [instance: d7e8dcdb-41cb-46fb-8b61-d251e7c2d372] Powered on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 791.644253] env[63515]: DEBUG nova.compute.manager [None req-9afb130a-5980-40d4-acae-35d697720c32 tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] [instance: d7e8dcdb-41cb-46fb-8b61-d251e7c2d372] Checking state {{(pid=63515) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 791.647395] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a2c184b-cf32-435f-9eb2-d283292bd841 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.663927] env[63515]: DEBUG nova.compute.manager [None req-66c3e202-5a3e-433e-b543-4bfb02c8ad30 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775-project-member] [instance: 4e453127-1f3e-40ea-819f-6678479826c8] Start spawning the instance on the hypervisor. {{(pid=63515) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 791.695984] env[63515]: DEBUG nova.virt.hardware [None req-66c3e202-5a3e-433e-b543-4bfb02c8ad30 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T02:52:14Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T02:51:56Z,direct_url=,disk_format='vmdk',id=8a120570-cb06-4099-b262-554ca0ad15c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b569255cc43e42e7ae3f2b4ad37c6ef4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T02:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 791.696293] env[63515]: DEBUG nova.virt.hardware [None req-66c3e202-5a3e-433e-b543-4bfb02c8ad30 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775-project-member] Flavor limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 791.696472] env[63515]: DEBUG nova.virt.hardware [None req-66c3e202-5a3e-433e-b543-4bfb02c8ad30 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775-project-member] Image limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 791.696656] env[63515]: DEBUG nova.virt.hardware [None req-66c3e202-5a3e-433e-b543-4bfb02c8ad30 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775-project-member] Flavor pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 791.696800] env[63515]: DEBUG nova.virt.hardware [None req-66c3e202-5a3e-433e-b543-4bfb02c8ad30 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775-project-member] Image pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 791.697027] env[63515]: DEBUG nova.virt.hardware [None req-66c3e202-5a3e-433e-b543-4bfb02c8ad30 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 791.697240] env[63515]: DEBUG nova.virt.hardware [None req-66c3e202-5a3e-433e-b543-4bfb02c8ad30 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 791.697347] env[63515]: DEBUG nova.virt.hardware [None req-66c3e202-5a3e-433e-b543-4bfb02c8ad30 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 791.701609] env[63515]: DEBUG nova.virt.hardware [None req-66c3e202-5a3e-433e-b543-4bfb02c8ad30 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775-project-member] Got 1 possible topologies {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 791.701814] env[63515]: DEBUG nova.virt.hardware [None req-66c3e202-5a3e-433e-b543-4bfb02c8ad30 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 791.702229] env[63515]: DEBUG nova.virt.hardware [None req-66c3e202-5a3e-433e-b543-4bfb02c8ad30 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 791.703234] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97b2d742-a19c-4983-a871-3ea3222f5e68 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.714210] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a42b99ae-e153-45a2-9037-662530cefa5c {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.851915] env[63515]: DEBUG oslo_vmware.api [None req-4b5baf79-d4c3-467e-88d4-a131ea6620d7 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]5223fd03-1001-f386-18ee-873361ebbaef, 'name': SearchDatastore_Task, 'duration_secs': 0.009579} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 791.856356] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-21b3a427-f974-4f2b-8b39-cb1de4b74630 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.860520] env[63515]: DEBUG oslo_vmware.api [None req-4b5baf79-d4c3-467e-88d4-a131ea6620d7 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] Waiting for the task: (returnval){ [ 791.860520] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]522d58a5-4a45-72f6-89a9-9d16a6fabf4a" [ 791.860520] env[63515]: _type = "Task" [ 791.860520] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 791.878207] env[63515]: DEBUG oslo_vmware.api [None req-4b5baf79-d4c3-467e-88d4-a131ea6620d7 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]522d58a5-4a45-72f6-89a9-9d16a6fabf4a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 791.881178] env[63515]: DEBUG oslo_vmware.api [None req-83fbd51b-fc37-49eb-8423-cdaa019d3b0c tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Task: {'id': task-1111244, 'name': PowerOnVM_Task} progress is 33%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 791.905266] env[63515]: DEBUG oslo_vmware.api [None req-67e41316-801d-46e9-947d-28c19f5a3017 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Task: {'id': task-1111242, 'name': CloneVM_Task} progress is 100%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 791.993804] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1111245, 'name': CreateVM_Task, 'duration_secs': 0.990552} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 791.994205] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 192137b0-03e5-4bc4-b911-4b4f1a874f74] Created VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 791.996885] env[63515]: DEBUG oslo_concurrency.lockutils [None req-b53e671a-9cf4-4e34-a2ac-4321dce89b1c tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 791.997074] env[63515]: DEBUG oslo_concurrency.lockutils [None req-b53e671a-9cf4-4e34-a2ac-4321dce89b1c tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 791.997381] env[63515]: DEBUG oslo_concurrency.lockutils [None req-b53e671a-9cf4-4e34-a2ac-4321dce89b1c tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 791.998034] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4d4a80aa-23dd-4284-bed4-86adf8351fd2 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.000252] env[63515]: DEBUG oslo_concurrency.lockutils [None req-65badfd4-59e2-4bf6-bac8-39c1a78dd350 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 792.003118] env[63515]: DEBUG oslo_vmware.api [None req-b53e671a-9cf4-4e34-a2ac-4321dce89b1c tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] Waiting for the task: (returnval){ [ 792.003118] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]5235fcfb-7e27-26a2-ad26-ce65a51f849d" [ 792.003118] env[63515]: _type = "Task" [ 792.003118] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 792.010752] env[63515]: DEBUG oslo_vmware.api [None req-b53e671a-9cf4-4e34-a2ac-4321dce89b1c tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]5235fcfb-7e27-26a2-ad26-ce65a51f849d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 792.053047] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a66f299-fb95-495f-b5d4-ad76d4329832 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.061688] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97715f0b-47cc-462c-a4fc-1ad8be217311 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.094989] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2cbeee75-cfa1-4695-9c63-b1c19f798611 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.103073] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ec084b5-1ff4-4ec3-9e77-31278c53fd06 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.124738] env[63515]: DEBUG nova.compute.provider_tree [None req-95950a02-20c9-4a3f-9ea8-1f7f19725cc1 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 792.130100] env[63515]: DEBUG oslo_vmware.api [None req-fa0cfe6d-3452-46d4-9987-2cb422956d62 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Task: {'id': task-1111248, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 792.135071] env[63515]: DEBUG oslo_vmware.api [None req-8d4c1f89-f73c-481b-8b79-61983aeba07b tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Task: {'id': task-1111239, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 792.373782] env[63515]: DEBUG oslo_vmware.api [None req-83fbd51b-fc37-49eb-8423-cdaa019d3b0c tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Task: {'id': task-1111244, 'name': PowerOnVM_Task} progress is 100%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 792.377567] env[63515]: DEBUG oslo_vmware.api [None req-4b5baf79-d4c3-467e-88d4-a131ea6620d7 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]522d58a5-4a45-72f6-89a9-9d16a6fabf4a, 'name': SearchDatastore_Task, 'duration_secs': 0.028947} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 792.377931] env[63515]: DEBUG oslo_concurrency.lockutils [None req-4b5baf79-d4c3-467e-88d4-a131ea6620d7 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 792.378233] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-4b5baf79-d4c3-467e-88d4-a131ea6620d7 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk to [datastore1] e666825c-ff4e-4a0e-93c0-43c00f167bbb/e666825c-ff4e-4a0e-93c0-43c00f167bbb.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 792.378509] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-10757175-0768-425c-a897-70cd1f6e2144 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.384997] env[63515]: DEBUG oslo_vmware.api [None req-4b5baf79-d4c3-467e-88d4-a131ea6620d7 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] Waiting for the task: (returnval){ [ 792.384997] env[63515]: value = "task-1111249" [ 792.384997] env[63515]: _type = "Task" [ 792.384997] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 792.394219] env[63515]: DEBUG oslo_vmware.api [None req-4b5baf79-d4c3-467e-88d4-a131ea6620d7 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] Task: {'id': task-1111249, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 792.406721] env[63515]: DEBUG oslo_vmware.api [None req-67e41316-801d-46e9-947d-28c19f5a3017 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Task: {'id': task-1111242, 'name': CloneVM_Task} progress is 100%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 792.516665] env[63515]: DEBUG oslo_vmware.api [None req-b53e671a-9cf4-4e34-a2ac-4321dce89b1c tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]5235fcfb-7e27-26a2-ad26-ce65a51f849d, 'name': SearchDatastore_Task, 'duration_secs': 0.010281} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 792.517180] env[63515]: DEBUG oslo_concurrency.lockutils [None req-b53e671a-9cf4-4e34-a2ac-4321dce89b1c tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 792.517561] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-b53e671a-9cf4-4e34-a2ac-4321dce89b1c tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] [instance: 192137b0-03e5-4bc4-b911-4b4f1a874f74] Processing image 8a120570-cb06-4099-b262-554ca0ad15c5 {{(pid=63515) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 792.517939] env[63515]: DEBUG oslo_concurrency.lockutils [None req-b53e671a-9cf4-4e34-a2ac-4321dce89b1c tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 792.518227] env[63515]: DEBUG oslo_concurrency.lockutils [None req-b53e671a-9cf4-4e34-a2ac-4321dce89b1c tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 792.518534] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-b53e671a-9cf4-4e34-a2ac-4321dce89b1c tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 792.518943] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-99d1edbe-44d9-469a-b86e-913310eb801f {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.530113] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-b53e671a-9cf4-4e34-a2ac-4321dce89b1c tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 792.530212] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-b53e671a-9cf4-4e34-a2ac-4321dce89b1c tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63515) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 792.531383] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cdc1274f-d95a-421c-b5d9-c97433d1fd7a {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.538431] env[63515]: DEBUG oslo_vmware.api [None req-b53e671a-9cf4-4e34-a2ac-4321dce89b1c tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] Waiting for the task: (returnval){ [ 792.538431] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]52655665-ee99-1f93-1517-214d0452dc08" [ 792.538431] env[63515]: _type = "Task" [ 792.538431] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 792.547582] env[63515]: DEBUG oslo_vmware.api [None req-b53e671a-9cf4-4e34-a2ac-4321dce89b1c tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52655665-ee99-1f93-1517-214d0452dc08, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 792.549432] env[63515]: DEBUG nova.compute.manager [req-4ddcc7f3-c294-4cce-a47e-93ba85699d21 req-d3a311ae-23de-4b86-aa2a-2cf47d00d50f service nova] [instance: 4e453127-1f3e-40ea-819f-6678479826c8] Received event network-vif-plugged-3d5ad33f-8649-42c8-b899-ca20c9a2636e {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 792.549646] env[63515]: DEBUG oslo_concurrency.lockutils [req-4ddcc7f3-c294-4cce-a47e-93ba85699d21 req-d3a311ae-23de-4b86-aa2a-2cf47d00d50f service nova] Acquiring lock "4e453127-1f3e-40ea-819f-6678479826c8-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 792.549853] env[63515]: DEBUG oslo_concurrency.lockutils [req-4ddcc7f3-c294-4cce-a47e-93ba85699d21 req-d3a311ae-23de-4b86-aa2a-2cf47d00d50f service nova] Lock "4e453127-1f3e-40ea-819f-6678479826c8-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 792.550039] env[63515]: DEBUG oslo_concurrency.lockutils [req-4ddcc7f3-c294-4cce-a47e-93ba85699d21 req-d3a311ae-23de-4b86-aa2a-2cf47d00d50f service nova] Lock "4e453127-1f3e-40ea-819f-6678479826c8-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 792.550218] env[63515]: DEBUG nova.compute.manager [req-4ddcc7f3-c294-4cce-a47e-93ba85699d21 req-d3a311ae-23de-4b86-aa2a-2cf47d00d50f service nova] [instance: 4e453127-1f3e-40ea-819f-6678479826c8] No waiting events found dispatching network-vif-plugged-3d5ad33f-8649-42c8-b899-ca20c9a2636e {{(pid=63515) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 792.550409] env[63515]: WARNING nova.compute.manager [req-4ddcc7f3-c294-4cce-a47e-93ba85699d21 req-d3a311ae-23de-4b86-aa2a-2cf47d00d50f service nova] [instance: 4e453127-1f3e-40ea-819f-6678479826c8] Received unexpected event network-vif-plugged-3d5ad33f-8649-42c8-b899-ca20c9a2636e for instance with vm_state building and task_state spawning. [ 792.623746] env[63515]: DEBUG oslo_vmware.api [None req-fa0cfe6d-3452-46d4-9987-2cb422956d62 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Task: {'id': task-1111248, 'name': PowerOnVM_Task, 'duration_secs': 0.790915} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 792.628698] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-fa0cfe6d-3452-46d4-9987-2cb422956d62 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] [instance: 63ce797d-7180-4209-ac2c-81978bf7607a] Powered on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 792.628698] env[63515]: INFO nova.compute.manager [None req-fa0cfe6d-3452-46d4-9987-2cb422956d62 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] [instance: 63ce797d-7180-4209-ac2c-81978bf7607a] Took 8.09 seconds to spawn the instance on the hypervisor. [ 792.628698] env[63515]: DEBUG nova.compute.manager [None req-fa0cfe6d-3452-46d4-9987-2cb422956d62 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] [instance: 63ce797d-7180-4209-ac2c-81978bf7607a] Checking state {{(pid=63515) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 792.628947] env[63515]: DEBUG oslo_vmware.api [None req-8d4c1f89-f73c-481b-8b79-61983aeba07b tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Task: {'id': task-1111239, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 792.629525] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3162b88c-bf38-44d4-8f2f-805df9ad5ae7 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.635735] env[63515]: DEBUG nova.scheduler.client.report [None req-95950a02-20c9-4a3f-9ea8-1f7f19725cc1 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 792.652286] env[63515]: DEBUG nova.network.neutron [None req-66c3e202-5a3e-433e-b543-4bfb02c8ad30 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775-project-member] [instance: 4e453127-1f3e-40ea-819f-6678479826c8] Successfully updated port: 3d5ad33f-8649-42c8-b899-ca20c9a2636e {{(pid=63515) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 792.871976] env[63515]: DEBUG oslo_vmware.api [None req-83fbd51b-fc37-49eb-8423-cdaa019d3b0c tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Task: {'id': task-1111244, 'name': PowerOnVM_Task, 'duration_secs': 1.536345} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 792.872417] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-83fbd51b-fc37-49eb-8423-cdaa019d3b0c tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] [instance: 5cbce760-0163-4b27-8ae3-e46c926c8916] Powered on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 792.872748] env[63515]: INFO nova.compute.manager [None req-83fbd51b-fc37-49eb-8423-cdaa019d3b0c tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] [instance: 5cbce760-0163-4b27-8ae3-e46c926c8916] Took 10.89 seconds to spawn the instance on the hypervisor. [ 792.873338] env[63515]: DEBUG nova.compute.manager [None req-83fbd51b-fc37-49eb-8423-cdaa019d3b0c tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] [instance: 5cbce760-0163-4b27-8ae3-e46c926c8916] Checking state {{(pid=63515) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 792.874017] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21bd8546-2cdf-4319-96b1-468652670903 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.894532] env[63515]: DEBUG oslo_vmware.api [None req-4b5baf79-d4c3-467e-88d4-a131ea6620d7 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] Task: {'id': task-1111249, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 792.906777] env[63515]: DEBUG oslo_vmware.api [None req-67e41316-801d-46e9-947d-28c19f5a3017 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Task: {'id': task-1111242, 'name': CloneVM_Task} progress is 100%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 793.051185] env[63515]: DEBUG oslo_vmware.api [None req-b53e671a-9cf4-4e34-a2ac-4321dce89b1c tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52655665-ee99-1f93-1517-214d0452dc08, 'name': SearchDatastore_Task, 'duration_secs': 0.024384} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 793.051810] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-41b68c32-7beb-4721-9cd4-2466a5256d70 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.056978] env[63515]: DEBUG oslo_vmware.api [None req-b53e671a-9cf4-4e34-a2ac-4321dce89b1c tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] Waiting for the task: (returnval){ [ 793.056978] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]52943790-e2e0-a889-6de0-4ddf5dbe49b2" [ 793.056978] env[63515]: _type = "Task" [ 793.056978] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 793.064471] env[63515]: DEBUG oslo_vmware.api [None req-b53e671a-9cf4-4e34-a2ac-4321dce89b1c tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52943790-e2e0-a889-6de0-4ddf5dbe49b2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 793.121159] env[63515]: DEBUG oslo_vmware.api [None req-8d4c1f89-f73c-481b-8b79-61983aeba07b tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Task: {'id': task-1111239, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 793.140287] env[63515]: DEBUG oslo_concurrency.lockutils [None req-95950a02-20c9-4a3f-9ea8-1f7f19725cc1 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.499s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 793.141122] env[63515]: DEBUG nova.compute.manager [None req-95950a02-20c9-4a3f-9ea8-1f7f19725cc1 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: 4ef8066a-b68c-457b-9964-b1c34bab0fc3] Start building networks asynchronously for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 793.144146] env[63515]: DEBUG oslo_concurrency.lockutils [None req-aef604e2-0f8c-4f11-a420-cb96f46c502e tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 28.111s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 793.144877] env[63515]: DEBUG nova.objects.instance [None req-aef604e2-0f8c-4f11-a420-cb96f46c502e tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Lazy-loading 'resources' on Instance uuid 001a48e1-b237-41f8-a6f4-2fe98d173512 {{(pid=63515) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 793.152894] env[63515]: INFO nova.compute.manager [None req-fa0cfe6d-3452-46d4-9987-2cb422956d62 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] [instance: 63ce797d-7180-4209-ac2c-81978bf7607a] Took 37.92 seconds to build instance. [ 793.157632] env[63515]: DEBUG oslo_concurrency.lockutils [None req-66c3e202-5a3e-433e-b543-4bfb02c8ad30 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775-project-member] Acquiring lock "refresh_cache-4e453127-1f3e-40ea-819f-6678479826c8" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 793.157632] env[63515]: DEBUG oslo_concurrency.lockutils [None req-66c3e202-5a3e-433e-b543-4bfb02c8ad30 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775-project-member] Acquired lock "refresh_cache-4e453127-1f3e-40ea-819f-6678479826c8" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 793.157702] env[63515]: DEBUG nova.network.neutron [None req-66c3e202-5a3e-433e-b543-4bfb02c8ad30 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775-project-member] [instance: 4e453127-1f3e-40ea-819f-6678479826c8] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 793.395313] env[63515]: INFO nova.compute.manager [None req-83fbd51b-fc37-49eb-8423-cdaa019d3b0c tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] [instance: 5cbce760-0163-4b27-8ae3-e46c926c8916] Took 43.81 seconds to build instance. [ 793.401344] env[63515]: DEBUG oslo_vmware.api [None req-4b5baf79-d4c3-467e-88d4-a131ea6620d7 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] Task: {'id': task-1111249, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.570881} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 793.404502] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-4b5baf79-d4c3-467e-88d4-a131ea6620d7 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk to [datastore1] e666825c-ff4e-4a0e-93c0-43c00f167bbb/e666825c-ff4e-4a0e-93c0-43c00f167bbb.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 793.404720] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-4b5baf79-d4c3-467e-88d4-a131ea6620d7 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] [instance: e666825c-ff4e-4a0e-93c0-43c00f167bbb] Extending root virtual disk to 1048576 {{(pid=63515) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 793.404968] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a6d49a92-f773-4ef4-9533-394ae5ad6d76 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.413582] env[63515]: DEBUG oslo_vmware.api [None req-67e41316-801d-46e9-947d-28c19f5a3017 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Task: {'id': task-1111242, 'name': CloneVM_Task} progress is 100%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 793.414981] env[63515]: DEBUG oslo_vmware.api [None req-4b5baf79-d4c3-467e-88d4-a131ea6620d7 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] Waiting for the task: (returnval){ [ 793.414981] env[63515]: value = "task-1111250" [ 793.414981] env[63515]: _type = "Task" [ 793.414981] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 793.424507] env[63515]: DEBUG oslo_vmware.api [None req-4b5baf79-d4c3-467e-88d4-a131ea6620d7 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] Task: {'id': task-1111250, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 793.566801] env[63515]: DEBUG oslo_vmware.api [None req-b53e671a-9cf4-4e34-a2ac-4321dce89b1c tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52943790-e2e0-a889-6de0-4ddf5dbe49b2, 'name': SearchDatastore_Task, 'duration_secs': 0.009606} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 793.567090] env[63515]: DEBUG oslo_concurrency.lockutils [None req-b53e671a-9cf4-4e34-a2ac-4321dce89b1c tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 793.567358] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-b53e671a-9cf4-4e34-a2ac-4321dce89b1c tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk to [datastore1] 192137b0-03e5-4bc4-b911-4b4f1a874f74/192137b0-03e5-4bc4-b911-4b4f1a874f74.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 793.567652] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-fd25628a-6223-4b0c-8b47-a3dfa8eab0b4 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.574291] env[63515]: DEBUG oslo_vmware.api [None req-b53e671a-9cf4-4e34-a2ac-4321dce89b1c tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] Waiting for the task: (returnval){ [ 793.574291] env[63515]: value = "task-1111251" [ 793.574291] env[63515]: _type = "Task" [ 793.574291] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 793.581668] env[63515]: DEBUG oslo_vmware.api [None req-b53e671a-9cf4-4e34-a2ac-4321dce89b1c tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] Task: {'id': task-1111251, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 793.623299] env[63515]: DEBUG oslo_vmware.api [None req-8d4c1f89-f73c-481b-8b79-61983aeba07b tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Task: {'id': task-1111239, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 793.648160] env[63515]: DEBUG nova.compute.utils [None req-95950a02-20c9-4a3f-9ea8-1f7f19725cc1 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Using /dev/sd instead of None {{(pid=63515) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 793.653711] env[63515]: DEBUG nova.compute.manager [None req-95950a02-20c9-4a3f-9ea8-1f7f19725cc1 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: 4ef8066a-b68c-457b-9964-b1c34bab0fc3] Allocating IP information in the background. {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 793.653711] env[63515]: DEBUG nova.network.neutron [None req-95950a02-20c9-4a3f-9ea8-1f7f19725cc1 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: 4ef8066a-b68c-457b-9964-b1c34bab0fc3] allocate_for_instance() {{(pid=63515) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 793.655992] env[63515]: DEBUG oslo_concurrency.lockutils [None req-fa0cfe6d-3452-46d4-9987-2cb422956d62 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Lock "63ce797d-7180-4209-ac2c-81978bf7607a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 95.525s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 793.726828] env[63515]: DEBUG nova.network.neutron [None req-66c3e202-5a3e-433e-b543-4bfb02c8ad30 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775-project-member] [instance: 4e453127-1f3e-40ea-819f-6678479826c8] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 793.766090] env[63515]: DEBUG nova.policy [None req-95950a02-20c9-4a3f-9ea8-1f7f19725cc1 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f58ac812970845dc825202b842feb5d0', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ea099ce07bfb4a8da014d9303cf552eb', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63515) authorize /opt/stack/nova/nova/policy.py:201}} [ 793.897074] env[63515]: DEBUG oslo_concurrency.lockutils [None req-83fbd51b-fc37-49eb-8423-cdaa019d3b0c tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Lock "5cbce760-0163-4b27-8ae3-e46c926c8916" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 96.800s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 793.923847] env[63515]: DEBUG oslo_vmware.api [None req-67e41316-801d-46e9-947d-28c19f5a3017 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Task: {'id': task-1111242, 'name': CloneVM_Task, 'duration_secs': 3.152987} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 793.927121] env[63515]: INFO nova.virt.vmwareapi.vmops [None req-67e41316-801d-46e9-947d-28c19f5a3017 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: 84dbf321-2c70-4e08-b430-cb5a06fc6829] Created linked-clone VM from snapshot [ 793.928239] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-398b3c57-4658-4803-9d71-16721d13fe2e {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.934449] env[63515]: DEBUG oslo_vmware.api [None req-4b5baf79-d4c3-467e-88d4-a131ea6620d7 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] Task: {'id': task-1111250, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.06748} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 793.935234] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-4b5baf79-d4c3-467e-88d4-a131ea6620d7 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] [instance: e666825c-ff4e-4a0e-93c0-43c00f167bbb] Extended root virtual disk {{(pid=63515) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 793.936226] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4ddc07e-09db-491e-8534-7f21a7f5d35e {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.945598] env[63515]: DEBUG nova.virt.vmwareapi.images [None req-67e41316-801d-46e9-947d-28c19f5a3017 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: 84dbf321-2c70-4e08-b430-cb5a06fc6829] Uploading image caa06f85-6406-4076-afa2-f2a71b73190c {{(pid=63515) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 793.971931] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-4b5baf79-d4c3-467e-88d4-a131ea6620d7 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] [instance: e666825c-ff4e-4a0e-93c0-43c00f167bbb] Reconfiguring VM instance instance-0000003c to attach disk [datastore1] e666825c-ff4e-4a0e-93c0-43c00f167bbb/e666825c-ff4e-4a0e-93c0-43c00f167bbb.vmdk or device None with type sparse {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 793.977470] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-fbfbdf7a-7ff7-4e03-834b-4c57e819bf20 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.002993] env[63515]: DEBUG oslo_vmware.api [None req-4b5baf79-d4c3-467e-88d4-a131ea6620d7 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] Waiting for the task: (returnval){ [ 794.002993] env[63515]: value = "task-1111252" [ 794.002993] env[63515]: _type = "Task" [ 794.002993] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 794.008980] env[63515]: DEBUG oslo_vmware.rw_handles [None req-67e41316-801d-46e9-947d-28c19f5a3017 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 794.008980] env[63515]: value = "vm-243466" [ 794.008980] env[63515]: _type = "VirtualMachine" [ 794.008980] env[63515]: }. {{(pid=63515) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 794.009309] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-6a5e4d16-eda3-4464-a869-c61bedb65648 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.017973] env[63515]: DEBUG oslo_vmware.api [None req-4b5baf79-d4c3-467e-88d4-a131ea6620d7 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] Task: {'id': task-1111252, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 794.024495] env[63515]: DEBUG oslo_vmware.rw_handles [None req-67e41316-801d-46e9-947d-28c19f5a3017 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Lease: (returnval){ [ 794.024495] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]52bec6f9-98eb-8de6-bdfd-1ac2a2a738a3" [ 794.024495] env[63515]: _type = "HttpNfcLease" [ 794.024495] env[63515]: } obtained for exporting VM: (result){ [ 794.024495] env[63515]: value = "vm-243466" [ 794.024495] env[63515]: _type = "VirtualMachine" [ 794.024495] env[63515]: }. {{(pid=63515) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 794.025281] env[63515]: DEBUG oslo_vmware.api [None req-67e41316-801d-46e9-947d-28c19f5a3017 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Waiting for the lease: (returnval){ [ 794.025281] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]52bec6f9-98eb-8de6-bdfd-1ac2a2a738a3" [ 794.025281] env[63515]: _type = "HttpNfcLease" [ 794.025281] env[63515]: } to be ready. {{(pid=63515) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 794.033493] env[63515]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 794.033493] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]52bec6f9-98eb-8de6-bdfd-1ac2a2a738a3" [ 794.033493] env[63515]: _type = "HttpNfcLease" [ 794.033493] env[63515]: } is initializing. {{(pid=63515) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 794.064112] env[63515]: DEBUG nova.network.neutron [None req-66c3e202-5a3e-433e-b543-4bfb02c8ad30 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775-project-member] [instance: 4e453127-1f3e-40ea-819f-6678479826c8] Updating instance_info_cache with network_info: [{"id": "3d5ad33f-8649-42c8-b899-ca20c9a2636e", "address": "fa:16:3e:c2:77:9b", "network": {"id": "5f57f1e7-21bb-425f-aab4-7d69b7ec94f7", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationNegativeTestJSON-177649896-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "60d9dcd0f30e48e7965ecfc1a72f5375", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3e05affa-2640-435e-a124-0ee8a6ab1152", "external-id": "nsx-vlan-transportzone-839", "segmentation_id": 839, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3d5ad33f-86", "ovs_interfaceid": "3d5ad33f-8649-42c8-b899-ca20c9a2636e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 794.083802] env[63515]: DEBUG oslo_vmware.api [None req-b53e671a-9cf4-4e34-a2ac-4321dce89b1c tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] Task: {'id': task-1111251, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.469192} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 794.084264] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-b53e671a-9cf4-4e34-a2ac-4321dce89b1c tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk to [datastore1] 192137b0-03e5-4bc4-b911-4b4f1a874f74/192137b0-03e5-4bc4-b911-4b4f1a874f74.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 794.084373] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-b53e671a-9cf4-4e34-a2ac-4321dce89b1c tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] [instance: 192137b0-03e5-4bc4-b911-4b4f1a874f74] Extending root virtual disk to 1048576 {{(pid=63515) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 794.084649] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d7cb7541-357b-49fc-879b-311f73a5ae42 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.094974] env[63515]: DEBUG oslo_vmware.api [None req-b53e671a-9cf4-4e34-a2ac-4321dce89b1c tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] Waiting for the task: (returnval){ [ 794.094974] env[63515]: value = "task-1111254" [ 794.094974] env[63515]: _type = "Task" [ 794.094974] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 794.104378] env[63515]: DEBUG oslo_vmware.api [None req-b53e671a-9cf4-4e34-a2ac-4321dce89b1c tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] Task: {'id': task-1111254, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 794.114538] env[63515]: INFO nova.compute.manager [None req-4f34aec7-5320-4ef9-b5f8-71a5209ff0fa tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] [instance: 63ce797d-7180-4209-ac2c-81978bf7607a] Rescuing [ 794.114862] env[63515]: DEBUG oslo_concurrency.lockutils [None req-4f34aec7-5320-4ef9-b5f8-71a5209ff0fa tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Acquiring lock "refresh_cache-63ce797d-7180-4209-ac2c-81978bf7607a" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 794.114954] env[63515]: DEBUG oslo_concurrency.lockutils [None req-4f34aec7-5320-4ef9-b5f8-71a5209ff0fa tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Acquired lock "refresh_cache-63ce797d-7180-4209-ac2c-81978bf7607a" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 794.115230] env[63515]: DEBUG nova.network.neutron [None req-4f34aec7-5320-4ef9-b5f8-71a5209ff0fa tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] [instance: 63ce797d-7180-4209-ac2c-81978bf7607a] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 794.126123] env[63515]: DEBUG oslo_vmware.api [None req-8d4c1f89-f73c-481b-8b79-61983aeba07b tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Task: {'id': task-1111239, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 794.148586] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29dec84d-6530-4cd6-8d7c-62d949f9d61b {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.155881] env[63515]: DEBUG nova.compute.manager [None req-95950a02-20c9-4a3f-9ea8-1f7f19725cc1 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: 4ef8066a-b68c-457b-9964-b1c34bab0fc3] Start building block device mappings for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 794.160789] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09a25e82-3e06-4c9c-847c-f583ee636ab3 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.193806] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da04c04e-5405-4993-b7c3-b7cc6b09b316 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.201556] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-311874cc-ee6b-4a2a-94e9-7cb661b53a9f {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.216165] env[63515]: DEBUG nova.compute.provider_tree [None req-aef604e2-0f8c-4f11-a420-cb96f46c502e tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 794.307700] env[63515]: DEBUG nova.network.neutron [None req-95950a02-20c9-4a3f-9ea8-1f7f19725cc1 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: 4ef8066a-b68c-457b-9964-b1c34bab0fc3] Successfully created port: f2d49ff8-dbd8-41a2-9006-c71265c36b87 {{(pid=63515) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 794.512946] env[63515]: DEBUG oslo_vmware.api [None req-4b5baf79-d4c3-467e-88d4-a131ea6620d7 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] Task: {'id': task-1111252, 'name': ReconfigVM_Task, 'duration_secs': 0.318057} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 794.513275] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-4b5baf79-d4c3-467e-88d4-a131ea6620d7 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] [instance: e666825c-ff4e-4a0e-93c0-43c00f167bbb] Reconfigured VM instance instance-0000003c to attach disk [datastore1] e666825c-ff4e-4a0e-93c0-43c00f167bbb/e666825c-ff4e-4a0e-93c0-43c00f167bbb.vmdk or device None with type sparse {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 794.513918] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-78ec242d-71f9-4a52-a6f3-9bf7f2abbba1 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.521051] env[63515]: DEBUG oslo_vmware.api [None req-4b5baf79-d4c3-467e-88d4-a131ea6620d7 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] Waiting for the task: (returnval){ [ 794.521051] env[63515]: value = "task-1111255" [ 794.521051] env[63515]: _type = "Task" [ 794.521051] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 794.531477] env[63515]: DEBUG oslo_vmware.api [None req-4b5baf79-d4c3-467e-88d4-a131ea6620d7 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] Task: {'id': task-1111255, 'name': Rename_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 794.534606] env[63515]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 794.534606] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]52bec6f9-98eb-8de6-bdfd-1ac2a2a738a3" [ 794.534606] env[63515]: _type = "HttpNfcLease" [ 794.534606] env[63515]: } is ready. {{(pid=63515) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 794.534851] env[63515]: DEBUG oslo_vmware.rw_handles [None req-67e41316-801d-46e9-947d-28c19f5a3017 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 794.534851] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]52bec6f9-98eb-8de6-bdfd-1ac2a2a738a3" [ 794.534851] env[63515]: _type = "HttpNfcLease" [ 794.534851] env[63515]: }. {{(pid=63515) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 794.535530] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30d81405-dbb7-4be2-9847-405391231b99 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.542310] env[63515]: DEBUG oslo_vmware.rw_handles [None req-67e41316-801d-46e9-947d-28c19f5a3017 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52db4223-71af-a208-688f-066abc855425/disk-0.vmdk from lease info. {{(pid=63515) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 794.542494] env[63515]: DEBUG oslo_vmware.rw_handles [None req-67e41316-801d-46e9-947d-28c19f5a3017 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52db4223-71af-a208-688f-066abc855425/disk-0.vmdk for reading. {{(pid=63515) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 794.601414] env[63515]: DEBUG oslo_concurrency.lockutils [None req-66c3e202-5a3e-433e-b543-4bfb02c8ad30 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775-project-member] Releasing lock "refresh_cache-4e453127-1f3e-40ea-819f-6678479826c8" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 794.601759] env[63515]: DEBUG nova.compute.manager [None req-66c3e202-5a3e-433e-b543-4bfb02c8ad30 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775-project-member] [instance: 4e453127-1f3e-40ea-819f-6678479826c8] Instance network_info: |[{"id": "3d5ad33f-8649-42c8-b899-ca20c9a2636e", "address": "fa:16:3e:c2:77:9b", "network": {"id": "5f57f1e7-21bb-425f-aab4-7d69b7ec94f7", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationNegativeTestJSON-177649896-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "60d9dcd0f30e48e7965ecfc1a72f5375", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3e05affa-2640-435e-a124-0ee8a6ab1152", "external-id": "nsx-vlan-transportzone-839", "segmentation_id": 839, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3d5ad33f-86", "ovs_interfaceid": "3d5ad33f-8649-42c8-b899-ca20c9a2636e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 794.603371] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-66c3e202-5a3e-433e-b543-4bfb02c8ad30 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775-project-member] [instance: 4e453127-1f3e-40ea-819f-6678479826c8] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c2:77:9b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '3e05affa-2640-435e-a124-0ee8a6ab1152', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '3d5ad33f-8649-42c8-b899-ca20c9a2636e', 'vif_model': 'vmxnet3'}] {{(pid=63515) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 794.611133] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-66c3e202-5a3e-433e-b543-4bfb02c8ad30 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775-project-member] Creating folder: Project (60d9dcd0f30e48e7965ecfc1a72f5375). Parent ref: group-v243370. {{(pid=63515) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 794.615512] env[63515]: DEBUG nova.compute.manager [req-34c3608b-1cb1-42d3-ba2b-8dce58efb176 req-cb35b2ae-ed87-4bd7-ad1f-05fed52d0d47 service nova] [instance: 4e453127-1f3e-40ea-819f-6678479826c8] Received event network-changed-3d5ad33f-8649-42c8-b899-ca20c9a2636e {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 794.615703] env[63515]: DEBUG nova.compute.manager [req-34c3608b-1cb1-42d3-ba2b-8dce58efb176 req-cb35b2ae-ed87-4bd7-ad1f-05fed52d0d47 service nova] [instance: 4e453127-1f3e-40ea-819f-6678479826c8] Refreshing instance network info cache due to event network-changed-3d5ad33f-8649-42c8-b899-ca20c9a2636e. {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 794.615916] env[63515]: DEBUG oslo_concurrency.lockutils [req-34c3608b-1cb1-42d3-ba2b-8dce58efb176 req-cb35b2ae-ed87-4bd7-ad1f-05fed52d0d47 service nova] Acquiring lock "refresh_cache-4e453127-1f3e-40ea-819f-6678479826c8" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 794.616076] env[63515]: DEBUG oslo_concurrency.lockutils [req-34c3608b-1cb1-42d3-ba2b-8dce58efb176 req-cb35b2ae-ed87-4bd7-ad1f-05fed52d0d47 service nova] Acquired lock "refresh_cache-4e453127-1f3e-40ea-819f-6678479826c8" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 794.616246] env[63515]: DEBUG nova.network.neutron [req-34c3608b-1cb1-42d3-ba2b-8dce58efb176 req-cb35b2ae-ed87-4bd7-ad1f-05fed52d0d47 service nova] [instance: 4e453127-1f3e-40ea-819f-6678479826c8] Refreshing network info cache for port 3d5ad33f-8649-42c8-b899-ca20c9a2636e {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 794.619519] env[63515]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-28d7889f-dc1d-4d27-9c4f-303179b3170e {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.634144] env[63515]: DEBUG oslo_vmware.api [None req-b53e671a-9cf4-4e34-a2ac-4321dce89b1c tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] Task: {'id': task-1111254, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.070015} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 794.634713] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-b53e671a-9cf4-4e34-a2ac-4321dce89b1c tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] [instance: 192137b0-03e5-4bc4-b911-4b4f1a874f74] Extended root virtual disk {{(pid=63515) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 794.635684] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1739c419-a54b-4143-8171-e93c3c4a5535 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.643203] env[63515]: INFO nova.virt.vmwareapi.vm_util [None req-66c3e202-5a3e-433e-b543-4bfb02c8ad30 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775-project-member] Created folder: Project (60d9dcd0f30e48e7965ecfc1a72f5375) in parent group-v243370. [ 794.643406] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-66c3e202-5a3e-433e-b543-4bfb02c8ad30 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775-project-member] Creating folder: Instances. Parent ref: group-v243468. {{(pid=63515) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 794.643655] env[63515]: DEBUG oslo_vmware.api [None req-8d4c1f89-f73c-481b-8b79-61983aeba07b tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Task: {'id': task-1111239, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 794.644183] env[63515]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-9127e73f-fa29-4a68-9409-2bf1b3fc9a00 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.663191] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-b53e671a-9cf4-4e34-a2ac-4321dce89b1c tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] [instance: 192137b0-03e5-4bc4-b911-4b4f1a874f74] Reconfiguring VM instance instance-00000039 to attach disk [datastore1] 192137b0-03e5-4bc4-b911-4b4f1a874f74/192137b0-03e5-4bc4-b911-4b4f1a874f74.vmdk or device None with type sparse {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 794.666776] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7f7774ef-ffc3-47cf-b5ad-51470db3f418 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.680337] env[63515]: INFO nova.virt.vmwareapi.vm_util [None req-66c3e202-5a3e-433e-b543-4bfb02c8ad30 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775-project-member] Created folder: Instances in parent group-v243468. [ 794.680582] env[63515]: DEBUG oslo.service.loopingcall [None req-66c3e202-5a3e-433e-b543-4bfb02c8ad30 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 794.685673] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4e453127-1f3e-40ea-819f-6678479826c8] Creating VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 794.686725] env[63515]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e10feffc-5562-46ff-8949-099ac213fada {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.702432] env[63515]: DEBUG oslo_vmware.api [None req-b53e671a-9cf4-4e34-a2ac-4321dce89b1c tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] Waiting for the task: (returnval){ [ 794.702432] env[63515]: value = "task-1111258" [ 794.702432] env[63515]: _type = "Task" [ 794.702432] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 794.707772] env[63515]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 794.707772] env[63515]: value = "task-1111259" [ 794.707772] env[63515]: _type = "Task" [ 794.707772] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 794.710920] env[63515]: DEBUG oslo_vmware.api [None req-b53e671a-9cf4-4e34-a2ac-4321dce89b1c tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] Task: {'id': task-1111258, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 794.720280] env[63515]: DEBUG nova.scheduler.client.report [None req-aef604e2-0f8c-4f11-a420-cb96f46c502e tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 794.724234] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1111259, 'name': CreateVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 794.770493] env[63515]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-d9bda70a-5c68-4c9f-bbd4-be39528be34b {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.916214] env[63515]: DEBUG nova.network.neutron [None req-4f34aec7-5320-4ef9-b5f8-71a5209ff0fa tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] [instance: 63ce797d-7180-4209-ac2c-81978bf7607a] Updating instance_info_cache with network_info: [{"id": "25ea9d0d-3405-42da-a576-a2a1f37fa75a", "address": "fa:16:3e:35:5c:ab", "network": {"id": "850bfa95-3df6-4019-8260-ade20ec58799", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-832000501-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "08c9aa51b1864410809f8877c7cd49c4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3cf748a8-7ae0-4dca-817d-e727c30d72f4", "external-id": "nsx-vlan-transportzone-853", "segmentation_id": 853, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap25ea9d0d-34", "ovs_interfaceid": "25ea9d0d-3405-42da-a576-a2a1f37fa75a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 795.032421] env[63515]: DEBUG oslo_vmware.api [None req-4b5baf79-d4c3-467e-88d4-a131ea6620d7 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] Task: {'id': task-1111255, 'name': Rename_Task, 'duration_secs': 0.141518} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 795.032804] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-4b5baf79-d4c3-467e-88d4-a131ea6620d7 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] [instance: e666825c-ff4e-4a0e-93c0-43c00f167bbb] Powering on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 795.033106] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-bf5cdcd0-0bf7-4e30-a5f3-5bcc869e322f {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.040266] env[63515]: DEBUG oslo_vmware.api [None req-4b5baf79-d4c3-467e-88d4-a131ea6620d7 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] Waiting for the task: (returnval){ [ 795.040266] env[63515]: value = "task-1111260" [ 795.040266] env[63515]: _type = "Task" [ 795.040266] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 795.049308] env[63515]: DEBUG oslo_vmware.api [None req-4b5baf79-d4c3-467e-88d4-a131ea6620d7 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] Task: {'id': task-1111260, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 795.134371] env[63515]: DEBUG oslo_vmware.api [None req-8d4c1f89-f73c-481b-8b79-61983aeba07b tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Task: {'id': task-1111239, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 795.202828] env[63515]: DEBUG nova.compute.manager [None req-95950a02-20c9-4a3f-9ea8-1f7f19725cc1 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: 4ef8066a-b68c-457b-9964-b1c34bab0fc3] Start spawning the instance on the hypervisor. {{(pid=63515) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 795.221779] env[63515]: DEBUG oslo_vmware.api [None req-b53e671a-9cf4-4e34-a2ac-4321dce89b1c tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] Task: {'id': task-1111258, 'name': ReconfigVM_Task, 'duration_secs': 0.516845} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 795.222784] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-b53e671a-9cf4-4e34-a2ac-4321dce89b1c tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] [instance: 192137b0-03e5-4bc4-b911-4b4f1a874f74] Reconfigured VM instance instance-00000039 to attach disk [datastore1] 192137b0-03e5-4bc4-b911-4b4f1a874f74/192137b0-03e5-4bc4-b911-4b4f1a874f74.vmdk or device None with type sparse {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 795.224454] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d6196095-ac86-455f-bc42-fae4b7ad6cb5 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.230084] env[63515]: DEBUG oslo_concurrency.lockutils [None req-aef604e2-0f8c-4f11-a420-cb96f46c502e tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.086s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 795.233580] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1111259, 'name': CreateVM_Task, 'duration_secs': 0.378787} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 795.234231] env[63515]: DEBUG oslo_concurrency.lockutils [None req-2e592662-996b-4dfa-8941-00965c1317fc tempest-ServersTestBootFromVolume-2042560423 tempest-ServersTestBootFromVolume-2042560423-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 29.659s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 795.234704] env[63515]: DEBUG nova.objects.instance [None req-2e592662-996b-4dfa-8941-00965c1317fc tempest-ServersTestBootFromVolume-2042560423 tempest-ServersTestBootFromVolume-2042560423-project-member] Lazy-loading 'resources' on Instance uuid 90f4930b-aaa0-4c4b-9ab8-92aed45e200b {{(pid=63515) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 795.240499] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4e453127-1f3e-40ea-819f-6678479826c8] Created VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 795.244165] env[63515]: DEBUG oslo_concurrency.lockutils [None req-66c3e202-5a3e-433e-b543-4bfb02c8ad30 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 795.244420] env[63515]: DEBUG oslo_concurrency.lockutils [None req-66c3e202-5a3e-433e-b543-4bfb02c8ad30 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 795.245018] env[63515]: DEBUG oslo_concurrency.lockutils [None req-66c3e202-5a3e-433e-b543-4bfb02c8ad30 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 795.246865] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dd83dbd0-f89d-489f-902b-06a7012244ad {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.249293] env[63515]: DEBUG oslo_vmware.api [None req-b53e671a-9cf4-4e34-a2ac-4321dce89b1c tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] Waiting for the task: (returnval){ [ 795.249293] env[63515]: value = "task-1111261" [ 795.249293] env[63515]: _type = "Task" [ 795.249293] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 795.258263] env[63515]: DEBUG oslo_vmware.api [None req-66c3e202-5a3e-433e-b543-4bfb02c8ad30 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775-project-member] Waiting for the task: (returnval){ [ 795.258263] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]526ea9ad-a670-0542-c904-5d27753aff88" [ 795.258263] env[63515]: _type = "Task" [ 795.258263] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 795.260569] env[63515]: DEBUG nova.virt.hardware [None req-95950a02-20c9-4a3f-9ea8-1f7f19725cc1 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T02:52:14Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T02:51:56Z,direct_url=,disk_format='vmdk',id=8a120570-cb06-4099-b262-554ca0ad15c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b569255cc43e42e7ae3f2b4ad37c6ef4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T02:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 795.260911] env[63515]: DEBUG nova.virt.hardware [None req-95950a02-20c9-4a3f-9ea8-1f7f19725cc1 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Flavor limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 795.261109] env[63515]: DEBUG nova.virt.hardware [None req-95950a02-20c9-4a3f-9ea8-1f7f19725cc1 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Image limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 795.261351] env[63515]: DEBUG nova.virt.hardware [None req-95950a02-20c9-4a3f-9ea8-1f7f19725cc1 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Flavor pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 795.261612] env[63515]: DEBUG nova.virt.hardware [None req-95950a02-20c9-4a3f-9ea8-1f7f19725cc1 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Image pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 795.261839] env[63515]: DEBUG nova.virt.hardware [None req-95950a02-20c9-4a3f-9ea8-1f7f19725cc1 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 795.262137] env[63515]: DEBUG nova.virt.hardware [None req-95950a02-20c9-4a3f-9ea8-1f7f19725cc1 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 795.262388] env[63515]: DEBUG nova.virt.hardware [None req-95950a02-20c9-4a3f-9ea8-1f7f19725cc1 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 795.262576] env[63515]: DEBUG nova.virt.hardware [None req-95950a02-20c9-4a3f-9ea8-1f7f19725cc1 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Got 1 possible topologies {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 795.262757] env[63515]: DEBUG nova.virt.hardware [None req-95950a02-20c9-4a3f-9ea8-1f7f19725cc1 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 795.265256] env[63515]: DEBUG nova.virt.hardware [None req-95950a02-20c9-4a3f-9ea8-1f7f19725cc1 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 795.265548] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb448022-2d37-498a-9ffe-3dab17357daf {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.276891] env[63515]: DEBUG oslo_vmware.api [None req-b53e671a-9cf4-4e34-a2ac-4321dce89b1c tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] Task: {'id': task-1111261, 'name': Rename_Task} progress is 14%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 795.283842] env[63515]: DEBUG oslo_vmware.api [None req-66c3e202-5a3e-433e-b543-4bfb02c8ad30 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]526ea9ad-a670-0542-c904-5d27753aff88, 'name': SearchDatastore_Task, 'duration_secs': 0.011164} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 795.286320] env[63515]: DEBUG oslo_concurrency.lockutils [None req-66c3e202-5a3e-433e-b543-4bfb02c8ad30 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 795.286780] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-66c3e202-5a3e-433e-b543-4bfb02c8ad30 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775-project-member] [instance: 4e453127-1f3e-40ea-819f-6678479826c8] Processing image 8a120570-cb06-4099-b262-554ca0ad15c5 {{(pid=63515) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 795.287146] env[63515]: DEBUG oslo_concurrency.lockutils [None req-66c3e202-5a3e-433e-b543-4bfb02c8ad30 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 795.287449] env[63515]: DEBUG oslo_concurrency.lockutils [None req-66c3e202-5a3e-433e-b543-4bfb02c8ad30 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 795.287672] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-66c3e202-5a3e-433e-b543-4bfb02c8ad30 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 795.289678] env[63515]: INFO nova.scheduler.client.report [None req-aef604e2-0f8c-4f11-a420-cb96f46c502e tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Deleted allocations for instance 001a48e1-b237-41f8-a6f4-2fe98d173512 [ 795.290853] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-18bad18f-ae0b-4034-a2c6-58514584d3ff {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.294777] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0413d3e4-f7fe-4c6f-bda8-ff3c91f55fc5 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.319493] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-66c3e202-5a3e-433e-b543-4bfb02c8ad30 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 795.320169] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-66c3e202-5a3e-433e-b543-4bfb02c8ad30 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63515) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 795.320660] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-57bc3e75-d56e-474f-91ee-bdf12ae15b35 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.327129] env[63515]: DEBUG oslo_vmware.api [None req-66c3e202-5a3e-433e-b543-4bfb02c8ad30 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775-project-member] Waiting for the task: (returnval){ [ 795.327129] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]529acc84-cc83-3fa4-f4bc-03c5d45d66f5" [ 795.327129] env[63515]: _type = "Task" [ 795.327129] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 795.335953] env[63515]: DEBUG oslo_vmware.api [None req-66c3e202-5a3e-433e-b543-4bfb02c8ad30 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]529acc84-cc83-3fa4-f4bc-03c5d45d66f5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 795.419955] env[63515]: DEBUG oslo_concurrency.lockutils [None req-4f34aec7-5320-4ef9-b5f8-71a5209ff0fa tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Releasing lock "refresh_cache-63ce797d-7180-4209-ac2c-81978bf7607a" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 795.529339] env[63515]: DEBUG nova.network.neutron [req-34c3608b-1cb1-42d3-ba2b-8dce58efb176 req-cb35b2ae-ed87-4bd7-ad1f-05fed52d0d47 service nova] [instance: 4e453127-1f3e-40ea-819f-6678479826c8] Updated VIF entry in instance network info cache for port 3d5ad33f-8649-42c8-b899-ca20c9a2636e. {{(pid=63515) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 795.530097] env[63515]: DEBUG nova.network.neutron [req-34c3608b-1cb1-42d3-ba2b-8dce58efb176 req-cb35b2ae-ed87-4bd7-ad1f-05fed52d0d47 service nova] [instance: 4e453127-1f3e-40ea-819f-6678479826c8] Updating instance_info_cache with network_info: [{"id": "3d5ad33f-8649-42c8-b899-ca20c9a2636e", "address": "fa:16:3e:c2:77:9b", "network": {"id": "5f57f1e7-21bb-425f-aab4-7d69b7ec94f7", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationNegativeTestJSON-177649896-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "60d9dcd0f30e48e7965ecfc1a72f5375", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3e05affa-2640-435e-a124-0ee8a6ab1152", "external-id": "nsx-vlan-transportzone-839", "segmentation_id": 839, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3d5ad33f-86", "ovs_interfaceid": "3d5ad33f-8649-42c8-b899-ca20c9a2636e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 795.555974] env[63515]: DEBUG oslo_vmware.api [None req-4b5baf79-d4c3-467e-88d4-a131ea6620d7 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] Task: {'id': task-1111260, 'name': PowerOnVM_Task} progress is 78%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 795.634127] env[63515]: DEBUG oslo_vmware.api [None req-8d4c1f89-f73c-481b-8b79-61983aeba07b tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Task: {'id': task-1111239, 'name': ReconfigVM_Task} progress is 18%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 795.763829] env[63515]: DEBUG oslo_vmware.api [None req-b53e671a-9cf4-4e34-a2ac-4321dce89b1c tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] Task: {'id': task-1111261, 'name': Rename_Task, 'duration_secs': 0.173087} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 795.765626] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-b53e671a-9cf4-4e34-a2ac-4321dce89b1c tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] [instance: 192137b0-03e5-4bc4-b911-4b4f1a874f74] Powering on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 795.765626] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-70e4ad48-1c19-4fca-a3d0-0b72f3e30738 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.773845] env[63515]: DEBUG oslo_vmware.api [None req-b53e671a-9cf4-4e34-a2ac-4321dce89b1c tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] Waiting for the task: (returnval){ [ 795.773845] env[63515]: value = "task-1111262" [ 795.773845] env[63515]: _type = "Task" [ 795.773845] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 795.788550] env[63515]: DEBUG oslo_vmware.api [None req-b53e671a-9cf4-4e34-a2ac-4321dce89b1c tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] Task: {'id': task-1111262, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 795.811804] env[63515]: DEBUG oslo_concurrency.lockutils [None req-aef604e2-0f8c-4f11-a420-cb96f46c502e tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Lock "001a48e1-b237-41f8-a6f4-2fe98d173512" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 33.875s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 795.842320] env[63515]: DEBUG oslo_vmware.api [None req-66c3e202-5a3e-433e-b543-4bfb02c8ad30 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]529acc84-cc83-3fa4-f4bc-03c5d45d66f5, 'name': SearchDatastore_Task, 'duration_secs': 0.011609} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 795.848966] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-50890e11-3249-42ef-ba68-a37bad463a94 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.855972] env[63515]: DEBUG oslo_vmware.api [None req-66c3e202-5a3e-433e-b543-4bfb02c8ad30 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775-project-member] Waiting for the task: (returnval){ [ 795.855972] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]52bbfca6-6144-a750-b51e-a611b64c4f82" [ 795.855972] env[63515]: _type = "Task" [ 795.855972] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 795.867265] env[63515]: DEBUG oslo_vmware.api [None req-66c3e202-5a3e-433e-b543-4bfb02c8ad30 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52bbfca6-6144-a750-b51e-a611b64c4f82, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 795.962684] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-4f34aec7-5320-4ef9-b5f8-71a5209ff0fa tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] [instance: 63ce797d-7180-4209-ac2c-81978bf7607a] Powering off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 795.963043] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9b79fdaa-3a8c-4a0a-92af-2b535d7a0fb0 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.971607] env[63515]: DEBUG oslo_vmware.api [None req-4f34aec7-5320-4ef9-b5f8-71a5209ff0fa tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Waiting for the task: (returnval){ [ 795.971607] env[63515]: value = "task-1111263" [ 795.971607] env[63515]: _type = "Task" [ 795.971607] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 795.988199] env[63515]: DEBUG oslo_vmware.api [None req-4f34aec7-5320-4ef9-b5f8-71a5209ff0fa tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Task: {'id': task-1111263, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 796.033593] env[63515]: DEBUG oslo_concurrency.lockutils [req-34c3608b-1cb1-42d3-ba2b-8dce58efb176 req-cb35b2ae-ed87-4bd7-ad1f-05fed52d0d47 service nova] Releasing lock "refresh_cache-4e453127-1f3e-40ea-819f-6678479826c8" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 796.055056] env[63515]: DEBUG oslo_vmware.api [None req-4b5baf79-d4c3-467e-88d4-a131ea6620d7 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] Task: {'id': task-1111260, 'name': PowerOnVM_Task, 'duration_secs': 0.847458} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 796.055442] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-4b5baf79-d4c3-467e-88d4-a131ea6620d7 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] [instance: e666825c-ff4e-4a0e-93c0-43c00f167bbb] Powered on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 796.055767] env[63515]: INFO nova.compute.manager [None req-4b5baf79-d4c3-467e-88d4-a131ea6620d7 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] [instance: e666825c-ff4e-4a0e-93c0-43c00f167bbb] Took 8.95 seconds to spawn the instance on the hypervisor. [ 796.057030] env[63515]: DEBUG nova.compute.manager [None req-4b5baf79-d4c3-467e-88d4-a131ea6620d7 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] [instance: e666825c-ff4e-4a0e-93c0-43c00f167bbb] Checking state {{(pid=63515) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 796.057030] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d8820c6-dee8-44a6-ad37-19e2fe886ee9 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.131426] env[63515]: DEBUG nova.network.neutron [None req-95950a02-20c9-4a3f-9ea8-1f7f19725cc1 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: 4ef8066a-b68c-457b-9964-b1c34bab0fc3] Successfully updated port: f2d49ff8-dbd8-41a2-9006-c71265c36b87 {{(pid=63515) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 796.143958] env[63515]: DEBUG oslo_vmware.api [None req-8d4c1f89-f73c-481b-8b79-61983aeba07b tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Task: {'id': task-1111239, 'name': ReconfigVM_Task} progress is 18%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 796.292296] env[63515]: DEBUG oslo_vmware.api [None req-b53e671a-9cf4-4e34-a2ac-4321dce89b1c tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] Task: {'id': task-1111262, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 796.326040] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-748eeca5-6501-4444-b2ef-0d54072b8f1e {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.335840] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10f95f59-ca50-4785-85a7-f5ca5bb59569 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.370859] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10f6b303-b817-4e36-a2ad-1d63b541ee4b {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.379548] env[63515]: DEBUG oslo_vmware.api [None req-66c3e202-5a3e-433e-b543-4bfb02c8ad30 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52bbfca6-6144-a750-b51e-a611b64c4f82, 'name': SearchDatastore_Task, 'duration_secs': 0.016279} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 796.381868] env[63515]: DEBUG oslo_concurrency.lockutils [None req-66c3e202-5a3e-433e-b543-4bfb02c8ad30 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 796.382379] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-66c3e202-5a3e-433e-b543-4bfb02c8ad30 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk to [datastore1] 4e453127-1f3e-40ea-819f-6678479826c8/4e453127-1f3e-40ea-819f-6678479826c8.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 796.383108] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c58b00ba-3696-4076-9110-95862246e348 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.386642] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ddc6f72-ec50-4d3f-81cb-a9a9a653f536 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.407300] env[63515]: DEBUG nova.compute.provider_tree [None req-2e592662-996b-4dfa-8941-00965c1317fc tempest-ServersTestBootFromVolume-2042560423 tempest-ServersTestBootFromVolume-2042560423-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 796.412043] env[63515]: DEBUG oslo_vmware.api [None req-66c3e202-5a3e-433e-b543-4bfb02c8ad30 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775-project-member] Waiting for the task: (returnval){ [ 796.412043] env[63515]: value = "task-1111264" [ 796.412043] env[63515]: _type = "Task" [ 796.412043] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 796.424423] env[63515]: DEBUG oslo_vmware.api [None req-66c3e202-5a3e-433e-b543-4bfb02c8ad30 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775-project-member] Task: {'id': task-1111264, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 796.483702] env[63515]: DEBUG oslo_vmware.api [None req-4f34aec7-5320-4ef9-b5f8-71a5209ff0fa tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Task: {'id': task-1111263, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 796.586860] env[63515]: INFO nova.compute.manager [None req-4b5baf79-d4c3-467e-88d4-a131ea6620d7 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] [instance: e666825c-ff4e-4a0e-93c0-43c00f167bbb] Took 38.73 seconds to build instance. [ 796.641303] env[63515]: DEBUG oslo_concurrency.lockutils [None req-95950a02-20c9-4a3f-9ea8-1f7f19725cc1 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Acquiring lock "refresh_cache-4ef8066a-b68c-457b-9964-b1c34bab0fc3" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 796.641603] env[63515]: DEBUG oslo_concurrency.lockutils [None req-95950a02-20c9-4a3f-9ea8-1f7f19725cc1 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Acquired lock "refresh_cache-4ef8066a-b68c-457b-9964-b1c34bab0fc3" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 796.641712] env[63515]: DEBUG nova.network.neutron [None req-95950a02-20c9-4a3f-9ea8-1f7f19725cc1 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: 4ef8066a-b68c-457b-9964-b1c34bab0fc3] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 796.643460] env[63515]: DEBUG oslo_vmware.api [None req-8d4c1f89-f73c-481b-8b79-61983aeba07b tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Task: {'id': task-1111239, 'name': ReconfigVM_Task} progress is 18%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 796.706734] env[63515]: DEBUG nova.compute.manager [req-2b8b2982-f06d-4922-9b57-42c331dcce99 req-07c2e5a4-ac20-4f92-8b7b-8c6cd6db2027 service nova] [instance: 4ef8066a-b68c-457b-9964-b1c34bab0fc3] Received event network-vif-plugged-f2d49ff8-dbd8-41a2-9006-c71265c36b87 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 796.710337] env[63515]: DEBUG oslo_concurrency.lockutils [req-2b8b2982-f06d-4922-9b57-42c331dcce99 req-07c2e5a4-ac20-4f92-8b7b-8c6cd6db2027 service nova] Acquiring lock "4ef8066a-b68c-457b-9964-b1c34bab0fc3-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 796.710337] env[63515]: DEBUG oslo_concurrency.lockutils [req-2b8b2982-f06d-4922-9b57-42c331dcce99 req-07c2e5a4-ac20-4f92-8b7b-8c6cd6db2027 service nova] Lock "4ef8066a-b68c-457b-9964-b1c34bab0fc3-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 796.710337] env[63515]: DEBUG oslo_concurrency.lockutils [req-2b8b2982-f06d-4922-9b57-42c331dcce99 req-07c2e5a4-ac20-4f92-8b7b-8c6cd6db2027 service nova] Lock "4ef8066a-b68c-457b-9964-b1c34bab0fc3-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 796.710337] env[63515]: DEBUG nova.compute.manager [req-2b8b2982-f06d-4922-9b57-42c331dcce99 req-07c2e5a4-ac20-4f92-8b7b-8c6cd6db2027 service nova] [instance: 4ef8066a-b68c-457b-9964-b1c34bab0fc3] No waiting events found dispatching network-vif-plugged-f2d49ff8-dbd8-41a2-9006-c71265c36b87 {{(pid=63515) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 796.710337] env[63515]: WARNING nova.compute.manager [req-2b8b2982-f06d-4922-9b57-42c331dcce99 req-07c2e5a4-ac20-4f92-8b7b-8c6cd6db2027 service nova] [instance: 4ef8066a-b68c-457b-9964-b1c34bab0fc3] Received unexpected event network-vif-plugged-f2d49ff8-dbd8-41a2-9006-c71265c36b87 for instance with vm_state building and task_state spawning. [ 796.710337] env[63515]: DEBUG nova.compute.manager [req-2b8b2982-f06d-4922-9b57-42c331dcce99 req-07c2e5a4-ac20-4f92-8b7b-8c6cd6db2027 service nova] [instance: 4ef8066a-b68c-457b-9964-b1c34bab0fc3] Received event network-changed-f2d49ff8-dbd8-41a2-9006-c71265c36b87 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 796.710337] env[63515]: DEBUG nova.compute.manager [req-2b8b2982-f06d-4922-9b57-42c331dcce99 req-07c2e5a4-ac20-4f92-8b7b-8c6cd6db2027 service nova] [instance: 4ef8066a-b68c-457b-9964-b1c34bab0fc3] Refreshing instance network info cache due to event network-changed-f2d49ff8-dbd8-41a2-9006-c71265c36b87. {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 796.710337] env[63515]: DEBUG oslo_concurrency.lockutils [req-2b8b2982-f06d-4922-9b57-42c331dcce99 req-07c2e5a4-ac20-4f92-8b7b-8c6cd6db2027 service nova] Acquiring lock "refresh_cache-4ef8066a-b68c-457b-9964-b1c34bab0fc3" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 796.792985] env[63515]: DEBUG oslo_vmware.api [None req-b53e671a-9cf4-4e34-a2ac-4321dce89b1c tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] Task: {'id': task-1111262, 'name': PowerOnVM_Task, 'duration_secs': 0.958496} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 796.793487] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-b53e671a-9cf4-4e34-a2ac-4321dce89b1c tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] [instance: 192137b0-03e5-4bc4-b911-4b4f1a874f74] Powered on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 796.793814] env[63515]: DEBUG nova.compute.manager [None req-b53e671a-9cf4-4e34-a2ac-4321dce89b1c tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] [instance: 192137b0-03e5-4bc4-b911-4b4f1a874f74] Checking state {{(pid=63515) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 796.795875] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f48cc088-4433-43d6-8a92-cfabb23e76a6 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.911475] env[63515]: DEBUG nova.scheduler.client.report [None req-2e592662-996b-4dfa-8941-00965c1317fc tempest-ServersTestBootFromVolume-2042560423 tempest-ServersTestBootFromVolume-2042560423-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 796.931030] env[63515]: DEBUG oslo_vmware.api [None req-66c3e202-5a3e-433e-b543-4bfb02c8ad30 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775-project-member] Task: {'id': task-1111264, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 796.983615] env[63515]: DEBUG oslo_vmware.api [None req-4f34aec7-5320-4ef9-b5f8-71a5209ff0fa tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Task: {'id': task-1111263, 'name': PowerOffVM_Task, 'duration_secs': 0.633169} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 796.983615] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-4f34aec7-5320-4ef9-b5f8-71a5209ff0fa tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] [instance: 63ce797d-7180-4209-ac2c-81978bf7607a] Powered off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 796.984479] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8d191e0-3a29-43d3-b89a-4a283bfef6a3 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.004742] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d99ab7bb-62a1-4362-99ac-d7cd1c566d65 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.038373] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-4f34aec7-5320-4ef9-b5f8-71a5209ff0fa tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] [instance: 63ce797d-7180-4209-ac2c-81978bf7607a] Powering off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 797.038734] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-129ef280-a0ad-4d98-ad24-049a575c9a69 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.047294] env[63515]: DEBUG oslo_vmware.api [None req-4f34aec7-5320-4ef9-b5f8-71a5209ff0fa tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Waiting for the task: (returnval){ [ 797.047294] env[63515]: value = "task-1111265" [ 797.047294] env[63515]: _type = "Task" [ 797.047294] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 797.056224] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-4f34aec7-5320-4ef9-b5f8-71a5209ff0fa tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] [instance: 63ce797d-7180-4209-ac2c-81978bf7607a] VM already powered off {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 797.056421] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-4f34aec7-5320-4ef9-b5f8-71a5209ff0fa tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] [instance: 63ce797d-7180-4209-ac2c-81978bf7607a] Processing image 8a120570-cb06-4099-b262-554ca0ad15c5 {{(pid=63515) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 797.056669] env[63515]: DEBUG oslo_concurrency.lockutils [None req-4f34aec7-5320-4ef9-b5f8-71a5209ff0fa tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 797.056811] env[63515]: DEBUG oslo_concurrency.lockutils [None req-4f34aec7-5320-4ef9-b5f8-71a5209ff0fa tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 797.056986] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-4f34aec7-5320-4ef9-b5f8-71a5209ff0fa tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 797.057259] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e8318193-8757-4321-9e13-fbfb1545ebf3 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.069154] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-4f34aec7-5320-4ef9-b5f8-71a5209ff0fa tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 797.069466] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-4f34aec7-5320-4ef9-b5f8-71a5209ff0fa tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63515) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 797.070282] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-41293c3f-11e8-45f8-9437-9d052e586bee {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.076323] env[63515]: DEBUG oslo_vmware.api [None req-4f34aec7-5320-4ef9-b5f8-71a5209ff0fa tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Waiting for the task: (returnval){ [ 797.076323] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]520097e1-57a2-6c55-48c7-441c0de5db74" [ 797.076323] env[63515]: _type = "Task" [ 797.076323] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 797.084419] env[63515]: DEBUG oslo_vmware.api [None req-4f34aec7-5320-4ef9-b5f8-71a5209ff0fa tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]520097e1-57a2-6c55-48c7-441c0de5db74, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 797.089723] env[63515]: DEBUG oslo_concurrency.lockutils [None req-4b5baf79-d4c3-467e-88d4-a131ea6620d7 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] Lock "e666825c-ff4e-4a0e-93c0-43c00f167bbb" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 95.223s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 797.136364] env[63515]: DEBUG oslo_vmware.api [None req-8d4c1f89-f73c-481b-8b79-61983aeba07b tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Task: {'id': task-1111239, 'name': ReconfigVM_Task} progress is 18%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 797.179708] env[63515]: DEBUG nova.network.neutron [None req-95950a02-20c9-4a3f-9ea8-1f7f19725cc1 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: 4ef8066a-b68c-457b-9964-b1c34bab0fc3] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 797.188703] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc2f8fae-fa72-41c7-8e66-04dc63e6e30a {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.196332] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-32638298-0fa3-40f4-a1a6-2a70f7b4ad18 tempest-ServersAdminNegativeTestJSON-180276763 tempest-ServersAdminNegativeTestJSON-180276763-project-admin] [instance: e666825c-ff4e-4a0e-93c0-43c00f167bbb] Suspending the VM {{(pid=63515) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1162}} [ 797.197067] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-50f1ca37-9eb1-4297-b94b-9c873047eb92 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.210065] env[63515]: DEBUG oslo_vmware.api [None req-32638298-0fa3-40f4-a1a6-2a70f7b4ad18 tempest-ServersAdminNegativeTestJSON-180276763 tempest-ServersAdminNegativeTestJSON-180276763-project-admin] Waiting for the task: (returnval){ [ 797.210065] env[63515]: value = "task-1111266" [ 797.210065] env[63515]: _type = "Task" [ 797.210065] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 797.220509] env[63515]: DEBUG oslo_vmware.api [None req-32638298-0fa3-40f4-a1a6-2a70f7b4ad18 tempest-ServersAdminNegativeTestJSON-180276763 tempest-ServersAdminNegativeTestJSON-180276763-project-admin] Task: {'id': task-1111266, 'name': SuspendVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 797.318025] env[63515]: DEBUG oslo_concurrency.lockutils [None req-b53e671a-9cf4-4e34-a2ac-4321dce89b1c tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 797.417638] env[63515]: DEBUG nova.network.neutron [None req-95950a02-20c9-4a3f-9ea8-1f7f19725cc1 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: 4ef8066a-b68c-457b-9964-b1c34bab0fc3] Updating instance_info_cache with network_info: [{"id": "f2d49ff8-dbd8-41a2-9006-c71265c36b87", "address": "fa:16:3e:5b:79:66", "network": {"id": "3e5914c8-0cd4-49ee-81b4-1f3e280440bb", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-641632277-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ea099ce07bfb4a8da014d9303cf552eb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "225b6979-9329-403b-91fa-138bd41f6e83", "external-id": "nsx-vlan-transportzone-38", "segmentation_id": 38, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf2d49ff8-db", "ovs_interfaceid": "f2d49ff8-dbd8-41a2-9006-c71265c36b87", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 797.427040] env[63515]: DEBUG oslo_concurrency.lockutils [None req-2e592662-996b-4dfa-8941-00965c1317fc tempest-ServersTestBootFromVolume-2042560423 tempest-ServersTestBootFromVolume-2042560423-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.192s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 797.431173] env[63515]: DEBUG oslo_concurrency.lockutils [None req-655858bd-8f63-4e55-ac5e-dbecfbd88a62 tempest-ServersV294TestFqdnHostnames-2112993702 tempest-ServersV294TestFqdnHostnames-2112993702-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 31.261s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 797.431402] env[63515]: DEBUG nova.objects.instance [None req-655858bd-8f63-4e55-ac5e-dbecfbd88a62 tempest-ServersV294TestFqdnHostnames-2112993702 tempest-ServersV294TestFqdnHostnames-2112993702-project-member] Lazy-loading 'resources' on Instance uuid 396e49dd-48c0-496b-a1ec-190c33a22c5e {{(pid=63515) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 797.439866] env[63515]: DEBUG oslo_vmware.api [None req-66c3e202-5a3e-433e-b543-4bfb02c8ad30 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775-project-member] Task: {'id': task-1111264, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.682715} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 797.440204] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-66c3e202-5a3e-433e-b543-4bfb02c8ad30 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk to [datastore1] 4e453127-1f3e-40ea-819f-6678479826c8/4e453127-1f3e-40ea-819f-6678479826c8.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 797.440448] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-66c3e202-5a3e-433e-b543-4bfb02c8ad30 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775-project-member] [instance: 4e453127-1f3e-40ea-819f-6678479826c8] Extending root virtual disk to 1048576 {{(pid=63515) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 797.441157] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-8b64fad5-1155-402f-a13a-473ad3ebd6d4 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.449803] env[63515]: DEBUG oslo_vmware.api [None req-66c3e202-5a3e-433e-b543-4bfb02c8ad30 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775-project-member] Waiting for the task: (returnval){ [ 797.449803] env[63515]: value = "task-1111267" [ 797.449803] env[63515]: _type = "Task" [ 797.449803] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 797.461448] env[63515]: DEBUG oslo_vmware.api [None req-66c3e202-5a3e-433e-b543-4bfb02c8ad30 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775-project-member] Task: {'id': task-1111267, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 797.462751] env[63515]: INFO nova.scheduler.client.report [None req-2e592662-996b-4dfa-8941-00965c1317fc tempest-ServersTestBootFromVolume-2042560423 tempest-ServersTestBootFromVolume-2042560423-project-member] Deleted allocations for instance 90f4930b-aaa0-4c4b-9ab8-92aed45e200b [ 797.588819] env[63515]: DEBUG oslo_vmware.api [None req-4f34aec7-5320-4ef9-b5f8-71a5209ff0fa tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]520097e1-57a2-6c55-48c7-441c0de5db74, 'name': SearchDatastore_Task, 'duration_secs': 0.011986} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 797.590024] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-861f903a-023b-4587-af37-31247baad1b8 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.597580] env[63515]: DEBUG oslo_vmware.api [None req-4f34aec7-5320-4ef9-b5f8-71a5209ff0fa tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Waiting for the task: (returnval){ [ 797.597580] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]5257154b-4b4f-0749-ce3d-a18611e8e82b" [ 797.597580] env[63515]: _type = "Task" [ 797.597580] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 797.605909] env[63515]: DEBUG oslo_vmware.api [None req-4f34aec7-5320-4ef9-b5f8-71a5209ff0fa tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]5257154b-4b4f-0749-ce3d-a18611e8e82b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 797.638516] env[63515]: DEBUG oslo_vmware.api [None req-8d4c1f89-f73c-481b-8b79-61983aeba07b tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Task: {'id': task-1111239, 'name': ReconfigVM_Task} progress is 18%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 797.720267] env[63515]: DEBUG oslo_vmware.api [None req-32638298-0fa3-40f4-a1a6-2a70f7b4ad18 tempest-ServersAdminNegativeTestJSON-180276763 tempest-ServersAdminNegativeTestJSON-180276763-project-admin] Task: {'id': task-1111266, 'name': SuspendVM_Task} progress is 66%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 797.745347] env[63515]: DEBUG oslo_concurrency.lockutils [None req-b257f20f-3555-4579-b6f5-3780aaf93d24 tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] Acquiring lock "192137b0-03e5-4bc4-b911-4b4f1a874f74" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 797.745997] env[63515]: DEBUG oslo_concurrency.lockutils [None req-b257f20f-3555-4579-b6f5-3780aaf93d24 tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] Lock "192137b0-03e5-4bc4-b911-4b4f1a874f74" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 797.745997] env[63515]: DEBUG oslo_concurrency.lockutils [None req-b257f20f-3555-4579-b6f5-3780aaf93d24 tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] Acquiring lock "192137b0-03e5-4bc4-b911-4b4f1a874f74-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 797.745997] env[63515]: DEBUG oslo_concurrency.lockutils [None req-b257f20f-3555-4579-b6f5-3780aaf93d24 tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] Lock "192137b0-03e5-4bc4-b911-4b4f1a874f74-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 797.746167] env[63515]: DEBUG oslo_concurrency.lockutils [None req-b257f20f-3555-4579-b6f5-3780aaf93d24 tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] Lock "192137b0-03e5-4bc4-b911-4b4f1a874f74-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 797.748315] env[63515]: INFO nova.compute.manager [None req-b257f20f-3555-4579-b6f5-3780aaf93d24 tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] [instance: 192137b0-03e5-4bc4-b911-4b4f1a874f74] Terminating instance [ 797.750080] env[63515]: DEBUG oslo_concurrency.lockutils [None req-b257f20f-3555-4579-b6f5-3780aaf93d24 tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] Acquiring lock "refresh_cache-192137b0-03e5-4bc4-b911-4b4f1a874f74" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 797.750250] env[63515]: DEBUG oslo_concurrency.lockutils [None req-b257f20f-3555-4579-b6f5-3780aaf93d24 tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] Acquired lock "refresh_cache-192137b0-03e5-4bc4-b911-4b4f1a874f74" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 797.750437] env[63515]: DEBUG nova.network.neutron [None req-b257f20f-3555-4579-b6f5-3780aaf93d24 tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] [instance: 192137b0-03e5-4bc4-b911-4b4f1a874f74] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 797.926015] env[63515]: DEBUG oslo_concurrency.lockutils [None req-95950a02-20c9-4a3f-9ea8-1f7f19725cc1 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Releasing lock "refresh_cache-4ef8066a-b68c-457b-9964-b1c34bab0fc3" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 797.926390] env[63515]: DEBUG nova.compute.manager [None req-95950a02-20c9-4a3f-9ea8-1f7f19725cc1 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: 4ef8066a-b68c-457b-9964-b1c34bab0fc3] Instance network_info: |[{"id": "f2d49ff8-dbd8-41a2-9006-c71265c36b87", "address": "fa:16:3e:5b:79:66", "network": {"id": "3e5914c8-0cd4-49ee-81b4-1f3e280440bb", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-641632277-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ea099ce07bfb4a8da014d9303cf552eb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "225b6979-9329-403b-91fa-138bd41f6e83", "external-id": "nsx-vlan-transportzone-38", "segmentation_id": 38, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf2d49ff8-db", "ovs_interfaceid": "f2d49ff8-dbd8-41a2-9006-c71265c36b87", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 797.926719] env[63515]: DEBUG oslo_concurrency.lockutils [req-2b8b2982-f06d-4922-9b57-42c331dcce99 req-07c2e5a4-ac20-4f92-8b7b-8c6cd6db2027 service nova] Acquired lock "refresh_cache-4ef8066a-b68c-457b-9964-b1c34bab0fc3" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 797.926907] env[63515]: DEBUG nova.network.neutron [req-2b8b2982-f06d-4922-9b57-42c331dcce99 req-07c2e5a4-ac20-4f92-8b7b-8c6cd6db2027 service nova] [instance: 4ef8066a-b68c-457b-9964-b1c34bab0fc3] Refreshing network info cache for port f2d49ff8-dbd8-41a2-9006-c71265c36b87 {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 797.928473] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-95950a02-20c9-4a3f-9ea8-1f7f19725cc1 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: 4ef8066a-b68c-457b-9964-b1c34bab0fc3] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:5b:79:66', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '225b6979-9329-403b-91fa-138bd41f6e83', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f2d49ff8-dbd8-41a2-9006-c71265c36b87', 'vif_model': 'vmxnet3'}] {{(pid=63515) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 797.936573] env[63515]: DEBUG oslo.service.loopingcall [None req-95950a02-20c9-4a3f-9ea8-1f7f19725cc1 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 797.941830] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4ef8066a-b68c-457b-9964-b1c34bab0fc3] Creating VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 797.942554] env[63515]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a5175a22-2c09-408e-8ea9-f9df41f5278e {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.971133] env[63515]: DEBUG oslo_concurrency.lockutils [None req-2e592662-996b-4dfa-8941-00965c1317fc tempest-ServersTestBootFromVolume-2042560423 tempest-ServersTestBootFromVolume-2042560423-project-member] Lock "90f4930b-aaa0-4c4b-9ab8-92aed45e200b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 37.580s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 797.972495] env[63515]: DEBUG oslo_vmware.api [None req-66c3e202-5a3e-433e-b543-4bfb02c8ad30 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775-project-member] Task: {'id': task-1111267, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.163882} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 797.976214] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-66c3e202-5a3e-433e-b543-4bfb02c8ad30 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775-project-member] [instance: 4e453127-1f3e-40ea-819f-6678479826c8] Extended root virtual disk {{(pid=63515) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 797.976754] env[63515]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 797.976754] env[63515]: value = "task-1111268" [ 797.976754] env[63515]: _type = "Task" [ 797.976754] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 797.977804] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-648ab541-bffa-42bf-a7ad-88e98d960d49 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.008958] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-66c3e202-5a3e-433e-b543-4bfb02c8ad30 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775-project-member] [instance: 4e453127-1f3e-40ea-819f-6678479826c8] Reconfiguring VM instance instance-0000003d to attach disk [datastore1] 4e453127-1f3e-40ea-819f-6678479826c8/4e453127-1f3e-40ea-819f-6678479826c8.vmdk or device None with type sparse {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 798.014974] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-24237ab6-9a4a-435e-a555-35aafd2dde9b {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.029673] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1111268, 'name': CreateVM_Task} progress is 6%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.038624] env[63515]: DEBUG oslo_vmware.api [None req-66c3e202-5a3e-433e-b543-4bfb02c8ad30 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775-project-member] Waiting for the task: (returnval){ [ 798.038624] env[63515]: value = "task-1111269" [ 798.038624] env[63515]: _type = "Task" [ 798.038624] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 798.050026] env[63515]: DEBUG oslo_vmware.api [None req-66c3e202-5a3e-433e-b543-4bfb02c8ad30 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775-project-member] Task: {'id': task-1111269, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.109540] env[63515]: DEBUG oslo_vmware.api [None req-4f34aec7-5320-4ef9-b5f8-71a5209ff0fa tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]5257154b-4b4f-0749-ce3d-a18611e8e82b, 'name': SearchDatastore_Task, 'duration_secs': 0.017291} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 798.109804] env[63515]: DEBUG oslo_concurrency.lockutils [None req-4f34aec7-5320-4ef9-b5f8-71a5209ff0fa tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 798.109894] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-4f34aec7-5320-4ef9-b5f8-71a5209ff0fa tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Copying virtual disk from [datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk to [datastore1] 63ce797d-7180-4209-ac2c-81978bf7607a/8a120570-cb06-4099-b262-554ca0ad15c5-rescue.vmdk. {{(pid=63515) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 798.110184] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-22d305cf-720b-44eb-a974-74db18f14e24 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.120156] env[63515]: DEBUG oslo_vmware.api [None req-4f34aec7-5320-4ef9-b5f8-71a5209ff0fa tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Waiting for the task: (returnval){ [ 798.120156] env[63515]: value = "task-1111270" [ 798.120156] env[63515]: _type = "Task" [ 798.120156] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 798.137051] env[63515]: DEBUG oslo_vmware.api [None req-4f34aec7-5320-4ef9-b5f8-71a5209ff0fa tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Task: {'id': task-1111270, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.142424] env[63515]: DEBUG oslo_vmware.api [None req-8d4c1f89-f73c-481b-8b79-61983aeba07b tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Task: {'id': task-1111239, 'name': ReconfigVM_Task} progress is 18%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.223361] env[63515]: DEBUG oslo_vmware.api [None req-32638298-0fa3-40f4-a1a6-2a70f7b4ad18 tempest-ServersAdminNegativeTestJSON-180276763 tempest-ServersAdminNegativeTestJSON-180276763-project-admin] Task: {'id': task-1111266, 'name': SuspendVM_Task} progress is 66%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.244251] env[63515]: DEBUG nova.network.neutron [req-2b8b2982-f06d-4922-9b57-42c331dcce99 req-07c2e5a4-ac20-4f92-8b7b-8c6cd6db2027 service nova] [instance: 4ef8066a-b68c-457b-9964-b1c34bab0fc3] Updated VIF entry in instance network info cache for port f2d49ff8-dbd8-41a2-9006-c71265c36b87. {{(pid=63515) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 798.244655] env[63515]: DEBUG nova.network.neutron [req-2b8b2982-f06d-4922-9b57-42c331dcce99 req-07c2e5a4-ac20-4f92-8b7b-8c6cd6db2027 service nova] [instance: 4ef8066a-b68c-457b-9964-b1c34bab0fc3] Updating instance_info_cache with network_info: [{"id": "f2d49ff8-dbd8-41a2-9006-c71265c36b87", "address": "fa:16:3e:5b:79:66", "network": {"id": "3e5914c8-0cd4-49ee-81b4-1f3e280440bb", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-641632277-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ea099ce07bfb4a8da014d9303cf552eb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "225b6979-9329-403b-91fa-138bd41f6e83", "external-id": "nsx-vlan-transportzone-38", "segmentation_id": 38, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf2d49ff8-db", "ovs_interfaceid": "f2d49ff8-dbd8-41a2-9006-c71265c36b87", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 798.268767] env[63515]: DEBUG nova.network.neutron [None req-b257f20f-3555-4579-b6f5-3780aaf93d24 tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] [instance: 192137b0-03e5-4bc4-b911-4b4f1a874f74] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 798.316358] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9db0e394-874f-420a-8e85-41a32af66a2e {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.324815] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e49fbb84-90a1-46bf-9345-1e774d7ad73b {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.361355] env[63515]: DEBUG nova.network.neutron [None req-b257f20f-3555-4579-b6f5-3780aaf93d24 tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] [instance: 192137b0-03e5-4bc4-b911-4b4f1a874f74] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 798.363051] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f81d70a-fc30-4742-8eb9-8eef1904733b {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.371715] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9131e6c3-7ed2-4b7e-b4d0-4ec2243e72c2 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.388993] env[63515]: DEBUG nova.compute.provider_tree [None req-655858bd-8f63-4e55-ac5e-dbecfbd88a62 tempest-ServersV294TestFqdnHostnames-2112993702 tempest-ServersV294TestFqdnHostnames-2112993702-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 798.493214] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1111268, 'name': CreateVM_Task} progress is 99%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.548589] env[63515]: DEBUG oslo_vmware.api [None req-66c3e202-5a3e-433e-b543-4bfb02c8ad30 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775-project-member] Task: {'id': task-1111269, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.635967] env[63515]: DEBUG oslo_vmware.api [None req-4f34aec7-5320-4ef9-b5f8-71a5209ff0fa tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Task: {'id': task-1111270, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.641584] env[63515]: DEBUG oslo_vmware.api [None req-8d4c1f89-f73c-481b-8b79-61983aeba07b tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Task: {'id': task-1111239, 'name': ReconfigVM_Task} progress is 18%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.720557] env[63515]: DEBUG oslo_vmware.api [None req-32638298-0fa3-40f4-a1a6-2a70f7b4ad18 tempest-ServersAdminNegativeTestJSON-180276763 tempest-ServersAdminNegativeTestJSON-180276763-project-admin] Task: {'id': task-1111266, 'name': SuspendVM_Task} progress is 66%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.749596] env[63515]: DEBUG oslo_concurrency.lockutils [req-2b8b2982-f06d-4922-9b57-42c331dcce99 req-07c2e5a4-ac20-4f92-8b7b-8c6cd6db2027 service nova] Releasing lock "refresh_cache-4ef8066a-b68c-457b-9964-b1c34bab0fc3" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 798.866771] env[63515]: DEBUG oslo_concurrency.lockutils [None req-b257f20f-3555-4579-b6f5-3780aaf93d24 tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] Releasing lock "refresh_cache-192137b0-03e5-4bc4-b911-4b4f1a874f74" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 798.867302] env[63515]: DEBUG nova.compute.manager [None req-b257f20f-3555-4579-b6f5-3780aaf93d24 tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] [instance: 192137b0-03e5-4bc4-b911-4b4f1a874f74] Start destroying the instance on the hypervisor. {{(pid=63515) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 798.867565] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-b257f20f-3555-4579-b6f5-3780aaf93d24 tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] [instance: 192137b0-03e5-4bc4-b911-4b4f1a874f74] Destroying instance {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 798.868589] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ea6b3b7-d697-4fc3-ad3b-71ccd5d6201a {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.877831] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-b257f20f-3555-4579-b6f5-3780aaf93d24 tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] [instance: 192137b0-03e5-4bc4-b911-4b4f1a874f74] Powering off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 798.878125] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-976e4ee8-d47d-4960-8f59-ba98aabea600 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.885040] env[63515]: DEBUG oslo_vmware.api [None req-b257f20f-3555-4579-b6f5-3780aaf93d24 tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] Waiting for the task: (returnval){ [ 798.885040] env[63515]: value = "task-1111271" [ 798.885040] env[63515]: _type = "Task" [ 798.885040] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 798.894941] env[63515]: DEBUG nova.scheduler.client.report [None req-655858bd-8f63-4e55-ac5e-dbecfbd88a62 tempest-ServersV294TestFqdnHostnames-2112993702 tempest-ServersV294TestFqdnHostnames-2112993702-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 798.898346] env[63515]: DEBUG oslo_vmware.api [None req-b257f20f-3555-4579-b6f5-3780aaf93d24 tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] Task: {'id': task-1111271, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.990346] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1111268, 'name': CreateVM_Task, 'duration_secs': 0.529299} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 798.990550] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4ef8066a-b68c-457b-9964-b1c34bab0fc3] Created VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 798.991217] env[63515]: DEBUG oslo_concurrency.lockutils [None req-95950a02-20c9-4a3f-9ea8-1f7f19725cc1 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 798.991391] env[63515]: DEBUG oslo_concurrency.lockutils [None req-95950a02-20c9-4a3f-9ea8-1f7f19725cc1 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 798.991710] env[63515]: DEBUG oslo_concurrency.lockutils [None req-95950a02-20c9-4a3f-9ea8-1f7f19725cc1 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 798.991966] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-053e5227-71fc-4e04-88d4-1b81d37c1248 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.996588] env[63515]: DEBUG oslo_vmware.api [None req-95950a02-20c9-4a3f-9ea8-1f7f19725cc1 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Waiting for the task: (returnval){ [ 798.996588] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]52fe42d8-2988-7e24-a6a1-29272f5ffa4d" [ 798.996588] env[63515]: _type = "Task" [ 798.996588] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 799.004511] env[63515]: DEBUG oslo_vmware.api [None req-95950a02-20c9-4a3f-9ea8-1f7f19725cc1 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52fe42d8-2988-7e24-a6a1-29272f5ffa4d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 799.048889] env[63515]: DEBUG oslo_vmware.api [None req-66c3e202-5a3e-433e-b543-4bfb02c8ad30 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775-project-member] Task: {'id': task-1111269, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 799.145570] env[63515]: DEBUG oslo_vmware.api [None req-8d4c1f89-f73c-481b-8b79-61983aeba07b tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Task: {'id': task-1111239, 'name': ReconfigVM_Task} progress is 18%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 799.145906] env[63515]: DEBUG oslo_vmware.api [None req-4f34aec7-5320-4ef9-b5f8-71a5209ff0fa tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Task: {'id': task-1111270, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 799.221664] env[63515]: DEBUG oslo_vmware.api [None req-32638298-0fa3-40f4-a1a6-2a70f7b4ad18 tempest-ServersAdminNegativeTestJSON-180276763 tempest-ServersAdminNegativeTestJSON-180276763-project-admin] Task: {'id': task-1111266, 'name': SuspendVM_Task} progress is 66%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 799.401296] env[63515]: DEBUG oslo_concurrency.lockutils [None req-655858bd-8f63-4e55-ac5e-dbecfbd88a62 tempest-ServersV294TestFqdnHostnames-2112993702 tempest-ServersV294TestFqdnHostnames-2112993702-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.970s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 799.403392] env[63515]: DEBUG oslo_vmware.api [None req-b257f20f-3555-4579-b6f5-3780aaf93d24 tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] Task: {'id': task-1111271, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 799.403812] env[63515]: DEBUG oslo_concurrency.lockutils [None req-41b4a543-33b4-43d9-863c-4344bf9d0a46 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 32.636s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 799.407284] env[63515]: INFO nova.compute.claims [None req-41b4a543-33b4-43d9-863c-4344bf9d0a46 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: d926c699-a64a-4942-9ef4-f0166414661d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 799.425818] env[63515]: INFO nova.scheduler.client.report [None req-655858bd-8f63-4e55-ac5e-dbecfbd88a62 tempest-ServersV294TestFqdnHostnames-2112993702 tempest-ServersV294TestFqdnHostnames-2112993702-project-member] Deleted allocations for instance 396e49dd-48c0-496b-a1ec-190c33a22c5e [ 799.506973] env[63515]: DEBUG oslo_vmware.api [None req-95950a02-20c9-4a3f-9ea8-1f7f19725cc1 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52fe42d8-2988-7e24-a6a1-29272f5ffa4d, 'name': SearchDatastore_Task, 'duration_secs': 0.011655} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 799.507315] env[63515]: DEBUG oslo_concurrency.lockutils [None req-95950a02-20c9-4a3f-9ea8-1f7f19725cc1 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 799.507572] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-95950a02-20c9-4a3f-9ea8-1f7f19725cc1 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: 4ef8066a-b68c-457b-9964-b1c34bab0fc3] Processing image 8a120570-cb06-4099-b262-554ca0ad15c5 {{(pid=63515) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 799.507823] env[63515]: DEBUG oslo_concurrency.lockutils [None req-95950a02-20c9-4a3f-9ea8-1f7f19725cc1 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 799.507971] env[63515]: DEBUG oslo_concurrency.lockutils [None req-95950a02-20c9-4a3f-9ea8-1f7f19725cc1 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 799.508172] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-95950a02-20c9-4a3f-9ea8-1f7f19725cc1 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 799.508450] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ba9b24ba-d33b-42c5-abe2-6e30ab8c4850 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.516877] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-95950a02-20c9-4a3f-9ea8-1f7f19725cc1 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 799.517087] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-95950a02-20c9-4a3f-9ea8-1f7f19725cc1 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63515) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 799.517891] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d55471c2-cad0-4ede-9650-d4cfb55bdb97 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.522981] env[63515]: DEBUG oslo_vmware.api [None req-95950a02-20c9-4a3f-9ea8-1f7f19725cc1 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Waiting for the task: (returnval){ [ 799.522981] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]527daaad-9bb3-70d8-8744-2dca9bdcdd4b" [ 799.522981] env[63515]: _type = "Task" [ 799.522981] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 799.532748] env[63515]: DEBUG oslo_vmware.api [None req-95950a02-20c9-4a3f-9ea8-1f7f19725cc1 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]527daaad-9bb3-70d8-8744-2dca9bdcdd4b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 799.549023] env[63515]: DEBUG oslo_vmware.api [None req-66c3e202-5a3e-433e-b543-4bfb02c8ad30 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775-project-member] Task: {'id': task-1111269, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 799.638948] env[63515]: DEBUG oslo_vmware.api [None req-4f34aec7-5320-4ef9-b5f8-71a5209ff0fa tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Task: {'id': task-1111270, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 799.644160] env[63515]: DEBUG oslo_vmware.api [None req-8d4c1f89-f73c-481b-8b79-61983aeba07b tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Task: {'id': task-1111239, 'name': ReconfigVM_Task} progress is 18%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 799.721560] env[63515]: DEBUG oslo_vmware.api [None req-32638298-0fa3-40f4-a1a6-2a70f7b4ad18 tempest-ServersAdminNegativeTestJSON-180276763 tempest-ServersAdminNegativeTestJSON-180276763-project-admin] Task: {'id': task-1111266, 'name': SuspendVM_Task} progress is 66%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 799.902677] env[63515]: DEBUG oslo_vmware.api [None req-b257f20f-3555-4579-b6f5-3780aaf93d24 tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] Task: {'id': task-1111271, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 799.933694] env[63515]: DEBUG oslo_concurrency.lockutils [None req-655858bd-8f63-4e55-ac5e-dbecfbd88a62 tempest-ServersV294TestFqdnHostnames-2112993702 tempest-ServersV294TestFqdnHostnames-2112993702-project-member] Lock "396e49dd-48c0-496b-a1ec-190c33a22c5e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 37.084s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 800.033406] env[63515]: DEBUG oslo_vmware.api [None req-95950a02-20c9-4a3f-9ea8-1f7f19725cc1 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]527daaad-9bb3-70d8-8744-2dca9bdcdd4b, 'name': SearchDatastore_Task, 'duration_secs': 0.009487} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 800.034123] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-add8b402-9207-4208-a49d-4a1b60ac37d8 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.039471] env[63515]: DEBUG oslo_vmware.api [None req-95950a02-20c9-4a3f-9ea8-1f7f19725cc1 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Waiting for the task: (returnval){ [ 800.039471] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]52eb87bd-fffa-c462-43d0-c864038753e2" [ 800.039471] env[63515]: _type = "Task" [ 800.039471] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 800.050069] env[63515]: DEBUG oslo_vmware.api [None req-95950a02-20c9-4a3f-9ea8-1f7f19725cc1 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52eb87bd-fffa-c462-43d0-c864038753e2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 800.053215] env[63515]: DEBUG oslo_vmware.api [None req-66c3e202-5a3e-433e-b543-4bfb02c8ad30 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775-project-member] Task: {'id': task-1111269, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 800.145599] env[63515]: DEBUG oslo_vmware.api [None req-4f34aec7-5320-4ef9-b5f8-71a5209ff0fa tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Task: {'id': task-1111270, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 800.149388] env[63515]: DEBUG oslo_vmware.api [None req-8d4c1f89-f73c-481b-8b79-61983aeba07b tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Task: {'id': task-1111239, 'name': ReconfigVM_Task} progress is 18%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 800.224535] env[63515]: DEBUG oslo_vmware.api [None req-32638298-0fa3-40f4-a1a6-2a70f7b4ad18 tempest-ServersAdminNegativeTestJSON-180276763 tempest-ServersAdminNegativeTestJSON-180276763-project-admin] Task: {'id': task-1111266, 'name': SuspendVM_Task} progress is 66%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 800.402869] env[63515]: DEBUG oslo_vmware.api [None req-b257f20f-3555-4579-b6f5-3780aaf93d24 tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] Task: {'id': task-1111271, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 800.557996] env[63515]: DEBUG oslo_vmware.api [None req-95950a02-20c9-4a3f-9ea8-1f7f19725cc1 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52eb87bd-fffa-c462-43d0-c864038753e2, 'name': SearchDatastore_Task, 'duration_secs': 0.010717} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 800.563148] env[63515]: DEBUG oslo_concurrency.lockutils [None req-95950a02-20c9-4a3f-9ea8-1f7f19725cc1 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 800.563452] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-95950a02-20c9-4a3f-9ea8-1f7f19725cc1 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk to [datastore2] 4ef8066a-b68c-457b-9964-b1c34bab0fc3/4ef8066a-b68c-457b-9964-b1c34bab0fc3.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 800.563728] env[63515]: DEBUG oslo_vmware.api [None req-66c3e202-5a3e-433e-b543-4bfb02c8ad30 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775-project-member] Task: {'id': task-1111269, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 800.564159] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-40930f52-0073-4599-bb19-dad52cadc860 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.572648] env[63515]: DEBUG oslo_vmware.api [None req-95950a02-20c9-4a3f-9ea8-1f7f19725cc1 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Waiting for the task: (returnval){ [ 800.572648] env[63515]: value = "task-1111272" [ 800.572648] env[63515]: _type = "Task" [ 800.572648] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 800.583633] env[63515]: DEBUG oslo_vmware.api [None req-95950a02-20c9-4a3f-9ea8-1f7f19725cc1 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Task: {'id': task-1111272, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 800.643304] env[63515]: DEBUG oslo_vmware.api [None req-4f34aec7-5320-4ef9-b5f8-71a5209ff0fa tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Task: {'id': task-1111270, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 800.648623] env[63515]: DEBUG oslo_vmware.api [None req-8d4c1f89-f73c-481b-8b79-61983aeba07b tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Task: {'id': task-1111239, 'name': ReconfigVM_Task} progress is 18%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 800.724864] env[63515]: DEBUG oslo_vmware.api [None req-32638298-0fa3-40f4-a1a6-2a70f7b4ad18 tempest-ServersAdminNegativeTestJSON-180276763 tempest-ServersAdminNegativeTestJSON-180276763-project-admin] Task: {'id': task-1111266, 'name': SuspendVM_Task} progress is 66%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 800.770552] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e31e7586-542a-4de4-a3e5-f96b3ce11f9d {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.778731] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa2ce253-f2f8-4744-bab6-3a0fb8c5ce60 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.811618] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e897f735-88ae-41d2-94a3-0a4dca1c39d6 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.820592] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a169a54-71f8-468f-a2a8-49dfaa88c496 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.836131] env[63515]: DEBUG nova.compute.provider_tree [None req-41b4a543-33b4-43d9-863c-4344bf9d0a46 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 800.900186] env[63515]: DEBUG oslo_vmware.api [None req-b257f20f-3555-4579-b6f5-3780aaf93d24 tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] Task: {'id': task-1111271, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 801.057368] env[63515]: DEBUG oslo_vmware.api [None req-66c3e202-5a3e-433e-b543-4bfb02c8ad30 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775-project-member] Task: {'id': task-1111269, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 801.082907] env[63515]: DEBUG oslo_vmware.api [None req-95950a02-20c9-4a3f-9ea8-1f7f19725cc1 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Task: {'id': task-1111272, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 801.138628] env[63515]: DEBUG oslo_vmware.api [None req-4f34aec7-5320-4ef9-b5f8-71a5209ff0fa tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Task: {'id': task-1111270, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 801.147966] env[63515]: DEBUG oslo_vmware.api [None req-8d4c1f89-f73c-481b-8b79-61983aeba07b tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Task: {'id': task-1111239, 'name': ReconfigVM_Task} progress is 18%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 801.225439] env[63515]: DEBUG oslo_vmware.api [None req-32638298-0fa3-40f4-a1a6-2a70f7b4ad18 tempest-ServersAdminNegativeTestJSON-180276763 tempest-ServersAdminNegativeTestJSON-180276763-project-admin] Task: {'id': task-1111266, 'name': SuspendVM_Task} progress is 66%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 801.340193] env[63515]: DEBUG nova.scheduler.client.report [None req-41b4a543-33b4-43d9-863c-4344bf9d0a46 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 801.400895] env[63515]: DEBUG oslo_vmware.api [None req-b257f20f-3555-4579-b6f5-3780aaf93d24 tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] Task: {'id': task-1111271, 'name': PowerOffVM_Task, 'duration_secs': 2.272009} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 801.401873] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-b257f20f-3555-4579-b6f5-3780aaf93d24 tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] [instance: 192137b0-03e5-4bc4-b911-4b4f1a874f74] Powered off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 801.402045] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-b257f20f-3555-4579-b6f5-3780aaf93d24 tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] [instance: 192137b0-03e5-4bc4-b911-4b4f1a874f74] Unregistering the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 801.402434] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b954b755-41a5-4eb8-983a-2754ec21dc52 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.470362] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-b257f20f-3555-4579-b6f5-3780aaf93d24 tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] [instance: 192137b0-03e5-4bc4-b911-4b4f1a874f74] Unregistered the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 801.470668] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-b257f20f-3555-4579-b6f5-3780aaf93d24 tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] [instance: 192137b0-03e5-4bc4-b911-4b4f1a874f74] Deleting contents of the VM from datastore datastore1 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 801.470876] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-b257f20f-3555-4579-b6f5-3780aaf93d24 tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] Deleting the datastore file [datastore1] 192137b0-03e5-4bc4-b911-4b4f1a874f74 {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 801.471186] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ce955d09-02af-4848-a45c-d82f07b698d9 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.478158] env[63515]: DEBUG oslo_vmware.api [None req-b257f20f-3555-4579-b6f5-3780aaf93d24 tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] Waiting for the task: (returnval){ [ 801.478158] env[63515]: value = "task-1111274" [ 801.478158] env[63515]: _type = "Task" [ 801.478158] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 801.486573] env[63515]: DEBUG oslo_vmware.api [None req-b257f20f-3555-4579-b6f5-3780aaf93d24 tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] Task: {'id': task-1111274, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 801.570614] env[63515]: DEBUG oslo_vmware.api [None req-66c3e202-5a3e-433e-b543-4bfb02c8ad30 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775-project-member] Task: {'id': task-1111269, 'name': ReconfigVM_Task} progress is 99%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 801.587026] env[63515]: DEBUG oslo_vmware.api [None req-95950a02-20c9-4a3f-9ea8-1f7f19725cc1 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Task: {'id': task-1111272, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 801.640059] env[63515]: DEBUG oslo_vmware.api [None req-4f34aec7-5320-4ef9-b5f8-71a5209ff0fa tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Task: {'id': task-1111270, 'name': CopyVirtualDisk_Task, 'duration_secs': 3.318585} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 801.643534] env[63515]: INFO nova.virt.vmwareapi.ds_util [None req-4f34aec7-5320-4ef9-b5f8-71a5209ff0fa tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Copied virtual disk from [datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk to [datastore1] 63ce797d-7180-4209-ac2c-81978bf7607a/8a120570-cb06-4099-b262-554ca0ad15c5-rescue.vmdk. [ 801.644483] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a471aad8-4f2e-44b9-b085-6848b66e8ad1 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.654858] env[63515]: DEBUG oslo_vmware.api [None req-8d4c1f89-f73c-481b-8b79-61983aeba07b tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Task: {'id': task-1111239, 'name': ReconfigVM_Task} progress is 99%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 801.679165] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-4f34aec7-5320-4ef9-b5f8-71a5209ff0fa tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] [instance: 63ce797d-7180-4209-ac2c-81978bf7607a] Reconfiguring VM instance instance-0000003b to attach disk [datastore1] 63ce797d-7180-4209-ac2c-81978bf7607a/8a120570-cb06-4099-b262-554ca0ad15c5-rescue.vmdk or device None with type thin {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 801.680196] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cf0f8d8b-a891-4762-90e2-05d42a931258 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.700337] env[63515]: DEBUG oslo_vmware.api [None req-4f34aec7-5320-4ef9-b5f8-71a5209ff0fa tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Waiting for the task: (returnval){ [ 801.700337] env[63515]: value = "task-1111275" [ 801.700337] env[63515]: _type = "Task" [ 801.700337] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 801.709674] env[63515]: DEBUG oslo_vmware.api [None req-4f34aec7-5320-4ef9-b5f8-71a5209ff0fa tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Task: {'id': task-1111275, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 801.724434] env[63515]: DEBUG oslo_vmware.api [None req-32638298-0fa3-40f4-a1a6-2a70f7b4ad18 tempest-ServersAdminNegativeTestJSON-180276763 tempest-ServersAdminNegativeTestJSON-180276763-project-admin] Task: {'id': task-1111266, 'name': SuspendVM_Task, 'duration_secs': 4.473382} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 801.724818] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-32638298-0fa3-40f4-a1a6-2a70f7b4ad18 tempest-ServersAdminNegativeTestJSON-180276763 tempest-ServersAdminNegativeTestJSON-180276763-project-admin] [instance: e666825c-ff4e-4a0e-93c0-43c00f167bbb] Suspended the VM {{(pid=63515) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1166}} [ 801.725050] env[63515]: DEBUG nova.compute.manager [None req-32638298-0fa3-40f4-a1a6-2a70f7b4ad18 tempest-ServersAdminNegativeTestJSON-180276763 tempest-ServersAdminNegativeTestJSON-180276763-project-admin] [instance: e666825c-ff4e-4a0e-93c0-43c00f167bbb] Checking state {{(pid=63515) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 801.725969] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe0e43b0-29c0-428a-b832-26d203356b93 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.845542] env[63515]: DEBUG oslo_concurrency.lockutils [None req-41b4a543-33b4-43d9-863c-4344bf9d0a46 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.442s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 801.847397] env[63515]: DEBUG nova.compute.manager [None req-41b4a543-33b4-43d9-863c-4344bf9d0a46 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: d926c699-a64a-4942-9ef4-f0166414661d] Start building networks asynchronously for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 801.850843] env[63515]: DEBUG oslo_concurrency.lockutils [None req-65f1d514-d43e-4354-a5f1-c564a5380800 tempest-ServersAaction247Test-2019176257 tempest-ServersAaction247Test-2019176257-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 31.581s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 801.851094] env[63515]: DEBUG nova.objects.instance [None req-65f1d514-d43e-4354-a5f1-c564a5380800 tempest-ServersAaction247Test-2019176257 tempest-ServersAaction247Test-2019176257-project-member] Lazy-loading 'resources' on Instance uuid 5d50fbd6-a7fb-422c-9dd7-df7140fd0aae {{(pid=63515) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 801.991804] env[63515]: DEBUG oslo_vmware.api [None req-b257f20f-3555-4579-b6f5-3780aaf93d24 tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] Task: {'id': task-1111274, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.244454} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 801.991804] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-b257f20f-3555-4579-b6f5-3780aaf93d24 tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] Deleted the datastore file {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 801.991804] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-b257f20f-3555-4579-b6f5-3780aaf93d24 tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] [instance: 192137b0-03e5-4bc4-b911-4b4f1a874f74] Deleted contents of the VM from datastore datastore1 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 801.991804] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-b257f20f-3555-4579-b6f5-3780aaf93d24 tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] [instance: 192137b0-03e5-4bc4-b911-4b4f1a874f74] Instance destroyed {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 801.991804] env[63515]: INFO nova.compute.manager [None req-b257f20f-3555-4579-b6f5-3780aaf93d24 tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] [instance: 192137b0-03e5-4bc4-b911-4b4f1a874f74] Took 3.12 seconds to destroy the instance on the hypervisor. [ 801.991804] env[63515]: DEBUG oslo.service.loopingcall [None req-b257f20f-3555-4579-b6f5-3780aaf93d24 tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 801.991804] env[63515]: DEBUG nova.compute.manager [-] [instance: 192137b0-03e5-4bc4-b911-4b4f1a874f74] Deallocating network for instance {{(pid=63515) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 801.992117] env[63515]: DEBUG nova.network.neutron [-] [instance: 192137b0-03e5-4bc4-b911-4b4f1a874f74] deallocate_for_instance() {{(pid=63515) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 802.037654] env[63515]: DEBUG nova.network.neutron [-] [instance: 192137b0-03e5-4bc4-b911-4b4f1a874f74] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 802.062576] env[63515]: DEBUG oslo_vmware.api [None req-66c3e202-5a3e-433e-b543-4bfb02c8ad30 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775-project-member] Task: {'id': task-1111269, 'name': ReconfigVM_Task, 'duration_secs': 3.620973} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 802.063118] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-66c3e202-5a3e-433e-b543-4bfb02c8ad30 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775-project-member] [instance: 4e453127-1f3e-40ea-819f-6678479826c8] Reconfigured VM instance instance-0000003d to attach disk [datastore1] 4e453127-1f3e-40ea-819f-6678479826c8/4e453127-1f3e-40ea-819f-6678479826c8.vmdk or device None with type sparse {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 802.063904] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-81f39d47-6e8d-4fa4-a7ee-a111c04304c1 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.071326] env[63515]: DEBUG oslo_vmware.api [None req-66c3e202-5a3e-433e-b543-4bfb02c8ad30 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775-project-member] Waiting for the task: (returnval){ [ 802.071326] env[63515]: value = "task-1111276" [ 802.071326] env[63515]: _type = "Task" [ 802.071326] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 802.087115] env[63515]: DEBUG oslo_vmware.api [None req-95950a02-20c9-4a3f-9ea8-1f7f19725cc1 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Task: {'id': task-1111272, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.123358} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 802.091072] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-95950a02-20c9-4a3f-9ea8-1f7f19725cc1 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk to [datastore2] 4ef8066a-b68c-457b-9964-b1c34bab0fc3/4ef8066a-b68c-457b-9964-b1c34bab0fc3.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 802.091476] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-95950a02-20c9-4a3f-9ea8-1f7f19725cc1 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: 4ef8066a-b68c-457b-9964-b1c34bab0fc3] Extending root virtual disk to 1048576 {{(pid=63515) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 802.092171] env[63515]: DEBUG oslo_vmware.api [None req-66c3e202-5a3e-433e-b543-4bfb02c8ad30 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775-project-member] Task: {'id': task-1111276, 'name': Rename_Task} progress is 5%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 802.092364] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-411df5c7-b879-48c6-a4ab-71a8e3dea133 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.100707] env[63515]: DEBUG oslo_vmware.api [None req-95950a02-20c9-4a3f-9ea8-1f7f19725cc1 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Waiting for the task: (returnval){ [ 802.100707] env[63515]: value = "task-1111277" [ 802.100707] env[63515]: _type = "Task" [ 802.100707] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 802.110267] env[63515]: DEBUG oslo_vmware.api [None req-95950a02-20c9-4a3f-9ea8-1f7f19725cc1 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Task: {'id': task-1111277, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 802.151021] env[63515]: DEBUG oslo_vmware.api [None req-8d4c1f89-f73c-481b-8b79-61983aeba07b tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Task: {'id': task-1111239, 'name': ReconfigVM_Task} progress is 99%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 802.211896] env[63515]: DEBUG oslo_vmware.api [None req-4f34aec7-5320-4ef9-b5f8-71a5209ff0fa tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Task: {'id': task-1111275, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 802.355024] env[63515]: DEBUG nova.compute.utils [None req-41b4a543-33b4-43d9-863c-4344bf9d0a46 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Using /dev/sd instead of None {{(pid=63515) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 802.359187] env[63515]: DEBUG nova.compute.manager [None req-41b4a543-33b4-43d9-863c-4344bf9d0a46 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: d926c699-a64a-4942-9ef4-f0166414661d] Allocating IP information in the background. {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 802.359394] env[63515]: DEBUG nova.network.neutron [None req-41b4a543-33b4-43d9-863c-4344bf9d0a46 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: d926c699-a64a-4942-9ef4-f0166414661d] allocate_for_instance() {{(pid=63515) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 802.488620] env[63515]: DEBUG nova.policy [None req-41b4a543-33b4-43d9-863c-4344bf9d0a46 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b5de3e36ae88482eb795894592e76c7c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f9c98e9f6020475490aaa7e76d907ab7', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63515) authorize /opt/stack/nova/nova/policy.py:201}} [ 802.539966] env[63515]: DEBUG nova.network.neutron [-] [instance: 192137b0-03e5-4bc4-b911-4b4f1a874f74] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 802.586263] env[63515]: DEBUG oslo_vmware.api [None req-66c3e202-5a3e-433e-b543-4bfb02c8ad30 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775-project-member] Task: {'id': task-1111276, 'name': Rename_Task, 'duration_secs': 0.177955} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 802.587132] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-66c3e202-5a3e-433e-b543-4bfb02c8ad30 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775-project-member] [instance: 4e453127-1f3e-40ea-819f-6678479826c8] Powering on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 802.587423] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ed1b0dcf-e357-4007-a863-e1fb5949909a {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.596736] env[63515]: DEBUG oslo_vmware.api [None req-66c3e202-5a3e-433e-b543-4bfb02c8ad30 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775-project-member] Waiting for the task: (returnval){ [ 802.596736] env[63515]: value = "task-1111278" [ 802.596736] env[63515]: _type = "Task" [ 802.596736] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 802.609931] env[63515]: DEBUG oslo_vmware.api [None req-66c3e202-5a3e-433e-b543-4bfb02c8ad30 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775-project-member] Task: {'id': task-1111278, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 802.618503] env[63515]: DEBUG oslo_vmware.api [None req-95950a02-20c9-4a3f-9ea8-1f7f19725cc1 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Task: {'id': task-1111277, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.070118} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 802.618802] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-95950a02-20c9-4a3f-9ea8-1f7f19725cc1 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: 4ef8066a-b68c-457b-9964-b1c34bab0fc3] Extended root virtual disk {{(pid=63515) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 802.619635] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7c18a71-87ce-46b0-8999-cce53b2e96b2 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.645963] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-95950a02-20c9-4a3f-9ea8-1f7f19725cc1 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: 4ef8066a-b68c-457b-9964-b1c34bab0fc3] Reconfiguring VM instance instance-0000003e to attach disk [datastore2] 4ef8066a-b68c-457b-9964-b1c34bab0fc3/4ef8066a-b68c-457b-9964-b1c34bab0fc3.vmdk or device None with type sparse {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 802.649435] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-04f20a39-5b39-4bcb-8b36-8aa59ef7ed27 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.673694] env[63515]: DEBUG oslo_vmware.api [None req-8d4c1f89-f73c-481b-8b79-61983aeba07b tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Task: {'id': task-1111239, 'name': ReconfigVM_Task, 'duration_secs': 13.002523} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 802.675116] env[63515]: DEBUG oslo_concurrency.lockutils [None req-8d4c1f89-f73c-481b-8b79-61983aeba07b tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Releasing lock "ed550b10-d58f-45b8-b766-198f431c3788" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 802.675339] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-8d4c1f89-f73c-481b-8b79-61983aeba07b tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: ed550b10-d58f-45b8-b766-198f431c3788] Reconfigured VM to detach interface {{(pid=63515) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1978}} [ 802.677674] env[63515]: DEBUG oslo_vmware.api [None req-95950a02-20c9-4a3f-9ea8-1f7f19725cc1 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Waiting for the task: (returnval){ [ 802.677674] env[63515]: value = "task-1111279" [ 802.677674] env[63515]: _type = "Task" [ 802.677674] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 802.686305] env[63515]: DEBUG oslo_vmware.api [None req-95950a02-20c9-4a3f-9ea8-1f7f19725cc1 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Task: {'id': task-1111279, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 802.710899] env[63515]: DEBUG oslo_vmware.api [None req-4f34aec7-5320-4ef9-b5f8-71a5209ff0fa tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Task: {'id': task-1111275, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 802.821753] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-239ddf1e-a490-4828-bca4-0e6240c6553c {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.830126] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0226306-6a72-4cb1-91cd-f830475af0b8 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.861668] env[63515]: DEBUG nova.compute.manager [None req-41b4a543-33b4-43d9-863c-4344bf9d0a46 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: d926c699-a64a-4942-9ef4-f0166414661d] Start building block device mappings for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 802.865404] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bbafaa94-fabb-46dc-a5fd-069e13ad8e53 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.873628] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79c7ecb1-91c4-4cdc-8497-c0feb240a75f {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.889576] env[63515]: DEBUG nova.compute.provider_tree [None req-65f1d514-d43e-4354-a5f1-c564a5380800 tempest-ServersAaction247Test-2019176257 tempest-ServersAaction247Test-2019176257-project-member] Updating inventory in ProviderTree for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 803.045030] env[63515]: INFO nova.compute.manager [-] [instance: 192137b0-03e5-4bc4-b911-4b4f1a874f74] Took 1.05 seconds to deallocate network for instance. [ 803.112030] env[63515]: DEBUG oslo_vmware.api [None req-66c3e202-5a3e-433e-b543-4bfb02c8ad30 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775-project-member] Task: {'id': task-1111278, 'name': PowerOnVM_Task, 'duration_secs': 0.493799} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 803.112030] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-66c3e202-5a3e-433e-b543-4bfb02c8ad30 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775-project-member] [instance: 4e453127-1f3e-40ea-819f-6678479826c8] Powered on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 803.112030] env[63515]: INFO nova.compute.manager [None req-66c3e202-5a3e-433e-b543-4bfb02c8ad30 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775-project-member] [instance: 4e453127-1f3e-40ea-819f-6678479826c8] Took 11.45 seconds to spawn the instance on the hypervisor. [ 803.112030] env[63515]: DEBUG nova.compute.manager [None req-66c3e202-5a3e-433e-b543-4bfb02c8ad30 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775-project-member] [instance: 4e453127-1f3e-40ea-819f-6678479826c8] Checking state {{(pid=63515) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 803.112030] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4f4981a-3cea-4307-87d4-33ea22af0196 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.143416] env[63515]: DEBUG nova.network.neutron [None req-41b4a543-33b4-43d9-863c-4344bf9d0a46 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: d926c699-a64a-4942-9ef4-f0166414661d] Successfully created port: fdec91d2-43e8-4d40-9aa9-885a82fd9b42 {{(pid=63515) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 803.191638] env[63515]: DEBUG oslo_vmware.api [None req-95950a02-20c9-4a3f-9ea8-1f7f19725cc1 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Task: {'id': task-1111279, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 803.215750] env[63515]: DEBUG oslo_vmware.api [None req-4f34aec7-5320-4ef9-b5f8-71a5209ff0fa tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Task: {'id': task-1111275, 'name': ReconfigVM_Task, 'duration_secs': 1.449618} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 803.215897] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-4f34aec7-5320-4ef9-b5f8-71a5209ff0fa tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] [instance: 63ce797d-7180-4209-ac2c-81978bf7607a] Reconfigured VM instance instance-0000003b to attach disk [datastore1] 63ce797d-7180-4209-ac2c-81978bf7607a/8a120570-cb06-4099-b262-554ca0ad15c5-rescue.vmdk or device None with type thin {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 803.217534] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c84dba36-53a4-446e-a23f-1d72f156d15a {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.251577] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-605d282c-d048-4f93-a796-8d930a45f2da {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.268057] env[63515]: DEBUG oslo_vmware.api [None req-4f34aec7-5320-4ef9-b5f8-71a5209ff0fa tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Waiting for the task: (returnval){ [ 803.268057] env[63515]: value = "task-1111280" [ 803.268057] env[63515]: _type = "Task" [ 803.268057] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 803.277721] env[63515]: DEBUG oslo_vmware.api [None req-4f34aec7-5320-4ef9-b5f8-71a5209ff0fa tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Task: {'id': task-1111280, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 803.420838] env[63515]: ERROR nova.scheduler.client.report [None req-65f1d514-d43e-4354-a5f1-c564a5380800 tempest-ServersAaction247Test-2019176257 tempest-ServersAaction247Test-2019176257-project-member] [req-e69cbfc9-674e-486e-ae3e-dc9355b0c0b5] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 2bd86232-2b6f-44d5-9057-1a3a6b27185a. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-e69cbfc9-674e-486e-ae3e-dc9355b0c0b5"}]} [ 803.438877] env[63515]: DEBUG nova.scheduler.client.report [None req-65f1d514-d43e-4354-a5f1-c564a5380800 tempest-ServersAaction247Test-2019176257 tempest-ServersAaction247Test-2019176257-project-member] Refreshing inventories for resource provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 803.454930] env[63515]: DEBUG nova.scheduler.client.report [None req-65f1d514-d43e-4354-a5f1-c564a5380800 tempest-ServersAaction247Test-2019176257 tempest-ServersAaction247Test-2019176257-project-member] Updating ProviderTree inventory for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 803.455193] env[63515]: DEBUG nova.compute.provider_tree [None req-65f1d514-d43e-4354-a5f1-c564a5380800 tempest-ServersAaction247Test-2019176257 tempest-ServersAaction247Test-2019176257-project-member] Updating inventory in ProviderTree for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 803.470240] env[63515]: DEBUG nova.scheduler.client.report [None req-65f1d514-d43e-4354-a5f1-c564a5380800 tempest-ServersAaction247Test-2019176257 tempest-ServersAaction247Test-2019176257-project-member] Refreshing aggregate associations for resource provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a, aggregates: None {{(pid=63515) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 803.490081] env[63515]: DEBUG nova.scheduler.client.report [None req-65f1d514-d43e-4354-a5f1-c564a5380800 tempest-ServersAaction247Test-2019176257 tempest-ServersAaction247Test-2019176257-project-member] Refreshing trait associations for resource provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a, traits: COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=63515) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 803.554067] env[63515]: DEBUG oslo_concurrency.lockutils [None req-b257f20f-3555-4579-b6f5-3780aaf93d24 tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 803.576711] env[63515]: DEBUG nova.compute.manager [req-e11a9d1e-ba6d-4fd8-8082-eee79f2f4061 req-5b3eeeac-ef38-4ed9-9459-a5435215427a service nova] [instance: ed550b10-d58f-45b8-b766-198f431c3788] Received event network-vif-deleted-e24932f9-9063-4a16-8498-950bdfeeaa84 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 803.579221] env[63515]: INFO nova.compute.manager [req-e11a9d1e-ba6d-4fd8-8082-eee79f2f4061 req-5b3eeeac-ef38-4ed9-9459-a5435215427a service nova] [instance: ed550b10-d58f-45b8-b766-198f431c3788] Neutron deleted interface e24932f9-9063-4a16-8498-950bdfeeaa84; detaching it from the instance and deleting it from the info cache [ 803.579221] env[63515]: DEBUG nova.network.neutron [req-e11a9d1e-ba6d-4fd8-8082-eee79f2f4061 req-5b3eeeac-ef38-4ed9-9459-a5435215427a service nova] [instance: ed550b10-d58f-45b8-b766-198f431c3788] Updating instance_info_cache with network_info: [{"id": "ad9dda29-9326-472b-8c82-294a33a7faf0", "address": "fa:16:3e:26:e3:72", "network": {"id": "88dfb29d-41e0-4df7-8f70-e52cda05620c", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1355578400-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.251", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7d4ada9ec35f42b19c6480a9101d21a4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "357d2811-e990-4985-9f9e-b158d10d3699", "external-id": "nsx-vlan-transportzone-641", "segmentation_id": 641, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapad9dda29-93", "ovs_interfaceid": "ad9dda29-9326-472b-8c82-294a33a7faf0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 803.636224] env[63515]: INFO nova.compute.manager [None req-66c3e202-5a3e-433e-b543-4bfb02c8ad30 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775-project-member] [instance: 4e453127-1f3e-40ea-819f-6678479826c8] Took 44.03 seconds to build instance. [ 803.696017] env[63515]: DEBUG oslo_vmware.api [None req-95950a02-20c9-4a3f-9ea8-1f7f19725cc1 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Task: {'id': task-1111279, 'name': ReconfigVM_Task, 'duration_secs': 0.609673} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 803.696017] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-95950a02-20c9-4a3f-9ea8-1f7f19725cc1 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: 4ef8066a-b68c-457b-9964-b1c34bab0fc3] Reconfigured VM instance instance-0000003e to attach disk [datastore2] 4ef8066a-b68c-457b-9964-b1c34bab0fc3/4ef8066a-b68c-457b-9964-b1c34bab0fc3.vmdk or device None with type sparse {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 803.696017] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6b957c19-1713-48be-9484-e73367b2f640 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.702152] env[63515]: DEBUG oslo_vmware.api [None req-95950a02-20c9-4a3f-9ea8-1f7f19725cc1 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Waiting for the task: (returnval){ [ 803.702152] env[63515]: value = "task-1111281" [ 803.702152] env[63515]: _type = "Task" [ 803.702152] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 803.716127] env[63515]: DEBUG oslo_vmware.api [None req-95950a02-20c9-4a3f-9ea8-1f7f19725cc1 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Task: {'id': task-1111281, 'name': Rename_Task} progress is 5%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 803.779578] env[63515]: DEBUG oslo_vmware.api [None req-4f34aec7-5320-4ef9-b5f8-71a5209ff0fa tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Task: {'id': task-1111280, 'name': ReconfigVM_Task, 'duration_secs': 0.157095} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 803.782903] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-4f34aec7-5320-4ef9-b5f8-71a5209ff0fa tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] [instance: 63ce797d-7180-4209-ac2c-81978bf7607a] Powering on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 803.783505] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-506827e5-05df-49f2-8d51-1c419a1785c5 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.794969] env[63515]: DEBUG oslo_vmware.api [None req-4f34aec7-5320-4ef9-b5f8-71a5209ff0fa tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Waiting for the task: (returnval){ [ 803.794969] env[63515]: value = "task-1111282" [ 803.794969] env[63515]: _type = "Task" [ 803.794969] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 803.803706] env[63515]: DEBUG oslo_vmware.api [None req-4f34aec7-5320-4ef9-b5f8-71a5209ff0fa tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Task: {'id': task-1111282, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 803.878310] env[63515]: DEBUG nova.compute.manager [None req-41b4a543-33b4-43d9-863c-4344bf9d0a46 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: d926c699-a64a-4942-9ef4-f0166414661d] Start spawning the instance on the hypervisor. {{(pid=63515) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 803.882728] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba13d629-0632-4a72-9998-8db38bfedc7d {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.890932] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f494b7ff-ef0c-4a61-a31c-091465908d6d {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.925599] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c15812b-8ff4-4886-9902-a52cd580c2b5 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.933487] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bee0b292-ca11-446e-9dbd-bf737810e3d9 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.940487] env[63515]: DEBUG nova.virt.hardware [None req-41b4a543-33b4-43d9-863c-4344bf9d0a46 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T02:52:14Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T02:51:56Z,direct_url=,disk_format='vmdk',id=8a120570-cb06-4099-b262-554ca0ad15c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b569255cc43e42e7ae3f2b4ad37c6ef4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T02:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 803.940742] env[63515]: DEBUG nova.virt.hardware [None req-41b4a543-33b4-43d9-863c-4344bf9d0a46 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Flavor limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 803.940908] env[63515]: DEBUG nova.virt.hardware [None req-41b4a543-33b4-43d9-863c-4344bf9d0a46 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Image limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 803.941298] env[63515]: DEBUG nova.virt.hardware [None req-41b4a543-33b4-43d9-863c-4344bf9d0a46 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Flavor pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 803.941501] env[63515]: DEBUG nova.virt.hardware [None req-41b4a543-33b4-43d9-863c-4344bf9d0a46 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Image pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 803.941697] env[63515]: DEBUG nova.virt.hardware [None req-41b4a543-33b4-43d9-863c-4344bf9d0a46 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 803.941879] env[63515]: DEBUG nova.virt.hardware [None req-41b4a543-33b4-43d9-863c-4344bf9d0a46 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 803.942202] env[63515]: DEBUG nova.virt.hardware [None req-41b4a543-33b4-43d9-863c-4344bf9d0a46 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 803.942438] env[63515]: DEBUG nova.virt.hardware [None req-41b4a543-33b4-43d9-863c-4344bf9d0a46 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Got 1 possible topologies {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 803.942622] env[63515]: DEBUG nova.virt.hardware [None req-41b4a543-33b4-43d9-863c-4344bf9d0a46 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 803.942801] env[63515]: DEBUG nova.virt.hardware [None req-41b4a543-33b4-43d9-863c-4344bf9d0a46 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 803.944265] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c718121e-815f-4b8c-8858-fd2d653ee1be {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.959990] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b313cd05-47b5-4860-9596-ee7333b493b3 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.964241] env[63515]: DEBUG nova.compute.provider_tree [None req-65f1d514-d43e-4354-a5f1-c564a5380800 tempest-ServersAaction247Test-2019176257 tempest-ServersAaction247Test-2019176257-project-member] Updating inventory in ProviderTree for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 804.084430] env[63515]: DEBUG oslo_concurrency.lockutils [req-e11a9d1e-ba6d-4fd8-8082-eee79f2f4061 req-5b3eeeac-ef38-4ed9-9459-a5435215427a service nova] Acquiring lock "ed550b10-d58f-45b8-b766-198f431c3788" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 804.084644] env[63515]: DEBUG oslo_concurrency.lockutils [req-e11a9d1e-ba6d-4fd8-8082-eee79f2f4061 req-5b3eeeac-ef38-4ed9-9459-a5435215427a service nova] Acquired lock "ed550b10-d58f-45b8-b766-198f431c3788" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 804.085641] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b754603b-941b-48c7-a30b-6d18c6057e92 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.103530] env[63515]: DEBUG oslo_concurrency.lockutils [req-e11a9d1e-ba6d-4fd8-8082-eee79f2f4061 req-5b3eeeac-ef38-4ed9-9459-a5435215427a service nova] Releasing lock "ed550b10-d58f-45b8-b766-198f431c3788" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 804.103966] env[63515]: WARNING nova.compute.manager [req-e11a9d1e-ba6d-4fd8-8082-eee79f2f4061 req-5b3eeeac-ef38-4ed9-9459-a5435215427a service nova] [instance: ed550b10-d58f-45b8-b766-198f431c3788] Detach interface failed, port_id=e24932f9-9063-4a16-8498-950bdfeeaa84, reason: No device with interface-id e24932f9-9063-4a16-8498-950bdfeeaa84 exists on VM: nova.exception.NotFound: No device with interface-id e24932f9-9063-4a16-8498-950bdfeeaa84 exists on VM [ 804.141514] env[63515]: DEBUG oslo_concurrency.lockutils [None req-66c3e202-5a3e-433e-b543-4bfb02c8ad30 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775-project-member] Lock "4e453127-1f3e-40ea-819f-6678479826c8" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 90.694s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 804.212759] env[63515]: DEBUG oslo_vmware.api [None req-95950a02-20c9-4a3f-9ea8-1f7f19725cc1 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Task: {'id': task-1111281, 'name': Rename_Task, 'duration_secs': 0.215228} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 804.213173] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-95950a02-20c9-4a3f-9ea8-1f7f19725cc1 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: 4ef8066a-b68c-457b-9964-b1c34bab0fc3] Powering on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 804.213456] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5e18bd28-2bb8-4772-9ecf-a516060fe262 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.220749] env[63515]: DEBUG oslo_vmware.api [None req-95950a02-20c9-4a3f-9ea8-1f7f19725cc1 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Waiting for the task: (returnval){ [ 804.220749] env[63515]: value = "task-1111283" [ 804.220749] env[63515]: _type = "Task" [ 804.220749] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 804.230366] env[63515]: DEBUG oslo_vmware.api [None req-95950a02-20c9-4a3f-9ea8-1f7f19725cc1 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Task: {'id': task-1111283, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 804.293815] env[63515]: DEBUG oslo_concurrency.lockutils [None req-8d4c1f89-f73c-481b-8b79-61983aeba07b tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Acquiring lock "refresh_cache-ed550b10-d58f-45b8-b766-198f431c3788" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 804.294021] env[63515]: DEBUG oslo_concurrency.lockutils [None req-8d4c1f89-f73c-481b-8b79-61983aeba07b tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Acquired lock "refresh_cache-ed550b10-d58f-45b8-b766-198f431c3788" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 804.294209] env[63515]: DEBUG nova.network.neutron [None req-8d4c1f89-f73c-481b-8b79-61983aeba07b tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: ed550b10-d58f-45b8-b766-198f431c3788] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 804.306323] env[63515]: DEBUG oslo_vmware.api [None req-4f34aec7-5320-4ef9-b5f8-71a5209ff0fa tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Task: {'id': task-1111282, 'name': PowerOnVM_Task, 'duration_secs': 0.438668} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 804.306689] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-4f34aec7-5320-4ef9-b5f8-71a5209ff0fa tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] [instance: 63ce797d-7180-4209-ac2c-81978bf7607a] Powered on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 804.309958] env[63515]: DEBUG nova.compute.manager [None req-4f34aec7-5320-4ef9-b5f8-71a5209ff0fa tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] [instance: 63ce797d-7180-4209-ac2c-81978bf7607a] Checking state {{(pid=63515) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 804.311304] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4035175-d92a-42f5-b1ce-7ffa75c76008 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.563706] env[63515]: DEBUG nova.scheduler.client.report [None req-65f1d514-d43e-4354-a5f1-c564a5380800 tempest-ServersAaction247Test-2019176257 tempest-ServersAaction247Test-2019176257-project-member] Updated inventory for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a with generation 81 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 804.564059] env[63515]: DEBUG nova.compute.provider_tree [None req-65f1d514-d43e-4354-a5f1-c564a5380800 tempest-ServersAaction247Test-2019176257 tempest-ServersAaction247Test-2019176257-project-member] Updating resource provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a generation from 81 to 82 during operation: update_inventory {{(pid=63515) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 804.564293] env[63515]: DEBUG nova.compute.provider_tree [None req-65f1d514-d43e-4354-a5f1-c564a5380800 tempest-ServersAaction247Test-2019176257 tempest-ServersAaction247Test-2019176257-project-member] Updating inventory in ProviderTree for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 804.733027] env[63515]: DEBUG oslo_vmware.api [None req-95950a02-20c9-4a3f-9ea8-1f7f19725cc1 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Task: {'id': task-1111283, 'name': PowerOnVM_Task} progress is 89%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 804.766959] env[63515]: DEBUG oslo_vmware.rw_handles [None req-67e41316-801d-46e9-947d-28c19f5a3017 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52db4223-71af-a208-688f-066abc855425/disk-0.vmdk. {{(pid=63515) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 804.767927] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2484ebd-0336-4f49-8bc4-1430a48ca02b {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.774657] env[63515]: DEBUG oslo_vmware.rw_handles [None req-67e41316-801d-46e9-947d-28c19f5a3017 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52db4223-71af-a208-688f-066abc855425/disk-0.vmdk is in state: ready. {{(pid=63515) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 804.774836] env[63515]: ERROR oslo_vmware.rw_handles [None req-67e41316-801d-46e9-947d-28c19f5a3017 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52db4223-71af-a208-688f-066abc855425/disk-0.vmdk due to incomplete transfer. [ 804.775103] env[63515]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-751c704b-2481-4b4d-aa1f-bba1225a6c17 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.781826] env[63515]: DEBUG oslo_vmware.rw_handles [None req-67e41316-801d-46e9-947d-28c19f5a3017 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52db4223-71af-a208-688f-066abc855425/disk-0.vmdk. {{(pid=63515) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 804.781973] env[63515]: DEBUG nova.virt.vmwareapi.images [None req-67e41316-801d-46e9-947d-28c19f5a3017 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: 84dbf321-2c70-4e08-b430-cb5a06fc6829] Uploaded image caa06f85-6406-4076-afa2-f2a71b73190c to the Glance image server {{(pid=63515) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 804.786030] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-67e41316-801d-46e9-947d-28c19f5a3017 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: 84dbf321-2c70-4e08-b430-cb5a06fc6829] Destroying the VM {{(pid=63515) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 804.786030] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-f80f8698-8c4b-448a-92a6-de412a709fd5 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.792160] env[63515]: DEBUG oslo_vmware.api [None req-67e41316-801d-46e9-947d-28c19f5a3017 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Waiting for the task: (returnval){ [ 804.792160] env[63515]: value = "task-1111284" [ 804.792160] env[63515]: _type = "Task" [ 804.792160] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 804.798635] env[63515]: DEBUG oslo_vmware.api [None req-67e41316-801d-46e9-947d-28c19f5a3017 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Task: {'id': task-1111284, 'name': Destroy_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.073096] env[63515]: DEBUG oslo_concurrency.lockutils [None req-65f1d514-d43e-4354-a5f1-c564a5380800 tempest-ServersAaction247Test-2019176257 tempest-ServersAaction247Test-2019176257-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 3.222s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 805.075420] env[63515]: DEBUG oslo_concurrency.lockutils [None req-db31b1fd-34e1-40f4-b7c8-92b451eb1400 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 33.875s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 805.075701] env[63515]: DEBUG nova.objects.instance [None req-db31b1fd-34e1-40f4-b7c8-92b451eb1400 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] [instance: 8eb17506-ff93-4d25-b9af-ec5886569e65] Trying to apply a migration context that does not seem to be set for this instance {{(pid=63515) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 805.087235] env[63515]: DEBUG oslo_concurrency.lockutils [None req-811d8802-f522-4005-9c27-973badf5075e tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Acquiring lock "ed550b10-d58f-45b8-b766-198f431c3788" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 805.087235] env[63515]: DEBUG oslo_concurrency.lockutils [None req-811d8802-f522-4005-9c27-973badf5075e tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Lock "ed550b10-d58f-45b8-b766-198f431c3788" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 805.087235] env[63515]: DEBUG oslo_concurrency.lockutils [None req-811d8802-f522-4005-9c27-973badf5075e tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Acquiring lock "ed550b10-d58f-45b8-b766-198f431c3788-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 805.087235] env[63515]: DEBUG oslo_concurrency.lockutils [None req-811d8802-f522-4005-9c27-973badf5075e tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Lock "ed550b10-d58f-45b8-b766-198f431c3788-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 805.087235] env[63515]: DEBUG oslo_concurrency.lockutils [None req-811d8802-f522-4005-9c27-973badf5075e tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Lock "ed550b10-d58f-45b8-b766-198f431c3788-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 805.091141] env[63515]: INFO nova.compute.manager [None req-811d8802-f522-4005-9c27-973badf5075e tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: ed550b10-d58f-45b8-b766-198f431c3788] Terminating instance [ 805.095381] env[63515]: DEBUG nova.compute.manager [None req-811d8802-f522-4005-9c27-973badf5075e tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: ed550b10-d58f-45b8-b766-198f431c3788] Start destroying the instance on the hypervisor. {{(pid=63515) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 805.095632] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-811d8802-f522-4005-9c27-973badf5075e tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: ed550b10-d58f-45b8-b766-198f431c3788] Destroying instance {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 805.096815] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b2614b3-39a5-4e40-869a-91558ff84a73 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.112430] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-811d8802-f522-4005-9c27-973badf5075e tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: ed550b10-d58f-45b8-b766-198f431c3788] Powering off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 805.112729] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-96462108-104c-48be-8849-a238d5cb5c63 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.117275] env[63515]: INFO nova.scheduler.client.report [None req-65f1d514-d43e-4354-a5f1-c564a5380800 tempest-ServersAaction247Test-2019176257 tempest-ServersAaction247Test-2019176257-project-member] Deleted allocations for instance 5d50fbd6-a7fb-422c-9dd7-df7140fd0aae [ 805.124164] env[63515]: DEBUG oslo_vmware.api [None req-811d8802-f522-4005-9c27-973badf5075e tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Waiting for the task: (returnval){ [ 805.124164] env[63515]: value = "task-1111285" [ 805.124164] env[63515]: _type = "Task" [ 805.124164] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 805.139371] env[63515]: DEBUG oslo_vmware.api [None req-811d8802-f522-4005-9c27-973badf5075e tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Task: {'id': task-1111285, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.233320] env[63515]: DEBUG oslo_vmware.api [None req-95950a02-20c9-4a3f-9ea8-1f7f19725cc1 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Task: {'id': task-1111283, 'name': PowerOnVM_Task, 'duration_secs': 0.547613} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 805.233320] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-95950a02-20c9-4a3f-9ea8-1f7f19725cc1 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: 4ef8066a-b68c-457b-9964-b1c34bab0fc3] Powered on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 805.234478] env[63515]: INFO nova.compute.manager [None req-95950a02-20c9-4a3f-9ea8-1f7f19725cc1 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: 4ef8066a-b68c-457b-9964-b1c34bab0fc3] Took 10.03 seconds to spawn the instance on the hypervisor. [ 805.235037] env[63515]: DEBUG nova.compute.manager [None req-95950a02-20c9-4a3f-9ea8-1f7f19725cc1 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: 4ef8066a-b68c-457b-9964-b1c34bab0fc3] Checking state {{(pid=63515) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 805.236336] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72d84be3-9f53-42ee-b512-015d1d9e609f {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.302963] env[63515]: DEBUG oslo_vmware.api [None req-67e41316-801d-46e9-947d-28c19f5a3017 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Task: {'id': task-1111284, 'name': Destroy_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.633192] env[63515]: DEBUG oslo_concurrency.lockutils [None req-65f1d514-d43e-4354-a5f1-c564a5380800 tempest-ServersAaction247Test-2019176257 tempest-ServersAaction247Test-2019176257-project-member] Lock "5d50fbd6-a7fb-422c-9dd7-df7140fd0aae" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 39.086s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 805.637186] env[63515]: DEBUG nova.network.neutron [None req-8d4c1f89-f73c-481b-8b79-61983aeba07b tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: ed550b10-d58f-45b8-b766-198f431c3788] Updating instance_info_cache with network_info: [{"id": "ad9dda29-9326-472b-8c82-294a33a7faf0", "address": "fa:16:3e:26:e3:72", "network": {"id": "88dfb29d-41e0-4df7-8f70-e52cda05620c", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1355578400-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.251", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7d4ada9ec35f42b19c6480a9101d21a4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "357d2811-e990-4985-9f9e-b158d10d3699", "external-id": "nsx-vlan-transportzone-641", "segmentation_id": 641, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapad9dda29-93", "ovs_interfaceid": "ad9dda29-9326-472b-8c82-294a33a7faf0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 805.647109] env[63515]: DEBUG oslo_vmware.api [None req-811d8802-f522-4005-9c27-973badf5075e tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Task: {'id': task-1111285, 'name': PowerOffVM_Task, 'duration_secs': 0.415993} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 805.647109] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-811d8802-f522-4005-9c27-973badf5075e tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: ed550b10-d58f-45b8-b766-198f431c3788] Powered off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 805.647109] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-811d8802-f522-4005-9c27-973badf5075e tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: ed550b10-d58f-45b8-b766-198f431c3788] Unregistering the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 805.647671] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-5ce1bc3e-be03-4cf5-aab8-bdabfd4f242a {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.659405] env[63515]: DEBUG nova.network.neutron [None req-41b4a543-33b4-43d9-863c-4344bf9d0a46 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: d926c699-a64a-4942-9ef4-f0166414661d] Successfully updated port: fdec91d2-43e8-4d40-9aa9-885a82fd9b42 {{(pid=63515) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 805.760795] env[63515]: INFO nova.compute.manager [None req-95950a02-20c9-4a3f-9ea8-1f7f19725cc1 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: 4ef8066a-b68c-457b-9964-b1c34bab0fc3] Took 43.28 seconds to build instance. [ 805.804732] env[63515]: DEBUG oslo_vmware.api [None req-67e41316-801d-46e9-947d-28c19f5a3017 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Task: {'id': task-1111284, 'name': Destroy_Task} progress is 100%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 806.078387] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-811d8802-f522-4005-9c27-973badf5075e tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: ed550b10-d58f-45b8-b766-198f431c3788] Unregistered the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 806.078674] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-811d8802-f522-4005-9c27-973badf5075e tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: ed550b10-d58f-45b8-b766-198f431c3788] Deleting contents of the VM from datastore datastore2 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 806.080575] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-811d8802-f522-4005-9c27-973badf5075e tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Deleting the datastore file [datastore2] ed550b10-d58f-45b8-b766-198f431c3788 {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 806.080896] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-8ab917e8-90fa-42ac-a381-00e8c64412a8 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.088612] env[63515]: DEBUG oslo_vmware.api [None req-811d8802-f522-4005-9c27-973badf5075e tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Waiting for the task: (returnval){ [ 806.088612] env[63515]: value = "task-1111287" [ 806.088612] env[63515]: _type = "Task" [ 806.088612] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 806.089870] env[63515]: DEBUG oslo_concurrency.lockutils [None req-db31b1fd-34e1-40f4-b7c8-92b451eb1400 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.015s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 806.094678] env[63515]: DEBUG oslo_concurrency.lockutils [None req-6d0510cf-97fb-4582-ac56-e14cb4885964 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 31.406s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 806.099138] env[63515]: DEBUG nova.objects.instance [None req-6d0510cf-97fb-4582-ac56-e14cb4885964 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Lazy-loading 'resources' on Instance uuid 38d6d030-06b0-4185-904d-44a038b3a752 {{(pid=63515) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 806.108013] env[63515]: DEBUG oslo_vmware.api [None req-811d8802-f522-4005-9c27-973badf5075e tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Task: {'id': task-1111287, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 806.144976] env[63515]: DEBUG oslo_concurrency.lockutils [None req-8d4c1f89-f73c-481b-8b79-61983aeba07b tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Releasing lock "refresh_cache-ed550b10-d58f-45b8-b766-198f431c3788" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 806.163969] env[63515]: DEBUG oslo_concurrency.lockutils [None req-41b4a543-33b4-43d9-863c-4344bf9d0a46 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Acquiring lock "refresh_cache-d926c699-a64a-4942-9ef4-f0166414661d" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 806.163969] env[63515]: DEBUG oslo_concurrency.lockutils [None req-41b4a543-33b4-43d9-863c-4344bf9d0a46 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Acquired lock "refresh_cache-d926c699-a64a-4942-9ef4-f0166414661d" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 806.163969] env[63515]: DEBUG nova.network.neutron [None req-41b4a543-33b4-43d9-863c-4344bf9d0a46 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: d926c699-a64a-4942-9ef4-f0166414661d] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 806.264775] env[63515]: DEBUG oslo_concurrency.lockutils [None req-95950a02-20c9-4a3f-9ea8-1f7f19725cc1 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Lock "4ef8066a-b68c-457b-9964-b1c34bab0fc3" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 86.695s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 806.299999] env[63515]: DEBUG nova.compute.manager [req-cd8c5911-6cac-4ae2-8bcd-aaa354051453 req-b2bc7daf-61ae-4cee-9f42-a05188dc5cc6 service nova] [instance: d926c699-a64a-4942-9ef4-f0166414661d] Received event network-vif-plugged-fdec91d2-43e8-4d40-9aa9-885a82fd9b42 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 806.300419] env[63515]: DEBUG oslo_concurrency.lockutils [req-cd8c5911-6cac-4ae2-8bcd-aaa354051453 req-b2bc7daf-61ae-4cee-9f42-a05188dc5cc6 service nova] Acquiring lock "d926c699-a64a-4942-9ef4-f0166414661d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 806.300759] env[63515]: DEBUG oslo_concurrency.lockutils [req-cd8c5911-6cac-4ae2-8bcd-aaa354051453 req-b2bc7daf-61ae-4cee-9f42-a05188dc5cc6 service nova] Lock "d926c699-a64a-4942-9ef4-f0166414661d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 806.301074] env[63515]: DEBUG oslo_concurrency.lockutils [req-cd8c5911-6cac-4ae2-8bcd-aaa354051453 req-b2bc7daf-61ae-4cee-9f42-a05188dc5cc6 service nova] Lock "d926c699-a64a-4942-9ef4-f0166414661d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 806.301384] env[63515]: DEBUG nova.compute.manager [req-cd8c5911-6cac-4ae2-8bcd-aaa354051453 req-b2bc7daf-61ae-4cee-9f42-a05188dc5cc6 service nova] [instance: d926c699-a64a-4942-9ef4-f0166414661d] No waiting events found dispatching network-vif-plugged-fdec91d2-43e8-4d40-9aa9-885a82fd9b42 {{(pid=63515) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 806.301674] env[63515]: WARNING nova.compute.manager [req-cd8c5911-6cac-4ae2-8bcd-aaa354051453 req-b2bc7daf-61ae-4cee-9f42-a05188dc5cc6 service nova] [instance: d926c699-a64a-4942-9ef4-f0166414661d] Received unexpected event network-vif-plugged-fdec91d2-43e8-4d40-9aa9-885a82fd9b42 for instance with vm_state building and task_state spawning. [ 806.315230] env[63515]: DEBUG oslo_vmware.api [None req-67e41316-801d-46e9-947d-28c19f5a3017 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Task: {'id': task-1111284, 'name': Destroy_Task, 'duration_secs': 1.015225} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 806.315230] env[63515]: INFO nova.virt.vmwareapi.vm_util [None req-67e41316-801d-46e9-947d-28c19f5a3017 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: 84dbf321-2c70-4e08-b430-cb5a06fc6829] Destroyed the VM [ 806.315230] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-67e41316-801d-46e9-947d-28c19f5a3017 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: 84dbf321-2c70-4e08-b430-cb5a06fc6829] Deleting Snapshot of the VM instance {{(pid=63515) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 806.315230] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-e9878614-664f-442b-b2c6-1ca6281e6ccf {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.321199] env[63515]: DEBUG oslo_vmware.api [None req-67e41316-801d-46e9-947d-28c19f5a3017 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Waiting for the task: (returnval){ [ 806.321199] env[63515]: value = "task-1111288" [ 806.321199] env[63515]: _type = "Task" [ 806.321199] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 806.331294] env[63515]: DEBUG nova.compute.manager [req-62338844-6c45-412c-a658-a7d8e42d81aa req-ebf6076f-8633-4a5b-a7b2-63eac44f3933 service nova] [instance: 4e453127-1f3e-40ea-819f-6678479826c8] Received event network-changed-3d5ad33f-8649-42c8-b899-ca20c9a2636e {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 806.331882] env[63515]: DEBUG nova.compute.manager [req-62338844-6c45-412c-a658-a7d8e42d81aa req-ebf6076f-8633-4a5b-a7b2-63eac44f3933 service nova] [instance: 4e453127-1f3e-40ea-819f-6678479826c8] Refreshing instance network info cache due to event network-changed-3d5ad33f-8649-42c8-b899-ca20c9a2636e. {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 806.332511] env[63515]: DEBUG oslo_concurrency.lockutils [req-62338844-6c45-412c-a658-a7d8e42d81aa req-ebf6076f-8633-4a5b-a7b2-63eac44f3933 service nova] Acquiring lock "refresh_cache-4e453127-1f3e-40ea-819f-6678479826c8" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 806.332822] env[63515]: DEBUG oslo_concurrency.lockutils [req-62338844-6c45-412c-a658-a7d8e42d81aa req-ebf6076f-8633-4a5b-a7b2-63eac44f3933 service nova] Acquired lock "refresh_cache-4e453127-1f3e-40ea-819f-6678479826c8" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 806.333134] env[63515]: DEBUG nova.network.neutron [req-62338844-6c45-412c-a658-a7d8e42d81aa req-ebf6076f-8633-4a5b-a7b2-63eac44f3933 service nova] [instance: 4e453127-1f3e-40ea-819f-6678479826c8] Refreshing network info cache for port 3d5ad33f-8649-42c8-b899-ca20c9a2636e {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 806.342651] env[63515]: DEBUG oslo_vmware.api [None req-67e41316-801d-46e9-947d-28c19f5a3017 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Task: {'id': task-1111288, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 806.603183] env[63515]: DEBUG oslo_vmware.api [None req-811d8802-f522-4005-9c27-973badf5075e tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Task: {'id': task-1111287, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.256047} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 806.603915] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-811d8802-f522-4005-9c27-973badf5075e tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Deleted the datastore file {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 806.603983] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-811d8802-f522-4005-9c27-973badf5075e tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: ed550b10-d58f-45b8-b766-198f431c3788] Deleted contents of the VM from datastore datastore2 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 806.604288] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-811d8802-f522-4005-9c27-973badf5075e tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: ed550b10-d58f-45b8-b766-198f431c3788] Instance destroyed {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 806.604601] env[63515]: INFO nova.compute.manager [None req-811d8802-f522-4005-9c27-973badf5075e tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: ed550b10-d58f-45b8-b766-198f431c3788] Took 1.51 seconds to destroy the instance on the hypervisor. [ 806.604951] env[63515]: DEBUG oslo.service.loopingcall [None req-811d8802-f522-4005-9c27-973badf5075e tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 806.605238] env[63515]: DEBUG nova.compute.manager [-] [instance: ed550b10-d58f-45b8-b766-198f431c3788] Deallocating network for instance {{(pid=63515) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 806.605503] env[63515]: DEBUG nova.network.neutron [-] [instance: ed550b10-d58f-45b8-b766-198f431c3788] deallocate_for_instance() {{(pid=63515) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 806.647274] env[63515]: DEBUG oslo_concurrency.lockutils [None req-8d4c1f89-f73c-481b-8b79-61983aeba07b tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Lock "interface-ed550b10-d58f-45b8-b766-198f431c3788-e24932f9-9063-4a16-8498-950bdfeeaa84" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 17.618s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 806.773114] env[63515]: DEBUG nova.network.neutron [None req-41b4a543-33b4-43d9-863c-4344bf9d0a46 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: d926c699-a64a-4942-9ef4-f0166414661d] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 806.834215] env[63515]: DEBUG oslo_vmware.api [None req-67e41316-801d-46e9-947d-28c19f5a3017 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Task: {'id': task-1111288, 'name': RemoveSnapshot_Task} progress is 100%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 806.972667] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39f55c71-f49a-4c73-9f31-c0c3a9d177d0 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.978952] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-efc88c70-55a2-4f31-b706-06a28a8b1f00 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.016428] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86bbc6f5-95fe-40d7-8d92-650e1c7f5517 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.024506] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2b4876d-b8ad-4343-afa1-999429cac517 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.044218] env[63515]: DEBUG nova.compute.provider_tree [None req-6d0510cf-97fb-4582-ac56-e14cb4885964 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Updating inventory in ProviderTree for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 807.242593] env[63515]: DEBUG nova.network.neutron [None req-41b4a543-33b4-43d9-863c-4344bf9d0a46 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: d926c699-a64a-4942-9ef4-f0166414661d] Updating instance_info_cache with network_info: [{"id": "fdec91d2-43e8-4d40-9aa9-885a82fd9b42", "address": "fa:16:3e:5c:b7:75", "network": {"id": "15fdaabe-6793-400f-9737-2cdff8f07238", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1175204866-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f9c98e9f6020475490aaa7e76d907ab7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7d377d75-3add-4a15-8691-74b2eb010924", "external-id": "nsx-vlan-transportzone-71", "segmentation_id": 71, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfdec91d2-43", "ovs_interfaceid": "fdec91d2-43e8-4d40-9aa9-885a82fd9b42", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 807.268860] env[63515]: DEBUG nova.network.neutron [req-62338844-6c45-412c-a658-a7d8e42d81aa req-ebf6076f-8633-4a5b-a7b2-63eac44f3933 service nova] [instance: 4e453127-1f3e-40ea-819f-6678479826c8] Updated VIF entry in instance network info cache for port 3d5ad33f-8649-42c8-b899-ca20c9a2636e. {{(pid=63515) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 807.269279] env[63515]: DEBUG nova.network.neutron [req-62338844-6c45-412c-a658-a7d8e42d81aa req-ebf6076f-8633-4a5b-a7b2-63eac44f3933 service nova] [instance: 4e453127-1f3e-40ea-819f-6678479826c8] Updating instance_info_cache with network_info: [{"id": "3d5ad33f-8649-42c8-b899-ca20c9a2636e", "address": "fa:16:3e:c2:77:9b", "network": {"id": "5f57f1e7-21bb-425f-aab4-7d69b7ec94f7", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationNegativeTestJSON-177649896-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.242", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "60d9dcd0f30e48e7965ecfc1a72f5375", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3e05affa-2640-435e-a124-0ee8a6ab1152", "external-id": "nsx-vlan-transportzone-839", "segmentation_id": 839, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3d5ad33f-86", "ovs_interfaceid": "3d5ad33f-8649-42c8-b899-ca20c9a2636e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 807.336682] env[63515]: DEBUG oslo_vmware.api [None req-67e41316-801d-46e9-947d-28c19f5a3017 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Task: {'id': task-1111288, 'name': RemoveSnapshot_Task, 'duration_secs': 0.875769} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 807.337041] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-67e41316-801d-46e9-947d-28c19f5a3017 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: 84dbf321-2c70-4e08-b430-cb5a06fc6829] Deleted Snapshot of the VM instance {{(pid=63515) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 807.337410] env[63515]: INFO nova.compute.manager [None req-67e41316-801d-46e9-947d-28c19f5a3017 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: 84dbf321-2c70-4e08-b430-cb5a06fc6829] Took 19.04 seconds to snapshot the instance on the hypervisor. [ 807.564650] env[63515]: DEBUG oslo_concurrency.lockutils [None req-1f202073-ad2a-4c82-aa50-bf79e5416807 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Acquiring lock "84dbf321-2c70-4e08-b430-cb5a06fc6829" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 807.564896] env[63515]: DEBUG oslo_concurrency.lockutils [None req-1f202073-ad2a-4c82-aa50-bf79e5416807 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Lock "84dbf321-2c70-4e08-b430-cb5a06fc6829" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 807.565131] env[63515]: DEBUG oslo_concurrency.lockutils [None req-1f202073-ad2a-4c82-aa50-bf79e5416807 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Acquiring lock "84dbf321-2c70-4e08-b430-cb5a06fc6829-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 807.565367] env[63515]: DEBUG oslo_concurrency.lockutils [None req-1f202073-ad2a-4c82-aa50-bf79e5416807 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Lock "84dbf321-2c70-4e08-b430-cb5a06fc6829-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 807.568240] env[63515]: DEBUG oslo_concurrency.lockutils [None req-1f202073-ad2a-4c82-aa50-bf79e5416807 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Lock "84dbf321-2c70-4e08-b430-cb5a06fc6829-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 807.570072] env[63515]: INFO nova.compute.manager [None req-1f202073-ad2a-4c82-aa50-bf79e5416807 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: 84dbf321-2c70-4e08-b430-cb5a06fc6829] Terminating instance [ 807.572894] env[63515]: DEBUG nova.compute.manager [None req-1f202073-ad2a-4c82-aa50-bf79e5416807 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: 84dbf321-2c70-4e08-b430-cb5a06fc6829] Start destroying the instance on the hypervisor. {{(pid=63515) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 807.574453] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-1f202073-ad2a-4c82-aa50-bf79e5416807 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: 84dbf321-2c70-4e08-b430-cb5a06fc6829] Destroying instance {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 807.575182] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b8b27cc-a0a5-4e7d-aae6-68c96cde16c6 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.580810] env[63515]: DEBUG nova.scheduler.client.report [None req-6d0510cf-97fb-4582-ac56-e14cb4885964 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Updated inventory for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a with generation 82 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 807.582551] env[63515]: DEBUG nova.compute.provider_tree [None req-6d0510cf-97fb-4582-ac56-e14cb4885964 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Updating resource provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a generation from 82 to 83 during operation: update_inventory {{(pid=63515) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 807.582551] env[63515]: DEBUG nova.compute.provider_tree [None req-6d0510cf-97fb-4582-ac56-e14cb4885964 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Updating inventory in ProviderTree for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 807.593954] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-1f202073-ad2a-4c82-aa50-bf79e5416807 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: 84dbf321-2c70-4e08-b430-cb5a06fc6829] Unregistering the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 807.594201] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9acc3a61-24d4-4c2c-b18b-1d3ead98a0b6 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.669721] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-1f202073-ad2a-4c82-aa50-bf79e5416807 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: 84dbf321-2c70-4e08-b430-cb5a06fc6829] Unregistered the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 807.669968] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-1f202073-ad2a-4c82-aa50-bf79e5416807 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: 84dbf321-2c70-4e08-b430-cb5a06fc6829] Deleting contents of the VM from datastore datastore2 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 807.670172] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-1f202073-ad2a-4c82-aa50-bf79e5416807 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Deleting the datastore file [datastore2] 84dbf321-2c70-4e08-b430-cb5a06fc6829 {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 807.670450] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-6a2a33cf-0779-44ef-ad7d-bdbb7e644998 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.677454] env[63515]: DEBUG oslo_vmware.api [None req-1f202073-ad2a-4c82-aa50-bf79e5416807 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Waiting for the task: (returnval){ [ 807.677454] env[63515]: value = "task-1111290" [ 807.677454] env[63515]: _type = "Task" [ 807.677454] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 807.687771] env[63515]: DEBUG oslo_vmware.api [None req-1f202073-ad2a-4c82-aa50-bf79e5416807 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Task: {'id': task-1111290, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.745453] env[63515]: DEBUG oslo_concurrency.lockutils [None req-41b4a543-33b4-43d9-863c-4344bf9d0a46 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Releasing lock "refresh_cache-d926c699-a64a-4942-9ef4-f0166414661d" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 807.745796] env[63515]: DEBUG nova.compute.manager [None req-41b4a543-33b4-43d9-863c-4344bf9d0a46 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: d926c699-a64a-4942-9ef4-f0166414661d] Instance network_info: |[{"id": "fdec91d2-43e8-4d40-9aa9-885a82fd9b42", "address": "fa:16:3e:5c:b7:75", "network": {"id": "15fdaabe-6793-400f-9737-2cdff8f07238", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1175204866-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f9c98e9f6020475490aaa7e76d907ab7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7d377d75-3add-4a15-8691-74b2eb010924", "external-id": "nsx-vlan-transportzone-71", "segmentation_id": 71, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfdec91d2-43", "ovs_interfaceid": "fdec91d2-43e8-4d40-9aa9-885a82fd9b42", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 807.747042] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-41b4a543-33b4-43d9-863c-4344bf9d0a46 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: d926c699-a64a-4942-9ef4-f0166414661d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:5c:b7:75', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '7d377d75-3add-4a15-8691-74b2eb010924', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'fdec91d2-43e8-4d40-9aa9-885a82fd9b42', 'vif_model': 'vmxnet3'}] {{(pid=63515) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 807.757918] env[63515]: DEBUG oslo.service.loopingcall [None req-41b4a543-33b4-43d9-863c-4344bf9d0a46 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 807.757918] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d926c699-a64a-4942-9ef4-f0166414661d] Creating VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 807.757918] env[63515]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-356d190a-f970-48fa-9477-07719c14e82f {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.774116] env[63515]: DEBUG oslo_concurrency.lockutils [req-62338844-6c45-412c-a658-a7d8e42d81aa req-ebf6076f-8633-4a5b-a7b2-63eac44f3933 service nova] Releasing lock "refresh_cache-4e453127-1f3e-40ea-819f-6678479826c8" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 807.781234] env[63515]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 807.781234] env[63515]: value = "task-1111291" [ 807.781234] env[63515]: _type = "Task" [ 807.781234] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 807.795835] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1111291, 'name': CreateVM_Task} progress is 6%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.842781] env[63515]: DEBUG nova.compute.manager [None req-67e41316-801d-46e9-947d-28c19f5a3017 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: 84dbf321-2c70-4e08-b430-cb5a06fc6829] Instance disappeared during snapshot {{(pid=63515) _snapshot_instance /opt/stack/nova/nova/compute/manager.py:4494}} [ 807.851997] env[63515]: DEBUG oslo_concurrency.lockutils [None req-6974871c-fbdf-4315-ae5e-0f262a5a9f50 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Acquiring lock "4ef8066a-b68c-457b-9964-b1c34bab0fc3" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 807.852885] env[63515]: DEBUG oslo_concurrency.lockutils [None req-6974871c-fbdf-4315-ae5e-0f262a5a9f50 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Lock "4ef8066a-b68c-457b-9964-b1c34bab0fc3" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.001s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 807.858176] env[63515]: DEBUG nova.compute.manager [None req-67e41316-801d-46e9-947d-28c19f5a3017 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Image not found during clean up caa06f85-6406-4076-afa2-f2a71b73190c {{(pid=63515) _snapshot_instance /opt/stack/nova/nova/compute/manager.py:4500}} [ 808.089187] env[63515]: DEBUG oslo_concurrency.lockutils [None req-6d0510cf-97fb-4582-ac56-e14cb4885964 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.995s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 808.094266] env[63515]: DEBUG oslo_concurrency.lockutils [None req-67d44811-1ff4-4a5b-a720-2ef182e4f0da tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 32.548s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 808.094879] env[63515]: DEBUG nova.objects.instance [None req-67d44811-1ff4-4a5b-a720-2ef182e4f0da tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] Lazy-loading 'resources' on Instance uuid 30b8b872-40ca-4297-b98e-a64c83a75483 {{(pid=63515) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 808.116881] env[63515]: INFO nova.scheduler.client.report [None req-6d0510cf-97fb-4582-ac56-e14cb4885964 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Deleted allocations for instance 38d6d030-06b0-4185-904d-44a038b3a752 [ 808.118095] env[63515]: DEBUG nova.network.neutron [-] [instance: ed550b10-d58f-45b8-b766-198f431c3788] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 808.195614] env[63515]: DEBUG oslo_vmware.api [None req-1f202073-ad2a-4c82-aa50-bf79e5416807 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Task: {'id': task-1111290, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.462049} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 808.195993] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-1f202073-ad2a-4c82-aa50-bf79e5416807 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Deleted the datastore file {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 808.196284] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-1f202073-ad2a-4c82-aa50-bf79e5416807 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: 84dbf321-2c70-4e08-b430-cb5a06fc6829] Deleted contents of the VM from datastore datastore2 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 808.196836] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-1f202073-ad2a-4c82-aa50-bf79e5416807 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: 84dbf321-2c70-4e08-b430-cb5a06fc6829] Instance destroyed {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 808.196836] env[63515]: INFO nova.compute.manager [None req-1f202073-ad2a-4c82-aa50-bf79e5416807 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: 84dbf321-2c70-4e08-b430-cb5a06fc6829] Took 0.62 seconds to destroy the instance on the hypervisor. [ 808.197165] env[63515]: DEBUG oslo.service.loopingcall [None req-1f202073-ad2a-4c82-aa50-bf79e5416807 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 808.197449] env[63515]: DEBUG nova.compute.manager [-] [instance: 84dbf321-2c70-4e08-b430-cb5a06fc6829] Deallocating network for instance {{(pid=63515) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 808.197599] env[63515]: DEBUG nova.network.neutron [-] [instance: 84dbf321-2c70-4e08-b430-cb5a06fc6829] deallocate_for_instance() {{(pid=63515) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 808.293437] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1111291, 'name': CreateVM_Task} progress is 99%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.357908] env[63515]: DEBUG nova.compute.utils [None req-6974871c-fbdf-4315-ae5e-0f262a5a9f50 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Using /dev/sd instead of None {{(pid=63515) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 808.622855] env[63515]: INFO nova.compute.manager [-] [instance: ed550b10-d58f-45b8-b766-198f431c3788] Took 2.02 seconds to deallocate network for instance. [ 808.630835] env[63515]: DEBUG oslo_concurrency.lockutils [None req-6d0510cf-97fb-4582-ac56-e14cb4885964 tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Lock "38d6d030-06b0-4185-904d-44a038b3a752" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 36.836s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 808.795359] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1111291, 'name': CreateVM_Task, 'duration_secs': 0.530419} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 808.795359] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d926c699-a64a-4942-9ef4-f0166414661d] Created VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 808.795736] env[63515]: DEBUG oslo_concurrency.lockutils [None req-41b4a543-33b4-43d9-863c-4344bf9d0a46 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 808.795907] env[63515]: DEBUG oslo_concurrency.lockutils [None req-41b4a543-33b4-43d9-863c-4344bf9d0a46 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 808.796248] env[63515]: DEBUG oslo_concurrency.lockutils [None req-41b4a543-33b4-43d9-863c-4344bf9d0a46 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 808.796526] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-79560a32-afed-44a4-9570-2e5fd39659cf {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.803704] env[63515]: DEBUG oslo_vmware.api [None req-41b4a543-33b4-43d9-863c-4344bf9d0a46 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Waiting for the task: (returnval){ [ 808.803704] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]52a46a34-6c56-161d-0fe4-1830e92d4e73" [ 808.803704] env[63515]: _type = "Task" [ 808.803704] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 808.813441] env[63515]: DEBUG oslo_vmware.api [None req-41b4a543-33b4-43d9-863c-4344bf9d0a46 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52a46a34-6c56-161d-0fe4-1830e92d4e73, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.818581] env[63515]: DEBUG oslo_concurrency.lockutils [None req-9dd4330d-d35d-47d5-b468-11f24e31a49a tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Acquiring lock "70df69e5-687b-44fb-b6fc-cdb08e21dda0" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 808.818705] env[63515]: DEBUG oslo_concurrency.lockutils [None req-9dd4330d-d35d-47d5-b468-11f24e31a49a tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Lock "70df69e5-687b-44fb-b6fc-cdb08e21dda0" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 808.860221] env[63515]: DEBUG oslo_concurrency.lockutils [None req-6974871c-fbdf-4315-ae5e-0f262a5a9f50 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Lock "4ef8066a-b68c-457b-9964-b1c34bab0fc3" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.007s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 808.981256] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae9f137a-9bac-4ae5-a21f-92835f313908 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.989914] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5df4d42-f8f7-4177-81f0-3580498ec16b {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.026339] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51e79ab7-1efd-4639-89be-18fc5030ce89 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.033888] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf93c12a-e565-455e-be1c-b9df8017612d {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.048894] env[63515]: DEBUG nova.network.neutron [-] [instance: 84dbf321-2c70-4e08-b430-cb5a06fc6829] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 809.048894] env[63515]: DEBUG nova.compute.provider_tree [None req-67d44811-1ff4-4a5b-a720-2ef182e4f0da tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 809.151421] env[63515]: DEBUG oslo_concurrency.lockutils [None req-811d8802-f522-4005-9c27-973badf5075e tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 809.316366] env[63515]: DEBUG oslo_vmware.api [None req-41b4a543-33b4-43d9-863c-4344bf9d0a46 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52a46a34-6c56-161d-0fe4-1830e92d4e73, 'name': SearchDatastore_Task, 'duration_secs': 0.026849} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 809.316697] env[63515]: DEBUG oslo_concurrency.lockutils [None req-41b4a543-33b4-43d9-863c-4344bf9d0a46 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 809.316942] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-41b4a543-33b4-43d9-863c-4344bf9d0a46 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: d926c699-a64a-4942-9ef4-f0166414661d] Processing image 8a120570-cb06-4099-b262-554ca0ad15c5 {{(pid=63515) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 809.317193] env[63515]: DEBUG oslo_concurrency.lockutils [None req-41b4a543-33b4-43d9-863c-4344bf9d0a46 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 809.317345] env[63515]: DEBUG oslo_concurrency.lockutils [None req-41b4a543-33b4-43d9-863c-4344bf9d0a46 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 809.317532] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-41b4a543-33b4-43d9-863c-4344bf9d0a46 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 809.317815] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ce55df65-0685-4faa-9514-5bcb09bed0ea {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.321383] env[63515]: DEBUG nova.compute.manager [None req-9dd4330d-d35d-47d5-b468-11f24e31a49a tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: 70df69e5-687b-44fb-b6fc-cdb08e21dda0] Starting instance... {{(pid=63515) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 809.329684] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-41b4a543-33b4-43d9-863c-4344bf9d0a46 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 809.329879] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-41b4a543-33b4-43d9-863c-4344bf9d0a46 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63515) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 809.330642] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8ac5dd0c-09d4-4d31-a2f3-bf9de48409e4 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.338737] env[63515]: DEBUG oslo_vmware.api [None req-41b4a543-33b4-43d9-863c-4344bf9d0a46 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Waiting for the task: (returnval){ [ 809.338737] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]526801c9-8b16-5048-486d-e6f45dac77f7" [ 809.338737] env[63515]: _type = "Task" [ 809.338737] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 809.347551] env[63515]: DEBUG oslo_vmware.api [None req-41b4a543-33b4-43d9-863c-4344bf9d0a46 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]526801c9-8b16-5048-486d-e6f45dac77f7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.489429] env[63515]: DEBUG nova.compute.manager [req-7d21d1db-6b29-4a6d-8a97-2c7956e5fded req-53bde895-3e21-4815-9c85-5ea69fbb972b service nova] [instance: d926c699-a64a-4942-9ef4-f0166414661d] Received event network-changed-fdec91d2-43e8-4d40-9aa9-885a82fd9b42 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 809.489618] env[63515]: DEBUG nova.compute.manager [req-7d21d1db-6b29-4a6d-8a97-2c7956e5fded req-53bde895-3e21-4815-9c85-5ea69fbb972b service nova] [instance: d926c699-a64a-4942-9ef4-f0166414661d] Refreshing instance network info cache due to event network-changed-fdec91d2-43e8-4d40-9aa9-885a82fd9b42. {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 809.489845] env[63515]: DEBUG oslo_concurrency.lockutils [req-7d21d1db-6b29-4a6d-8a97-2c7956e5fded req-53bde895-3e21-4815-9c85-5ea69fbb972b service nova] Acquiring lock "refresh_cache-d926c699-a64a-4942-9ef4-f0166414661d" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 809.490696] env[63515]: DEBUG oslo_concurrency.lockutils [req-7d21d1db-6b29-4a6d-8a97-2c7956e5fded req-53bde895-3e21-4815-9c85-5ea69fbb972b service nova] Acquired lock "refresh_cache-d926c699-a64a-4942-9ef4-f0166414661d" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 809.490696] env[63515]: DEBUG nova.network.neutron [req-7d21d1db-6b29-4a6d-8a97-2c7956e5fded req-53bde895-3e21-4815-9c85-5ea69fbb972b service nova] [instance: d926c699-a64a-4942-9ef4-f0166414661d] Refreshing network info cache for port fdec91d2-43e8-4d40-9aa9-885a82fd9b42 {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 809.538462] env[63515]: DEBUG oslo_concurrency.lockutils [None req-7bca9c72-42f5-4d65-99a2-4374b66c9c04 tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] Acquiring lock "17275472-2921-49c9-b4e0-5485649ebd1c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 809.538759] env[63515]: DEBUG oslo_concurrency.lockutils [None req-7bca9c72-42f5-4d65-99a2-4374b66c9c04 tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] Lock "17275472-2921-49c9-b4e0-5485649ebd1c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 809.550362] env[63515]: INFO nova.compute.manager [-] [instance: 84dbf321-2c70-4e08-b430-cb5a06fc6829] Took 1.35 seconds to deallocate network for instance. [ 809.551439] env[63515]: DEBUG nova.scheduler.client.report [None req-67d44811-1ff4-4a5b-a720-2ef182e4f0da tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 809.847968] env[63515]: DEBUG oslo_concurrency.lockutils [None req-9dd4330d-d35d-47d5-b468-11f24e31a49a tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 809.853962] env[63515]: DEBUG oslo_vmware.api [None req-41b4a543-33b4-43d9-863c-4344bf9d0a46 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]526801c9-8b16-5048-486d-e6f45dac77f7, 'name': SearchDatastore_Task, 'duration_secs': 0.024341} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 809.857682] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ebe49d31-be0f-4943-8127-eef9869d5482 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.864689] env[63515]: DEBUG oslo_vmware.api [None req-41b4a543-33b4-43d9-863c-4344bf9d0a46 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Waiting for the task: (returnval){ [ 809.864689] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]5257316e-762b-ca73-ec6b-040818bb78f7" [ 809.864689] env[63515]: _type = "Task" [ 809.864689] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 809.875795] env[63515]: DEBUG oslo_vmware.api [None req-41b4a543-33b4-43d9-863c-4344bf9d0a46 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]5257316e-762b-ca73-ec6b-040818bb78f7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.961596] env[63515]: DEBUG oslo_concurrency.lockutils [None req-6974871c-fbdf-4315-ae5e-0f262a5a9f50 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Acquiring lock "4ef8066a-b68c-457b-9964-b1c34bab0fc3" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 809.961700] env[63515]: DEBUG oslo_concurrency.lockutils [None req-6974871c-fbdf-4315-ae5e-0f262a5a9f50 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Lock "4ef8066a-b68c-457b-9964-b1c34bab0fc3" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 809.961944] env[63515]: INFO nova.compute.manager [None req-6974871c-fbdf-4315-ae5e-0f262a5a9f50 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: 4ef8066a-b68c-457b-9964-b1c34bab0fc3] Attaching volume 6112b2eb-44db-4a56-8191-4ff5c6765dea to /dev/sdb [ 810.000311] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6595782-417e-49ea-8d57-c41b17f852f5 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.010021] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5223dbc-0046-4536-aa75-49db3fa0a24e {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.027358] env[63515]: DEBUG nova.virt.block_device [None req-6974871c-fbdf-4315-ae5e-0f262a5a9f50 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: 4ef8066a-b68c-457b-9964-b1c34bab0fc3] Updating existing volume attachment record: f084ffb5-6ca3-4c5e-9b61-37bce219661a {{(pid=63515) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 810.041150] env[63515]: DEBUG nova.compute.manager [None req-7bca9c72-42f5-4d65-99a2-4374b66c9c04 tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] [instance: 17275472-2921-49c9-b4e0-5485649ebd1c] Starting instance... {{(pid=63515) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 810.058755] env[63515]: DEBUG oslo_concurrency.lockutils [None req-67d44811-1ff4-4a5b-a720-2ef182e4f0da tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.964s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 810.062568] env[63515]: DEBUG oslo_concurrency.lockutils [None req-9044ca9f-2eb8-445e-a5e5-e15900a98ac3 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 34.333s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 810.062860] env[63515]: DEBUG nova.objects.instance [None req-9044ca9f-2eb8-445e-a5e5-e15900a98ac3 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] Lazy-loading 'resources' on Instance uuid 8eb17506-ff93-4d25-b9af-ec5886569e65 {{(pid=63515) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 810.064681] env[63515]: DEBUG oslo_concurrency.lockutils [None req-1f202073-ad2a-4c82-aa50-bf79e5416807 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 810.086607] env[63515]: INFO nova.scheduler.client.report [None req-67d44811-1ff4-4a5b-a720-2ef182e4f0da tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] Deleted allocations for instance 30b8b872-40ca-4297-b98e-a64c83a75483 [ 810.327965] env[63515]: INFO nova.compute.manager [None req-b3d6c90d-bdbe-42fc-8e37-795c665e1a37 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] [instance: 5cbce760-0163-4b27-8ae3-e46c926c8916] Rescuing [ 810.328257] env[63515]: DEBUG oslo_concurrency.lockutils [None req-b3d6c90d-bdbe-42fc-8e37-795c665e1a37 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Acquiring lock "refresh_cache-5cbce760-0163-4b27-8ae3-e46c926c8916" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 810.328413] env[63515]: DEBUG oslo_concurrency.lockutils [None req-b3d6c90d-bdbe-42fc-8e37-795c665e1a37 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Acquired lock "refresh_cache-5cbce760-0163-4b27-8ae3-e46c926c8916" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 810.328584] env[63515]: DEBUG nova.network.neutron [None req-b3d6c90d-bdbe-42fc-8e37-795c665e1a37 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] [instance: 5cbce760-0163-4b27-8ae3-e46c926c8916] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 810.357878] env[63515]: DEBUG oslo_concurrency.lockutils [None req-08dbd0f6-25be-4803-9b5a-523d28be6f7f tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Acquiring lock "6688cdaa-29ba-413a-8131-4f834cdb70e4" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 810.358474] env[63515]: DEBUG oslo_concurrency.lockutils [None req-08dbd0f6-25be-4803-9b5a-523d28be6f7f tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Lock "6688cdaa-29ba-413a-8131-4f834cdb70e4" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 810.358687] env[63515]: DEBUG oslo_concurrency.lockutils [None req-08dbd0f6-25be-4803-9b5a-523d28be6f7f tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Acquiring lock "6688cdaa-29ba-413a-8131-4f834cdb70e4-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 810.358904] env[63515]: DEBUG oslo_concurrency.lockutils [None req-08dbd0f6-25be-4803-9b5a-523d28be6f7f tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Lock "6688cdaa-29ba-413a-8131-4f834cdb70e4-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 810.359097] env[63515]: DEBUG oslo_concurrency.lockutils [None req-08dbd0f6-25be-4803-9b5a-523d28be6f7f tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Lock "6688cdaa-29ba-413a-8131-4f834cdb70e4-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 810.362806] env[63515]: INFO nova.compute.manager [None req-08dbd0f6-25be-4803-9b5a-523d28be6f7f tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] [instance: 6688cdaa-29ba-413a-8131-4f834cdb70e4] Terminating instance [ 810.366028] env[63515]: DEBUG nova.compute.manager [None req-08dbd0f6-25be-4803-9b5a-523d28be6f7f tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] [instance: 6688cdaa-29ba-413a-8131-4f834cdb70e4] Start destroying the instance on the hypervisor. {{(pid=63515) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 810.366028] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-08dbd0f6-25be-4803-9b5a-523d28be6f7f tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] [instance: 6688cdaa-29ba-413a-8131-4f834cdb70e4] Destroying instance {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 810.366948] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51b2e4d8-73eb-483b-967e-d3ac47650c53 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.388932] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-08dbd0f6-25be-4803-9b5a-523d28be6f7f tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] [instance: 6688cdaa-29ba-413a-8131-4f834cdb70e4] Powering off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 810.390111] env[63515]: DEBUG oslo_vmware.api [None req-41b4a543-33b4-43d9-863c-4344bf9d0a46 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]5257316e-762b-ca73-ec6b-040818bb78f7, 'name': SearchDatastore_Task, 'duration_secs': 0.016349} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 810.391807] env[63515]: DEBUG nova.network.neutron [req-7d21d1db-6b29-4a6d-8a97-2c7956e5fded req-53bde895-3e21-4815-9c85-5ea69fbb972b service nova] [instance: d926c699-a64a-4942-9ef4-f0166414661d] Updated VIF entry in instance network info cache for port fdec91d2-43e8-4d40-9aa9-885a82fd9b42. {{(pid=63515) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 810.392709] env[63515]: DEBUG nova.network.neutron [req-7d21d1db-6b29-4a6d-8a97-2c7956e5fded req-53bde895-3e21-4815-9c85-5ea69fbb972b service nova] [instance: d926c699-a64a-4942-9ef4-f0166414661d] Updating instance_info_cache with network_info: [{"id": "fdec91d2-43e8-4d40-9aa9-885a82fd9b42", "address": "fa:16:3e:5c:b7:75", "network": {"id": "15fdaabe-6793-400f-9737-2cdff8f07238", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1175204866-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f9c98e9f6020475490aaa7e76d907ab7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7d377d75-3add-4a15-8691-74b2eb010924", "external-id": "nsx-vlan-transportzone-71", "segmentation_id": 71, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfdec91d2-43", "ovs_interfaceid": "fdec91d2-43e8-4d40-9aa9-885a82fd9b42", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 810.393360] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1e88e7d6-7300-451d-b692-6f3fb9149dbd {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.395928] env[63515]: DEBUG oslo_concurrency.lockutils [None req-41b4a543-33b4-43d9-863c-4344bf9d0a46 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 810.395928] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-41b4a543-33b4-43d9-863c-4344bf9d0a46 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk to [datastore2] d926c699-a64a-4942-9ef4-f0166414661d/d926c699-a64a-4942-9ef4-f0166414661d.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 810.396453] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-dcf9d683-2413-4542-8e1b-a0d7ca2ce1ea {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.404261] env[63515]: DEBUG oslo_vmware.api [None req-08dbd0f6-25be-4803-9b5a-523d28be6f7f tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Waiting for the task: (returnval){ [ 810.404261] env[63515]: value = "task-1111294" [ 810.404261] env[63515]: _type = "Task" [ 810.404261] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 810.408176] env[63515]: DEBUG oslo_vmware.api [None req-41b4a543-33b4-43d9-863c-4344bf9d0a46 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Waiting for the task: (returnval){ [ 810.408176] env[63515]: value = "task-1111295" [ 810.408176] env[63515]: _type = "Task" [ 810.408176] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 810.421644] env[63515]: DEBUG oslo_vmware.api [None req-08dbd0f6-25be-4803-9b5a-523d28be6f7f tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Task: {'id': task-1111294, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.425090] env[63515]: DEBUG oslo_vmware.api [None req-41b4a543-33b4-43d9-863c-4344bf9d0a46 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Task: {'id': task-1111295, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.574675] env[63515]: DEBUG oslo_concurrency.lockutils [None req-7bca9c72-42f5-4d65-99a2-4374b66c9c04 tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 810.597827] env[63515]: DEBUG oslo_concurrency.lockutils [None req-67d44811-1ff4-4a5b-a720-2ef182e4f0da tempest-SecurityGroupsTestJSON-1730559586 tempest-SecurityGroupsTestJSON-1730559586-project-member] Lock "30b8b872-40ca-4297-b98e-a64c83a75483" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 38.017s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 810.760993] env[63515]: DEBUG oslo_concurrency.lockutils [None req-73bb59a8-34a2-4cd0-acb3-804cea19b215 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] Acquiring lock "e666825c-ff4e-4a0e-93c0-43c00f167bbb" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 810.761189] env[63515]: DEBUG oslo_concurrency.lockutils [None req-73bb59a8-34a2-4cd0-acb3-804cea19b215 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] Lock "e666825c-ff4e-4a0e-93c0-43c00f167bbb" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 810.761410] env[63515]: DEBUG oslo_concurrency.lockutils [None req-73bb59a8-34a2-4cd0-acb3-804cea19b215 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] Acquiring lock "e666825c-ff4e-4a0e-93c0-43c00f167bbb-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 810.761603] env[63515]: DEBUG oslo_concurrency.lockutils [None req-73bb59a8-34a2-4cd0-acb3-804cea19b215 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] Lock "e666825c-ff4e-4a0e-93c0-43c00f167bbb-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 810.761778] env[63515]: DEBUG oslo_concurrency.lockutils [None req-73bb59a8-34a2-4cd0-acb3-804cea19b215 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] Lock "e666825c-ff4e-4a0e-93c0-43c00f167bbb-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 810.764350] env[63515]: INFO nova.compute.manager [None req-73bb59a8-34a2-4cd0-acb3-804cea19b215 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] [instance: e666825c-ff4e-4a0e-93c0-43c00f167bbb] Terminating instance [ 810.767131] env[63515]: DEBUG nova.compute.manager [None req-73bb59a8-34a2-4cd0-acb3-804cea19b215 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] [instance: e666825c-ff4e-4a0e-93c0-43c00f167bbb] Start destroying the instance on the hypervisor. {{(pid=63515) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 810.767371] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-73bb59a8-34a2-4cd0-acb3-804cea19b215 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] [instance: e666825c-ff4e-4a0e-93c0-43c00f167bbb] Destroying instance {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 810.768881] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78f0b63e-73d6-4236-be18-25c026a801ea {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.786787] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-73bb59a8-34a2-4cd0-acb3-804cea19b215 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] [instance: e666825c-ff4e-4a0e-93c0-43c00f167bbb] Unregistering the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 810.786787] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-11754874-c393-4a9b-9615-270d8643cef5 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.875410] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-73bb59a8-34a2-4cd0-acb3-804cea19b215 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] [instance: e666825c-ff4e-4a0e-93c0-43c00f167bbb] Unregistered the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 810.875410] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-73bb59a8-34a2-4cd0-acb3-804cea19b215 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] [instance: e666825c-ff4e-4a0e-93c0-43c00f167bbb] Deleting contents of the VM from datastore datastore1 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 810.876925] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-73bb59a8-34a2-4cd0-acb3-804cea19b215 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] Deleting the datastore file [datastore1] e666825c-ff4e-4a0e-93c0-43c00f167bbb {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 810.876925] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a55822c6-c964-4eb5-ba82-de4ddee8a7d1 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.893832] env[63515]: DEBUG oslo_vmware.api [None req-73bb59a8-34a2-4cd0-acb3-804cea19b215 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] Waiting for the task: (returnval){ [ 810.893832] env[63515]: value = "task-1111298" [ 810.893832] env[63515]: _type = "Task" [ 810.893832] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 810.900316] env[63515]: DEBUG oslo_concurrency.lockutils [req-7d21d1db-6b29-4a6d-8a97-2c7956e5fded req-53bde895-3e21-4815-9c85-5ea69fbb972b service nova] Releasing lock "refresh_cache-d926c699-a64a-4942-9ef4-f0166414661d" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 810.900690] env[63515]: DEBUG nova.compute.manager [req-7d21d1db-6b29-4a6d-8a97-2c7956e5fded req-53bde895-3e21-4815-9c85-5ea69fbb972b service nova] [instance: ed550b10-d58f-45b8-b766-198f431c3788] Received event network-vif-deleted-ad9dda29-9326-472b-8c82-294a33a7faf0 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 810.908453] env[63515]: DEBUG oslo_vmware.api [None req-73bb59a8-34a2-4cd0-acb3-804cea19b215 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] Task: {'id': task-1111298, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.922433] env[63515]: DEBUG oslo_vmware.api [None req-08dbd0f6-25be-4803-9b5a-523d28be6f7f tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Task: {'id': task-1111294, 'name': PowerOffVM_Task, 'duration_secs': 0.249612} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 810.929199] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-08dbd0f6-25be-4803-9b5a-523d28be6f7f tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] [instance: 6688cdaa-29ba-413a-8131-4f834cdb70e4] Powered off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 810.929199] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-08dbd0f6-25be-4803-9b5a-523d28be6f7f tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] [instance: 6688cdaa-29ba-413a-8131-4f834cdb70e4] Unregistering the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 810.929317] env[63515]: DEBUG oslo_vmware.api [None req-41b4a543-33b4-43d9-863c-4344bf9d0a46 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Task: {'id': task-1111295, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.929794] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-7956f3f8-9b17-4e3d-9b82-10c50e5a86a2 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.996753] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f459593a-369e-479d-bc53-323f43c45c78 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.007577] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-08dbd0f6-25be-4803-9b5a-523d28be6f7f tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] [instance: 6688cdaa-29ba-413a-8131-4f834cdb70e4] Unregistered the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 811.007941] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-08dbd0f6-25be-4803-9b5a-523d28be6f7f tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] [instance: 6688cdaa-29ba-413a-8131-4f834cdb70e4] Deleting contents of the VM from datastore datastore1 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 811.008224] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-08dbd0f6-25be-4803-9b5a-523d28be6f7f tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Deleting the datastore file [datastore1] 6688cdaa-29ba-413a-8131-4f834cdb70e4 {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 811.009437] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c8f1c000-d466-4da4-82bd-5ec09cb08db8 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.019160] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-799048d9-7523-4bce-8be0-8a9e8dcbbd26 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.024149] env[63515]: DEBUG oslo_vmware.api [None req-08dbd0f6-25be-4803-9b5a-523d28be6f7f tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Waiting for the task: (returnval){ [ 811.024149] env[63515]: value = "task-1111300" [ 811.024149] env[63515]: _type = "Task" [ 811.024149] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 811.065559] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc6705db-97ac-4d22-8d8d-ed048dafb93e {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.073786] env[63515]: DEBUG oslo_vmware.api [None req-08dbd0f6-25be-4803-9b5a-523d28be6f7f tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Task: {'id': task-1111300, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.082648] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d60d9bfd-d616-4060-bb04-d17b4a19355a {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.104809] env[63515]: DEBUG nova.compute.provider_tree [None req-9044ca9f-2eb8-445e-a5e5-e15900a98ac3 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 811.190399] env[63515]: DEBUG nova.network.neutron [None req-b3d6c90d-bdbe-42fc-8e37-795c665e1a37 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] [instance: 5cbce760-0163-4b27-8ae3-e46c926c8916] Updating instance_info_cache with network_info: [{"id": "5fa807c4-5ce1-492a-bea3-d8f7a65d898f", "address": "fa:16:3e:23:bb:83", "network": {"id": "850bfa95-3df6-4019-8260-ade20ec58799", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-832000501-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "08c9aa51b1864410809f8877c7cd49c4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3cf748a8-7ae0-4dca-817d-e727c30d72f4", "external-id": "nsx-vlan-transportzone-853", "segmentation_id": 853, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5fa807c4-5c", "ovs_interfaceid": "5fa807c4-5ce1-492a-bea3-d8f7a65d898f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 811.402861] env[63515]: DEBUG oslo_vmware.api [None req-73bb59a8-34a2-4cd0-acb3-804cea19b215 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] Task: {'id': task-1111298, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.192935} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 811.403279] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-73bb59a8-34a2-4cd0-acb3-804cea19b215 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] Deleted the datastore file {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 811.403602] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-73bb59a8-34a2-4cd0-acb3-804cea19b215 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] [instance: e666825c-ff4e-4a0e-93c0-43c00f167bbb] Deleted contents of the VM from datastore datastore1 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 811.403866] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-73bb59a8-34a2-4cd0-acb3-804cea19b215 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] [instance: e666825c-ff4e-4a0e-93c0-43c00f167bbb] Instance destroyed {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 811.404111] env[63515]: INFO nova.compute.manager [None req-73bb59a8-34a2-4cd0-acb3-804cea19b215 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] [instance: e666825c-ff4e-4a0e-93c0-43c00f167bbb] Took 0.64 seconds to destroy the instance on the hypervisor. [ 811.404443] env[63515]: DEBUG oslo.service.loopingcall [None req-73bb59a8-34a2-4cd0-acb3-804cea19b215 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 811.404708] env[63515]: DEBUG nova.compute.manager [-] [instance: e666825c-ff4e-4a0e-93c0-43c00f167bbb] Deallocating network for instance {{(pid=63515) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 811.404840] env[63515]: DEBUG nova.network.neutron [-] [instance: e666825c-ff4e-4a0e-93c0-43c00f167bbb] deallocate_for_instance() {{(pid=63515) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 811.423212] env[63515]: DEBUG oslo_vmware.api [None req-41b4a543-33b4-43d9-863c-4344bf9d0a46 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Task: {'id': task-1111295, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.534762} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 811.423547] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-41b4a543-33b4-43d9-863c-4344bf9d0a46 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk to [datastore2] d926c699-a64a-4942-9ef4-f0166414661d/d926c699-a64a-4942-9ef4-f0166414661d.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 811.423816] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-41b4a543-33b4-43d9-863c-4344bf9d0a46 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: d926c699-a64a-4942-9ef4-f0166414661d] Extending root virtual disk to 1048576 {{(pid=63515) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 811.424166] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-37af2b39-ad9d-4fbe-a671-c1fd1972cd3a {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.435060] env[63515]: DEBUG oslo_vmware.api [None req-41b4a543-33b4-43d9-863c-4344bf9d0a46 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Waiting for the task: (returnval){ [ 811.435060] env[63515]: value = "task-1111301" [ 811.435060] env[63515]: _type = "Task" [ 811.435060] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 811.447037] env[63515]: DEBUG oslo_vmware.api [None req-41b4a543-33b4-43d9-863c-4344bf9d0a46 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Task: {'id': task-1111301, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.536225] env[63515]: DEBUG oslo_vmware.api [None req-08dbd0f6-25be-4803-9b5a-523d28be6f7f tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Task: {'id': task-1111300, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.373924} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 811.536608] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-08dbd0f6-25be-4803-9b5a-523d28be6f7f tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Deleted the datastore file {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 811.536837] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-08dbd0f6-25be-4803-9b5a-523d28be6f7f tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] [instance: 6688cdaa-29ba-413a-8131-4f834cdb70e4] Deleted contents of the VM from datastore datastore1 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 811.537283] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-08dbd0f6-25be-4803-9b5a-523d28be6f7f tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] [instance: 6688cdaa-29ba-413a-8131-4f834cdb70e4] Instance destroyed {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 811.537718] env[63515]: INFO nova.compute.manager [None req-08dbd0f6-25be-4803-9b5a-523d28be6f7f tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] [instance: 6688cdaa-29ba-413a-8131-4f834cdb70e4] Took 1.17 seconds to destroy the instance on the hypervisor. [ 811.538058] env[63515]: DEBUG oslo.service.loopingcall [None req-08dbd0f6-25be-4803-9b5a-523d28be6f7f tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 811.538388] env[63515]: DEBUG nova.compute.manager [-] [instance: 6688cdaa-29ba-413a-8131-4f834cdb70e4] Deallocating network for instance {{(pid=63515) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 811.538532] env[63515]: DEBUG nova.network.neutron [-] [instance: 6688cdaa-29ba-413a-8131-4f834cdb70e4] deallocate_for_instance() {{(pid=63515) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 811.609484] env[63515]: DEBUG nova.scheduler.client.report [None req-9044ca9f-2eb8-445e-a5e5-e15900a98ac3 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 811.644133] env[63515]: DEBUG oslo_concurrency.lockutils [None req-fe95cb73-4a2a-4d94-83ac-9bcd96be3709 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Acquiring lock "19475235-dbec-40a5-bd8f-8070616b7ddb" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 811.644445] env[63515]: DEBUG oslo_concurrency.lockutils [None req-fe95cb73-4a2a-4d94-83ac-9bcd96be3709 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Lock "19475235-dbec-40a5-bd8f-8070616b7ddb" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 811.693047] env[63515]: DEBUG oslo_concurrency.lockutils [None req-b3d6c90d-bdbe-42fc-8e37-795c665e1a37 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Releasing lock "refresh_cache-5cbce760-0163-4b27-8ae3-e46c926c8916" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 811.873131] env[63515]: DEBUG nova.compute.manager [req-cdf87d59-d7c7-417b-8d30-953d98b8863c req-d364f253-c596-4d9f-b2ea-ef65a94e854c service nova] [instance: 4e453127-1f3e-40ea-819f-6678479826c8] Received event network-changed-3d5ad33f-8649-42c8-b899-ca20c9a2636e {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 811.873131] env[63515]: DEBUG nova.compute.manager [req-cdf87d59-d7c7-417b-8d30-953d98b8863c req-d364f253-c596-4d9f-b2ea-ef65a94e854c service nova] [instance: 4e453127-1f3e-40ea-819f-6678479826c8] Refreshing instance network info cache due to event network-changed-3d5ad33f-8649-42c8-b899-ca20c9a2636e. {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 811.873131] env[63515]: DEBUG oslo_concurrency.lockutils [req-cdf87d59-d7c7-417b-8d30-953d98b8863c req-d364f253-c596-4d9f-b2ea-ef65a94e854c service nova] Acquiring lock "refresh_cache-4e453127-1f3e-40ea-819f-6678479826c8" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 811.873131] env[63515]: DEBUG oslo_concurrency.lockutils [req-cdf87d59-d7c7-417b-8d30-953d98b8863c req-d364f253-c596-4d9f-b2ea-ef65a94e854c service nova] Acquired lock "refresh_cache-4e453127-1f3e-40ea-819f-6678479826c8" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 811.873131] env[63515]: DEBUG nova.network.neutron [req-cdf87d59-d7c7-417b-8d30-953d98b8863c req-d364f253-c596-4d9f-b2ea-ef65a94e854c service nova] [instance: 4e453127-1f3e-40ea-819f-6678479826c8] Refreshing network info cache for port 3d5ad33f-8649-42c8-b899-ca20c9a2636e {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 811.956486] env[63515]: DEBUG oslo_vmware.api [None req-41b4a543-33b4-43d9-863c-4344bf9d0a46 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Task: {'id': task-1111301, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.082513} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 811.956486] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-41b4a543-33b4-43d9-863c-4344bf9d0a46 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: d926c699-a64a-4942-9ef4-f0166414661d] Extended root virtual disk {{(pid=63515) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 811.956486] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2276de69-6ed4-465d-bc36-61d3585bfee7 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.962182] env[63515]: DEBUG nova.compute.manager [req-1993e18c-a0dd-4da7-9a0d-a8a401d079e4 req-f363afdf-efee-473f-b19d-31303a68d03f service nova] [instance: 84dbf321-2c70-4e08-b430-cb5a06fc6829] Received event network-vif-deleted-ad6685ff-3a87-4cc5-bdfa-5eb3bb2f0590 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 811.986699] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-41b4a543-33b4-43d9-863c-4344bf9d0a46 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: d926c699-a64a-4942-9ef4-f0166414661d] Reconfiguring VM instance instance-0000003f to attach disk [datastore2] d926c699-a64a-4942-9ef4-f0166414661d/d926c699-a64a-4942-9ef4-f0166414661d.vmdk or device None with type sparse {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 811.986699] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-87243419-4786-4283-bff8-ead4236e332d {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.008938] env[63515]: DEBUG oslo_vmware.api [None req-41b4a543-33b4-43d9-863c-4344bf9d0a46 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Waiting for the task: (returnval){ [ 812.008938] env[63515]: value = "task-1111302" [ 812.008938] env[63515]: _type = "Task" [ 812.008938] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 812.019906] env[63515]: DEBUG oslo_vmware.api [None req-41b4a543-33b4-43d9-863c-4344bf9d0a46 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Task: {'id': task-1111302, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 812.117023] env[63515]: DEBUG oslo_concurrency.lockutils [None req-9044ca9f-2eb8-445e-a5e5-e15900a98ac3 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.053s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 812.117831] env[63515]: DEBUG oslo_concurrency.lockutils [None req-78f15907-fa94-44dd-a2b3-ccdc7dec0a0f tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 35.344s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 812.118842] env[63515]: DEBUG nova.objects.instance [None req-78f15907-fa94-44dd-a2b3-ccdc7dec0a0f tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Lazy-loading 'resources' on Instance uuid 87c468d9-9594-4804-b461-527f01f6118f {{(pid=63515) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 812.147499] env[63515]: DEBUG nova.compute.manager [None req-fe95cb73-4a2a-4d94-83ac-9bcd96be3709 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: 19475235-dbec-40a5-bd8f-8070616b7ddb] Starting instance... {{(pid=63515) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 812.150990] env[63515]: INFO nova.scheduler.client.report [None req-9044ca9f-2eb8-445e-a5e5-e15900a98ac3 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] Deleted allocations for instance 8eb17506-ff93-4d25-b9af-ec5886569e65 [ 812.232629] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-b3d6c90d-bdbe-42fc-8e37-795c665e1a37 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] [instance: 5cbce760-0163-4b27-8ae3-e46c926c8916] Powering off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 812.232629] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e70b0f0b-e4d2-4441-9a6c-7720c41cba6e {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.241317] env[63515]: DEBUG oslo_vmware.api [None req-b3d6c90d-bdbe-42fc-8e37-795c665e1a37 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Waiting for the task: (returnval){ [ 812.241317] env[63515]: value = "task-1111303" [ 812.241317] env[63515]: _type = "Task" [ 812.241317] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 812.247668] env[63515]: DEBUG nova.network.neutron [-] [instance: e666825c-ff4e-4a0e-93c0-43c00f167bbb] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 812.252491] env[63515]: DEBUG oslo_vmware.api [None req-b3d6c90d-bdbe-42fc-8e37-795c665e1a37 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Task: {'id': task-1111303, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 812.422146] env[63515]: DEBUG nova.network.neutron [-] [instance: 6688cdaa-29ba-413a-8131-4f834cdb70e4] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 812.521160] env[63515]: DEBUG oslo_vmware.api [None req-41b4a543-33b4-43d9-863c-4344bf9d0a46 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Task: {'id': task-1111302, 'name': ReconfigVM_Task} progress is 99%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 812.622388] env[63515]: DEBUG nova.objects.instance [None req-78f15907-fa94-44dd-a2b3-ccdc7dec0a0f tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Lazy-loading 'numa_topology' on Instance uuid 87c468d9-9594-4804-b461-527f01f6118f {{(pid=63515) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 812.662069] env[63515]: DEBUG oslo_concurrency.lockutils [None req-9044ca9f-2eb8-445e-a5e5-e15900a98ac3 tempest-ServerShowV254Test-953193898 tempest-ServerShowV254Test-953193898-project-member] Lock "8eb17506-ff93-4d25-b9af-ec5886569e65" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 40.686s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 812.679123] env[63515]: DEBUG oslo_concurrency.lockutils [None req-fe95cb73-4a2a-4d94-83ac-9bcd96be3709 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 812.754079] env[63515]: DEBUG nova.network.neutron [req-cdf87d59-d7c7-417b-8d30-953d98b8863c req-d364f253-c596-4d9f-b2ea-ef65a94e854c service nova] [instance: 4e453127-1f3e-40ea-819f-6678479826c8] Updated VIF entry in instance network info cache for port 3d5ad33f-8649-42c8-b899-ca20c9a2636e. {{(pid=63515) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 812.754662] env[63515]: DEBUG nova.network.neutron [req-cdf87d59-d7c7-417b-8d30-953d98b8863c req-d364f253-c596-4d9f-b2ea-ef65a94e854c service nova] [instance: 4e453127-1f3e-40ea-819f-6678479826c8] Updating instance_info_cache with network_info: [{"id": "3d5ad33f-8649-42c8-b899-ca20c9a2636e", "address": "fa:16:3e:c2:77:9b", "network": {"id": "5f57f1e7-21bb-425f-aab4-7d69b7ec94f7", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationNegativeTestJSON-177649896-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "60d9dcd0f30e48e7965ecfc1a72f5375", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3e05affa-2640-435e-a124-0ee8a6ab1152", "external-id": "nsx-vlan-transportzone-839", "segmentation_id": 839, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3d5ad33f-86", "ovs_interfaceid": "3d5ad33f-8649-42c8-b899-ca20c9a2636e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 812.756747] env[63515]: INFO nova.compute.manager [-] [instance: e666825c-ff4e-4a0e-93c0-43c00f167bbb] Took 1.35 seconds to deallocate network for instance. [ 812.762901] env[63515]: DEBUG oslo_vmware.api [None req-b3d6c90d-bdbe-42fc-8e37-795c665e1a37 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Task: {'id': task-1111303, 'name': PowerOffVM_Task, 'duration_secs': 0.194812} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 812.762901] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-b3d6c90d-bdbe-42fc-8e37-795c665e1a37 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] [instance: 5cbce760-0163-4b27-8ae3-e46c926c8916] Powered off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 812.766183] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c806945-32cd-497f-8857-98dc456c6566 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.792181] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b65b563-a8e6-4f3f-b343-c32004b66b60 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.838884] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-b3d6c90d-bdbe-42fc-8e37-795c665e1a37 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] [instance: 5cbce760-0163-4b27-8ae3-e46c926c8916] Powering off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 812.839231] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-38a023ac-bc24-43ce-a6d5-f5ac4879366e {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.854188] env[63515]: DEBUG oslo_vmware.api [None req-b3d6c90d-bdbe-42fc-8e37-795c665e1a37 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Waiting for the task: (returnval){ [ 812.854188] env[63515]: value = "task-1111305" [ 812.854188] env[63515]: _type = "Task" [ 812.854188] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 812.863652] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-b3d6c90d-bdbe-42fc-8e37-795c665e1a37 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] [instance: 5cbce760-0163-4b27-8ae3-e46c926c8916] VM already powered off {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 812.863891] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-b3d6c90d-bdbe-42fc-8e37-795c665e1a37 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] [instance: 5cbce760-0163-4b27-8ae3-e46c926c8916] Processing image 8a120570-cb06-4099-b262-554ca0ad15c5 {{(pid=63515) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 812.864188] env[63515]: DEBUG oslo_concurrency.lockutils [None req-b3d6c90d-bdbe-42fc-8e37-795c665e1a37 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 812.864358] env[63515]: DEBUG oslo_concurrency.lockutils [None req-b3d6c90d-bdbe-42fc-8e37-795c665e1a37 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 812.864558] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-b3d6c90d-bdbe-42fc-8e37-795c665e1a37 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 812.864900] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-19c74baa-4f72-4fa9-a278-01317d67aa1f {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.876881] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-b3d6c90d-bdbe-42fc-8e37-795c665e1a37 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 812.877131] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-b3d6c90d-bdbe-42fc-8e37-795c665e1a37 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63515) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 812.878700] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9d497623-808d-4bbf-bbc4-7a8814b5b747 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.886744] env[63515]: DEBUG oslo_vmware.api [None req-b3d6c90d-bdbe-42fc-8e37-795c665e1a37 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Waiting for the task: (returnval){ [ 812.886744] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]527ac196-5db0-1b15-f68b-e12e9dd750f1" [ 812.886744] env[63515]: _type = "Task" [ 812.886744] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 812.896373] env[63515]: DEBUG oslo_vmware.api [None req-b3d6c90d-bdbe-42fc-8e37-795c665e1a37 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]527ac196-5db0-1b15-f68b-e12e9dd750f1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 812.925088] env[63515]: INFO nova.compute.manager [-] [instance: 6688cdaa-29ba-413a-8131-4f834cdb70e4] Took 1.39 seconds to deallocate network for instance. [ 813.020385] env[63515]: DEBUG oslo_vmware.api [None req-41b4a543-33b4-43d9-863c-4344bf9d0a46 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Task: {'id': task-1111302, 'name': ReconfigVM_Task, 'duration_secs': 0.53846} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 813.020774] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-41b4a543-33b4-43d9-863c-4344bf9d0a46 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: d926c699-a64a-4942-9ef4-f0166414661d] Reconfigured VM instance instance-0000003f to attach disk [datastore2] d926c699-a64a-4942-9ef4-f0166414661d/d926c699-a64a-4942-9ef4-f0166414661d.vmdk or device None with type sparse {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 813.021949] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-7d621ad4-5fdc-427e-a96f-cedec23e8f5e {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.031114] env[63515]: DEBUG oslo_vmware.api [None req-41b4a543-33b4-43d9-863c-4344bf9d0a46 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Waiting for the task: (returnval){ [ 813.031114] env[63515]: value = "task-1111306" [ 813.031114] env[63515]: _type = "Task" [ 813.031114] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 813.043974] env[63515]: DEBUG oslo_vmware.api [None req-41b4a543-33b4-43d9-863c-4344bf9d0a46 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Task: {'id': task-1111306, 'name': Rename_Task} progress is 5%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.125738] env[63515]: DEBUG nova.objects.base [None req-78f15907-fa94-44dd-a2b3-ccdc7dec0a0f tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Object Instance<87c468d9-9594-4804-b461-527f01f6118f> lazy-loaded attributes: resources,numa_topology {{(pid=63515) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 813.263376] env[63515]: DEBUG oslo_concurrency.lockutils [req-cdf87d59-d7c7-417b-8d30-953d98b8863c req-d364f253-c596-4d9f-b2ea-ef65a94e854c service nova] Releasing lock "refresh_cache-4e453127-1f3e-40ea-819f-6678479826c8" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 813.274238] env[63515]: DEBUG oslo_concurrency.lockutils [None req-73bb59a8-34a2-4cd0-acb3-804cea19b215 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 813.400317] env[63515]: DEBUG oslo_vmware.api [None req-b3d6c90d-bdbe-42fc-8e37-795c665e1a37 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]527ac196-5db0-1b15-f68b-e12e9dd750f1, 'name': SearchDatastore_Task, 'duration_secs': 0.03202} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 813.400906] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8259ffd1-d104-4cce-9964-6c76cc07dd3e {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.408037] env[63515]: DEBUG oslo_vmware.api [None req-b3d6c90d-bdbe-42fc-8e37-795c665e1a37 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Waiting for the task: (returnval){ [ 813.408037] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]52eac3cd-4f4e-8646-b98e-e7905ec06114" [ 813.408037] env[63515]: _type = "Task" [ 813.408037] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 813.417884] env[63515]: DEBUG oslo_vmware.api [None req-b3d6c90d-bdbe-42fc-8e37-795c665e1a37 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52eac3cd-4f4e-8646-b98e-e7905ec06114, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.431983] env[63515]: DEBUG oslo_concurrency.lockutils [None req-08dbd0f6-25be-4803-9b5a-523d28be6f7f tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 813.450878] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-acee7baf-d9da-4496-aa42-0f4322f0f4bb {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.460572] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb74e71d-a511-4b1d-82cf-c19cfd4e9dc6 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.496211] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5918902a-310e-47be-aad6-4dbe41802558 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.504726] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39a4917d-2b53-4033-b599-d16d3f2aa4ab {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.519805] env[63515]: DEBUG nova.compute.provider_tree [None req-78f15907-fa94-44dd-a2b3-ccdc7dec0a0f tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 813.541641] env[63515]: DEBUG oslo_vmware.api [None req-41b4a543-33b4-43d9-863c-4344bf9d0a46 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Task: {'id': task-1111306, 'name': Rename_Task, 'duration_secs': 0.143029} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 813.541931] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-41b4a543-33b4-43d9-863c-4344bf9d0a46 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: d926c699-a64a-4942-9ef4-f0166414661d] Powering on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 813.542190] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f440a5ba-db05-4eae-b284-d2a8b354decc {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.550138] env[63515]: DEBUG oslo_vmware.api [None req-41b4a543-33b4-43d9-863c-4344bf9d0a46 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Waiting for the task: (returnval){ [ 813.550138] env[63515]: value = "task-1111307" [ 813.550138] env[63515]: _type = "Task" [ 813.550138] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 813.558458] env[63515]: DEBUG oslo_vmware.api [None req-41b4a543-33b4-43d9-863c-4344bf9d0a46 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Task: {'id': task-1111307, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.730790] env[63515]: DEBUG oslo_concurrency.lockutils [None req-5a74a6fe-271a-406d-93f5-9b098b35317a tempest-FloatingIPsAssociationNegativeTestJSON-2063029775 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775-project-member] Acquiring lock "4e453127-1f3e-40ea-819f-6678479826c8" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 813.730790] env[63515]: DEBUG oslo_concurrency.lockutils [None req-5a74a6fe-271a-406d-93f5-9b098b35317a tempest-FloatingIPsAssociationNegativeTestJSON-2063029775 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775-project-member] Lock "4e453127-1f3e-40ea-819f-6678479826c8" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 813.730790] env[63515]: DEBUG oslo_concurrency.lockutils [None req-5a74a6fe-271a-406d-93f5-9b098b35317a tempest-FloatingIPsAssociationNegativeTestJSON-2063029775 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775-project-member] Acquiring lock "4e453127-1f3e-40ea-819f-6678479826c8-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 813.730790] env[63515]: DEBUG oslo_concurrency.lockutils [None req-5a74a6fe-271a-406d-93f5-9b098b35317a tempest-FloatingIPsAssociationNegativeTestJSON-2063029775 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775-project-member] Lock "4e453127-1f3e-40ea-819f-6678479826c8-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 813.730790] env[63515]: DEBUG oslo_concurrency.lockutils [None req-5a74a6fe-271a-406d-93f5-9b098b35317a tempest-FloatingIPsAssociationNegativeTestJSON-2063029775 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775-project-member] Lock "4e453127-1f3e-40ea-819f-6678479826c8-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 813.732806] env[63515]: INFO nova.compute.manager [None req-5a74a6fe-271a-406d-93f5-9b098b35317a tempest-FloatingIPsAssociationNegativeTestJSON-2063029775 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775-project-member] [instance: 4e453127-1f3e-40ea-819f-6678479826c8] Terminating instance [ 813.734987] env[63515]: DEBUG nova.compute.manager [None req-5a74a6fe-271a-406d-93f5-9b098b35317a tempest-FloatingIPsAssociationNegativeTestJSON-2063029775 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775-project-member] [instance: 4e453127-1f3e-40ea-819f-6678479826c8] Start destroying the instance on the hypervisor. {{(pid=63515) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 813.735295] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-5a74a6fe-271a-406d-93f5-9b098b35317a tempest-FloatingIPsAssociationNegativeTestJSON-2063029775 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775-project-member] [instance: 4e453127-1f3e-40ea-819f-6678479826c8] Destroying instance {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 813.736246] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c3487ee-ed2e-46e4-84f0-05ebf0636b74 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.746153] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-5a74a6fe-271a-406d-93f5-9b098b35317a tempest-FloatingIPsAssociationNegativeTestJSON-2063029775 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775-project-member] [instance: 4e453127-1f3e-40ea-819f-6678479826c8] Powering off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 813.746153] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-60e42754-7f40-4bca-8408-7e3c9fd332be {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.754423] env[63515]: DEBUG oslo_vmware.api [None req-5a74a6fe-271a-406d-93f5-9b098b35317a tempest-FloatingIPsAssociationNegativeTestJSON-2063029775 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775-project-member] Waiting for the task: (returnval){ [ 813.754423] env[63515]: value = "task-1111308" [ 813.754423] env[63515]: _type = "Task" [ 813.754423] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 813.765462] env[63515]: DEBUG oslo_vmware.api [None req-5a74a6fe-271a-406d-93f5-9b098b35317a tempest-FloatingIPsAssociationNegativeTestJSON-2063029775 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775-project-member] Task: {'id': task-1111308, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.920698] env[63515]: DEBUG oslo_vmware.api [None req-b3d6c90d-bdbe-42fc-8e37-795c665e1a37 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52eac3cd-4f4e-8646-b98e-e7905ec06114, 'name': SearchDatastore_Task, 'duration_secs': 0.012581} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 813.921631] env[63515]: DEBUG oslo_concurrency.lockutils [None req-b3d6c90d-bdbe-42fc-8e37-795c665e1a37 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 813.922189] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-b3d6c90d-bdbe-42fc-8e37-795c665e1a37 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Copying virtual disk from [datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk to [datastore1] 5cbce760-0163-4b27-8ae3-e46c926c8916/8a120570-cb06-4099-b262-554ca0ad15c5-rescue.vmdk. {{(pid=63515) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 813.922985] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-1a915940-0cc8-4761-b857-eb459db38eec {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.933016] env[63515]: DEBUG oslo_vmware.api [None req-b3d6c90d-bdbe-42fc-8e37-795c665e1a37 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Waiting for the task: (returnval){ [ 813.933016] env[63515]: value = "task-1111309" [ 813.933016] env[63515]: _type = "Task" [ 813.933016] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 813.942266] env[63515]: DEBUG oslo_vmware.api [None req-b3d6c90d-bdbe-42fc-8e37-795c665e1a37 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Task: {'id': task-1111309, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.023050] env[63515]: DEBUG nova.scheduler.client.report [None req-78f15907-fa94-44dd-a2b3-ccdc7dec0a0f tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 814.063113] env[63515]: DEBUG oslo_vmware.api [None req-41b4a543-33b4-43d9-863c-4344bf9d0a46 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Task: {'id': task-1111307, 'name': PowerOnVM_Task} progress is 89%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.069746] env[63515]: DEBUG nova.compute.manager [req-157fdbf6-60a9-4624-bbd6-0204fcea66c5 req-cef5f78c-dc56-4623-bf3f-b084b6093694 service nova] [instance: 6688cdaa-29ba-413a-8131-4f834cdb70e4] Received event network-vif-deleted-6ccc2ab3-1e4f-484b-984b-5e0ebb251bd4 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 814.157942] env[63515]: DEBUG nova.compute.manager [req-bea984bb-c26a-411d-8187-58b895b44ba5 req-9a955740-3fdf-447b-a94e-cd5aeb2774f3 service nova] [instance: e666825c-ff4e-4a0e-93c0-43c00f167bbb] Received event network-vif-deleted-7872b4c3-d54f-4613-85d2-f568ec3cb47f {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 814.267113] env[63515]: DEBUG oslo_vmware.api [None req-5a74a6fe-271a-406d-93f5-9b098b35317a tempest-FloatingIPsAssociationNegativeTestJSON-2063029775 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775-project-member] Task: {'id': task-1111308, 'name': PowerOffVM_Task, 'duration_secs': 0.231444} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 814.267113] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-5a74a6fe-271a-406d-93f5-9b098b35317a tempest-FloatingIPsAssociationNegativeTestJSON-2063029775 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775-project-member] [instance: 4e453127-1f3e-40ea-819f-6678479826c8] Powered off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 814.267113] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-5a74a6fe-271a-406d-93f5-9b098b35317a tempest-FloatingIPsAssociationNegativeTestJSON-2063029775 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775-project-member] [instance: 4e453127-1f3e-40ea-819f-6678479826c8] Unregistering the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 814.267113] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-6eb3ab28-75b4-4b05-a88c-112eb6ad9491 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.372647] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-5a74a6fe-271a-406d-93f5-9b098b35317a tempest-FloatingIPsAssociationNegativeTestJSON-2063029775 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775-project-member] [instance: 4e453127-1f3e-40ea-819f-6678479826c8] Unregistered the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 814.372907] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-5a74a6fe-271a-406d-93f5-9b098b35317a tempest-FloatingIPsAssociationNegativeTestJSON-2063029775 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775-project-member] [instance: 4e453127-1f3e-40ea-819f-6678479826c8] Deleting contents of the VM from datastore datastore1 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 814.373113] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-5a74a6fe-271a-406d-93f5-9b098b35317a tempest-FloatingIPsAssociationNegativeTestJSON-2063029775 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775-project-member] Deleting the datastore file [datastore1] 4e453127-1f3e-40ea-819f-6678479826c8 {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 814.373403] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-37916d56-271d-46db-bfa4-30d6e0c5c35a {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.383412] env[63515]: DEBUG oslo_vmware.api [None req-5a74a6fe-271a-406d-93f5-9b098b35317a tempest-FloatingIPsAssociationNegativeTestJSON-2063029775 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775-project-member] Waiting for the task: (returnval){ [ 814.383412] env[63515]: value = "task-1111311" [ 814.383412] env[63515]: _type = "Task" [ 814.383412] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 814.391456] env[63515]: DEBUG oslo_vmware.api [None req-5a74a6fe-271a-406d-93f5-9b098b35317a tempest-FloatingIPsAssociationNegativeTestJSON-2063029775 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775-project-member] Task: {'id': task-1111311, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.444342] env[63515]: DEBUG oslo_vmware.api [None req-b3d6c90d-bdbe-42fc-8e37-795c665e1a37 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Task: {'id': task-1111309, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.529584] env[63515]: DEBUG oslo_concurrency.lockutils [None req-78f15907-fa94-44dd-a2b3-ccdc7dec0a0f tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.412s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 814.532332] env[63515]: DEBUG oslo_concurrency.lockutils [None req-ca1a9b4c-30ff-4beb-9aec-3fa3a8a9ff12 tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 36.680s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 814.532570] env[63515]: DEBUG nova.objects.instance [None req-ca1a9b4c-30ff-4beb-9aec-3fa3a8a9ff12 tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Lazy-loading 'resources' on Instance uuid b4477e66-ae12-4929-90ed-b7b652e0f207 {{(pid=63515) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 814.562858] env[63515]: DEBUG oslo_vmware.api [None req-41b4a543-33b4-43d9-863c-4344bf9d0a46 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Task: {'id': task-1111307, 'name': PowerOnVM_Task, 'duration_secs': 0.636568} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 814.563165] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-41b4a543-33b4-43d9-863c-4344bf9d0a46 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: d926c699-a64a-4942-9ef4-f0166414661d] Powered on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 814.564947] env[63515]: INFO nova.compute.manager [None req-41b4a543-33b4-43d9-863c-4344bf9d0a46 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: d926c699-a64a-4942-9ef4-f0166414661d] Took 10.68 seconds to spawn the instance on the hypervisor. [ 814.564947] env[63515]: DEBUG nova.compute.manager [None req-41b4a543-33b4-43d9-863c-4344bf9d0a46 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: d926c699-a64a-4942-9ef4-f0166414661d] Checking state {{(pid=63515) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 814.565542] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0dd9eb17-eb2d-42eb-8865-f32e7e130f20 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.897266] env[63515]: DEBUG oslo_vmware.api [None req-5a74a6fe-271a-406d-93f5-9b098b35317a tempest-FloatingIPsAssociationNegativeTestJSON-2063029775 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775-project-member] Task: {'id': task-1111311, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.232038} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 814.897266] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-5a74a6fe-271a-406d-93f5-9b098b35317a tempest-FloatingIPsAssociationNegativeTestJSON-2063029775 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775-project-member] Deleted the datastore file {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 814.897266] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-5a74a6fe-271a-406d-93f5-9b098b35317a tempest-FloatingIPsAssociationNegativeTestJSON-2063029775 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775-project-member] [instance: 4e453127-1f3e-40ea-819f-6678479826c8] Deleted contents of the VM from datastore datastore1 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 814.897266] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-5a74a6fe-271a-406d-93f5-9b098b35317a tempest-FloatingIPsAssociationNegativeTestJSON-2063029775 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775-project-member] [instance: 4e453127-1f3e-40ea-819f-6678479826c8] Instance destroyed {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 814.897266] env[63515]: INFO nova.compute.manager [None req-5a74a6fe-271a-406d-93f5-9b098b35317a tempest-FloatingIPsAssociationNegativeTestJSON-2063029775 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775-project-member] [instance: 4e453127-1f3e-40ea-819f-6678479826c8] Took 1.16 seconds to destroy the instance on the hypervisor. [ 814.897266] env[63515]: DEBUG oslo.service.loopingcall [None req-5a74a6fe-271a-406d-93f5-9b098b35317a tempest-FloatingIPsAssociationNegativeTestJSON-2063029775 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 814.897266] env[63515]: DEBUG nova.compute.manager [-] [instance: 4e453127-1f3e-40ea-819f-6678479826c8] Deallocating network for instance {{(pid=63515) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 814.897266] env[63515]: DEBUG nova.network.neutron [-] [instance: 4e453127-1f3e-40ea-819f-6678479826c8] deallocate_for_instance() {{(pid=63515) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 814.944626] env[63515]: DEBUG oslo_vmware.api [None req-b3d6c90d-bdbe-42fc-8e37-795c665e1a37 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Task: {'id': task-1111309, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.561918} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 814.944920] env[63515]: INFO nova.virt.vmwareapi.ds_util [None req-b3d6c90d-bdbe-42fc-8e37-795c665e1a37 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Copied virtual disk from [datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk to [datastore1] 5cbce760-0163-4b27-8ae3-e46c926c8916/8a120570-cb06-4099-b262-554ca0ad15c5-rescue.vmdk. [ 814.945762] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68e2aa23-ec75-405e-8baa-6c1d0cc77685 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.975455] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-b3d6c90d-bdbe-42fc-8e37-795c665e1a37 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] [instance: 5cbce760-0163-4b27-8ae3-e46c926c8916] Reconfiguring VM instance instance-0000003a to attach disk [datastore1] 5cbce760-0163-4b27-8ae3-e46c926c8916/8a120570-cb06-4099-b262-554ca0ad15c5-rescue.vmdk or device None with type thin {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 814.975925] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-be01034b-a61e-4dbf-ac1c-afce4e6710b5 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.997261] env[63515]: DEBUG oslo_vmware.api [None req-b3d6c90d-bdbe-42fc-8e37-795c665e1a37 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Waiting for the task: (returnval){ [ 814.997261] env[63515]: value = "task-1111312" [ 814.997261] env[63515]: _type = "Task" [ 814.997261] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 815.007406] env[63515]: DEBUG oslo_vmware.api [None req-b3d6c90d-bdbe-42fc-8e37-795c665e1a37 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Task: {'id': task-1111312, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.043685] env[63515]: DEBUG oslo_concurrency.lockutils [None req-78f15907-fa94-44dd-a2b3-ccdc7dec0a0f tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Lock "87c468d9-9594-4804-b461-527f01f6118f" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 61.423s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 815.044611] env[63515]: DEBUG oslo_concurrency.lockutils [None req-ce131312-5bf5-4677-8923-2d91405ade5d tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Lock "87c468d9-9594-4804-b461-527f01f6118f" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 36.136s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 815.048032] env[63515]: INFO nova.compute.manager [None req-ce131312-5bf5-4677-8923-2d91405ade5d tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] [instance: 87c468d9-9594-4804-b461-527f01f6118f] Unshelving [ 815.073334] env[63515]: DEBUG oslo_concurrency.lockutils [None req-f30dfc84-d4b1-4816-9900-ab0daf6eae41 tempest-ServersTestJSON-324454119 tempest-ServersTestJSON-324454119-project-member] Acquiring lock "70ed982f-affd-4dd1-bc90-c64e7c6d49d2" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 815.073659] env[63515]: DEBUG oslo_concurrency.lockutils [None req-f30dfc84-d4b1-4816-9900-ab0daf6eae41 tempest-ServersTestJSON-324454119 tempest-ServersTestJSON-324454119-project-member] Lock "70ed982f-affd-4dd1-bc90-c64e7c6d49d2" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 815.073887] env[63515]: DEBUG oslo_concurrency.lockutils [None req-f30dfc84-d4b1-4816-9900-ab0daf6eae41 tempest-ServersTestJSON-324454119 tempest-ServersTestJSON-324454119-project-member] Acquiring lock "70ed982f-affd-4dd1-bc90-c64e7c6d49d2-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 815.074084] env[63515]: DEBUG oslo_concurrency.lockutils [None req-f30dfc84-d4b1-4816-9900-ab0daf6eae41 tempest-ServersTestJSON-324454119 tempest-ServersTestJSON-324454119-project-member] Lock "70ed982f-affd-4dd1-bc90-c64e7c6d49d2-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 815.074261] env[63515]: DEBUG oslo_concurrency.lockutils [None req-f30dfc84-d4b1-4816-9900-ab0daf6eae41 tempest-ServersTestJSON-324454119 tempest-ServersTestJSON-324454119-project-member] Lock "70ed982f-affd-4dd1-bc90-c64e7c6d49d2-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 815.095856] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-6974871c-fbdf-4315-ae5e-0f262a5a9f50 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: 4ef8066a-b68c-457b-9964-b1c34bab0fc3] Volume attach. Driver type: vmdk {{(pid=63515) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 815.096421] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-6974871c-fbdf-4315-ae5e-0f262a5a9f50 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: 4ef8066a-b68c-457b-9964-b1c34bab0fc3] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-243474', 'volume_id': '6112b2eb-44db-4a56-8191-4ff5c6765dea', 'name': 'volume-6112b2eb-44db-4a56-8191-4ff5c6765dea', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '4ef8066a-b68c-457b-9964-b1c34bab0fc3', 'attached_at': '', 'detached_at': '', 'volume_id': '6112b2eb-44db-4a56-8191-4ff5c6765dea', 'serial': '6112b2eb-44db-4a56-8191-4ff5c6765dea'} {{(pid=63515) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 815.097548] env[63515]: INFO nova.compute.manager [None req-f30dfc84-d4b1-4816-9900-ab0daf6eae41 tempest-ServersTestJSON-324454119 tempest-ServersTestJSON-324454119-project-member] [instance: 70ed982f-affd-4dd1-bc90-c64e7c6d49d2] Terminating instance [ 815.104360] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-660ffbaf-ab70-44a8-9db9-bd1f37083a41 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.113935] env[63515]: DEBUG nova.compute.manager [None req-f30dfc84-d4b1-4816-9900-ab0daf6eae41 tempest-ServersTestJSON-324454119 tempest-ServersTestJSON-324454119-project-member] [instance: 70ed982f-affd-4dd1-bc90-c64e7c6d49d2] Start destroying the instance on the hypervisor. {{(pid=63515) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 815.114358] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-f30dfc84-d4b1-4816-9900-ab0daf6eae41 tempest-ServersTestJSON-324454119 tempest-ServersTestJSON-324454119-project-member] [instance: 70ed982f-affd-4dd1-bc90-c64e7c6d49d2] Destroying instance {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 815.116164] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0fe363bc-ed38-4e1a-84f7-fba4ad9186e7 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.121550] env[63515]: INFO nova.compute.manager [None req-41b4a543-33b4-43d9-863c-4344bf9d0a46 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: d926c699-a64a-4942-9ef4-f0166414661d] Took 48.37 seconds to build instance. [ 815.146160] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-790930d6-5a03-48d2-8b72-dc3da4368b40 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.148952] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-f30dfc84-d4b1-4816-9900-ab0daf6eae41 tempest-ServersTestJSON-324454119 tempest-ServersTestJSON-324454119-project-member] [instance: 70ed982f-affd-4dd1-bc90-c64e7c6d49d2] Powering off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 815.150840] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-dc403229-262e-407a-90b9-8f68260c4387 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.177184] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-6974871c-fbdf-4315-ae5e-0f262a5a9f50 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: 4ef8066a-b68c-457b-9964-b1c34bab0fc3] Reconfiguring VM instance instance-0000003e to attach disk [datastore2] volume-6112b2eb-44db-4a56-8191-4ff5c6765dea/volume-6112b2eb-44db-4a56-8191-4ff5c6765dea.vmdk or device None with type thin {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 815.185026] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5db88a2a-79af-4fd8-8988-d79c20fd9c75 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.195087] env[63515]: DEBUG oslo_vmware.api [None req-f30dfc84-d4b1-4816-9900-ab0daf6eae41 tempest-ServersTestJSON-324454119 tempest-ServersTestJSON-324454119-project-member] Waiting for the task: (returnval){ [ 815.195087] env[63515]: value = "task-1111313" [ 815.195087] env[63515]: _type = "Task" [ 815.195087] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 815.203679] env[63515]: DEBUG oslo_vmware.api [None req-6974871c-fbdf-4315-ae5e-0f262a5a9f50 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Waiting for the task: (returnval){ [ 815.203679] env[63515]: value = "task-1111314" [ 815.203679] env[63515]: _type = "Task" [ 815.203679] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 815.206908] env[63515]: DEBUG oslo_vmware.api [None req-f30dfc84-d4b1-4816-9900-ab0daf6eae41 tempest-ServersTestJSON-324454119 tempest-ServersTestJSON-324454119-project-member] Task: {'id': task-1111313, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.216616] env[63515]: DEBUG oslo_vmware.api [None req-6974871c-fbdf-4315-ae5e-0f262a5a9f50 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Task: {'id': task-1111314, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.492190] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7625b88-62d7-4f86-b39c-f0c6d21d86a3 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.502959] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46ac5b14-c0ac-4848-8ad7-b7d06937d701 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.511433] env[63515]: DEBUG oslo_vmware.api [None req-b3d6c90d-bdbe-42fc-8e37-795c665e1a37 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Task: {'id': task-1111312, 'name': ReconfigVM_Task, 'duration_secs': 0.406139} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 815.540029] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-b3d6c90d-bdbe-42fc-8e37-795c665e1a37 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] [instance: 5cbce760-0163-4b27-8ae3-e46c926c8916] Reconfigured VM instance instance-0000003a to attach disk [datastore1] 5cbce760-0163-4b27-8ae3-e46c926c8916/8a120570-cb06-4099-b262-554ca0ad15c5-rescue.vmdk or device None with type thin {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 815.540620] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-875539cd-4fb0-4e41-b1ac-a482c1fda7e2 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.544373] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be68e6fc-7206-46bc-ac23-92149db4254d {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.577456] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f81180cb-55b2-4b6e-abc9-72b6f08e09ca {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.589834] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50d4ac94-55f8-4a09-ac27-40fa996966cc {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.600476] env[63515]: DEBUG oslo_vmware.api [None req-b3d6c90d-bdbe-42fc-8e37-795c665e1a37 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Waiting for the task: (returnval){ [ 815.600476] env[63515]: value = "task-1111315" [ 815.600476] env[63515]: _type = "Task" [ 815.600476] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 815.609556] env[63515]: DEBUG nova.compute.provider_tree [None req-ca1a9b4c-30ff-4beb-9aec-3fa3a8a9ff12 tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 815.621151] env[63515]: DEBUG oslo_vmware.api [None req-b3d6c90d-bdbe-42fc-8e37-795c665e1a37 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Task: {'id': task-1111315, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.627530] env[63515]: DEBUG oslo_concurrency.lockutils [None req-41b4a543-33b4-43d9-863c-4344bf9d0a46 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Lock "d926c699-a64a-4942-9ef4-f0166414661d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 53.071s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 815.705781] env[63515]: DEBUG oslo_vmware.api [None req-f30dfc84-d4b1-4816-9900-ab0daf6eae41 tempest-ServersTestJSON-324454119 tempest-ServersTestJSON-324454119-project-member] Task: {'id': task-1111313, 'name': PowerOffVM_Task, 'duration_secs': 0.307404} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 815.706116] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-f30dfc84-d4b1-4816-9900-ab0daf6eae41 tempest-ServersTestJSON-324454119 tempest-ServersTestJSON-324454119-project-member] [instance: 70ed982f-affd-4dd1-bc90-c64e7c6d49d2] Powered off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 815.706295] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-f30dfc84-d4b1-4816-9900-ab0daf6eae41 tempest-ServersTestJSON-324454119 tempest-ServersTestJSON-324454119-project-member] [instance: 70ed982f-affd-4dd1-bc90-c64e7c6d49d2] Unregistering the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 815.706575] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-4f505352-88f4-455c-8b4f-ea7efca488ec {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.719341] env[63515]: DEBUG oslo_vmware.api [None req-6974871c-fbdf-4315-ae5e-0f262a5a9f50 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Task: {'id': task-1111314, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.786246] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-f30dfc84-d4b1-4816-9900-ab0daf6eae41 tempest-ServersTestJSON-324454119 tempest-ServersTestJSON-324454119-project-member] [instance: 70ed982f-affd-4dd1-bc90-c64e7c6d49d2] Unregistered the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 815.786735] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-f30dfc84-d4b1-4816-9900-ab0daf6eae41 tempest-ServersTestJSON-324454119 tempest-ServersTestJSON-324454119-project-member] [instance: 70ed982f-affd-4dd1-bc90-c64e7c6d49d2] Deleting contents of the VM from datastore datastore2 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 815.787037] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-f30dfc84-d4b1-4816-9900-ab0daf6eae41 tempest-ServersTestJSON-324454119 tempest-ServersTestJSON-324454119-project-member] Deleting the datastore file [datastore2] 70ed982f-affd-4dd1-bc90-c64e7c6d49d2 {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 815.787441] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-65e78915-b431-41b5-87e2-fd560e458753 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.796840] env[63515]: DEBUG oslo_vmware.api [None req-f30dfc84-d4b1-4816-9900-ab0daf6eae41 tempest-ServersTestJSON-324454119 tempest-ServersTestJSON-324454119-project-member] Waiting for the task: (returnval){ [ 815.796840] env[63515]: value = "task-1111317" [ 815.796840] env[63515]: _type = "Task" [ 815.796840] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 815.809426] env[63515]: DEBUG oslo_vmware.api [None req-f30dfc84-d4b1-4816-9900-ab0daf6eae41 tempest-ServersTestJSON-324454119 tempest-ServersTestJSON-324454119-project-member] Task: {'id': task-1111317, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.832968] env[63515]: DEBUG nova.network.neutron [-] [instance: 4e453127-1f3e-40ea-819f-6678479826c8] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 816.072275] env[63515]: DEBUG nova.compute.utils [None req-ce131312-5bf5-4677-8923-2d91405ade5d tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Using /dev/sd instead of None {{(pid=63515) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 816.098340] env[63515]: DEBUG nova.compute.manager [req-2f2c315c-85b2-4784-8da0-040aad543548 req-ccf097eb-08fd-4144-9462-e45a5912117d service nova] [instance: 4e453127-1f3e-40ea-819f-6678479826c8] Received event network-vif-deleted-3d5ad33f-8649-42c8-b899-ca20c9a2636e {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 816.116890] env[63515]: DEBUG nova.scheduler.client.report [None req-ca1a9b4c-30ff-4beb-9aec-3fa3a8a9ff12 tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 816.125116] env[63515]: DEBUG oslo_vmware.api [None req-b3d6c90d-bdbe-42fc-8e37-795c665e1a37 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Task: {'id': task-1111315, 'name': ReconfigVM_Task, 'duration_secs': 0.179752} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 816.125564] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-b3d6c90d-bdbe-42fc-8e37-795c665e1a37 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] [instance: 5cbce760-0163-4b27-8ae3-e46c926c8916] Powering on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 816.129103] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-90213f8b-073f-4ac0-8e60-f050f460aa78 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.134020] env[63515]: DEBUG oslo_vmware.api [None req-b3d6c90d-bdbe-42fc-8e37-795c665e1a37 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Waiting for the task: (returnval){ [ 816.134020] env[63515]: value = "task-1111318" [ 816.134020] env[63515]: _type = "Task" [ 816.134020] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 816.145296] env[63515]: DEBUG oslo_vmware.api [None req-b3d6c90d-bdbe-42fc-8e37-795c665e1a37 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Task: {'id': task-1111318, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.227689] env[63515]: DEBUG oslo_vmware.api [None req-6974871c-fbdf-4315-ae5e-0f262a5a9f50 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Task: {'id': task-1111314, 'name': ReconfigVM_Task, 'duration_secs': 1.016209} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 816.228505] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-6974871c-fbdf-4315-ae5e-0f262a5a9f50 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: 4ef8066a-b68c-457b-9964-b1c34bab0fc3] Reconfigured VM instance instance-0000003e to attach disk [datastore2] volume-6112b2eb-44db-4a56-8191-4ff5c6765dea/volume-6112b2eb-44db-4a56-8191-4ff5c6765dea.vmdk or device None with type thin {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 816.232915] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3c4c6a8f-3bd6-423a-8591-c779e921cbf6 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.250747] env[63515]: DEBUG oslo_vmware.api [None req-6974871c-fbdf-4315-ae5e-0f262a5a9f50 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Waiting for the task: (returnval){ [ 816.250747] env[63515]: value = "task-1111319" [ 816.250747] env[63515]: _type = "Task" [ 816.250747] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 816.259801] env[63515]: DEBUG oslo_vmware.api [None req-6974871c-fbdf-4315-ae5e-0f262a5a9f50 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Task: {'id': task-1111319, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.308557] env[63515]: DEBUG oslo_vmware.api [None req-f30dfc84-d4b1-4816-9900-ab0daf6eae41 tempest-ServersTestJSON-324454119 tempest-ServersTestJSON-324454119-project-member] Task: {'id': task-1111317, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.168297} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 816.308741] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-f30dfc84-d4b1-4816-9900-ab0daf6eae41 tempest-ServersTestJSON-324454119 tempest-ServersTestJSON-324454119-project-member] Deleted the datastore file {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 816.308922] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-f30dfc84-d4b1-4816-9900-ab0daf6eae41 tempest-ServersTestJSON-324454119 tempest-ServersTestJSON-324454119-project-member] [instance: 70ed982f-affd-4dd1-bc90-c64e7c6d49d2] Deleted contents of the VM from datastore datastore2 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 816.308985] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-f30dfc84-d4b1-4816-9900-ab0daf6eae41 tempest-ServersTestJSON-324454119 tempest-ServersTestJSON-324454119-project-member] [instance: 70ed982f-affd-4dd1-bc90-c64e7c6d49d2] Instance destroyed {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 816.309177] env[63515]: INFO nova.compute.manager [None req-f30dfc84-d4b1-4816-9900-ab0daf6eae41 tempest-ServersTestJSON-324454119 tempest-ServersTestJSON-324454119-project-member] [instance: 70ed982f-affd-4dd1-bc90-c64e7c6d49d2] Took 1.20 seconds to destroy the instance on the hypervisor. [ 816.309424] env[63515]: DEBUG oslo.service.loopingcall [None req-f30dfc84-d4b1-4816-9900-ab0daf6eae41 tempest-ServersTestJSON-324454119 tempest-ServersTestJSON-324454119-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 816.309623] env[63515]: DEBUG nova.compute.manager [-] [instance: 70ed982f-affd-4dd1-bc90-c64e7c6d49d2] Deallocating network for instance {{(pid=63515) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 816.309735] env[63515]: DEBUG nova.network.neutron [-] [instance: 70ed982f-affd-4dd1-bc90-c64e7c6d49d2] deallocate_for_instance() {{(pid=63515) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 816.336791] env[63515]: INFO nova.compute.manager [-] [instance: 4e453127-1f3e-40ea-819f-6678479826c8] Took 1.44 seconds to deallocate network for instance. [ 816.579040] env[63515]: INFO nova.virt.block_device [None req-ce131312-5bf5-4677-8923-2d91405ade5d tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] [instance: 87c468d9-9594-4804-b461-527f01f6118f] Booting with volume 67893a17-820c-4c63-b387-06761ba8e0d5 at /dev/sdb [ 816.627120] env[63515]: DEBUG oslo_concurrency.lockutils [None req-ca1a9b4c-30ff-4beb-9aec-3fa3a8a9ff12 tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.095s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 816.634477] env[63515]: DEBUG oslo_concurrency.lockutils [None req-f8bace5a-3f38-4490-8a6d-e315321b7f63 tempest-ServersTestManualDisk-1114050206 tempest-ServersTestManualDisk-1114050206-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 37.591s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 816.634564] env[63515]: DEBUG nova.objects.instance [None req-f8bace5a-3f38-4490-8a6d-e315321b7f63 tempest-ServersTestManualDisk-1114050206 tempest-ServersTestManualDisk-1114050206-project-member] Lazy-loading 'resources' on Instance uuid e91aa479-1540-4950-851b-b2409e5f89f1 {{(pid=63515) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 816.636477] env[63515]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-fff79248-cb4e-4858-8b3d-20ee1d2c2acc {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.655929] env[63515]: INFO nova.scheduler.client.report [None req-ca1a9b4c-30ff-4beb-9aec-3fa3a8a9ff12 tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Deleted allocations for instance b4477e66-ae12-4929-90ed-b7b652e0f207 [ 816.657330] env[63515]: DEBUG oslo_vmware.api [None req-b3d6c90d-bdbe-42fc-8e37-795c665e1a37 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Task: {'id': task-1111318, 'name': PowerOnVM_Task, 'duration_secs': 0.464225} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 816.661538] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-b3d6c90d-bdbe-42fc-8e37-795c665e1a37 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] [instance: 5cbce760-0163-4b27-8ae3-e46c926c8916] Powered on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 816.667522] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dadbd573-0af7-4896-9220-23a632d48a31 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.680266] env[63515]: DEBUG nova.compute.manager [None req-b3d6c90d-bdbe-42fc-8e37-795c665e1a37 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] [instance: 5cbce760-0163-4b27-8ae3-e46c926c8916] Checking state {{(pid=63515) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 816.684140] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db9ce494-a1cd-4fa4-b566-2fd7f2bdac95 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.712165] env[63515]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-44020b7a-6297-4471-ba07-31e7c9aabfd8 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.725751] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-869061bf-929c-4b0e-b5d5-2b7d89b80c57 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.766952] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4e5f940-4836-4141-b819-df52780570b5 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.779308] env[63515]: DEBUG oslo_vmware.api [None req-6974871c-fbdf-4315-ae5e-0f262a5a9f50 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Task: {'id': task-1111319, 'name': ReconfigVM_Task, 'duration_secs': 0.297273} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 816.780467] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-6974871c-fbdf-4315-ae5e-0f262a5a9f50 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: 4ef8066a-b68c-457b-9964-b1c34bab0fc3] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-243474', 'volume_id': '6112b2eb-44db-4a56-8191-4ff5c6765dea', 'name': 'volume-6112b2eb-44db-4a56-8191-4ff5c6765dea', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '4ef8066a-b68c-457b-9964-b1c34bab0fc3', 'attached_at': '', 'detached_at': '', 'volume_id': '6112b2eb-44db-4a56-8191-4ff5c6765dea', 'serial': '6112b2eb-44db-4a56-8191-4ff5c6765dea'} {{(pid=63515) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 816.785215] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-365d836b-4846-4277-a7e4-6b17093f53bb {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.802544] env[63515]: DEBUG nova.virt.block_device [None req-ce131312-5bf5-4677-8923-2d91405ade5d tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] [instance: 87c468d9-9594-4804-b461-527f01f6118f] Updating existing volume attachment record: 67842379-9552-4358-b269-226527568f79 {{(pid=63515) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 816.843717] env[63515]: DEBUG oslo_concurrency.lockutils [None req-5a74a6fe-271a-406d-93f5-9b098b35317a tempest-FloatingIPsAssociationNegativeTestJSON-2063029775 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 817.166353] env[63515]: DEBUG oslo_concurrency.lockutils [None req-ca1a9b4c-30ff-4beb-9aec-3fa3a8a9ff12 tempest-ServerShowV247Test-1216424191 tempest-ServerShowV247Test-1216424191-project-member] Lock "b4477e66-ae12-4929-90ed-b7b652e0f207" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 43.124s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 817.257609] env[63515]: DEBUG nova.compute.manager [None req-4cdfd447-25bb-42ff-81df-85503f57f660 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: d926c699-a64a-4942-9ef4-f0166414661d] Stashing vm_state: active {{(pid=63515) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 817.482882] env[63515]: DEBUG nova.network.neutron [-] [instance: 70ed982f-affd-4dd1-bc90-c64e7c6d49d2] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 817.495972] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d40c272b-89e1-4ce9-9c66-e0326fc9d67b {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.504539] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-236355e1-924d-4622-9719-22a3b57f50df {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.557124] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1d5d48a-8f50-4993-ac2c-8607239b6533 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.564171] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8aef695-e984-4ebe-9e19-c8fd59510daf {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.580892] env[63515]: DEBUG nova.compute.provider_tree [None req-f8bace5a-3f38-4490-8a6d-e315321b7f63 tempest-ServersTestManualDisk-1114050206 tempest-ServersTestManualDisk-1114050206-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 817.782046] env[63515]: DEBUG oslo_concurrency.lockutils [None req-4cdfd447-25bb-42ff-81df-85503f57f660 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 817.824436] env[63515]: DEBUG nova.objects.instance [None req-6974871c-fbdf-4315-ae5e-0f262a5a9f50 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Lazy-loading 'flavor' on Instance uuid 4ef8066a-b68c-457b-9964-b1c34bab0fc3 {{(pid=63515) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 817.990615] env[63515]: INFO nova.compute.manager [-] [instance: 70ed982f-affd-4dd1-bc90-c64e7c6d49d2] Took 1.68 seconds to deallocate network for instance. [ 818.085165] env[63515]: DEBUG nova.scheduler.client.report [None req-f8bace5a-3f38-4490-8a6d-e315321b7f63 tempest-ServersTestManualDisk-1114050206 tempest-ServersTestManualDisk-1114050206-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 818.319968] env[63515]: DEBUG nova.compute.manager [req-40b67729-39a7-4361-9ec1-08b1fc12c3ea req-12af560f-0a29-4130-b0ec-f89a15acccc6 service nova] [instance: 70ed982f-affd-4dd1-bc90-c64e7c6d49d2] Received event network-vif-deleted-fe7c09da-d6bd-4f72-afa2-1a70bc663407 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 818.331720] env[63515]: DEBUG oslo_concurrency.lockutils [None req-6974871c-fbdf-4315-ae5e-0f262a5a9f50 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Lock "4ef8066a-b68c-457b-9964-b1c34bab0fc3" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 8.370s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 818.500686] env[63515]: DEBUG oslo_concurrency.lockutils [None req-f30dfc84-d4b1-4816-9900-ab0daf6eae41 tempest-ServersTestJSON-324454119 tempest-ServersTestJSON-324454119-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 818.549017] env[63515]: INFO nova.compute.manager [None req-477817e6-ef49-4acf-9d39-177b03b7a885 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] [instance: 5cbce760-0163-4b27-8ae3-e46c926c8916] Unrescuing [ 818.549340] env[63515]: DEBUG oslo_concurrency.lockutils [None req-477817e6-ef49-4acf-9d39-177b03b7a885 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Acquiring lock "refresh_cache-5cbce760-0163-4b27-8ae3-e46c926c8916" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 818.549497] env[63515]: DEBUG oslo_concurrency.lockutils [None req-477817e6-ef49-4acf-9d39-177b03b7a885 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Acquired lock "refresh_cache-5cbce760-0163-4b27-8ae3-e46c926c8916" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 818.549670] env[63515]: DEBUG nova.network.neutron [None req-477817e6-ef49-4acf-9d39-177b03b7a885 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] [instance: 5cbce760-0163-4b27-8ae3-e46c926c8916] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 818.597060] env[63515]: DEBUG oslo_concurrency.lockutils [None req-f8bace5a-3f38-4490-8a6d-e315321b7f63 tempest-ServersTestManualDisk-1114050206 tempest-ServersTestManualDisk-1114050206-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.963s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 818.600125] env[63515]: DEBUG oslo_concurrency.lockutils [None req-b19ee0f2-2420-400a-b8e9-86908fd35c93 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Acquiring lock "4ef8066a-b68c-457b-9964-b1c34bab0fc3" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 818.600388] env[63515]: DEBUG oslo_concurrency.lockutils [None req-b19ee0f2-2420-400a-b8e9-86908fd35c93 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Lock "4ef8066a-b68c-457b-9964-b1c34bab0fc3" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 818.600989] env[63515]: DEBUG oslo_concurrency.lockutils [None req-b19ee0f2-2420-400a-b8e9-86908fd35c93 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Acquiring lock "4ef8066a-b68c-457b-9964-b1c34bab0fc3-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 818.600989] env[63515]: DEBUG oslo_concurrency.lockutils [None req-b19ee0f2-2420-400a-b8e9-86908fd35c93 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Lock "4ef8066a-b68c-457b-9964-b1c34bab0fc3-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 818.600989] env[63515]: DEBUG oslo_concurrency.lockutils [None req-b19ee0f2-2420-400a-b8e9-86908fd35c93 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Lock "4ef8066a-b68c-457b-9964-b1c34bab0fc3-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 818.604495] env[63515]: DEBUG oslo_concurrency.lockutils [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 35.775s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 818.604495] env[63515]: DEBUG oslo_concurrency.lockutils [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 818.604495] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=63515) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 818.604495] env[63515]: DEBUG oslo_concurrency.lockutils [None req-49d5860e-9bde-4499-902f-a44ae9cea035 tempest-ServerMetadataNegativeTestJSON-502053314 tempest-ServerMetadataNegativeTestJSON-502053314-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 32.510s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 818.605959] env[63515]: INFO nova.compute.claims [None req-49d5860e-9bde-4499-902f-a44ae9cea035 tempest-ServerMetadataNegativeTestJSON-502053314 tempest-ServerMetadataNegativeTestJSON-502053314-project-member] [instance: 9d18c7b6-ce8e-4042-9e70-696bb7f57cb3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 818.609293] env[63515]: INFO nova.compute.manager [None req-b19ee0f2-2420-400a-b8e9-86908fd35c93 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: 4ef8066a-b68c-457b-9964-b1c34bab0fc3] Terminating instance [ 818.613020] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54f4d95d-7bc2-42db-8249-aabc702f9ce3 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.614377] env[63515]: DEBUG nova.compute.manager [None req-b19ee0f2-2420-400a-b8e9-86908fd35c93 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: 4ef8066a-b68c-457b-9964-b1c34bab0fc3] Start destroying the instance on the hypervisor. {{(pid=63515) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 818.614585] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-b19ee0f2-2420-400a-b8e9-86908fd35c93 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: 4ef8066a-b68c-457b-9964-b1c34bab0fc3] Powering off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 818.614860] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-61ee63e9-8ed2-4c73-878d-1c47b5ac3482 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.625617] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-109dca2f-0285-4a0c-b592-fd94765d73e1 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.631031] env[63515]: DEBUG oslo_vmware.api [None req-b19ee0f2-2420-400a-b8e9-86908fd35c93 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Waiting for the task: (returnval){ [ 818.631031] env[63515]: value = "task-1111323" [ 818.631031] env[63515]: _type = "Task" [ 818.631031] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 818.646245] env[63515]: INFO nova.scheduler.client.report [None req-f8bace5a-3f38-4490-8a6d-e315321b7f63 tempest-ServersTestManualDisk-1114050206 tempest-ServersTestManualDisk-1114050206-project-member] Deleted allocations for instance e91aa479-1540-4950-851b-b2409e5f89f1 [ 818.647324] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75e4125f-4601-47b5-9604-9cb105ed2dc6 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.656936] env[63515]: DEBUG oslo_vmware.api [None req-b19ee0f2-2420-400a-b8e9-86908fd35c93 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Task: {'id': task-1111323, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.667625] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31f70f66-3776-48c0-9f47-5783c7c08e91 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.701773] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180501MB free_disk=171GB free_vcpus=48 pci_devices=None {{(pid=63515) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 818.702091] env[63515]: DEBUG oslo_concurrency.lockutils [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 819.146409] env[63515]: DEBUG oslo_vmware.api [None req-b19ee0f2-2420-400a-b8e9-86908fd35c93 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Task: {'id': task-1111323, 'name': PowerOffVM_Task, 'duration_secs': 0.231097} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 819.146822] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-b19ee0f2-2420-400a-b8e9-86908fd35c93 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: 4ef8066a-b68c-457b-9964-b1c34bab0fc3] Powered off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 819.147126] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-b19ee0f2-2420-400a-b8e9-86908fd35c93 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: 4ef8066a-b68c-457b-9964-b1c34bab0fc3] Volume detach. Driver type: vmdk {{(pid=63515) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 819.147403] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-b19ee0f2-2420-400a-b8e9-86908fd35c93 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: 4ef8066a-b68c-457b-9964-b1c34bab0fc3] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-243474', 'volume_id': '6112b2eb-44db-4a56-8191-4ff5c6765dea', 'name': 'volume-6112b2eb-44db-4a56-8191-4ff5c6765dea', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '4ef8066a-b68c-457b-9964-b1c34bab0fc3', 'attached_at': '', 'detached_at': '', 'volume_id': '6112b2eb-44db-4a56-8191-4ff5c6765dea', 'serial': '6112b2eb-44db-4a56-8191-4ff5c6765dea'} {{(pid=63515) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 819.151402] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ade6b27-a83d-44d1-88ab-54db94e709ee {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.181752] env[63515]: DEBUG oslo_concurrency.lockutils [None req-f8bace5a-3f38-4490-8a6d-e315321b7f63 tempest-ServersTestManualDisk-1114050206 tempest-ServersTestManualDisk-1114050206-project-member] Lock "e91aa479-1540-4950-851b-b2409e5f89f1" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 43.783s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 819.182728] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce1a6865-f9a8-4a56-ac67-3cf8b6acb6b8 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.194255] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da18509b-a825-42e3-8745-512b066b61bc {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.225909] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80108755-4b8a-42c4-a61d-f6f8730cc24a {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.252373] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-b19ee0f2-2420-400a-b8e9-86908fd35c93 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] The volume has not been displaced from its original location: [datastore2] volume-6112b2eb-44db-4a56-8191-4ff5c6765dea/volume-6112b2eb-44db-4a56-8191-4ff5c6765dea.vmdk. No consolidation needed. {{(pid=63515) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 819.258574] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-b19ee0f2-2420-400a-b8e9-86908fd35c93 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: 4ef8066a-b68c-457b-9964-b1c34bab0fc3] Reconfiguring VM instance instance-0000003e to detach disk 2001 {{(pid=63515) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 819.261351] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c85b3663-3b5e-4c59-bbeb-bace1e17ba5c {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.281667] env[63515]: DEBUG oslo_vmware.api [None req-b19ee0f2-2420-400a-b8e9-86908fd35c93 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Waiting for the task: (returnval){ [ 819.281667] env[63515]: value = "task-1111325" [ 819.281667] env[63515]: _type = "Task" [ 819.281667] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 819.293412] env[63515]: DEBUG oslo_vmware.api [None req-b19ee0f2-2420-400a-b8e9-86908fd35c93 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Task: {'id': task-1111325, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.430672] env[63515]: DEBUG nova.network.neutron [None req-477817e6-ef49-4acf-9d39-177b03b7a885 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] [instance: 5cbce760-0163-4b27-8ae3-e46c926c8916] Updating instance_info_cache with network_info: [{"id": "5fa807c4-5ce1-492a-bea3-d8f7a65d898f", "address": "fa:16:3e:23:bb:83", "network": {"id": "850bfa95-3df6-4019-8260-ade20ec58799", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-832000501-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "08c9aa51b1864410809f8877c7cd49c4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3cf748a8-7ae0-4dca-817d-e727c30d72f4", "external-id": "nsx-vlan-transportzone-853", "segmentation_id": 853, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5fa807c4-5c", "ovs_interfaceid": "5fa807c4-5ce1-492a-bea3-d8f7a65d898f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 819.796338] env[63515]: DEBUG oslo_vmware.api [None req-b19ee0f2-2420-400a-b8e9-86908fd35c93 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Task: {'id': task-1111325, 'name': ReconfigVM_Task, 'duration_secs': 0.406005} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 819.796738] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-b19ee0f2-2420-400a-b8e9-86908fd35c93 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: 4ef8066a-b68c-457b-9964-b1c34bab0fc3] Reconfigured VM instance instance-0000003e to detach disk 2001 {{(pid=63515) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 819.801918] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-37fe27f5-1871-4f02-af5d-642851974782 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.823110] env[63515]: DEBUG oslo_vmware.api [None req-b19ee0f2-2420-400a-b8e9-86908fd35c93 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Waiting for the task: (returnval){ [ 819.823110] env[63515]: value = "task-1111326" [ 819.823110] env[63515]: _type = "Task" [ 819.823110] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 819.837418] env[63515]: DEBUG oslo_vmware.api [None req-b19ee0f2-2420-400a-b8e9-86908fd35c93 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Task: {'id': task-1111326, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.935047] env[63515]: DEBUG oslo_concurrency.lockutils [None req-477817e6-ef49-4acf-9d39-177b03b7a885 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Releasing lock "refresh_cache-5cbce760-0163-4b27-8ae3-e46c926c8916" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 819.935047] env[63515]: DEBUG nova.objects.instance [None req-477817e6-ef49-4acf-9d39-177b03b7a885 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Lazy-loading 'flavor' on Instance uuid 5cbce760-0163-4b27-8ae3-e46c926c8916 {{(pid=63515) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 819.997464] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84b65a84-cece-46c4-8fdd-b0b87bb5e458 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.010080] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13e30696-5270-4616-975f-d9e1bd6af7b8 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.060187] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6d93f9a-04c2-482d-adcf-29f7ccb484c5 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.072360] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb6b919d-f153-449b-8355-e1871955ac03 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.094360] env[63515]: DEBUG nova.compute.provider_tree [None req-49d5860e-9bde-4499-902f-a44ae9cea035 tempest-ServerMetadataNegativeTestJSON-502053314 tempest-ServerMetadataNegativeTestJSON-502053314-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 820.335871] env[63515]: DEBUG oslo_vmware.api [None req-b19ee0f2-2420-400a-b8e9-86908fd35c93 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Task: {'id': task-1111326, 'name': ReconfigVM_Task, 'duration_secs': 0.434102} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 820.337583] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-b19ee0f2-2420-400a-b8e9-86908fd35c93 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: 4ef8066a-b68c-457b-9964-b1c34bab0fc3] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-243474', 'volume_id': '6112b2eb-44db-4a56-8191-4ff5c6765dea', 'name': 'volume-6112b2eb-44db-4a56-8191-4ff5c6765dea', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '4ef8066a-b68c-457b-9964-b1c34bab0fc3', 'attached_at': '', 'detached_at': '', 'volume_id': '6112b2eb-44db-4a56-8191-4ff5c6765dea', 'serial': '6112b2eb-44db-4a56-8191-4ff5c6765dea'} {{(pid=63515) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 820.338059] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-b19ee0f2-2420-400a-b8e9-86908fd35c93 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: 4ef8066a-b68c-457b-9964-b1c34bab0fc3] Destroying instance {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 820.341018] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ec58b5d-cc68-41f3-946b-de8274af8805 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.348032] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-b19ee0f2-2420-400a-b8e9-86908fd35c93 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: 4ef8066a-b68c-457b-9964-b1c34bab0fc3] Unregistering the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 820.348325] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-746542e2-a1cf-453e-aee0-53f123f74582 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.440662] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-b19ee0f2-2420-400a-b8e9-86908fd35c93 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: 4ef8066a-b68c-457b-9964-b1c34bab0fc3] Unregistered the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 820.440964] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-b19ee0f2-2420-400a-b8e9-86908fd35c93 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: 4ef8066a-b68c-457b-9964-b1c34bab0fc3] Deleting contents of the VM from datastore datastore2 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 820.441143] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-b19ee0f2-2420-400a-b8e9-86908fd35c93 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Deleting the datastore file [datastore2] 4ef8066a-b68c-457b-9964-b1c34bab0fc3 {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 820.441766] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-56340a23-acc9-466a-91fe-693bd7fe304e {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.445222] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d18e3e51-25ae-4dd1-8691-d847a9cad3f5 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.472429] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-477817e6-ef49-4acf-9d39-177b03b7a885 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] [instance: 5cbce760-0163-4b27-8ae3-e46c926c8916] Powering off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 820.474556] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5f232dc2-345b-45ed-ba67-23d9da7a2929 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.475797] env[63515]: DEBUG oslo_vmware.api [None req-b19ee0f2-2420-400a-b8e9-86908fd35c93 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Waiting for the task: (returnval){ [ 820.475797] env[63515]: value = "task-1111328" [ 820.475797] env[63515]: _type = "Task" [ 820.475797] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 820.482101] env[63515]: DEBUG oslo_vmware.api [None req-477817e6-ef49-4acf-9d39-177b03b7a885 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Waiting for the task: (returnval){ [ 820.482101] env[63515]: value = "task-1111329" [ 820.482101] env[63515]: _type = "Task" [ 820.482101] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 820.485716] env[63515]: DEBUG oslo_vmware.api [None req-b19ee0f2-2420-400a-b8e9-86908fd35c93 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Task: {'id': task-1111328, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.495792] env[63515]: DEBUG oslo_vmware.api [None req-477817e6-ef49-4acf-9d39-177b03b7a885 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Task: {'id': task-1111329, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.601024] env[63515]: DEBUG nova.scheduler.client.report [None req-49d5860e-9bde-4499-902f-a44ae9cea035 tempest-ServerMetadataNegativeTestJSON-502053314 tempest-ServerMetadataNegativeTestJSON-502053314-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 820.988790] env[63515]: DEBUG oslo_vmware.api [None req-b19ee0f2-2420-400a-b8e9-86908fd35c93 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Task: {'id': task-1111328, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.180549} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 820.993441] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-b19ee0f2-2420-400a-b8e9-86908fd35c93 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Deleted the datastore file {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 820.993894] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-b19ee0f2-2420-400a-b8e9-86908fd35c93 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: 4ef8066a-b68c-457b-9964-b1c34bab0fc3] Deleted contents of the VM from datastore datastore2 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 820.994165] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-b19ee0f2-2420-400a-b8e9-86908fd35c93 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: 4ef8066a-b68c-457b-9964-b1c34bab0fc3] Instance destroyed {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 820.994464] env[63515]: INFO nova.compute.manager [None req-b19ee0f2-2420-400a-b8e9-86908fd35c93 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: 4ef8066a-b68c-457b-9964-b1c34bab0fc3] Took 2.38 seconds to destroy the instance on the hypervisor. [ 820.994765] env[63515]: DEBUG oslo.service.loopingcall [None req-b19ee0f2-2420-400a-b8e9-86908fd35c93 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 820.995028] env[63515]: DEBUG nova.compute.manager [-] [instance: 4ef8066a-b68c-457b-9964-b1c34bab0fc3] Deallocating network for instance {{(pid=63515) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 820.995328] env[63515]: DEBUG nova.network.neutron [-] [instance: 4ef8066a-b68c-457b-9964-b1c34bab0fc3] deallocate_for_instance() {{(pid=63515) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 821.004227] env[63515]: DEBUG oslo_vmware.api [None req-477817e6-ef49-4acf-9d39-177b03b7a885 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Task: {'id': task-1111329, 'name': PowerOffVM_Task, 'duration_secs': 0.277567} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 821.004698] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-477817e6-ef49-4acf-9d39-177b03b7a885 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] [instance: 5cbce760-0163-4b27-8ae3-e46c926c8916] Powered off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 821.016321] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-477817e6-ef49-4acf-9d39-177b03b7a885 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] [instance: 5cbce760-0163-4b27-8ae3-e46c926c8916] Reconfiguring VM instance instance-0000003a to detach disk 2001 {{(pid=63515) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 821.016747] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4e731279-669a-417f-8571-8539cd863b29 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.038686] env[63515]: DEBUG oslo_vmware.api [None req-477817e6-ef49-4acf-9d39-177b03b7a885 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Waiting for the task: (returnval){ [ 821.038686] env[63515]: value = "task-1111330" [ 821.038686] env[63515]: _type = "Task" [ 821.038686] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 821.048686] env[63515]: DEBUG oslo_vmware.api [None req-477817e6-ef49-4acf-9d39-177b03b7a885 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Task: {'id': task-1111330, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.103933] env[63515]: DEBUG oslo_concurrency.lockutils [None req-49d5860e-9bde-4499-902f-a44ae9cea035 tempest-ServerMetadataNegativeTestJSON-502053314 tempest-ServerMetadataNegativeTestJSON-502053314-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.499s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 821.104502] env[63515]: DEBUG nova.compute.manager [None req-49d5860e-9bde-4499-902f-a44ae9cea035 tempest-ServerMetadataNegativeTestJSON-502053314 tempest-ServerMetadataNegativeTestJSON-502053314-project-member] [instance: 9d18c7b6-ce8e-4042-9e70-696bb7f57cb3] Start building networks asynchronously for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 821.113738] env[63515]: DEBUG oslo_concurrency.lockutils [None req-65badfd4-59e2-4bf6-bac8-39c1a78dd350 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 29.113s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 821.114714] env[63515]: DEBUG oslo_concurrency.lockutils [None req-65badfd4-59e2-4bf6-bac8-39c1a78dd350 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 821.116843] env[63515]: DEBUG oslo_concurrency.lockutils [None req-b53e671a-9cf4-4e34-a2ac-4321dce89b1c tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 23.800s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 821.117102] env[63515]: DEBUG nova.objects.instance [None req-b53e671a-9cf4-4e34-a2ac-4321dce89b1c tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] [instance: 192137b0-03e5-4bc4-b911-4b4f1a874f74] Trying to apply a migration context that does not seem to be set for this instance {{(pid=63515) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 821.163476] env[63515]: INFO nova.scheduler.client.report [None req-65badfd4-59e2-4bf6-bac8-39c1a78dd350 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Deleted allocations for instance f1d01b75-ac9d-458d-8cc2-ae64cffca4e8 [ 821.462223] env[63515]: DEBUG nova.compute.manager [req-9a46d08e-fa6f-4801-8cae-dba788f432dd req-895f1890-d746-4cc2-8758-abc8d0e7275e service nova] [instance: 4ef8066a-b68c-457b-9964-b1c34bab0fc3] Received event network-vif-deleted-f2d49ff8-dbd8-41a2-9006-c71265c36b87 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 821.462223] env[63515]: INFO nova.compute.manager [req-9a46d08e-fa6f-4801-8cae-dba788f432dd req-895f1890-d746-4cc2-8758-abc8d0e7275e service nova] [instance: 4ef8066a-b68c-457b-9964-b1c34bab0fc3] Neutron deleted interface f2d49ff8-dbd8-41a2-9006-c71265c36b87; detaching it from the instance and deleting it from the info cache [ 821.462223] env[63515]: DEBUG nova.network.neutron [req-9a46d08e-fa6f-4801-8cae-dba788f432dd req-895f1890-d746-4cc2-8758-abc8d0e7275e service nova] [instance: 4ef8066a-b68c-457b-9964-b1c34bab0fc3] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 821.550811] env[63515]: DEBUG oslo_vmware.api [None req-477817e6-ef49-4acf-9d39-177b03b7a885 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Task: {'id': task-1111330, 'name': ReconfigVM_Task, 'duration_secs': 0.308968} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 821.551305] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-477817e6-ef49-4acf-9d39-177b03b7a885 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] [instance: 5cbce760-0163-4b27-8ae3-e46c926c8916] Reconfigured VM instance instance-0000003a to detach disk 2001 {{(pid=63515) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 821.551574] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-477817e6-ef49-4acf-9d39-177b03b7a885 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] [instance: 5cbce760-0163-4b27-8ae3-e46c926c8916] Powering on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 821.551821] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f05eb509-edc6-447e-9089-e5be3e407810 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.560048] env[63515]: DEBUG oslo_vmware.api [None req-477817e6-ef49-4acf-9d39-177b03b7a885 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Waiting for the task: (returnval){ [ 821.560048] env[63515]: value = "task-1111334" [ 821.560048] env[63515]: _type = "Task" [ 821.560048] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 821.569996] env[63515]: DEBUG oslo_vmware.api [None req-477817e6-ef49-4acf-9d39-177b03b7a885 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Task: {'id': task-1111334, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.613111] env[63515]: DEBUG nova.compute.utils [None req-49d5860e-9bde-4499-902f-a44ae9cea035 tempest-ServerMetadataNegativeTestJSON-502053314 tempest-ServerMetadataNegativeTestJSON-502053314-project-member] Using /dev/sd instead of None {{(pid=63515) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 821.614270] env[63515]: DEBUG nova.compute.manager [None req-49d5860e-9bde-4499-902f-a44ae9cea035 tempest-ServerMetadataNegativeTestJSON-502053314 tempest-ServerMetadataNegativeTestJSON-502053314-project-member] [instance: 9d18c7b6-ce8e-4042-9e70-696bb7f57cb3] Allocating IP information in the background. {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 821.614495] env[63515]: DEBUG nova.network.neutron [None req-49d5860e-9bde-4499-902f-a44ae9cea035 tempest-ServerMetadataNegativeTestJSON-502053314 tempest-ServerMetadataNegativeTestJSON-502053314-project-member] [instance: 9d18c7b6-ce8e-4042-9e70-696bb7f57cb3] allocate_for_instance() {{(pid=63515) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 821.665234] env[63515]: DEBUG nova.policy [None req-49d5860e-9bde-4499-902f-a44ae9cea035 tempest-ServerMetadataNegativeTestJSON-502053314 tempest-ServerMetadataNegativeTestJSON-502053314-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '1ccd71618d7d49099ec147878a60299b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'bc5e67e8d1364503bacb15aec38d90d3', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63515) authorize /opt/stack/nova/nova/policy.py:201}} [ 821.673222] env[63515]: DEBUG oslo_concurrency.lockutils [None req-65badfd4-59e2-4bf6-bac8-39c1a78dd350 tempest-MigrationsAdminTest-1743999803 tempest-MigrationsAdminTest-1743999803-project-member] Lock "f1d01b75-ac9d-458d-8cc2-ae64cffca4e8" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 32.763s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 821.933342] env[63515]: DEBUG nova.network.neutron [-] [instance: 4ef8066a-b68c-457b-9964-b1c34bab0fc3] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 821.964609] env[63515]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-bf2bfc7a-49cb-4cbc-8d74-019a954a1623 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.974364] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33fd7a14-2eb2-482b-a2ce-04bddfa8384e {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.003647] env[63515]: DEBUG nova.compute.manager [req-9a46d08e-fa6f-4801-8cae-dba788f432dd req-895f1890-d746-4cc2-8758-abc8d0e7275e service nova] [instance: 4ef8066a-b68c-457b-9964-b1c34bab0fc3] Detach interface failed, port_id=f2d49ff8-dbd8-41a2-9006-c71265c36b87, reason: Instance 4ef8066a-b68c-457b-9964-b1c34bab0fc3 could not be found. {{(pid=63515) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 822.073758] env[63515]: DEBUG oslo_vmware.api [None req-477817e6-ef49-4acf-9d39-177b03b7a885 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Task: {'id': task-1111334, 'name': PowerOnVM_Task, 'duration_secs': 0.374487} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 822.075136] env[63515]: DEBUG nova.network.neutron [None req-49d5860e-9bde-4499-902f-a44ae9cea035 tempest-ServerMetadataNegativeTestJSON-502053314 tempest-ServerMetadataNegativeTestJSON-502053314-project-member] [instance: 9d18c7b6-ce8e-4042-9e70-696bb7f57cb3] Successfully created port: ef0f7ff0-3d3e-4159-910f-c0c0e07aa269 {{(pid=63515) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 822.078623] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-477817e6-ef49-4acf-9d39-177b03b7a885 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] [instance: 5cbce760-0163-4b27-8ae3-e46c926c8916] Powered on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 822.079611] env[63515]: DEBUG nova.compute.manager [None req-477817e6-ef49-4acf-9d39-177b03b7a885 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] [instance: 5cbce760-0163-4b27-8ae3-e46c926c8916] Checking state {{(pid=63515) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 822.080296] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b622f837-caf5-406e-9b30-aad29ae89876 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.117449] env[63515]: DEBUG nova.compute.manager [None req-49d5860e-9bde-4499-902f-a44ae9cea035 tempest-ServerMetadataNegativeTestJSON-502053314 tempest-ServerMetadataNegativeTestJSON-502053314-project-member] [instance: 9d18c7b6-ce8e-4042-9e70-696bb7f57cb3] Start building block device mappings for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 822.136056] env[63515]: DEBUG oslo_concurrency.lockutils [None req-b53e671a-9cf4-4e34-a2ac-4321dce89b1c tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.018s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 822.137768] env[63515]: DEBUG oslo_concurrency.lockutils [None req-b257f20f-3555-4579-b6f5-3780aaf93d24 tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 18.584s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 822.138101] env[63515]: DEBUG nova.objects.instance [None req-b257f20f-3555-4579-b6f5-3780aaf93d24 tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] Lazy-loading 'resources' on Instance uuid 192137b0-03e5-4bc4-b911-4b4f1a874f74 {{(pid=63515) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 822.435590] env[63515]: INFO nova.compute.manager [-] [instance: 4ef8066a-b68c-457b-9964-b1c34bab0fc3] Took 1.44 seconds to deallocate network for instance. [ 822.455487] env[63515]: DEBUG oslo_concurrency.lockutils [None req-ce131312-5bf5-4677-8923-2d91405ade5d tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 822.960217] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca1910ea-4020-4b0b-ac30-ec6014746908 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.969724] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d742c2b-e656-40dc-98be-69822b4551b4 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.003866] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad3a779f-2e7f-4740-a74e-46739385a578 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.010819] env[63515]: INFO nova.compute.manager [None req-b19ee0f2-2420-400a-b8e9-86908fd35c93 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: 4ef8066a-b68c-457b-9964-b1c34bab0fc3] Took 0.57 seconds to detach 1 volumes for instance. [ 823.013541] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c69dc5af-1cda-40d0-b458-80f651d21303 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.032479] env[63515]: DEBUG nova.compute.provider_tree [None req-b257f20f-3555-4579-b6f5-3780aaf93d24 tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 823.127241] env[63515]: DEBUG nova.compute.manager [None req-49d5860e-9bde-4499-902f-a44ae9cea035 tempest-ServerMetadataNegativeTestJSON-502053314 tempest-ServerMetadataNegativeTestJSON-502053314-project-member] [instance: 9d18c7b6-ce8e-4042-9e70-696bb7f57cb3] Start spawning the instance on the hypervisor. {{(pid=63515) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 823.161094] env[63515]: DEBUG nova.virt.hardware [None req-49d5860e-9bde-4499-902f-a44ae9cea035 tempest-ServerMetadataNegativeTestJSON-502053314 tempest-ServerMetadataNegativeTestJSON-502053314-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T02:52:14Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T02:51:56Z,direct_url=,disk_format='vmdk',id=8a120570-cb06-4099-b262-554ca0ad15c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b569255cc43e42e7ae3f2b4ad37c6ef4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T02:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 823.161374] env[63515]: DEBUG nova.virt.hardware [None req-49d5860e-9bde-4499-902f-a44ae9cea035 tempest-ServerMetadataNegativeTestJSON-502053314 tempest-ServerMetadataNegativeTestJSON-502053314-project-member] Flavor limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 823.161538] env[63515]: DEBUG nova.virt.hardware [None req-49d5860e-9bde-4499-902f-a44ae9cea035 tempest-ServerMetadataNegativeTestJSON-502053314 tempest-ServerMetadataNegativeTestJSON-502053314-project-member] Image limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 823.161729] env[63515]: DEBUG nova.virt.hardware [None req-49d5860e-9bde-4499-902f-a44ae9cea035 tempest-ServerMetadataNegativeTestJSON-502053314 tempest-ServerMetadataNegativeTestJSON-502053314-project-member] Flavor pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 823.161877] env[63515]: DEBUG nova.virt.hardware [None req-49d5860e-9bde-4499-902f-a44ae9cea035 tempest-ServerMetadataNegativeTestJSON-502053314 tempest-ServerMetadataNegativeTestJSON-502053314-project-member] Image pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 823.162035] env[63515]: DEBUG nova.virt.hardware [None req-49d5860e-9bde-4499-902f-a44ae9cea035 tempest-ServerMetadataNegativeTestJSON-502053314 tempest-ServerMetadataNegativeTestJSON-502053314-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 823.162271] env[63515]: DEBUG nova.virt.hardware [None req-49d5860e-9bde-4499-902f-a44ae9cea035 tempest-ServerMetadataNegativeTestJSON-502053314 tempest-ServerMetadataNegativeTestJSON-502053314-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 823.162425] env[63515]: DEBUG nova.virt.hardware [None req-49d5860e-9bde-4499-902f-a44ae9cea035 tempest-ServerMetadataNegativeTestJSON-502053314 tempest-ServerMetadataNegativeTestJSON-502053314-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 823.162645] env[63515]: DEBUG nova.virt.hardware [None req-49d5860e-9bde-4499-902f-a44ae9cea035 tempest-ServerMetadataNegativeTestJSON-502053314 tempest-ServerMetadataNegativeTestJSON-502053314-project-member] Got 1 possible topologies {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 823.162906] env[63515]: DEBUG nova.virt.hardware [None req-49d5860e-9bde-4499-902f-a44ae9cea035 tempest-ServerMetadataNegativeTestJSON-502053314 tempest-ServerMetadataNegativeTestJSON-502053314-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 823.163018] env[63515]: DEBUG nova.virt.hardware [None req-49d5860e-9bde-4499-902f-a44ae9cea035 tempest-ServerMetadataNegativeTestJSON-502053314 tempest-ServerMetadataNegativeTestJSON-502053314-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 823.163885] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea982302-201d-45fd-bde6-cbfa789b0365 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.172142] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2113e8bf-0ae3-4d34-b0d8-f4e85b233f51 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.523168] env[63515]: DEBUG oslo_concurrency.lockutils [None req-b19ee0f2-2420-400a-b8e9-86908fd35c93 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 823.534700] env[63515]: DEBUG nova.scheduler.client.report [None req-b257f20f-3555-4579-b6f5-3780aaf93d24 tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 823.723751] env[63515]: DEBUG nova.network.neutron [None req-49d5860e-9bde-4499-902f-a44ae9cea035 tempest-ServerMetadataNegativeTestJSON-502053314 tempest-ServerMetadataNegativeTestJSON-502053314-project-member] [instance: 9d18c7b6-ce8e-4042-9e70-696bb7f57cb3] Successfully updated port: ef0f7ff0-3d3e-4159-910f-c0c0e07aa269 {{(pid=63515) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 823.888788] env[63515]: DEBUG nova.compute.manager [req-e17a6f8a-ff94-4c9e-96f4-d286a5ab9bb8 req-392a59b0-e10a-4492-8f39-173842798a5b service nova] [instance: 9d18c7b6-ce8e-4042-9e70-696bb7f57cb3] Received event network-vif-plugged-ef0f7ff0-3d3e-4159-910f-c0c0e07aa269 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 823.889041] env[63515]: DEBUG oslo_concurrency.lockutils [req-e17a6f8a-ff94-4c9e-96f4-d286a5ab9bb8 req-392a59b0-e10a-4492-8f39-173842798a5b service nova] Acquiring lock "9d18c7b6-ce8e-4042-9e70-696bb7f57cb3-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 823.889274] env[63515]: DEBUG oslo_concurrency.lockutils [req-e17a6f8a-ff94-4c9e-96f4-d286a5ab9bb8 req-392a59b0-e10a-4492-8f39-173842798a5b service nova] Lock "9d18c7b6-ce8e-4042-9e70-696bb7f57cb3-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 823.889450] env[63515]: DEBUG oslo_concurrency.lockutils [req-e17a6f8a-ff94-4c9e-96f4-d286a5ab9bb8 req-392a59b0-e10a-4492-8f39-173842798a5b service nova] Lock "9d18c7b6-ce8e-4042-9e70-696bb7f57cb3-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 823.889630] env[63515]: DEBUG nova.compute.manager [req-e17a6f8a-ff94-4c9e-96f4-d286a5ab9bb8 req-392a59b0-e10a-4492-8f39-173842798a5b service nova] [instance: 9d18c7b6-ce8e-4042-9e70-696bb7f57cb3] No waiting events found dispatching network-vif-plugged-ef0f7ff0-3d3e-4159-910f-c0c0e07aa269 {{(pid=63515) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 823.889799] env[63515]: WARNING nova.compute.manager [req-e17a6f8a-ff94-4c9e-96f4-d286a5ab9bb8 req-392a59b0-e10a-4492-8f39-173842798a5b service nova] [instance: 9d18c7b6-ce8e-4042-9e70-696bb7f57cb3] Received unexpected event network-vif-plugged-ef0f7ff0-3d3e-4159-910f-c0c0e07aa269 for instance with vm_state building and task_state spawning. [ 824.040380] env[63515]: DEBUG oslo_concurrency.lockutils [None req-b257f20f-3555-4579-b6f5-3780aaf93d24 tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.903s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 824.042722] env[63515]: DEBUG oslo_concurrency.lockutils [None req-811d8802-f522-4005-9c27-973badf5075e tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 14.892s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 824.042969] env[63515]: DEBUG nova.objects.instance [None req-811d8802-f522-4005-9c27-973badf5075e tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Lazy-loading 'resources' on Instance uuid ed550b10-d58f-45b8-b766-198f431c3788 {{(pid=63515) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 824.066456] env[63515]: INFO nova.scheduler.client.report [None req-b257f20f-3555-4579-b6f5-3780aaf93d24 tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] Deleted allocations for instance 192137b0-03e5-4bc4-b911-4b4f1a874f74 [ 824.227719] env[63515]: DEBUG oslo_concurrency.lockutils [None req-49d5860e-9bde-4499-902f-a44ae9cea035 tempest-ServerMetadataNegativeTestJSON-502053314 tempest-ServerMetadataNegativeTestJSON-502053314-project-member] Acquiring lock "refresh_cache-9d18c7b6-ce8e-4042-9e70-696bb7f57cb3" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 824.228264] env[63515]: DEBUG oslo_concurrency.lockutils [None req-49d5860e-9bde-4499-902f-a44ae9cea035 tempest-ServerMetadataNegativeTestJSON-502053314 tempest-ServerMetadataNegativeTestJSON-502053314-project-member] Acquired lock "refresh_cache-9d18c7b6-ce8e-4042-9e70-696bb7f57cb3" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 824.229704] env[63515]: DEBUG nova.network.neutron [None req-49d5860e-9bde-4499-902f-a44ae9cea035 tempest-ServerMetadataNegativeTestJSON-502053314 tempest-ServerMetadataNegativeTestJSON-502053314-project-member] [instance: 9d18c7b6-ce8e-4042-9e70-696bb7f57cb3] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 824.573716] env[63515]: DEBUG oslo_concurrency.lockutils [None req-b257f20f-3555-4579-b6f5-3780aaf93d24 tempest-ServerShowV257Test-917231397 tempest-ServerShowV257Test-917231397-project-member] Lock "192137b0-03e5-4bc4-b911-4b4f1a874f74" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 26.828s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 824.750853] env[63515]: DEBUG oslo_concurrency.lockutils [None req-957d8582-d135-4677-8fde-d84fe70e2761 tempest-ServerDiagnosticsV248Test-1736022867 tempest-ServerDiagnosticsV248Test-1736022867-project-member] Acquiring lock "705b7e9b-3ab6-40ec-8cf6-d86bf0a1faf0" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 824.750853] env[63515]: DEBUG oslo_concurrency.lockutils [None req-957d8582-d135-4677-8fde-d84fe70e2761 tempest-ServerDiagnosticsV248Test-1736022867 tempest-ServerDiagnosticsV248Test-1736022867-project-member] Lock "705b7e9b-3ab6-40ec-8cf6-d86bf0a1faf0" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 824.764453] env[63515]: DEBUG nova.network.neutron [None req-49d5860e-9bde-4499-902f-a44ae9cea035 tempest-ServerMetadataNegativeTestJSON-502053314 tempest-ServerMetadataNegativeTestJSON-502053314-project-member] [instance: 9d18c7b6-ce8e-4042-9e70-696bb7f57cb3] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 824.891823] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c4ecfec-2784-4bff-b3f1-a9ae1e873576 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.902311] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f97b5d0b-d8f8-420d-961c-4a7332fe5511 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.939070] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2cc26fe0-22e3-42df-9054-73b311af4882 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.947545] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84654b80-c8e5-4625-a19e-8c0b43532797 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.962355] env[63515]: DEBUG nova.compute.provider_tree [None req-811d8802-f522-4005-9c27-973badf5075e tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 824.970719] env[63515]: DEBUG nova.network.neutron [None req-49d5860e-9bde-4499-902f-a44ae9cea035 tempest-ServerMetadataNegativeTestJSON-502053314 tempest-ServerMetadataNegativeTestJSON-502053314-project-member] [instance: 9d18c7b6-ce8e-4042-9e70-696bb7f57cb3] Updating instance_info_cache with network_info: [{"id": "ef0f7ff0-3d3e-4159-910f-c0c0e07aa269", "address": "fa:16:3e:ce:ea:e5", "network": {"id": "9c7000fa-f63d-4a7c-ae36-36cbcaad350d", "bridge": "br-int", "label": "tempest-ServerMetadataNegativeTestJSON-1224376005-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bc5e67e8d1364503bacb15aec38d90d3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "eaba65c3-6925-4c7f-83b6-17cd1a328e27", "external-id": "nsx-vlan-transportzone-202", "segmentation_id": 202, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapef0f7ff0-3d", "ovs_interfaceid": "ef0f7ff0-3d3e-4159-910f-c0c0e07aa269", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 825.252321] env[63515]: DEBUG nova.compute.manager [None req-957d8582-d135-4677-8fde-d84fe70e2761 tempest-ServerDiagnosticsV248Test-1736022867 tempest-ServerDiagnosticsV248Test-1736022867-project-member] [instance: 705b7e9b-3ab6-40ec-8cf6-d86bf0a1faf0] Starting instance... {{(pid=63515) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 825.467374] env[63515]: DEBUG nova.scheduler.client.report [None req-811d8802-f522-4005-9c27-973badf5075e tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 825.474007] env[63515]: DEBUG oslo_concurrency.lockutils [None req-49d5860e-9bde-4499-902f-a44ae9cea035 tempest-ServerMetadataNegativeTestJSON-502053314 tempest-ServerMetadataNegativeTestJSON-502053314-project-member] Releasing lock "refresh_cache-9d18c7b6-ce8e-4042-9e70-696bb7f57cb3" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 825.474335] env[63515]: DEBUG nova.compute.manager [None req-49d5860e-9bde-4499-902f-a44ae9cea035 tempest-ServerMetadataNegativeTestJSON-502053314 tempest-ServerMetadataNegativeTestJSON-502053314-project-member] [instance: 9d18c7b6-ce8e-4042-9e70-696bb7f57cb3] Instance network_info: |[{"id": "ef0f7ff0-3d3e-4159-910f-c0c0e07aa269", "address": "fa:16:3e:ce:ea:e5", "network": {"id": "9c7000fa-f63d-4a7c-ae36-36cbcaad350d", "bridge": "br-int", "label": "tempest-ServerMetadataNegativeTestJSON-1224376005-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bc5e67e8d1364503bacb15aec38d90d3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "eaba65c3-6925-4c7f-83b6-17cd1a328e27", "external-id": "nsx-vlan-transportzone-202", "segmentation_id": 202, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapef0f7ff0-3d", "ovs_interfaceid": "ef0f7ff0-3d3e-4159-910f-c0c0e07aa269", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 825.474746] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-49d5860e-9bde-4499-902f-a44ae9cea035 tempest-ServerMetadataNegativeTestJSON-502053314 tempest-ServerMetadataNegativeTestJSON-502053314-project-member] [instance: 9d18c7b6-ce8e-4042-9e70-696bb7f57cb3] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ce:ea:e5', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'eaba65c3-6925-4c7f-83b6-17cd1a328e27', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ef0f7ff0-3d3e-4159-910f-c0c0e07aa269', 'vif_model': 'vmxnet3'}] {{(pid=63515) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 825.483664] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-49d5860e-9bde-4499-902f-a44ae9cea035 tempest-ServerMetadataNegativeTestJSON-502053314 tempest-ServerMetadataNegativeTestJSON-502053314-project-member] Creating folder: Project (bc5e67e8d1364503bacb15aec38d90d3). Parent ref: group-v243370. {{(pid=63515) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 825.484069] env[63515]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-5215eb98-a9fc-4a33-9c67-9f3e2e592025 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.495154] env[63515]: INFO nova.virt.vmwareapi.vm_util [None req-49d5860e-9bde-4499-902f-a44ae9cea035 tempest-ServerMetadataNegativeTestJSON-502053314 tempest-ServerMetadataNegativeTestJSON-502053314-project-member] Created folder: Project (bc5e67e8d1364503bacb15aec38d90d3) in parent group-v243370. [ 825.495593] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-49d5860e-9bde-4499-902f-a44ae9cea035 tempest-ServerMetadataNegativeTestJSON-502053314 tempest-ServerMetadataNegativeTestJSON-502053314-project-member] Creating folder: Instances. Parent ref: group-v243480. {{(pid=63515) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 825.496171] env[63515]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-9f1414c4-183d-435d-8b98-1601209a0194 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.504784] env[63515]: INFO nova.virt.vmwareapi.vm_util [None req-49d5860e-9bde-4499-902f-a44ae9cea035 tempest-ServerMetadataNegativeTestJSON-502053314 tempest-ServerMetadataNegativeTestJSON-502053314-project-member] Created folder: Instances in parent group-v243480. [ 825.506099] env[63515]: DEBUG oslo.service.loopingcall [None req-49d5860e-9bde-4499-902f-a44ae9cea035 tempest-ServerMetadataNegativeTestJSON-502053314 tempest-ServerMetadataNegativeTestJSON-502053314-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 825.507227] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9d18c7b6-ce8e-4042-9e70-696bb7f57cb3] Creating VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 825.507227] env[63515]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-9efc0d4d-3eee-4c17-91b4-d7b51d79a086 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.532438] env[63515]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 825.532438] env[63515]: value = "task-1111338" [ 825.532438] env[63515]: _type = "Task" [ 825.532438] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 825.541096] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1111338, 'name': CreateVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.777606] env[63515]: DEBUG oslo_concurrency.lockutils [None req-957d8582-d135-4677-8fde-d84fe70e2761 tempest-ServerDiagnosticsV248Test-1736022867 tempest-ServerDiagnosticsV248Test-1736022867-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 825.919111] env[63515]: DEBUG nova.compute.manager [req-b0a71a12-c9fc-4653-9f82-2e9c01d610bd req-1979a12f-dfbf-45dd-a023-48efabe60352 service nova] [instance: 9d18c7b6-ce8e-4042-9e70-696bb7f57cb3] Received event network-changed-ef0f7ff0-3d3e-4159-910f-c0c0e07aa269 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 825.919111] env[63515]: DEBUG nova.compute.manager [req-b0a71a12-c9fc-4653-9f82-2e9c01d610bd req-1979a12f-dfbf-45dd-a023-48efabe60352 service nova] [instance: 9d18c7b6-ce8e-4042-9e70-696bb7f57cb3] Refreshing instance network info cache due to event network-changed-ef0f7ff0-3d3e-4159-910f-c0c0e07aa269. {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 825.919111] env[63515]: DEBUG oslo_concurrency.lockutils [req-b0a71a12-c9fc-4653-9f82-2e9c01d610bd req-1979a12f-dfbf-45dd-a023-48efabe60352 service nova] Acquiring lock "refresh_cache-9d18c7b6-ce8e-4042-9e70-696bb7f57cb3" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 825.919111] env[63515]: DEBUG oslo_concurrency.lockutils [req-b0a71a12-c9fc-4653-9f82-2e9c01d610bd req-1979a12f-dfbf-45dd-a023-48efabe60352 service nova] Acquired lock "refresh_cache-9d18c7b6-ce8e-4042-9e70-696bb7f57cb3" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 825.919111] env[63515]: DEBUG nova.network.neutron [req-b0a71a12-c9fc-4653-9f82-2e9c01d610bd req-1979a12f-dfbf-45dd-a023-48efabe60352 service nova] [instance: 9d18c7b6-ce8e-4042-9e70-696bb7f57cb3] Refreshing network info cache for port ef0f7ff0-3d3e-4159-910f-c0c0e07aa269 {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 825.973598] env[63515]: DEBUG oslo_concurrency.lockutils [None req-811d8802-f522-4005-9c27-973badf5075e tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.930s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 825.982716] env[63515]: DEBUG oslo_concurrency.lockutils [None req-9dd4330d-d35d-47d5-b468-11f24e31a49a tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.131s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 825.983112] env[63515]: INFO nova.compute.claims [None req-9dd4330d-d35d-47d5-b468-11f24e31a49a tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: 70df69e5-687b-44fb-b6fc-cdb08e21dda0] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 825.997545] env[63515]: INFO nova.scheduler.client.report [None req-811d8802-f522-4005-9c27-973badf5075e tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Deleted allocations for instance ed550b10-d58f-45b8-b766-198f431c3788 [ 826.055086] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1111338, 'name': CreateVM_Task, 'duration_secs': 0.401221} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 826.055086] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9d18c7b6-ce8e-4042-9e70-696bb7f57cb3] Created VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 826.057028] env[63515]: DEBUG oslo_concurrency.lockutils [None req-49d5860e-9bde-4499-902f-a44ae9cea035 tempest-ServerMetadataNegativeTestJSON-502053314 tempest-ServerMetadataNegativeTestJSON-502053314-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 826.057028] env[63515]: DEBUG oslo_concurrency.lockutils [None req-49d5860e-9bde-4499-902f-a44ae9cea035 tempest-ServerMetadataNegativeTestJSON-502053314 tempest-ServerMetadataNegativeTestJSON-502053314-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 826.057028] env[63515]: DEBUG oslo_concurrency.lockutils [None req-49d5860e-9bde-4499-902f-a44ae9cea035 tempest-ServerMetadataNegativeTestJSON-502053314 tempest-ServerMetadataNegativeTestJSON-502053314-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 826.057028] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d7a347a2-8608-46a0-9354-70211441df28 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.064835] env[63515]: DEBUG oslo_vmware.api [None req-49d5860e-9bde-4499-902f-a44ae9cea035 tempest-ServerMetadataNegativeTestJSON-502053314 tempest-ServerMetadataNegativeTestJSON-502053314-project-member] Waiting for the task: (returnval){ [ 826.064835] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]524ade78-6d30-d9cf-3366-8fc948f56860" [ 826.064835] env[63515]: _type = "Task" [ 826.064835] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 826.073785] env[63515]: DEBUG oslo_vmware.api [None req-49d5860e-9bde-4499-902f-a44ae9cea035 tempest-ServerMetadataNegativeTestJSON-502053314 tempest-ServerMetadataNegativeTestJSON-502053314-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]524ade78-6d30-d9cf-3366-8fc948f56860, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.513074] env[63515]: DEBUG oslo_concurrency.lockutils [None req-811d8802-f522-4005-9c27-973badf5075e tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Lock "ed550b10-d58f-45b8-b766-198f431c3788" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 21.425s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 826.576699] env[63515]: DEBUG oslo_vmware.api [None req-49d5860e-9bde-4499-902f-a44ae9cea035 tempest-ServerMetadataNegativeTestJSON-502053314 tempest-ServerMetadataNegativeTestJSON-502053314-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]524ade78-6d30-d9cf-3366-8fc948f56860, 'name': SearchDatastore_Task, 'duration_secs': 0.00972} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 826.577342] env[63515]: DEBUG oslo_concurrency.lockutils [None req-49d5860e-9bde-4499-902f-a44ae9cea035 tempest-ServerMetadataNegativeTestJSON-502053314 tempest-ServerMetadataNegativeTestJSON-502053314-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 826.577427] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-49d5860e-9bde-4499-902f-a44ae9cea035 tempest-ServerMetadataNegativeTestJSON-502053314 tempest-ServerMetadataNegativeTestJSON-502053314-project-member] [instance: 9d18c7b6-ce8e-4042-9e70-696bb7f57cb3] Processing image 8a120570-cb06-4099-b262-554ca0ad15c5 {{(pid=63515) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 826.577628] env[63515]: DEBUG oslo_concurrency.lockutils [None req-49d5860e-9bde-4499-902f-a44ae9cea035 tempest-ServerMetadataNegativeTestJSON-502053314 tempest-ServerMetadataNegativeTestJSON-502053314-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 826.577791] env[63515]: DEBUG oslo_concurrency.lockutils [None req-49d5860e-9bde-4499-902f-a44ae9cea035 tempest-ServerMetadataNegativeTestJSON-502053314 tempest-ServerMetadataNegativeTestJSON-502053314-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 826.578068] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-49d5860e-9bde-4499-902f-a44ae9cea035 tempest-ServerMetadataNegativeTestJSON-502053314 tempest-ServerMetadataNegativeTestJSON-502053314-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 826.579025] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-333f9ea5-89dc-4b39-ad0e-190894f1f2ec {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.587080] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-49d5860e-9bde-4499-902f-a44ae9cea035 tempest-ServerMetadataNegativeTestJSON-502053314 tempest-ServerMetadataNegativeTestJSON-502053314-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 826.587174] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-49d5860e-9bde-4499-902f-a44ae9cea035 tempest-ServerMetadataNegativeTestJSON-502053314 tempest-ServerMetadataNegativeTestJSON-502053314-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63515) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 826.587888] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b477f83a-b0a7-464f-8a6a-b54a9c75fcc9 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.593561] env[63515]: DEBUG oslo_vmware.api [None req-49d5860e-9bde-4499-902f-a44ae9cea035 tempest-ServerMetadataNegativeTestJSON-502053314 tempest-ServerMetadataNegativeTestJSON-502053314-project-member] Waiting for the task: (returnval){ [ 826.593561] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]528a4da8-860a-6040-1208-4c311199def5" [ 826.593561] env[63515]: _type = "Task" [ 826.593561] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 826.604332] env[63515]: DEBUG oslo_vmware.api [None req-49d5860e-9bde-4499-902f-a44ae9cea035 tempest-ServerMetadataNegativeTestJSON-502053314 tempest-ServerMetadataNegativeTestJSON-502053314-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]528a4da8-860a-6040-1208-4c311199def5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.755499] env[63515]: DEBUG oslo_concurrency.lockutils [None req-eae7e17b-7393-4e13-b575-a8d38e798918 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Acquiring lock "2a29f37d-e786-48d3-b126-4854d6e7c67c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 826.755729] env[63515]: DEBUG oslo_concurrency.lockutils [None req-eae7e17b-7393-4e13-b575-a8d38e798918 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Lock "2a29f37d-e786-48d3-b126-4854d6e7c67c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 826.798315] env[63515]: DEBUG nova.network.neutron [req-b0a71a12-c9fc-4653-9f82-2e9c01d610bd req-1979a12f-dfbf-45dd-a023-48efabe60352 service nova] [instance: 9d18c7b6-ce8e-4042-9e70-696bb7f57cb3] Updated VIF entry in instance network info cache for port ef0f7ff0-3d3e-4159-910f-c0c0e07aa269. {{(pid=63515) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 826.798315] env[63515]: DEBUG nova.network.neutron [req-b0a71a12-c9fc-4653-9f82-2e9c01d610bd req-1979a12f-dfbf-45dd-a023-48efabe60352 service nova] [instance: 9d18c7b6-ce8e-4042-9e70-696bb7f57cb3] Updating instance_info_cache with network_info: [{"id": "ef0f7ff0-3d3e-4159-910f-c0c0e07aa269", "address": "fa:16:3e:ce:ea:e5", "network": {"id": "9c7000fa-f63d-4a7c-ae36-36cbcaad350d", "bridge": "br-int", "label": "tempest-ServerMetadataNegativeTestJSON-1224376005-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bc5e67e8d1364503bacb15aec38d90d3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "eaba65c3-6925-4c7f-83b6-17cd1a328e27", "external-id": "nsx-vlan-transportzone-202", "segmentation_id": 202, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapef0f7ff0-3d", "ovs_interfaceid": "ef0f7ff0-3d3e-4159-910f-c0c0e07aa269", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 827.110410] env[63515]: DEBUG oslo_vmware.api [None req-49d5860e-9bde-4499-902f-a44ae9cea035 tempest-ServerMetadataNegativeTestJSON-502053314 tempest-ServerMetadataNegativeTestJSON-502053314-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]528a4da8-860a-6040-1208-4c311199def5, 'name': SearchDatastore_Task, 'duration_secs': 0.008795} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 827.111451] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e70200d2-62d1-43f3-9d1f-52f98020edae {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.119180] env[63515]: DEBUG oslo_vmware.api [None req-49d5860e-9bde-4499-902f-a44ae9cea035 tempest-ServerMetadataNegativeTestJSON-502053314 tempest-ServerMetadataNegativeTestJSON-502053314-project-member] Waiting for the task: (returnval){ [ 827.119180] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]52ab65ca-7368-dbb0-568c-5427e19839b5" [ 827.119180] env[63515]: _type = "Task" [ 827.119180] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 827.131115] env[63515]: DEBUG oslo_vmware.api [None req-49d5860e-9bde-4499-902f-a44ae9cea035 tempest-ServerMetadataNegativeTestJSON-502053314 tempest-ServerMetadataNegativeTestJSON-502053314-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52ab65ca-7368-dbb0-568c-5427e19839b5, 'name': SearchDatastore_Task, 'duration_secs': 0.010342} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 827.131452] env[63515]: DEBUG oslo_concurrency.lockutils [None req-49d5860e-9bde-4499-902f-a44ae9cea035 tempest-ServerMetadataNegativeTestJSON-502053314 tempest-ServerMetadataNegativeTestJSON-502053314-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 827.131746] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-49d5860e-9bde-4499-902f-a44ae9cea035 tempest-ServerMetadataNegativeTestJSON-502053314 tempest-ServerMetadataNegativeTestJSON-502053314-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk to [datastore2] 9d18c7b6-ce8e-4042-9e70-696bb7f57cb3/9d18c7b6-ce8e-4042-9e70-696bb7f57cb3.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 827.131961] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-af28c893-9cad-4eff-be16-60f6159bfc2f {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.138366] env[63515]: DEBUG oslo_vmware.api [None req-49d5860e-9bde-4499-902f-a44ae9cea035 tempest-ServerMetadataNegativeTestJSON-502053314 tempest-ServerMetadataNegativeTestJSON-502053314-project-member] Waiting for the task: (returnval){ [ 827.138366] env[63515]: value = "task-1111340" [ 827.138366] env[63515]: _type = "Task" [ 827.138366] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 827.146035] env[63515]: DEBUG oslo_vmware.api [None req-49d5860e-9bde-4499-902f-a44ae9cea035 tempest-ServerMetadataNegativeTestJSON-502053314 tempest-ServerMetadataNegativeTestJSON-502053314-project-member] Task: {'id': task-1111340, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.220646] env[63515]: DEBUG oslo_concurrency.lockutils [None req-baaf7dbf-037b-4d17-a0eb-d252e746ead6 tempest-InstanceActionsTestJSON-247065106 tempest-InstanceActionsTestJSON-247065106-project-member] Acquiring lock "0ee9cce2-94b9-4038-ab38-1b3f96a864bc" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 827.221524] env[63515]: DEBUG oslo_concurrency.lockutils [None req-baaf7dbf-037b-4d17-a0eb-d252e746ead6 tempest-InstanceActionsTestJSON-247065106 tempest-InstanceActionsTestJSON-247065106-project-member] Lock "0ee9cce2-94b9-4038-ab38-1b3f96a864bc" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 827.262682] env[63515]: DEBUG nova.compute.manager [None req-eae7e17b-7393-4e13-b575-a8d38e798918 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] [instance: 2a29f37d-e786-48d3-b126-4854d6e7c67c] Starting instance... {{(pid=63515) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 827.304449] env[63515]: DEBUG oslo_concurrency.lockutils [req-b0a71a12-c9fc-4653-9f82-2e9c01d610bd req-1979a12f-dfbf-45dd-a023-48efabe60352 service nova] Releasing lock "refresh_cache-9d18c7b6-ce8e-4042-9e70-696bb7f57cb3" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 827.324368] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1c19454-acf8-4dbc-8917-a186158979ea {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.344153] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a30ee48-76d3-43e2-b5a7-76fc5e2929f5 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.376722] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0b16553-24e1-44a9-819e-55ba33980438 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.386049] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4652514-fb0a-4c88-8c17-6a725d055f8a {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.404014] env[63515]: DEBUG nova.compute.provider_tree [None req-9dd4330d-d35d-47d5-b468-11f24e31a49a tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 827.649729] env[63515]: DEBUG oslo_vmware.api [None req-49d5860e-9bde-4499-902f-a44ae9cea035 tempest-ServerMetadataNegativeTestJSON-502053314 tempest-ServerMetadataNegativeTestJSON-502053314-project-member] Task: {'id': task-1111340, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.482059} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 827.650068] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-49d5860e-9bde-4499-902f-a44ae9cea035 tempest-ServerMetadataNegativeTestJSON-502053314 tempest-ServerMetadataNegativeTestJSON-502053314-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk to [datastore2] 9d18c7b6-ce8e-4042-9e70-696bb7f57cb3/9d18c7b6-ce8e-4042-9e70-696bb7f57cb3.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 827.650268] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-49d5860e-9bde-4499-902f-a44ae9cea035 tempest-ServerMetadataNegativeTestJSON-502053314 tempest-ServerMetadataNegativeTestJSON-502053314-project-member] [instance: 9d18c7b6-ce8e-4042-9e70-696bb7f57cb3] Extending root virtual disk to 1048576 {{(pid=63515) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 827.650527] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-27990bb5-6ad6-4727-b732-c406a4321907 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.657747] env[63515]: DEBUG oslo_vmware.api [None req-49d5860e-9bde-4499-902f-a44ae9cea035 tempest-ServerMetadataNegativeTestJSON-502053314 tempest-ServerMetadataNegativeTestJSON-502053314-project-member] Waiting for the task: (returnval){ [ 827.657747] env[63515]: value = "task-1111341" [ 827.657747] env[63515]: _type = "Task" [ 827.657747] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 827.667669] env[63515]: DEBUG oslo_vmware.api [None req-49d5860e-9bde-4499-902f-a44ae9cea035 tempest-ServerMetadataNegativeTestJSON-502053314 tempest-ServerMetadataNegativeTestJSON-502053314-project-member] Task: {'id': task-1111341, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.725301] env[63515]: DEBUG nova.compute.manager [None req-baaf7dbf-037b-4d17-a0eb-d252e746ead6 tempest-InstanceActionsTestJSON-247065106 tempest-InstanceActionsTestJSON-247065106-project-member] [instance: 0ee9cce2-94b9-4038-ab38-1b3f96a864bc] Starting instance... {{(pid=63515) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 827.789336] env[63515]: DEBUG oslo_concurrency.lockutils [None req-eae7e17b-7393-4e13-b575-a8d38e798918 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 827.908552] env[63515]: DEBUG nova.scheduler.client.report [None req-9dd4330d-d35d-47d5-b468-11f24e31a49a tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 828.169700] env[63515]: DEBUG oslo_vmware.api [None req-49d5860e-9bde-4499-902f-a44ae9cea035 tempest-ServerMetadataNegativeTestJSON-502053314 tempest-ServerMetadataNegativeTestJSON-502053314-project-member] Task: {'id': task-1111341, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.140776} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 828.169993] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-49d5860e-9bde-4499-902f-a44ae9cea035 tempest-ServerMetadataNegativeTestJSON-502053314 tempest-ServerMetadataNegativeTestJSON-502053314-project-member] [instance: 9d18c7b6-ce8e-4042-9e70-696bb7f57cb3] Extended root virtual disk {{(pid=63515) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 828.170810] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc65e3f8-5676-4814-af95-c1d393e98f5f {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.194352] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-49d5860e-9bde-4499-902f-a44ae9cea035 tempest-ServerMetadataNegativeTestJSON-502053314 tempest-ServerMetadataNegativeTestJSON-502053314-project-member] [instance: 9d18c7b6-ce8e-4042-9e70-696bb7f57cb3] Reconfiguring VM instance instance-00000040 to attach disk [datastore2] 9d18c7b6-ce8e-4042-9e70-696bb7f57cb3/9d18c7b6-ce8e-4042-9e70-696bb7f57cb3.vmdk or device None with type sparse {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 828.195028] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e4fc8d49-869c-4ada-8e35-566450677fac {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.216179] env[63515]: DEBUG oslo_vmware.api [None req-49d5860e-9bde-4499-902f-a44ae9cea035 tempest-ServerMetadataNegativeTestJSON-502053314 tempest-ServerMetadataNegativeTestJSON-502053314-project-member] Waiting for the task: (returnval){ [ 828.216179] env[63515]: value = "task-1111342" [ 828.216179] env[63515]: _type = "Task" [ 828.216179] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 828.225658] env[63515]: DEBUG oslo_vmware.api [None req-49d5860e-9bde-4499-902f-a44ae9cea035 tempest-ServerMetadataNegativeTestJSON-502053314 tempest-ServerMetadataNegativeTestJSON-502053314-project-member] Task: {'id': task-1111342, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.245376] env[63515]: DEBUG oslo_concurrency.lockutils [None req-baaf7dbf-037b-4d17-a0eb-d252e746ead6 tempest-InstanceActionsTestJSON-247065106 tempest-InstanceActionsTestJSON-247065106-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 828.416036] env[63515]: DEBUG oslo_concurrency.lockutils [None req-9dd4330d-d35d-47d5-b468-11f24e31a49a tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.435s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 828.416036] env[63515]: DEBUG nova.compute.manager [None req-9dd4330d-d35d-47d5-b468-11f24e31a49a tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: 70df69e5-687b-44fb-b6fc-cdb08e21dda0] Start building networks asynchronously for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 828.416998] env[63515]: DEBUG oslo_concurrency.lockutils [None req-1f202073-ad2a-4c82-aa50-bf79e5416807 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 18.352s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 828.417252] env[63515]: DEBUG nova.objects.instance [None req-1f202073-ad2a-4c82-aa50-bf79e5416807 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Lazy-loading 'resources' on Instance uuid 84dbf321-2c70-4e08-b430-cb5a06fc6829 {{(pid=63515) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 828.727159] env[63515]: DEBUG oslo_vmware.api [None req-49d5860e-9bde-4499-902f-a44ae9cea035 tempest-ServerMetadataNegativeTestJSON-502053314 tempest-ServerMetadataNegativeTestJSON-502053314-project-member] Task: {'id': task-1111342, 'name': ReconfigVM_Task, 'duration_secs': 0.333475} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 828.727485] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-49d5860e-9bde-4499-902f-a44ae9cea035 tempest-ServerMetadataNegativeTestJSON-502053314 tempest-ServerMetadataNegativeTestJSON-502053314-project-member] [instance: 9d18c7b6-ce8e-4042-9e70-696bb7f57cb3] Reconfigured VM instance instance-00000040 to attach disk [datastore2] 9d18c7b6-ce8e-4042-9e70-696bb7f57cb3/9d18c7b6-ce8e-4042-9e70-696bb7f57cb3.vmdk or device None with type sparse {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 828.728289] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-0a10f9a2-01de-4b45-a6f7-d31e8332f2c5 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.734994] env[63515]: DEBUG oslo_vmware.api [None req-49d5860e-9bde-4499-902f-a44ae9cea035 tempest-ServerMetadataNegativeTestJSON-502053314 tempest-ServerMetadataNegativeTestJSON-502053314-project-member] Waiting for the task: (returnval){ [ 828.734994] env[63515]: value = "task-1111344" [ 828.734994] env[63515]: _type = "Task" [ 828.734994] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 828.743601] env[63515]: DEBUG oslo_vmware.api [None req-49d5860e-9bde-4499-902f-a44ae9cea035 tempest-ServerMetadataNegativeTestJSON-502053314 tempest-ServerMetadataNegativeTestJSON-502053314-project-member] Task: {'id': task-1111344, 'name': Rename_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.896820] env[63515]: DEBUG oslo_concurrency.lockutils [None req-b2ae28b1-9ab0-48f9-af42-2813a916c9a6 tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Acquiring lock "d7e8dcdb-41cb-46fb-8b61-d251e7c2d372" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 828.896820] env[63515]: DEBUG oslo_concurrency.lockutils [None req-b2ae28b1-9ab0-48f9-af42-2813a916c9a6 tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Lock "d7e8dcdb-41cb-46fb-8b61-d251e7c2d372" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 828.919954] env[63515]: DEBUG nova.compute.utils [None req-9dd4330d-d35d-47d5-b468-11f24e31a49a tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Using /dev/sd instead of None {{(pid=63515) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 828.921374] env[63515]: DEBUG nova.compute.manager [None req-9dd4330d-d35d-47d5-b468-11f24e31a49a tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: 70df69e5-687b-44fb-b6fc-cdb08e21dda0] Allocating IP information in the background. {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 828.921555] env[63515]: DEBUG nova.network.neutron [None req-9dd4330d-d35d-47d5-b468-11f24e31a49a tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: 70df69e5-687b-44fb-b6fc-cdb08e21dda0] allocate_for_instance() {{(pid=63515) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 828.949727] env[63515]: DEBUG oslo_concurrency.lockutils [None req-789266fe-c183-402c-aef4-222438c1eee4 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Acquiring lock "eab16df9-7bb5-4576-bca0-769a561c5fe9" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 828.949971] env[63515]: DEBUG oslo_concurrency.lockutils [None req-789266fe-c183-402c-aef4-222438c1eee4 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Lock "eab16df9-7bb5-4576-bca0-769a561c5fe9" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 828.965547] env[63515]: DEBUG nova.policy [None req-9dd4330d-d35d-47d5-b468-11f24e31a49a tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '002c1a9364184b82b649fc18d017d97f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2259667ddc0c42639ea1af3b6e6fa0e8', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63515) authorize /opt/stack/nova/nova/policy.py:201}} [ 829.201086] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f9cf225-72ed-4e57-b2aa-68c9725f3ead {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.210039] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71a64902-762a-4b26-824a-1c8825995226 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.246855] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f65829f-4d5c-4d98-930a-6a99b5e16431 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.259146] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93920f34-6367-4105-828a-226fb123544e {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.262765] env[63515]: DEBUG oslo_vmware.api [None req-49d5860e-9bde-4499-902f-a44ae9cea035 tempest-ServerMetadataNegativeTestJSON-502053314 tempest-ServerMetadataNegativeTestJSON-502053314-project-member] Task: {'id': task-1111344, 'name': Rename_Task, 'duration_secs': 0.15377} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 829.263173] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-49d5860e-9bde-4499-902f-a44ae9cea035 tempest-ServerMetadataNegativeTestJSON-502053314 tempest-ServerMetadataNegativeTestJSON-502053314-project-member] [instance: 9d18c7b6-ce8e-4042-9e70-696bb7f57cb3] Powering on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 829.263823] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-99bf4e5b-2738-4564-bcb0-26fee0f8514b {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.274134] env[63515]: DEBUG nova.compute.provider_tree [None req-1f202073-ad2a-4c82-aa50-bf79e5416807 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 829.276539] env[63515]: DEBUG oslo_vmware.api [None req-49d5860e-9bde-4499-902f-a44ae9cea035 tempest-ServerMetadataNegativeTestJSON-502053314 tempest-ServerMetadataNegativeTestJSON-502053314-project-member] Waiting for the task: (returnval){ [ 829.276539] env[63515]: value = "task-1111345" [ 829.276539] env[63515]: _type = "Task" [ 829.276539] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 829.285848] env[63515]: DEBUG oslo_vmware.api [None req-49d5860e-9bde-4499-902f-a44ae9cea035 tempest-ServerMetadataNegativeTestJSON-502053314 tempest-ServerMetadataNegativeTestJSON-502053314-project-member] Task: {'id': task-1111345, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.299914] env[63515]: DEBUG nova.network.neutron [None req-9dd4330d-d35d-47d5-b468-11f24e31a49a tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: 70df69e5-687b-44fb-b6fc-cdb08e21dda0] Successfully created port: 2639cfaa-8db0-4dcc-ac66-994ec5a84fdc {{(pid=63515) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 829.399795] env[63515]: INFO nova.compute.manager [None req-b2ae28b1-9ab0-48f9-af42-2813a916c9a6 tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] [instance: d7e8dcdb-41cb-46fb-8b61-d251e7c2d372] Detaching volume b1a5444a-1e2f-4443-8931-4a04ec5d9d68 [ 829.424908] env[63515]: DEBUG nova.compute.manager [None req-9dd4330d-d35d-47d5-b468-11f24e31a49a tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: 70df69e5-687b-44fb-b6fc-cdb08e21dda0] Start building block device mappings for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 829.440487] env[63515]: INFO nova.virt.block_device [None req-b2ae28b1-9ab0-48f9-af42-2813a916c9a6 tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] [instance: d7e8dcdb-41cb-46fb-8b61-d251e7c2d372] Attempting to driver detach volume b1a5444a-1e2f-4443-8931-4a04ec5d9d68 from mountpoint /dev/sdb [ 829.440929] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-b2ae28b1-9ab0-48f9-af42-2813a916c9a6 tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] [instance: d7e8dcdb-41cb-46fb-8b61-d251e7c2d372] Volume detach. Driver type: vmdk {{(pid=63515) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 829.441142] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-b2ae28b1-9ab0-48f9-af42-2813a916c9a6 tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] [instance: d7e8dcdb-41cb-46fb-8b61-d251e7c2d372] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-243453', 'volume_id': 'b1a5444a-1e2f-4443-8931-4a04ec5d9d68', 'name': 'volume-b1a5444a-1e2f-4443-8931-4a04ec5d9d68', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'd7e8dcdb-41cb-46fb-8b61-d251e7c2d372', 'attached_at': '', 'detached_at': '', 'volume_id': 'b1a5444a-1e2f-4443-8931-4a04ec5d9d68', 'serial': 'b1a5444a-1e2f-4443-8931-4a04ec5d9d68'} {{(pid=63515) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 829.443436] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee9e8826-9af3-48a1-8217-d2a238394ead {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.469965] env[63515]: DEBUG nova.compute.manager [None req-789266fe-c183-402c-aef4-222438c1eee4 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: eab16df9-7bb5-4576-bca0-769a561c5fe9] Starting instance... {{(pid=63515) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 829.475378] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f1cd1f6-81a6-4066-88de-a21660c75b0c {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.487681] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5fe35297-6a38-419f-a9ab-b7cb21f99dbe {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.508709] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1782e21-662d-447c-b09f-c16d591d3b79 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.531172] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-b2ae28b1-9ab0-48f9-af42-2813a916c9a6 tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] The volume has not been displaced from its original location: [datastore2] volume-b1a5444a-1e2f-4443-8931-4a04ec5d9d68/volume-b1a5444a-1e2f-4443-8931-4a04ec5d9d68.vmdk. No consolidation needed. {{(pid=63515) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 829.537128] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-b2ae28b1-9ab0-48f9-af42-2813a916c9a6 tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] [instance: d7e8dcdb-41cb-46fb-8b61-d251e7c2d372] Reconfiguring VM instance instance-0000002c to detach disk 2001 {{(pid=63515) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 829.537128] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6d876812-5cf7-4b89-8455-bcb42e19e249 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.558967] env[63515]: DEBUG oslo_vmware.api [None req-b2ae28b1-9ab0-48f9-af42-2813a916c9a6 tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Waiting for the task: (returnval){ [ 829.558967] env[63515]: value = "task-1111346" [ 829.558967] env[63515]: _type = "Task" [ 829.558967] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 829.569115] env[63515]: DEBUG oslo_vmware.api [None req-b2ae28b1-9ab0-48f9-af42-2813a916c9a6 tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Task: {'id': task-1111346, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.778622] env[63515]: DEBUG nova.scheduler.client.report [None req-1f202073-ad2a-4c82-aa50-bf79e5416807 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 829.792725] env[63515]: DEBUG oslo_vmware.api [None req-49d5860e-9bde-4499-902f-a44ae9cea035 tempest-ServerMetadataNegativeTestJSON-502053314 tempest-ServerMetadataNegativeTestJSON-502053314-project-member] Task: {'id': task-1111345, 'name': PowerOnVM_Task, 'duration_secs': 0.517492} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 829.793009] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-49d5860e-9bde-4499-902f-a44ae9cea035 tempest-ServerMetadataNegativeTestJSON-502053314 tempest-ServerMetadataNegativeTestJSON-502053314-project-member] [instance: 9d18c7b6-ce8e-4042-9e70-696bb7f57cb3] Powered on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 829.793263] env[63515]: INFO nova.compute.manager [None req-49d5860e-9bde-4499-902f-a44ae9cea035 tempest-ServerMetadataNegativeTestJSON-502053314 tempest-ServerMetadataNegativeTestJSON-502053314-project-member] [instance: 9d18c7b6-ce8e-4042-9e70-696bb7f57cb3] Took 6.67 seconds to spawn the instance on the hypervisor. [ 829.793459] env[63515]: DEBUG nova.compute.manager [None req-49d5860e-9bde-4499-902f-a44ae9cea035 tempest-ServerMetadataNegativeTestJSON-502053314 tempest-ServerMetadataNegativeTestJSON-502053314-project-member] [instance: 9d18c7b6-ce8e-4042-9e70-696bb7f57cb3] Checking state {{(pid=63515) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 829.794294] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e0eb856-1aa1-4be4-b387-4a4df65158fa {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.997278] env[63515]: DEBUG oslo_concurrency.lockutils [None req-789266fe-c183-402c-aef4-222438c1eee4 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 830.069735] env[63515]: DEBUG oslo_vmware.api [None req-b2ae28b1-9ab0-48f9-af42-2813a916c9a6 tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Task: {'id': task-1111346, 'name': ReconfigVM_Task, 'duration_secs': 0.276754} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 830.070125] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-b2ae28b1-9ab0-48f9-af42-2813a916c9a6 tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] [instance: d7e8dcdb-41cb-46fb-8b61-d251e7c2d372] Reconfigured VM instance instance-0000002c to detach disk 2001 {{(pid=63515) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 830.074724] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-fde74f21-0a0b-41ab-9cf0-9e487e7305ad {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.092063] env[63515]: DEBUG oslo_vmware.api [None req-b2ae28b1-9ab0-48f9-af42-2813a916c9a6 tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Waiting for the task: (returnval){ [ 830.092063] env[63515]: value = "task-1111347" [ 830.092063] env[63515]: _type = "Task" [ 830.092063] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 830.103325] env[63515]: DEBUG oslo_vmware.api [None req-b2ae28b1-9ab0-48f9-af42-2813a916c9a6 tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Task: {'id': task-1111347, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.288037] env[63515]: DEBUG oslo_concurrency.lockutils [None req-1f202073-ad2a-4c82-aa50-bf79e5416807 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.871s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 830.290670] env[63515]: DEBUG oslo_concurrency.lockutils [None req-7bca9c72-42f5-4d65-99a2-4374b66c9c04 tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.716s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 830.291815] env[63515]: INFO nova.compute.claims [None req-7bca9c72-42f5-4d65-99a2-4374b66c9c04 tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] [instance: 17275472-2921-49c9-b4e0-5485649ebd1c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 830.310552] env[63515]: INFO nova.scheduler.client.report [None req-1f202073-ad2a-4c82-aa50-bf79e5416807 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Deleted allocations for instance 84dbf321-2c70-4e08-b430-cb5a06fc6829 [ 830.316758] env[63515]: INFO nova.compute.manager [None req-49d5860e-9bde-4499-902f-a44ae9cea035 tempest-ServerMetadataNegativeTestJSON-502053314 tempest-ServerMetadataNegativeTestJSON-502053314-project-member] [instance: 9d18c7b6-ce8e-4042-9e70-696bb7f57cb3] Took 44.24 seconds to build instance. [ 830.435064] env[63515]: DEBUG nova.compute.manager [None req-9dd4330d-d35d-47d5-b468-11f24e31a49a tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: 70df69e5-687b-44fb-b6fc-cdb08e21dda0] Start spawning the instance on the hypervisor. {{(pid=63515) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 830.466494] env[63515]: DEBUG nova.virt.hardware [None req-9dd4330d-d35d-47d5-b468-11f24e31a49a tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T02:52:14Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T02:51:56Z,direct_url=,disk_format='vmdk',id=8a120570-cb06-4099-b262-554ca0ad15c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b569255cc43e42e7ae3f2b4ad37c6ef4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T02:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 830.466835] env[63515]: DEBUG nova.virt.hardware [None req-9dd4330d-d35d-47d5-b468-11f24e31a49a tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Flavor limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 830.467184] env[63515]: DEBUG nova.virt.hardware [None req-9dd4330d-d35d-47d5-b468-11f24e31a49a tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Image limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 830.467353] env[63515]: DEBUG nova.virt.hardware [None req-9dd4330d-d35d-47d5-b468-11f24e31a49a tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Flavor pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 830.467727] env[63515]: DEBUG nova.virt.hardware [None req-9dd4330d-d35d-47d5-b468-11f24e31a49a tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Image pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 830.467796] env[63515]: DEBUG nova.virt.hardware [None req-9dd4330d-d35d-47d5-b468-11f24e31a49a tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 830.468162] env[63515]: DEBUG nova.virt.hardware [None req-9dd4330d-d35d-47d5-b468-11f24e31a49a tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 830.468304] env[63515]: DEBUG nova.virt.hardware [None req-9dd4330d-d35d-47d5-b468-11f24e31a49a tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 830.468661] env[63515]: DEBUG nova.virt.hardware [None req-9dd4330d-d35d-47d5-b468-11f24e31a49a tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Got 1 possible topologies {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 830.468742] env[63515]: DEBUG nova.virt.hardware [None req-9dd4330d-d35d-47d5-b468-11f24e31a49a tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 830.468895] env[63515]: DEBUG nova.virt.hardware [None req-9dd4330d-d35d-47d5-b468-11f24e31a49a tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 830.470411] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be529e3f-47c1-4ba8-90bf-5c7ab62f5cf1 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.479652] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-662f8cf5-da7a-41c6-abb7-562c35e0967e {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.602733] env[63515]: DEBUG oslo_vmware.api [None req-b2ae28b1-9ab0-48f9-af42-2813a916c9a6 tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Task: {'id': task-1111347, 'name': ReconfigVM_Task, 'duration_secs': 0.142069} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 830.603081] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-b2ae28b1-9ab0-48f9-af42-2813a916c9a6 tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] [instance: d7e8dcdb-41cb-46fb-8b61-d251e7c2d372] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-243453', 'volume_id': 'b1a5444a-1e2f-4443-8931-4a04ec5d9d68', 'name': 'volume-b1a5444a-1e2f-4443-8931-4a04ec5d9d68', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'd7e8dcdb-41cb-46fb-8b61-d251e7c2d372', 'attached_at': '', 'detached_at': '', 'volume_id': 'b1a5444a-1e2f-4443-8931-4a04ec5d9d68', 'serial': 'b1a5444a-1e2f-4443-8931-4a04ec5d9d68'} {{(pid=63515) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 830.819981] env[63515]: DEBUG oslo_concurrency.lockutils [None req-49d5860e-9bde-4499-902f-a44ae9cea035 tempest-ServerMetadataNegativeTestJSON-502053314 tempest-ServerMetadataNegativeTestJSON-502053314-project-member] Lock "9d18c7b6-ce8e-4042-9e70-696bb7f57cb3" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 45.755s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 830.820488] env[63515]: DEBUG oslo_concurrency.lockutils [None req-1f202073-ad2a-4c82-aa50-bf79e5416807 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Lock "84dbf321-2c70-4e08-b430-cb5a06fc6829" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 23.256s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 830.971901] env[63515]: DEBUG nova.compute.manager [req-891d504b-70fa-4335-8a13-9e9ff0ca0853 req-1f235373-f1d9-49d9-be25-2517a96f6ed0 service nova] [instance: 70df69e5-687b-44fb-b6fc-cdb08e21dda0] Received event network-vif-plugged-2639cfaa-8db0-4dcc-ac66-994ec5a84fdc {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 830.973026] env[63515]: DEBUG oslo_concurrency.lockutils [req-891d504b-70fa-4335-8a13-9e9ff0ca0853 req-1f235373-f1d9-49d9-be25-2517a96f6ed0 service nova] Acquiring lock "70df69e5-687b-44fb-b6fc-cdb08e21dda0-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 830.973026] env[63515]: DEBUG oslo_concurrency.lockutils [req-891d504b-70fa-4335-8a13-9e9ff0ca0853 req-1f235373-f1d9-49d9-be25-2517a96f6ed0 service nova] Lock "70df69e5-687b-44fb-b6fc-cdb08e21dda0-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 830.973026] env[63515]: DEBUG oslo_concurrency.lockutils [req-891d504b-70fa-4335-8a13-9e9ff0ca0853 req-1f235373-f1d9-49d9-be25-2517a96f6ed0 service nova] Lock "70df69e5-687b-44fb-b6fc-cdb08e21dda0-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 830.973026] env[63515]: DEBUG nova.compute.manager [req-891d504b-70fa-4335-8a13-9e9ff0ca0853 req-1f235373-f1d9-49d9-be25-2517a96f6ed0 service nova] [instance: 70df69e5-687b-44fb-b6fc-cdb08e21dda0] No waiting events found dispatching network-vif-plugged-2639cfaa-8db0-4dcc-ac66-994ec5a84fdc {{(pid=63515) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 830.973280] env[63515]: WARNING nova.compute.manager [req-891d504b-70fa-4335-8a13-9e9ff0ca0853 req-1f235373-f1d9-49d9-be25-2517a96f6ed0 service nova] [instance: 70df69e5-687b-44fb-b6fc-cdb08e21dda0] Received unexpected event network-vif-plugged-2639cfaa-8db0-4dcc-ac66-994ec5a84fdc for instance with vm_state building and task_state spawning. [ 831.023522] env[63515]: DEBUG nova.network.neutron [None req-9dd4330d-d35d-47d5-b468-11f24e31a49a tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: 70df69e5-687b-44fb-b6fc-cdb08e21dda0] Successfully updated port: 2639cfaa-8db0-4dcc-ac66-994ec5a84fdc {{(pid=63515) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 831.156809] env[63515]: DEBUG nova.objects.instance [None req-b2ae28b1-9ab0-48f9-af42-2813a916c9a6 tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Lazy-loading 'flavor' on Instance uuid d7e8dcdb-41cb-46fb-8b61-d251e7c2d372 {{(pid=63515) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 831.528377] env[63515]: DEBUG oslo_concurrency.lockutils [None req-9dd4330d-d35d-47d5-b468-11f24e31a49a tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Acquiring lock "refresh_cache-70df69e5-687b-44fb-b6fc-cdb08e21dda0" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 831.528499] env[63515]: DEBUG oslo_concurrency.lockutils [None req-9dd4330d-d35d-47d5-b468-11f24e31a49a tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Acquired lock "refresh_cache-70df69e5-687b-44fb-b6fc-cdb08e21dda0" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 831.528591] env[63515]: DEBUG nova.network.neutron [None req-9dd4330d-d35d-47d5-b468-11f24e31a49a tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: 70df69e5-687b-44fb-b6fc-cdb08e21dda0] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 831.584850] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ffdeb8a8-3957-4d2c-857a-59496755ba89 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.594436] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5358ffa2-22c2-4dac-ada2-3e7707c7ce93 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.629737] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3fe36a3a-0d02-4c01-ae1f-2891f8a05b19 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.642272] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5a66b29-7721-4bc5-a362-3c2810e96c23 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.658946] env[63515]: DEBUG nova.compute.provider_tree [None req-7bca9c72-42f5-4d65-99a2-4374b66c9c04 tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 832.079778] env[63515]: DEBUG nova.network.neutron [None req-9dd4330d-d35d-47d5-b468-11f24e31a49a tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: 70df69e5-687b-44fb-b6fc-cdb08e21dda0] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 832.136180] env[63515]: DEBUG oslo_concurrency.lockutils [None req-8ba929a3-7856-4a09-b0c8-d9e96a5313ab tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Acquiring lock "d7e8dcdb-41cb-46fb-8b61-d251e7c2d372" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 832.165026] env[63515]: DEBUG nova.scheduler.client.report [None req-7bca9c72-42f5-4d65-99a2-4374b66c9c04 tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 832.167635] env[63515]: DEBUG oslo_concurrency.lockutils [None req-b2ae28b1-9ab0-48f9-af42-2813a916c9a6 tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Lock "d7e8dcdb-41cb-46fb-8b61-d251e7c2d372" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.271s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 832.169128] env[63515]: DEBUG oslo_concurrency.lockutils [None req-8ba929a3-7856-4a09-b0c8-d9e96a5313ab tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Lock "d7e8dcdb-41cb-46fb-8b61-d251e7c2d372" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.033s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 832.169322] env[63515]: DEBUG nova.compute.manager [None req-8ba929a3-7856-4a09-b0c8-d9e96a5313ab tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] [instance: d7e8dcdb-41cb-46fb-8b61-d251e7c2d372] Checking state {{(pid=63515) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 832.170182] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a77d0951-1269-4f56-851a-03c6ffaa4541 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.183224] env[63515]: DEBUG nova.compute.manager [None req-8ba929a3-7856-4a09-b0c8-d9e96a5313ab tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] [instance: d7e8dcdb-41cb-46fb-8b61-d251e7c2d372] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=63515) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 832.183427] env[63515]: DEBUG nova.objects.instance [None req-8ba929a3-7856-4a09-b0c8-d9e96a5313ab tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Lazy-loading 'flavor' on Instance uuid d7e8dcdb-41cb-46fb-8b61-d251e7c2d372 {{(pid=63515) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 832.248137] env[63515]: DEBUG nova.network.neutron [None req-9dd4330d-d35d-47d5-b468-11f24e31a49a tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: 70df69e5-687b-44fb-b6fc-cdb08e21dda0] Updating instance_info_cache with network_info: [{"id": "2639cfaa-8db0-4dcc-ac66-994ec5a84fdc", "address": "fa:16:3e:fc:b1:b0", "network": {"id": "320921bd-a6c9-4726-a801-19009d70d16e", "bridge": "br-int", "label": "tempest-ImagesTestJSON-304158429-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2259667ddc0c42639ea1af3b6e6fa0e8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2d94740a-bce8-4103-8ecf-230d02ec0a44", "external-id": "nsx-vlan-transportzone-149", "segmentation_id": 149, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2639cfaa-8d", "ovs_interfaceid": "2639cfaa-8db0-4dcc-ac66-994ec5a84fdc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 832.435653] env[63515]: DEBUG oslo_concurrency.lockutils [None req-ed70636f-053b-400c-8dba-ba2b774acd5f tempest-ServerMetadataNegativeTestJSON-502053314 tempest-ServerMetadataNegativeTestJSON-502053314-project-member] Acquiring lock "9d18c7b6-ce8e-4042-9e70-696bb7f57cb3" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 832.435968] env[63515]: DEBUG oslo_concurrency.lockutils [None req-ed70636f-053b-400c-8dba-ba2b774acd5f tempest-ServerMetadataNegativeTestJSON-502053314 tempest-ServerMetadataNegativeTestJSON-502053314-project-member] Lock "9d18c7b6-ce8e-4042-9e70-696bb7f57cb3" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 832.436230] env[63515]: DEBUG oslo_concurrency.lockutils [None req-ed70636f-053b-400c-8dba-ba2b774acd5f tempest-ServerMetadataNegativeTestJSON-502053314 tempest-ServerMetadataNegativeTestJSON-502053314-project-member] Acquiring lock "9d18c7b6-ce8e-4042-9e70-696bb7f57cb3-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 832.436458] env[63515]: DEBUG oslo_concurrency.lockutils [None req-ed70636f-053b-400c-8dba-ba2b774acd5f tempest-ServerMetadataNegativeTestJSON-502053314 tempest-ServerMetadataNegativeTestJSON-502053314-project-member] Lock "9d18c7b6-ce8e-4042-9e70-696bb7f57cb3-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 832.436696] env[63515]: DEBUG oslo_concurrency.lockutils [None req-ed70636f-053b-400c-8dba-ba2b774acd5f tempest-ServerMetadataNegativeTestJSON-502053314 tempest-ServerMetadataNegativeTestJSON-502053314-project-member] Lock "9d18c7b6-ce8e-4042-9e70-696bb7f57cb3-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 832.439083] env[63515]: INFO nova.compute.manager [None req-ed70636f-053b-400c-8dba-ba2b774acd5f tempest-ServerMetadataNegativeTestJSON-502053314 tempest-ServerMetadataNegativeTestJSON-502053314-project-member] [instance: 9d18c7b6-ce8e-4042-9e70-696bb7f57cb3] Terminating instance [ 832.441033] env[63515]: DEBUG nova.compute.manager [None req-ed70636f-053b-400c-8dba-ba2b774acd5f tempest-ServerMetadataNegativeTestJSON-502053314 tempest-ServerMetadataNegativeTestJSON-502053314-project-member] [instance: 9d18c7b6-ce8e-4042-9e70-696bb7f57cb3] Start destroying the instance on the hypervisor. {{(pid=63515) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 832.441238] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-ed70636f-053b-400c-8dba-ba2b774acd5f tempest-ServerMetadataNegativeTestJSON-502053314 tempest-ServerMetadataNegativeTestJSON-502053314-project-member] [instance: 9d18c7b6-ce8e-4042-9e70-696bb7f57cb3] Destroying instance {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 832.442089] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aea97a24-1c90-493a-b90b-184234535fa3 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.450178] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-ed70636f-053b-400c-8dba-ba2b774acd5f tempest-ServerMetadataNegativeTestJSON-502053314 tempest-ServerMetadataNegativeTestJSON-502053314-project-member] [instance: 9d18c7b6-ce8e-4042-9e70-696bb7f57cb3] Powering off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 832.450439] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-fddef94d-de97-4f7a-b08c-5bd3c2873d4e {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.458888] env[63515]: DEBUG oslo_vmware.api [None req-ed70636f-053b-400c-8dba-ba2b774acd5f tempest-ServerMetadataNegativeTestJSON-502053314 tempest-ServerMetadataNegativeTestJSON-502053314-project-member] Waiting for the task: (returnval){ [ 832.458888] env[63515]: value = "task-1111349" [ 832.458888] env[63515]: _type = "Task" [ 832.458888] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 832.467962] env[63515]: DEBUG oslo_vmware.api [None req-ed70636f-053b-400c-8dba-ba2b774acd5f tempest-ServerMetadataNegativeTestJSON-502053314 tempest-ServerMetadataNegativeTestJSON-502053314-project-member] Task: {'id': task-1111349, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.670397] env[63515]: DEBUG oslo_concurrency.lockutils [None req-7bca9c72-42f5-4d65-99a2-4374b66c9c04 tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.380s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 832.671048] env[63515]: DEBUG nova.compute.manager [None req-7bca9c72-42f5-4d65-99a2-4374b66c9c04 tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] [instance: 17275472-2921-49c9-b4e0-5485649ebd1c] Start building networks asynchronously for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 832.674326] env[63515]: DEBUG oslo_concurrency.lockutils [None req-fe95cb73-4a2a-4d94-83ac-9bcd96be3709 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.995s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 832.675805] env[63515]: INFO nova.compute.claims [None req-fe95cb73-4a2a-4d94-83ac-9bcd96be3709 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: 19475235-dbec-40a5-bd8f-8070616b7ddb] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 832.689590] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-8ba929a3-7856-4a09-b0c8-d9e96a5313ab tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] [instance: d7e8dcdb-41cb-46fb-8b61-d251e7c2d372] Powering off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 832.690816] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-11b56aac-302f-4bdd-874c-24152b113207 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.698422] env[63515]: DEBUG oslo_vmware.api [None req-8ba929a3-7856-4a09-b0c8-d9e96a5313ab tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Waiting for the task: (returnval){ [ 832.698422] env[63515]: value = "task-1111351" [ 832.698422] env[63515]: _type = "Task" [ 832.698422] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 832.707913] env[63515]: DEBUG oslo_vmware.api [None req-8ba929a3-7856-4a09-b0c8-d9e96a5313ab tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Task: {'id': task-1111351, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.750368] env[63515]: DEBUG oslo_concurrency.lockutils [None req-9dd4330d-d35d-47d5-b468-11f24e31a49a tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Releasing lock "refresh_cache-70df69e5-687b-44fb-b6fc-cdb08e21dda0" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 832.750735] env[63515]: DEBUG nova.compute.manager [None req-9dd4330d-d35d-47d5-b468-11f24e31a49a tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: 70df69e5-687b-44fb-b6fc-cdb08e21dda0] Instance network_info: |[{"id": "2639cfaa-8db0-4dcc-ac66-994ec5a84fdc", "address": "fa:16:3e:fc:b1:b0", "network": {"id": "320921bd-a6c9-4726-a801-19009d70d16e", "bridge": "br-int", "label": "tempest-ImagesTestJSON-304158429-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2259667ddc0c42639ea1af3b6e6fa0e8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2d94740a-bce8-4103-8ecf-230d02ec0a44", "external-id": "nsx-vlan-transportzone-149", "segmentation_id": 149, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2639cfaa-8d", "ovs_interfaceid": "2639cfaa-8db0-4dcc-ac66-994ec5a84fdc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 832.751521] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-9dd4330d-d35d-47d5-b468-11f24e31a49a tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: 70df69e5-687b-44fb-b6fc-cdb08e21dda0] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:fc:b1:b0', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '2d94740a-bce8-4103-8ecf-230d02ec0a44', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '2639cfaa-8db0-4dcc-ac66-994ec5a84fdc', 'vif_model': 'vmxnet3'}] {{(pid=63515) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 832.759142] env[63515]: DEBUG oslo.service.loopingcall [None req-9dd4330d-d35d-47d5-b468-11f24e31a49a tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 832.759371] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 70df69e5-687b-44fb-b6fc-cdb08e21dda0] Creating VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 832.759618] env[63515]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a9fddc7a-1b84-40d2-b27a-729ae0eee57a {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.780498] env[63515]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 832.780498] env[63515]: value = "task-1111352" [ 832.780498] env[63515]: _type = "Task" [ 832.780498] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 832.788814] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1111352, 'name': CreateVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.970300] env[63515]: DEBUG oslo_vmware.api [None req-ed70636f-053b-400c-8dba-ba2b774acd5f tempest-ServerMetadataNegativeTestJSON-502053314 tempest-ServerMetadataNegativeTestJSON-502053314-project-member] Task: {'id': task-1111349, 'name': PowerOffVM_Task, 'duration_secs': 0.215394} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 832.970625] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-ed70636f-053b-400c-8dba-ba2b774acd5f tempest-ServerMetadataNegativeTestJSON-502053314 tempest-ServerMetadataNegativeTestJSON-502053314-project-member] [instance: 9d18c7b6-ce8e-4042-9e70-696bb7f57cb3] Powered off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 832.970993] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-ed70636f-053b-400c-8dba-ba2b774acd5f tempest-ServerMetadataNegativeTestJSON-502053314 tempest-ServerMetadataNegativeTestJSON-502053314-project-member] [instance: 9d18c7b6-ce8e-4042-9e70-696bb7f57cb3] Unregistering the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 832.971081] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-73b8d82d-80b0-421d-b543-1cd875ad432b {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.995712] env[63515]: DEBUG nova.compute.manager [req-74836b77-ab91-4209-b1e0-62dc4568b5ac req-d7acf59d-1af7-4b8f-87e6-41e55817063b service nova] [instance: 70df69e5-687b-44fb-b6fc-cdb08e21dda0] Received event network-changed-2639cfaa-8db0-4dcc-ac66-994ec5a84fdc {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 832.995931] env[63515]: DEBUG nova.compute.manager [req-74836b77-ab91-4209-b1e0-62dc4568b5ac req-d7acf59d-1af7-4b8f-87e6-41e55817063b service nova] [instance: 70df69e5-687b-44fb-b6fc-cdb08e21dda0] Refreshing instance network info cache due to event network-changed-2639cfaa-8db0-4dcc-ac66-994ec5a84fdc. {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 832.996176] env[63515]: DEBUG oslo_concurrency.lockutils [req-74836b77-ab91-4209-b1e0-62dc4568b5ac req-d7acf59d-1af7-4b8f-87e6-41e55817063b service nova] Acquiring lock "refresh_cache-70df69e5-687b-44fb-b6fc-cdb08e21dda0" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 832.996323] env[63515]: DEBUG oslo_concurrency.lockutils [req-74836b77-ab91-4209-b1e0-62dc4568b5ac req-d7acf59d-1af7-4b8f-87e6-41e55817063b service nova] Acquired lock "refresh_cache-70df69e5-687b-44fb-b6fc-cdb08e21dda0" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 832.996487] env[63515]: DEBUG nova.network.neutron [req-74836b77-ab91-4209-b1e0-62dc4568b5ac req-d7acf59d-1af7-4b8f-87e6-41e55817063b service nova] [instance: 70df69e5-687b-44fb-b6fc-cdb08e21dda0] Refreshing network info cache for port 2639cfaa-8db0-4dcc-ac66-994ec5a84fdc {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 833.039390] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-ed70636f-053b-400c-8dba-ba2b774acd5f tempest-ServerMetadataNegativeTestJSON-502053314 tempest-ServerMetadataNegativeTestJSON-502053314-project-member] [instance: 9d18c7b6-ce8e-4042-9e70-696bb7f57cb3] Unregistered the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 833.039671] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-ed70636f-053b-400c-8dba-ba2b774acd5f tempest-ServerMetadataNegativeTestJSON-502053314 tempest-ServerMetadataNegativeTestJSON-502053314-project-member] [instance: 9d18c7b6-ce8e-4042-9e70-696bb7f57cb3] Deleting contents of the VM from datastore datastore2 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 833.039812] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-ed70636f-053b-400c-8dba-ba2b774acd5f tempest-ServerMetadataNegativeTestJSON-502053314 tempest-ServerMetadataNegativeTestJSON-502053314-project-member] Deleting the datastore file [datastore2] 9d18c7b6-ce8e-4042-9e70-696bb7f57cb3 {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 833.040410] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-9bb5ae2f-af3d-45e5-9d5d-213e496f4dde {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.048388] env[63515]: DEBUG oslo_vmware.api [None req-ed70636f-053b-400c-8dba-ba2b774acd5f tempest-ServerMetadataNegativeTestJSON-502053314 tempest-ServerMetadataNegativeTestJSON-502053314-project-member] Waiting for the task: (returnval){ [ 833.048388] env[63515]: value = "task-1111354" [ 833.048388] env[63515]: _type = "Task" [ 833.048388] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 833.056534] env[63515]: DEBUG oslo_vmware.api [None req-ed70636f-053b-400c-8dba-ba2b774acd5f tempest-ServerMetadataNegativeTestJSON-502053314 tempest-ServerMetadataNegativeTestJSON-502053314-project-member] Task: {'id': task-1111354, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.180515] env[63515]: DEBUG nova.compute.utils [None req-7bca9c72-42f5-4d65-99a2-4374b66c9c04 tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] Using /dev/sd instead of None {{(pid=63515) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 833.186050] env[63515]: DEBUG nova.compute.manager [None req-7bca9c72-42f5-4d65-99a2-4374b66c9c04 tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] [instance: 17275472-2921-49c9-b4e0-5485649ebd1c] Allocating IP information in the background. {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 833.186299] env[63515]: DEBUG nova.network.neutron [None req-7bca9c72-42f5-4d65-99a2-4374b66c9c04 tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] [instance: 17275472-2921-49c9-b4e0-5485649ebd1c] allocate_for_instance() {{(pid=63515) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 833.208011] env[63515]: DEBUG oslo_vmware.api [None req-8ba929a3-7856-4a09-b0c8-d9e96a5313ab tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Task: {'id': task-1111351, 'name': PowerOffVM_Task, 'duration_secs': 0.239136} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 833.208260] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-8ba929a3-7856-4a09-b0c8-d9e96a5313ab tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] [instance: d7e8dcdb-41cb-46fb-8b61-d251e7c2d372] Powered off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 833.208479] env[63515]: DEBUG nova.compute.manager [None req-8ba929a3-7856-4a09-b0c8-d9e96a5313ab tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] [instance: d7e8dcdb-41cb-46fb-8b61-d251e7c2d372] Checking state {{(pid=63515) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 833.209298] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2aff5877-eda3-414b-bf79-b1711eeb803a {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.254846] env[63515]: DEBUG nova.policy [None req-7bca9c72-42f5-4d65-99a2-4374b66c9c04 tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b0aec3663b22457f8b87f797ccb41af6', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '26a52fb1944a4f949f6779829e163eaa', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63515) authorize /opt/stack/nova/nova/policy.py:201}} [ 833.292674] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1111352, 'name': CreateVM_Task, 'duration_secs': 0.341381} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 833.292851] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 70df69e5-687b-44fb-b6fc-cdb08e21dda0] Created VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 833.293571] env[63515]: DEBUG oslo_concurrency.lockutils [None req-9dd4330d-d35d-47d5-b468-11f24e31a49a tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 833.293808] env[63515]: DEBUG oslo_concurrency.lockutils [None req-9dd4330d-d35d-47d5-b468-11f24e31a49a tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 833.294157] env[63515]: DEBUG oslo_concurrency.lockutils [None req-9dd4330d-d35d-47d5-b468-11f24e31a49a tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 833.294413] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7270ad9b-0e5b-48cd-967b-03ff280cbb71 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.299273] env[63515]: DEBUG oslo_vmware.api [None req-9dd4330d-d35d-47d5-b468-11f24e31a49a tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Waiting for the task: (returnval){ [ 833.299273] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]527ff8e8-3708-97f7-49da-5483041d70e1" [ 833.299273] env[63515]: _type = "Task" [ 833.299273] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 833.307486] env[63515]: DEBUG oslo_vmware.api [None req-9dd4330d-d35d-47d5-b468-11f24e31a49a tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]527ff8e8-3708-97f7-49da-5483041d70e1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.561248] env[63515]: DEBUG oslo_vmware.api [None req-ed70636f-053b-400c-8dba-ba2b774acd5f tempest-ServerMetadataNegativeTestJSON-502053314 tempest-ServerMetadataNegativeTestJSON-502053314-project-member] Task: {'id': task-1111354, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.133347} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 833.561523] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-ed70636f-053b-400c-8dba-ba2b774acd5f tempest-ServerMetadataNegativeTestJSON-502053314 tempest-ServerMetadataNegativeTestJSON-502053314-project-member] Deleted the datastore file {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 833.561793] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-ed70636f-053b-400c-8dba-ba2b774acd5f tempest-ServerMetadataNegativeTestJSON-502053314 tempest-ServerMetadataNegativeTestJSON-502053314-project-member] [instance: 9d18c7b6-ce8e-4042-9e70-696bb7f57cb3] Deleted contents of the VM from datastore datastore2 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 833.561916] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-ed70636f-053b-400c-8dba-ba2b774acd5f tempest-ServerMetadataNegativeTestJSON-502053314 tempest-ServerMetadataNegativeTestJSON-502053314-project-member] [instance: 9d18c7b6-ce8e-4042-9e70-696bb7f57cb3] Instance destroyed {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 833.562998] env[63515]: INFO nova.compute.manager [None req-ed70636f-053b-400c-8dba-ba2b774acd5f tempest-ServerMetadataNegativeTestJSON-502053314 tempest-ServerMetadataNegativeTestJSON-502053314-project-member] [instance: 9d18c7b6-ce8e-4042-9e70-696bb7f57cb3] Took 1.12 seconds to destroy the instance on the hypervisor. [ 833.563331] env[63515]: DEBUG oslo.service.loopingcall [None req-ed70636f-053b-400c-8dba-ba2b774acd5f tempest-ServerMetadataNegativeTestJSON-502053314 tempest-ServerMetadataNegativeTestJSON-502053314-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 833.563565] env[63515]: DEBUG nova.compute.manager [-] [instance: 9d18c7b6-ce8e-4042-9e70-696bb7f57cb3] Deallocating network for instance {{(pid=63515) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 833.563672] env[63515]: DEBUG nova.network.neutron [-] [instance: 9d18c7b6-ce8e-4042-9e70-696bb7f57cb3] deallocate_for_instance() {{(pid=63515) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 833.583815] env[63515]: DEBUG nova.network.neutron [None req-7bca9c72-42f5-4d65-99a2-4374b66c9c04 tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] [instance: 17275472-2921-49c9-b4e0-5485649ebd1c] Successfully created port: 5514ac51-9696-4c26-9abc-2cc0deef264e {{(pid=63515) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 833.686997] env[63515]: DEBUG nova.compute.manager [None req-7bca9c72-42f5-4d65-99a2-4374b66c9c04 tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] [instance: 17275472-2921-49c9-b4e0-5485649ebd1c] Start building block device mappings for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 833.722152] env[63515]: DEBUG oslo_concurrency.lockutils [None req-8ba929a3-7856-4a09-b0c8-d9e96a5313ab tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Lock "d7e8dcdb-41cb-46fb-8b61-d251e7c2d372" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.552s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 833.812930] env[63515]: DEBUG oslo_vmware.api [None req-9dd4330d-d35d-47d5-b468-11f24e31a49a tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]527ff8e8-3708-97f7-49da-5483041d70e1, 'name': SearchDatastore_Task, 'duration_secs': 0.00962} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 833.812930] env[63515]: DEBUG oslo_concurrency.lockutils [None req-9dd4330d-d35d-47d5-b468-11f24e31a49a tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 833.813104] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-9dd4330d-d35d-47d5-b468-11f24e31a49a tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: 70df69e5-687b-44fb-b6fc-cdb08e21dda0] Processing image 8a120570-cb06-4099-b262-554ca0ad15c5 {{(pid=63515) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 833.813273] env[63515]: DEBUG oslo_concurrency.lockutils [None req-9dd4330d-d35d-47d5-b468-11f24e31a49a tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 833.813437] env[63515]: DEBUG oslo_concurrency.lockutils [None req-9dd4330d-d35d-47d5-b468-11f24e31a49a tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 833.813638] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-9dd4330d-d35d-47d5-b468-11f24e31a49a tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 833.816139] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6796def6-5db0-4452-abe9-117db5995df9 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.828348] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-9dd4330d-d35d-47d5-b468-11f24e31a49a tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 833.828552] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-9dd4330d-d35d-47d5-b468-11f24e31a49a tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63515) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 833.829706] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d54abf9f-61ad-486d-824e-0c444fc45fff {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.838115] env[63515]: DEBUG oslo_vmware.api [None req-9dd4330d-d35d-47d5-b468-11f24e31a49a tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Waiting for the task: (returnval){ [ 833.838115] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]525e8ef9-4b58-fb3a-c44c-bfd3e17233b7" [ 833.838115] env[63515]: _type = "Task" [ 833.838115] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 833.848681] env[63515]: DEBUG oslo_vmware.api [None req-9dd4330d-d35d-47d5-b468-11f24e31a49a tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]525e8ef9-4b58-fb3a-c44c-bfd3e17233b7, 'name': SearchDatastore_Task, 'duration_secs': 0.00869} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 833.851516] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3517b133-ce5c-48ac-bddc-091d96ff46a0 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.857111] env[63515]: DEBUG oslo_vmware.api [None req-9dd4330d-d35d-47d5-b468-11f24e31a49a tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Waiting for the task: (returnval){ [ 833.857111] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]52ba93e0-f9d5-d571-ed58-02ed8161359f" [ 833.857111] env[63515]: _type = "Task" [ 833.857111] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 833.867431] env[63515]: DEBUG oslo_vmware.api [None req-9dd4330d-d35d-47d5-b468-11f24e31a49a tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52ba93e0-f9d5-d571-ed58-02ed8161359f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.895591] env[63515]: DEBUG nova.network.neutron [req-74836b77-ab91-4209-b1e0-62dc4568b5ac req-d7acf59d-1af7-4b8f-87e6-41e55817063b service nova] [instance: 70df69e5-687b-44fb-b6fc-cdb08e21dda0] Updated VIF entry in instance network info cache for port 2639cfaa-8db0-4dcc-ac66-994ec5a84fdc. {{(pid=63515) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 833.895847] env[63515]: DEBUG nova.network.neutron [req-74836b77-ab91-4209-b1e0-62dc4568b5ac req-d7acf59d-1af7-4b8f-87e6-41e55817063b service nova] [instance: 70df69e5-687b-44fb-b6fc-cdb08e21dda0] Updating instance_info_cache with network_info: [{"id": "2639cfaa-8db0-4dcc-ac66-994ec5a84fdc", "address": "fa:16:3e:fc:b1:b0", "network": {"id": "320921bd-a6c9-4726-a801-19009d70d16e", "bridge": "br-int", "label": "tempest-ImagesTestJSON-304158429-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2259667ddc0c42639ea1af3b6e6fa0e8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2d94740a-bce8-4103-8ecf-230d02ec0a44", "external-id": "nsx-vlan-transportzone-149", "segmentation_id": 149, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2639cfaa-8d", "ovs_interfaceid": "2639cfaa-8db0-4dcc-ac66-994ec5a84fdc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 833.931067] env[63515]: DEBUG nova.network.neutron [None req-7bca9c72-42f5-4d65-99a2-4374b66c9c04 tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] [instance: 17275472-2921-49c9-b4e0-5485649ebd1c] Successfully created port: db47ccfa-8557-4128-9e65-8a8ab57163bb {{(pid=63515) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 834.035111] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7cc33f36-3098-4a02-a3d3-c8138844de13 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.044407] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0278ba75-dbb8-4c47-bc48-ba5f980914b1 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.050237] env[63515]: DEBUG nova.compute.manager [req-63d6a364-5639-41d2-9607-77ed33f3963b req-cdd15b45-db0e-435d-b0ff-285ad3f5f21f service nova] [instance: 9d18c7b6-ce8e-4042-9e70-696bb7f57cb3] Received event network-vif-deleted-ef0f7ff0-3d3e-4159-910f-c0c0e07aa269 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 834.050450] env[63515]: INFO nova.compute.manager [req-63d6a364-5639-41d2-9607-77ed33f3963b req-cdd15b45-db0e-435d-b0ff-285ad3f5f21f service nova] [instance: 9d18c7b6-ce8e-4042-9e70-696bb7f57cb3] Neutron deleted interface ef0f7ff0-3d3e-4159-910f-c0c0e07aa269; detaching it from the instance and deleting it from the info cache [ 834.050632] env[63515]: DEBUG nova.network.neutron [req-63d6a364-5639-41d2-9607-77ed33f3963b req-cdd15b45-db0e-435d-b0ff-285ad3f5f21f service nova] [instance: 9d18c7b6-ce8e-4042-9e70-696bb7f57cb3] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 834.084677] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f577f3a2-7cc7-4577-9094-d584607cb269 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.094225] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe6c2679-5eb7-4136-96fe-8441f0336e28 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.109089] env[63515]: DEBUG nova.compute.provider_tree [None req-fe95cb73-4a2a-4d94-83ac-9bcd96be3709 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 834.369327] env[63515]: DEBUG oslo_vmware.api [None req-9dd4330d-d35d-47d5-b468-11f24e31a49a tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52ba93e0-f9d5-d571-ed58-02ed8161359f, 'name': SearchDatastore_Task, 'duration_secs': 0.009181} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.369327] env[63515]: DEBUG oslo_concurrency.lockutils [None req-9dd4330d-d35d-47d5-b468-11f24e31a49a tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 834.369595] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-9dd4330d-d35d-47d5-b468-11f24e31a49a tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk to [datastore2] 70df69e5-687b-44fb-b6fc-cdb08e21dda0/70df69e5-687b-44fb-b6fc-cdb08e21dda0.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 834.369823] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-cd0ed2f8-1dfb-4f1b-b8d4-4b78a0f5a3cc {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.378303] env[63515]: DEBUG oslo_vmware.api [None req-9dd4330d-d35d-47d5-b468-11f24e31a49a tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Waiting for the task: (returnval){ [ 834.378303] env[63515]: value = "task-1111355" [ 834.378303] env[63515]: _type = "Task" [ 834.378303] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 834.387669] env[63515]: DEBUG oslo_vmware.api [None req-9dd4330d-d35d-47d5-b468-11f24e31a49a tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Task: {'id': task-1111355, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.398773] env[63515]: DEBUG oslo_concurrency.lockutils [req-74836b77-ab91-4209-b1e0-62dc4568b5ac req-d7acf59d-1af7-4b8f-87e6-41e55817063b service nova] Releasing lock "refresh_cache-70df69e5-687b-44fb-b6fc-cdb08e21dda0" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 834.475156] env[63515]: DEBUG nova.network.neutron [-] [instance: 9d18c7b6-ce8e-4042-9e70-696bb7f57cb3] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 834.535802] env[63515]: DEBUG nova.network.neutron [None req-7bca9c72-42f5-4d65-99a2-4374b66c9c04 tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] [instance: 17275472-2921-49c9-b4e0-5485649ebd1c] Successfully created port: eadc0731-8390-4414-b5cb-0f6a0c660979 {{(pid=63515) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 834.555153] env[63515]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-81aa9586-0990-4c5d-816d-14c3f71ef427 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.566389] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-580e0c5f-cf14-415f-9263-2881db17539a {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.600286] env[63515]: DEBUG nova.compute.manager [req-63d6a364-5639-41d2-9607-77ed33f3963b req-cdd15b45-db0e-435d-b0ff-285ad3f5f21f service nova] [instance: 9d18c7b6-ce8e-4042-9e70-696bb7f57cb3] Detach interface failed, port_id=ef0f7ff0-3d3e-4159-910f-c0c0e07aa269, reason: Instance 9d18c7b6-ce8e-4042-9e70-696bb7f57cb3 could not be found. {{(pid=63515) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 834.611748] env[63515]: DEBUG nova.scheduler.client.report [None req-fe95cb73-4a2a-4d94-83ac-9bcd96be3709 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 834.700575] env[63515]: DEBUG nova.compute.manager [None req-7bca9c72-42f5-4d65-99a2-4374b66c9c04 tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] [instance: 17275472-2921-49c9-b4e0-5485649ebd1c] Start spawning the instance on the hypervisor. {{(pid=63515) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 834.734911] env[63515]: DEBUG nova.virt.hardware [None req-7bca9c72-42f5-4d65-99a2-4374b66c9c04 tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T02:52:14Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T02:51:56Z,direct_url=,disk_format='vmdk',id=8a120570-cb06-4099-b262-554ca0ad15c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b569255cc43e42e7ae3f2b4ad37c6ef4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T02:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 834.735238] env[63515]: DEBUG nova.virt.hardware [None req-7bca9c72-42f5-4d65-99a2-4374b66c9c04 tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] Flavor limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 834.735405] env[63515]: DEBUG nova.virt.hardware [None req-7bca9c72-42f5-4d65-99a2-4374b66c9c04 tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] Image limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 834.735604] env[63515]: DEBUG nova.virt.hardware [None req-7bca9c72-42f5-4d65-99a2-4374b66c9c04 tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] Flavor pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 834.735751] env[63515]: DEBUG nova.virt.hardware [None req-7bca9c72-42f5-4d65-99a2-4374b66c9c04 tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] Image pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 834.735905] env[63515]: DEBUG nova.virt.hardware [None req-7bca9c72-42f5-4d65-99a2-4374b66c9c04 tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 834.736353] env[63515]: DEBUG nova.virt.hardware [None req-7bca9c72-42f5-4d65-99a2-4374b66c9c04 tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 834.736353] env[63515]: DEBUG nova.virt.hardware [None req-7bca9c72-42f5-4d65-99a2-4374b66c9c04 tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 834.736525] env[63515]: DEBUG nova.virt.hardware [None req-7bca9c72-42f5-4d65-99a2-4374b66c9c04 tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] Got 1 possible topologies {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 834.736690] env[63515]: DEBUG nova.virt.hardware [None req-7bca9c72-42f5-4d65-99a2-4374b66c9c04 tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 834.736870] env[63515]: DEBUG nova.virt.hardware [None req-7bca9c72-42f5-4d65-99a2-4374b66c9c04 tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 834.737817] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6120dcc-54c8-4661-b1e1-5a8cf95eb0cb {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.749621] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d6f8eef-d645-421f-9a2d-0a2d6dd04f8e {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.760868] env[63515]: DEBUG nova.objects.instance [None req-c5d147d2-55a3-44f2-9b1a-4c348ae09dd7 tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Lazy-loading 'flavor' on Instance uuid d7e8dcdb-41cb-46fb-8b61-d251e7c2d372 {{(pid=63515) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 834.782799] env[63515]: DEBUG oslo_concurrency.lockutils [None req-c5d147d2-55a3-44f2-9b1a-4c348ae09dd7 tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Acquiring lock "refresh_cache-d7e8dcdb-41cb-46fb-8b61-d251e7c2d372" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 834.783145] env[63515]: DEBUG oslo_concurrency.lockutils [None req-c5d147d2-55a3-44f2-9b1a-4c348ae09dd7 tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Acquired lock "refresh_cache-d7e8dcdb-41cb-46fb-8b61-d251e7c2d372" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 834.783398] env[63515]: DEBUG nova.network.neutron [None req-c5d147d2-55a3-44f2-9b1a-4c348ae09dd7 tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] [instance: d7e8dcdb-41cb-46fb-8b61-d251e7c2d372] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 834.783721] env[63515]: DEBUG nova.objects.instance [None req-c5d147d2-55a3-44f2-9b1a-4c348ae09dd7 tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Lazy-loading 'info_cache' on Instance uuid d7e8dcdb-41cb-46fb-8b61-d251e7c2d372 {{(pid=63515) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 834.887833] env[63515]: DEBUG oslo_vmware.api [None req-9dd4330d-d35d-47d5-b468-11f24e31a49a tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Task: {'id': task-1111355, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.979159] env[63515]: INFO nova.compute.manager [-] [instance: 9d18c7b6-ce8e-4042-9e70-696bb7f57cb3] Took 1.42 seconds to deallocate network for instance. [ 835.118184] env[63515]: DEBUG oslo_concurrency.lockutils [None req-fe95cb73-4a2a-4d94-83ac-9bcd96be3709 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.443s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 835.118395] env[63515]: DEBUG nova.compute.manager [None req-fe95cb73-4a2a-4d94-83ac-9bcd96be3709 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: 19475235-dbec-40a5-bd8f-8070616b7ddb] Start building networks asynchronously for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 835.122112] env[63515]: DEBUG oslo_concurrency.lockutils [None req-73bb59a8-34a2-4cd0-acb3-804cea19b215 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 21.847s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 835.122112] env[63515]: DEBUG nova.objects.instance [None req-73bb59a8-34a2-4cd0-acb3-804cea19b215 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] Lazy-loading 'resources' on Instance uuid e666825c-ff4e-4a0e-93c0-43c00f167bbb {{(pid=63515) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 835.289176] env[63515]: DEBUG nova.objects.base [None req-c5d147d2-55a3-44f2-9b1a-4c348ae09dd7 tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Object Instance lazy-loaded attributes: flavor,info_cache {{(pid=63515) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 835.390430] env[63515]: DEBUG oslo_vmware.api [None req-9dd4330d-d35d-47d5-b468-11f24e31a49a tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Task: {'id': task-1111355, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.488099] env[63515]: DEBUG oslo_concurrency.lockutils [None req-ed70636f-053b-400c-8dba-ba2b774acd5f tempest-ServerMetadataNegativeTestJSON-502053314 tempest-ServerMetadataNegativeTestJSON-502053314-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 835.622914] env[63515]: DEBUG nova.compute.utils [None req-fe95cb73-4a2a-4d94-83ac-9bcd96be3709 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Using /dev/sd instead of None {{(pid=63515) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 835.624737] env[63515]: DEBUG nova.compute.manager [None req-fe95cb73-4a2a-4d94-83ac-9bcd96be3709 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: 19475235-dbec-40a5-bd8f-8070616b7ddb] Allocating IP information in the background. {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 835.624737] env[63515]: DEBUG nova.network.neutron [None req-fe95cb73-4a2a-4d94-83ac-9bcd96be3709 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: 19475235-dbec-40a5-bd8f-8070616b7ddb] allocate_for_instance() {{(pid=63515) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 835.667231] env[63515]: DEBUG nova.policy [None req-fe95cb73-4a2a-4d94-83ac-9bcd96be3709 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9b19bb2d32d84c019541c3b2e711a202', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '299fdeff647f486390366d5bbf911518', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63515) authorize /opt/stack/nova/nova/policy.py:201}} [ 835.890359] env[63515]: DEBUG oslo_vmware.api [None req-9dd4330d-d35d-47d5-b468-11f24e31a49a tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Task: {'id': task-1111355, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.468181} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 835.892926] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-9dd4330d-d35d-47d5-b468-11f24e31a49a tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk to [datastore2] 70df69e5-687b-44fb-b6fc-cdb08e21dda0/70df69e5-687b-44fb-b6fc-cdb08e21dda0.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 835.893170] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-9dd4330d-d35d-47d5-b468-11f24e31a49a tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: 70df69e5-687b-44fb-b6fc-cdb08e21dda0] Extending root virtual disk to 1048576 {{(pid=63515) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 835.893624] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-dc39564a-4366-48ed-8dbf-471bb0b80db5 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.900713] env[63515]: DEBUG oslo_vmware.api [None req-9dd4330d-d35d-47d5-b468-11f24e31a49a tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Waiting for the task: (returnval){ [ 835.900713] env[63515]: value = "task-1111357" [ 835.900713] env[63515]: _type = "Task" [ 835.900713] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.916800] env[63515]: DEBUG oslo_vmware.api [None req-9dd4330d-d35d-47d5-b468-11f24e31a49a tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Task: {'id': task-1111357, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.964870] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd5bea93-36f4-46f3-ba59-0979861dbd79 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.976235] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fec4ff48-a26d-4cf4-999c-caedce1ec12b {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.013739] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c59e240c-0fb2-4ad8-9079-294a58bf8b2f {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.028234] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c90234e8-593c-4b0c-8d23-8d5525eaa53a {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.044528] env[63515]: DEBUG nova.compute.provider_tree [None req-73bb59a8-34a2-4cd0-acb3-804cea19b215 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 836.130503] env[63515]: DEBUG nova.compute.manager [None req-fe95cb73-4a2a-4d94-83ac-9bcd96be3709 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: 19475235-dbec-40a5-bd8f-8070616b7ddb] Start building block device mappings for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 836.147227] env[63515]: DEBUG nova.network.neutron [None req-fe95cb73-4a2a-4d94-83ac-9bcd96be3709 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: 19475235-dbec-40a5-bd8f-8070616b7ddb] Successfully created port: 372ea923-ddc7-4b61-92db-c55f8f352b44 {{(pid=63515) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 836.162330] env[63515]: DEBUG nova.compute.manager [req-e58bd49f-f7d5-4e2a-b95a-1e445d88166d req-f7919e12-f3f8-42bb-84a0-1379c86063d9 service nova] [instance: 17275472-2921-49c9-b4e0-5485649ebd1c] Received event network-vif-plugged-5514ac51-9696-4c26-9abc-2cc0deef264e {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 836.162645] env[63515]: DEBUG oslo_concurrency.lockutils [req-e58bd49f-f7d5-4e2a-b95a-1e445d88166d req-f7919e12-f3f8-42bb-84a0-1379c86063d9 service nova] Acquiring lock "17275472-2921-49c9-b4e0-5485649ebd1c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 836.163358] env[63515]: DEBUG oslo_concurrency.lockutils [req-e58bd49f-f7d5-4e2a-b95a-1e445d88166d req-f7919e12-f3f8-42bb-84a0-1379c86063d9 service nova] Lock "17275472-2921-49c9-b4e0-5485649ebd1c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 836.163585] env[63515]: DEBUG oslo_concurrency.lockutils [req-e58bd49f-f7d5-4e2a-b95a-1e445d88166d req-f7919e12-f3f8-42bb-84a0-1379c86063d9 service nova] Lock "17275472-2921-49c9-b4e0-5485649ebd1c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 836.163748] env[63515]: DEBUG nova.compute.manager [req-e58bd49f-f7d5-4e2a-b95a-1e445d88166d req-f7919e12-f3f8-42bb-84a0-1379c86063d9 service nova] [instance: 17275472-2921-49c9-b4e0-5485649ebd1c] No waiting events found dispatching network-vif-plugged-5514ac51-9696-4c26-9abc-2cc0deef264e {{(pid=63515) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 836.163996] env[63515]: WARNING nova.compute.manager [req-e58bd49f-f7d5-4e2a-b95a-1e445d88166d req-f7919e12-f3f8-42bb-84a0-1379c86063d9 service nova] [instance: 17275472-2921-49c9-b4e0-5485649ebd1c] Received unexpected event network-vif-plugged-5514ac51-9696-4c26-9abc-2cc0deef264e for instance with vm_state building and task_state spawning. [ 836.218741] env[63515]: DEBUG nova.network.neutron [None req-c5d147d2-55a3-44f2-9b1a-4c348ae09dd7 tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] [instance: d7e8dcdb-41cb-46fb-8b61-d251e7c2d372] Updating instance_info_cache with network_info: [{"id": "9378f59d-f9bc-4c2e-8336-2c5710f987cf", "address": "fa:16:3e:7b:d9:e1", "network": {"id": "57de0b07-78e5-453d-83ba-b12b7dcaf6ce", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-272799484-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.170", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "12a90929c98e40d9a1263929e2b97532", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a1895250-76cc-41f7-b7f8-2e5679494607", "external-id": "nsx-vlan-transportzone-785", "segmentation_id": 785, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9378f59d-f9", "ovs_interfaceid": "9378f59d-f9bc-4c2e-8336-2c5710f987cf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 836.260630] env[63515]: DEBUG nova.network.neutron [None req-7bca9c72-42f5-4d65-99a2-4374b66c9c04 tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] [instance: 17275472-2921-49c9-b4e0-5485649ebd1c] Successfully updated port: 5514ac51-9696-4c26-9abc-2cc0deef264e {{(pid=63515) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 836.413839] env[63515]: DEBUG oslo_vmware.api [None req-9dd4330d-d35d-47d5-b468-11f24e31a49a tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Task: {'id': task-1111357, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.067405} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.415062] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-9dd4330d-d35d-47d5-b468-11f24e31a49a tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: 70df69e5-687b-44fb-b6fc-cdb08e21dda0] Extended root virtual disk {{(pid=63515) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 836.415523] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d36061c-8b14-4a58-bdd1-424356bd18a6 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.440984] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-9dd4330d-d35d-47d5-b468-11f24e31a49a tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: 70df69e5-687b-44fb-b6fc-cdb08e21dda0] Reconfiguring VM instance instance-00000041 to attach disk [datastore2] 70df69e5-687b-44fb-b6fc-cdb08e21dda0/70df69e5-687b-44fb-b6fc-cdb08e21dda0.vmdk or device None with type sparse {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 836.441510] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c872d417-c7e6-45cf-b1e2-b8e4b041f7ff {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.469146] env[63515]: DEBUG oslo_vmware.api [None req-9dd4330d-d35d-47d5-b468-11f24e31a49a tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Waiting for the task: (returnval){ [ 836.469146] env[63515]: value = "task-1111358" [ 836.469146] env[63515]: _type = "Task" [ 836.469146] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.479388] env[63515]: DEBUG oslo_vmware.api [None req-9dd4330d-d35d-47d5-b468-11f24e31a49a tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Task: {'id': task-1111358, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.548412] env[63515]: DEBUG nova.scheduler.client.report [None req-73bb59a8-34a2-4cd0-acb3-804cea19b215 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 836.723921] env[63515]: DEBUG oslo_concurrency.lockutils [None req-c5d147d2-55a3-44f2-9b1a-4c348ae09dd7 tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Releasing lock "refresh_cache-d7e8dcdb-41cb-46fb-8b61-d251e7c2d372" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 836.982620] env[63515]: DEBUG oslo_vmware.api [None req-9dd4330d-d35d-47d5-b468-11f24e31a49a tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Task: {'id': task-1111358, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.055754] env[63515]: DEBUG oslo_concurrency.lockutils [None req-73bb59a8-34a2-4cd0-acb3-804cea19b215 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.934s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 837.058901] env[63515]: DEBUG oslo_concurrency.lockutils [None req-08dbd0f6-25be-4803-9b5a-523d28be6f7f tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 23.626s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 837.058901] env[63515]: DEBUG nova.objects.instance [None req-08dbd0f6-25be-4803-9b5a-523d28be6f7f tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Lazy-loading 'resources' on Instance uuid 6688cdaa-29ba-413a-8131-4f834cdb70e4 {{(pid=63515) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 837.082054] env[63515]: INFO nova.scheduler.client.report [None req-73bb59a8-34a2-4cd0-acb3-804cea19b215 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] Deleted allocations for instance e666825c-ff4e-4a0e-93c0-43c00f167bbb [ 837.145102] env[63515]: DEBUG nova.compute.manager [None req-fe95cb73-4a2a-4d94-83ac-9bcd96be3709 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: 19475235-dbec-40a5-bd8f-8070616b7ddb] Start spawning the instance on the hypervisor. {{(pid=63515) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 837.169354] env[63515]: DEBUG nova.virt.hardware [None req-fe95cb73-4a2a-4d94-83ac-9bcd96be3709 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T02:52:14Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T02:51:56Z,direct_url=,disk_format='vmdk',id=8a120570-cb06-4099-b262-554ca0ad15c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b569255cc43e42e7ae3f2b4ad37c6ef4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T02:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 837.169646] env[63515]: DEBUG nova.virt.hardware [None req-fe95cb73-4a2a-4d94-83ac-9bcd96be3709 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Flavor limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 837.169832] env[63515]: DEBUG nova.virt.hardware [None req-fe95cb73-4a2a-4d94-83ac-9bcd96be3709 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Image limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 837.170029] env[63515]: DEBUG nova.virt.hardware [None req-fe95cb73-4a2a-4d94-83ac-9bcd96be3709 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Flavor pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 837.170185] env[63515]: DEBUG nova.virt.hardware [None req-fe95cb73-4a2a-4d94-83ac-9bcd96be3709 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Image pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 837.170337] env[63515]: DEBUG nova.virt.hardware [None req-fe95cb73-4a2a-4d94-83ac-9bcd96be3709 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 837.170550] env[63515]: DEBUG nova.virt.hardware [None req-fe95cb73-4a2a-4d94-83ac-9bcd96be3709 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 837.170716] env[63515]: DEBUG nova.virt.hardware [None req-fe95cb73-4a2a-4d94-83ac-9bcd96be3709 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 837.170884] env[63515]: DEBUG nova.virt.hardware [None req-fe95cb73-4a2a-4d94-83ac-9bcd96be3709 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Got 1 possible topologies {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 837.171068] env[63515]: DEBUG nova.virt.hardware [None req-fe95cb73-4a2a-4d94-83ac-9bcd96be3709 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 837.171250] env[63515]: DEBUG nova.virt.hardware [None req-fe95cb73-4a2a-4d94-83ac-9bcd96be3709 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 837.172106] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36628923-f1b1-42bd-9303-872852a093a2 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.180734] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a69f825-54df-42f6-9b8a-5a40ed73fbe7 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.226486] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-c5d147d2-55a3-44f2-9b1a-4c348ae09dd7 tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] [instance: d7e8dcdb-41cb-46fb-8b61-d251e7c2d372] Powering on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 837.226789] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e0acd6b4-590b-4323-afb3-b540eece92e6 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.234244] env[63515]: DEBUG oslo_vmware.api [None req-c5d147d2-55a3-44f2-9b1a-4c348ae09dd7 tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Waiting for the task: (returnval){ [ 837.234244] env[63515]: value = "task-1111360" [ 837.234244] env[63515]: _type = "Task" [ 837.234244] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.242668] env[63515]: DEBUG oslo_vmware.api [None req-c5d147d2-55a3-44f2-9b1a-4c348ae09dd7 tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Task: {'id': task-1111360, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.481503] env[63515]: DEBUG oslo_vmware.api [None req-9dd4330d-d35d-47d5-b468-11f24e31a49a tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Task: {'id': task-1111358, 'name': ReconfigVM_Task, 'duration_secs': 0.663547} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.481503] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-9dd4330d-d35d-47d5-b468-11f24e31a49a tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: 70df69e5-687b-44fb-b6fc-cdb08e21dda0] Reconfigured VM instance instance-00000041 to attach disk [datastore2] 70df69e5-687b-44fb-b6fc-cdb08e21dda0/70df69e5-687b-44fb-b6fc-cdb08e21dda0.vmdk or device None with type sparse {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 837.481873] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-bcbfcbaf-f33d-4c53-804c-a90a3431da91 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.489229] env[63515]: DEBUG oslo_vmware.api [None req-9dd4330d-d35d-47d5-b468-11f24e31a49a tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Waiting for the task: (returnval){ [ 837.489229] env[63515]: value = "task-1111361" [ 837.489229] env[63515]: _type = "Task" [ 837.489229] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.497501] env[63515]: DEBUG oslo_vmware.api [None req-9dd4330d-d35d-47d5-b468-11f24e31a49a tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Task: {'id': task-1111361, 'name': Rename_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.591900] env[63515]: DEBUG oslo_concurrency.lockutils [None req-73bb59a8-34a2-4cd0-acb3-804cea19b215 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] Lock "e666825c-ff4e-4a0e-93c0-43c00f167bbb" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 26.830s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 837.752160] env[63515]: DEBUG oslo_vmware.api [None req-c5d147d2-55a3-44f2-9b1a-4c348ae09dd7 tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Task: {'id': task-1111360, 'name': PowerOnVM_Task, 'duration_secs': 0.363697} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.755014] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-c5d147d2-55a3-44f2-9b1a-4c348ae09dd7 tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] [instance: d7e8dcdb-41cb-46fb-8b61-d251e7c2d372] Powered on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 837.755228] env[63515]: DEBUG nova.compute.manager [None req-c5d147d2-55a3-44f2-9b1a-4c348ae09dd7 tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] [instance: d7e8dcdb-41cb-46fb-8b61-d251e7c2d372] Checking state {{(pid=63515) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 837.756749] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57d889f6-415e-4488-a5aa-93e9bc97614c {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.788329] env[63515]: DEBUG nova.compute.manager [req-b0881f03-5f00-4364-8a46-c36c699afaf5 req-e9ed51ac-ff75-4066-9e76-affc3b5d1a4b service nova] [instance: 19475235-dbec-40a5-bd8f-8070616b7ddb] Received event network-vif-plugged-372ea923-ddc7-4b61-92db-c55f8f352b44 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 837.789222] env[63515]: DEBUG oslo_concurrency.lockutils [req-b0881f03-5f00-4364-8a46-c36c699afaf5 req-e9ed51ac-ff75-4066-9e76-affc3b5d1a4b service nova] Acquiring lock "19475235-dbec-40a5-bd8f-8070616b7ddb-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 837.789222] env[63515]: DEBUG oslo_concurrency.lockutils [req-b0881f03-5f00-4364-8a46-c36c699afaf5 req-e9ed51ac-ff75-4066-9e76-affc3b5d1a4b service nova] Lock "19475235-dbec-40a5-bd8f-8070616b7ddb-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 837.789222] env[63515]: DEBUG oslo_concurrency.lockutils [req-b0881f03-5f00-4364-8a46-c36c699afaf5 req-e9ed51ac-ff75-4066-9e76-affc3b5d1a4b service nova] Lock "19475235-dbec-40a5-bd8f-8070616b7ddb-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 837.789222] env[63515]: DEBUG nova.compute.manager [req-b0881f03-5f00-4364-8a46-c36c699afaf5 req-e9ed51ac-ff75-4066-9e76-affc3b5d1a4b service nova] [instance: 19475235-dbec-40a5-bd8f-8070616b7ddb] No waiting events found dispatching network-vif-plugged-372ea923-ddc7-4b61-92db-c55f8f352b44 {{(pid=63515) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 837.789342] env[63515]: WARNING nova.compute.manager [req-b0881f03-5f00-4364-8a46-c36c699afaf5 req-e9ed51ac-ff75-4066-9e76-affc3b5d1a4b service nova] [instance: 19475235-dbec-40a5-bd8f-8070616b7ddb] Received unexpected event network-vif-plugged-372ea923-ddc7-4b61-92db-c55f8f352b44 for instance with vm_state building and task_state spawning. [ 837.853112] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a469f37-302f-4412-8a32-6f5ced9bec67 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.861933] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-241898dc-7fd4-4d7b-bc8d-7c8e2731887d {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.894027] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36d8ad76-2412-4104-9dbc-b8181129b11a {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.899808] env[63515]: DEBUG nova.network.neutron [None req-fe95cb73-4a2a-4d94-83ac-9bcd96be3709 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: 19475235-dbec-40a5-bd8f-8070616b7ddb] Successfully updated port: 372ea923-ddc7-4b61-92db-c55f8f352b44 {{(pid=63515) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 837.908870] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e214b30e-c0b9-479c-99de-7f239c69022a {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.926440] env[63515]: DEBUG nova.compute.provider_tree [None req-08dbd0f6-25be-4803-9b5a-523d28be6f7f tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 837.999997] env[63515]: DEBUG oslo_vmware.api [None req-9dd4330d-d35d-47d5-b468-11f24e31a49a tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Task: {'id': task-1111361, 'name': Rename_Task, 'duration_secs': 0.151613} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 838.000422] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-9dd4330d-d35d-47d5-b468-11f24e31a49a tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: 70df69e5-687b-44fb-b6fc-cdb08e21dda0] Powering on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 838.000584] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-16d22d61-475d-46b6-8f6e-ae42dd8d2a37 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.007890] env[63515]: DEBUG oslo_vmware.api [None req-9dd4330d-d35d-47d5-b468-11f24e31a49a tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Waiting for the task: (returnval){ [ 838.007890] env[63515]: value = "task-1111362" [ 838.007890] env[63515]: _type = "Task" [ 838.007890] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 838.016371] env[63515]: DEBUG oslo_vmware.api [None req-9dd4330d-d35d-47d5-b468-11f24e31a49a tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Task: {'id': task-1111362, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.192348] env[63515]: DEBUG nova.compute.manager [req-44435d86-adee-4972-b4ce-a83dd0c0e589 req-f5cf36e3-d68e-4771-8116-bd5fc348b2e3 service nova] [instance: 17275472-2921-49c9-b4e0-5485649ebd1c] Received event network-changed-5514ac51-9696-4c26-9abc-2cc0deef264e {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 838.192550] env[63515]: DEBUG nova.compute.manager [req-44435d86-adee-4972-b4ce-a83dd0c0e589 req-f5cf36e3-d68e-4771-8116-bd5fc348b2e3 service nova] [instance: 17275472-2921-49c9-b4e0-5485649ebd1c] Refreshing instance network info cache due to event network-changed-5514ac51-9696-4c26-9abc-2cc0deef264e. {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 838.192773] env[63515]: DEBUG oslo_concurrency.lockutils [req-44435d86-adee-4972-b4ce-a83dd0c0e589 req-f5cf36e3-d68e-4771-8116-bd5fc348b2e3 service nova] Acquiring lock "refresh_cache-17275472-2921-49c9-b4e0-5485649ebd1c" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 838.192920] env[63515]: DEBUG oslo_concurrency.lockutils [req-44435d86-adee-4972-b4ce-a83dd0c0e589 req-f5cf36e3-d68e-4771-8116-bd5fc348b2e3 service nova] Acquired lock "refresh_cache-17275472-2921-49c9-b4e0-5485649ebd1c" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 838.193172] env[63515]: DEBUG nova.network.neutron [req-44435d86-adee-4972-b4ce-a83dd0c0e589 req-f5cf36e3-d68e-4771-8116-bd5fc348b2e3 service nova] [instance: 17275472-2921-49c9-b4e0-5485649ebd1c] Refreshing network info cache for port 5514ac51-9696-4c26-9abc-2cc0deef264e {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 838.402648] env[63515]: DEBUG oslo_concurrency.lockutils [None req-fe95cb73-4a2a-4d94-83ac-9bcd96be3709 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Acquiring lock "refresh_cache-19475235-dbec-40a5-bd8f-8070616b7ddb" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 838.402812] env[63515]: DEBUG oslo_concurrency.lockutils [None req-fe95cb73-4a2a-4d94-83ac-9bcd96be3709 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Acquired lock "refresh_cache-19475235-dbec-40a5-bd8f-8070616b7ddb" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 838.402974] env[63515]: DEBUG nova.network.neutron [None req-fe95cb73-4a2a-4d94-83ac-9bcd96be3709 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: 19475235-dbec-40a5-bd8f-8070616b7ddb] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 838.404447] env[63515]: DEBUG oslo_concurrency.lockutils [None req-7f7535aa-eede-4389-b9f3-fe528607f4b7 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] Acquiring lock "5b42f744-fdd6-45b1-8563-896869648c23" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 838.404519] env[63515]: DEBUG oslo_concurrency.lockutils [None req-7f7535aa-eede-4389-b9f3-fe528607f4b7 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] Lock "5b42f744-fdd6-45b1-8563-896869648c23" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 838.404721] env[63515]: DEBUG oslo_concurrency.lockutils [None req-7f7535aa-eede-4389-b9f3-fe528607f4b7 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] Acquiring lock "5b42f744-fdd6-45b1-8563-896869648c23-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 838.404918] env[63515]: DEBUG oslo_concurrency.lockutils [None req-7f7535aa-eede-4389-b9f3-fe528607f4b7 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] Lock "5b42f744-fdd6-45b1-8563-896869648c23-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 838.405689] env[63515]: DEBUG oslo_concurrency.lockutils [None req-7f7535aa-eede-4389-b9f3-fe528607f4b7 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] Lock "5b42f744-fdd6-45b1-8563-896869648c23-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 838.408035] env[63515]: INFO nova.compute.manager [None req-7f7535aa-eede-4389-b9f3-fe528607f4b7 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] [instance: 5b42f744-fdd6-45b1-8563-896869648c23] Terminating instance [ 838.410151] env[63515]: DEBUG nova.compute.manager [None req-7f7535aa-eede-4389-b9f3-fe528607f4b7 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] [instance: 5b42f744-fdd6-45b1-8563-896869648c23] Start destroying the instance on the hypervisor. {{(pid=63515) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 838.410433] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-7f7535aa-eede-4389-b9f3-fe528607f4b7 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] [instance: 5b42f744-fdd6-45b1-8563-896869648c23] Destroying instance {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 838.411490] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5f8811c-a84a-4444-975a-1288c2ad2cb1 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.421571] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-7f7535aa-eede-4389-b9f3-fe528607f4b7 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] [instance: 5b42f744-fdd6-45b1-8563-896869648c23] Powering off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 838.422179] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4b5cf6c9-a3cf-4b67-8022-8f4f6edb0115 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.429435] env[63515]: DEBUG nova.scheduler.client.report [None req-08dbd0f6-25be-4803-9b5a-523d28be6f7f tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 838.434373] env[63515]: DEBUG oslo_vmware.api [None req-7f7535aa-eede-4389-b9f3-fe528607f4b7 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] Waiting for the task: (returnval){ [ 838.434373] env[63515]: value = "task-1111363" [ 838.434373] env[63515]: _type = "Task" [ 838.434373] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 838.435335] env[63515]: DEBUG nova.network.neutron [None req-7bca9c72-42f5-4d65-99a2-4374b66c9c04 tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] [instance: 17275472-2921-49c9-b4e0-5485649ebd1c] Successfully updated port: db47ccfa-8557-4128-9e65-8a8ab57163bb {{(pid=63515) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 838.450699] env[63515]: DEBUG oslo_vmware.api [None req-7f7535aa-eede-4389-b9f3-fe528607f4b7 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] Task: {'id': task-1111363, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.520189] env[63515]: DEBUG oslo_vmware.api [None req-9dd4330d-d35d-47d5-b468-11f24e31a49a tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Task: {'id': task-1111362, 'name': PowerOnVM_Task, 'duration_secs': 0.474953} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 838.520569] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-9dd4330d-d35d-47d5-b468-11f24e31a49a tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: 70df69e5-687b-44fb-b6fc-cdb08e21dda0] Powered on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 838.520824] env[63515]: INFO nova.compute.manager [None req-9dd4330d-d35d-47d5-b468-11f24e31a49a tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: 70df69e5-687b-44fb-b6fc-cdb08e21dda0] Took 8.09 seconds to spawn the instance on the hypervisor. [ 838.521062] env[63515]: DEBUG nova.compute.manager [None req-9dd4330d-d35d-47d5-b468-11f24e31a49a tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: 70df69e5-687b-44fb-b6fc-cdb08e21dda0] Checking state {{(pid=63515) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 838.521972] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4459d12-486a-47f6-bef7-5bac16351033 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.744077] env[63515]: DEBUG nova.network.neutron [req-44435d86-adee-4972-b4ce-a83dd0c0e589 req-f5cf36e3-d68e-4771-8116-bd5fc348b2e3 service nova] [instance: 17275472-2921-49c9-b4e0-5485649ebd1c] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 838.822982] env[63515]: DEBUG nova.network.neutron [req-44435d86-adee-4972-b4ce-a83dd0c0e589 req-f5cf36e3-d68e-4771-8116-bd5fc348b2e3 service nova] [instance: 17275472-2921-49c9-b4e0-5485649ebd1c] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 838.938718] env[63515]: DEBUG oslo_concurrency.lockutils [None req-08dbd0f6-25be-4803-9b5a-523d28be6f7f tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.880s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 838.942988] env[63515]: DEBUG oslo_concurrency.lockutils [None req-5a74a6fe-271a-406d-93f5-9b098b35317a tempest-FloatingIPsAssociationNegativeTestJSON-2063029775 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 22.099s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 838.943245] env[63515]: DEBUG nova.objects.instance [None req-5a74a6fe-271a-406d-93f5-9b098b35317a tempest-FloatingIPsAssociationNegativeTestJSON-2063029775 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775-project-member] Lazy-loading 'resources' on Instance uuid 4e453127-1f3e-40ea-819f-6678479826c8 {{(pid=63515) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 838.954792] env[63515]: DEBUG oslo_vmware.api [None req-7f7535aa-eede-4389-b9f3-fe528607f4b7 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] Task: {'id': task-1111363, 'name': PowerOffVM_Task, 'duration_secs': 0.210015} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 838.956058] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-7f7535aa-eede-4389-b9f3-fe528607f4b7 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] [instance: 5b42f744-fdd6-45b1-8563-896869648c23] Powered off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 838.956058] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-7f7535aa-eede-4389-b9f3-fe528607f4b7 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] [instance: 5b42f744-fdd6-45b1-8563-896869648c23] Unregistering the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 838.956210] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-4936a377-7c11-45de-bcd3-7bb4dbf51742 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.960770] env[63515]: DEBUG nova.network.neutron [None req-fe95cb73-4a2a-4d94-83ac-9bcd96be3709 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: 19475235-dbec-40a5-bd8f-8070616b7ddb] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 838.964284] env[63515]: INFO nova.scheduler.client.report [None req-08dbd0f6-25be-4803-9b5a-523d28be6f7f tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Deleted allocations for instance 6688cdaa-29ba-413a-8131-4f834cdb70e4 [ 839.024535] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-7f7535aa-eede-4389-b9f3-fe528607f4b7 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] [instance: 5b42f744-fdd6-45b1-8563-896869648c23] Unregistered the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 839.024749] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-7f7535aa-eede-4389-b9f3-fe528607f4b7 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] [instance: 5b42f744-fdd6-45b1-8563-896869648c23] Deleting contents of the VM from datastore datastore1 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 839.028024] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-7f7535aa-eede-4389-b9f3-fe528607f4b7 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] Deleting the datastore file [datastore1] 5b42f744-fdd6-45b1-8563-896869648c23 {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 839.028024] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-9c6e1dcd-3cba-4264-a72c-2db95f0909c7 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.033404] env[63515]: DEBUG oslo_vmware.api [None req-7f7535aa-eede-4389-b9f3-fe528607f4b7 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] Waiting for the task: (returnval){ [ 839.033404] env[63515]: value = "task-1111365" [ 839.033404] env[63515]: _type = "Task" [ 839.033404] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.042846] env[63515]: INFO nova.compute.manager [None req-9dd4330d-d35d-47d5-b468-11f24e31a49a tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: 70df69e5-687b-44fb-b6fc-cdb08e21dda0] Took 29.22 seconds to build instance. [ 839.048209] env[63515]: DEBUG oslo_vmware.api [None req-7f7535aa-eede-4389-b9f3-fe528607f4b7 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] Task: {'id': task-1111365, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.139996] env[63515]: DEBUG nova.network.neutron [None req-fe95cb73-4a2a-4d94-83ac-9bcd96be3709 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: 19475235-dbec-40a5-bd8f-8070616b7ddb] Updating instance_info_cache with network_info: [{"id": "372ea923-ddc7-4b61-92db-c55f8f352b44", "address": "fa:16:3e:fa:c4:2a", "network": {"id": "581d448e-be1b-4d5b-8a6e-c27326aaffc5", "bridge": "br-int", "label": "tempest-ServersTestJSON-1769610073-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "299fdeff647f486390366d5bbf911518", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fb143ef7-8271-4a8a-a4aa-8eba9a89f6a1", "external-id": "nsx-vlan-transportzone-504", "segmentation_id": 504, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap372ea923-dd", "ovs_interfaceid": "372ea923-ddc7-4b61-92db-c55f8f352b44", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 839.325778] env[63515]: DEBUG oslo_concurrency.lockutils [req-44435d86-adee-4972-b4ce-a83dd0c0e589 req-f5cf36e3-d68e-4771-8116-bd5fc348b2e3 service nova] Releasing lock "refresh_cache-17275472-2921-49c9-b4e0-5485649ebd1c" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 839.474746] env[63515]: DEBUG oslo_concurrency.lockutils [None req-08dbd0f6-25be-4803-9b5a-523d28be6f7f tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Lock "6688cdaa-29ba-413a-8131-4f834cdb70e4" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 29.116s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 839.544776] env[63515]: DEBUG oslo_concurrency.lockutils [None req-9dd4330d-d35d-47d5-b468-11f24e31a49a tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Lock "70df69e5-687b-44fb-b6fc-cdb08e21dda0" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 30.726s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 839.545068] env[63515]: DEBUG oslo_vmware.api [None req-7f7535aa-eede-4389-b9f3-fe528607f4b7 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] Task: {'id': task-1111365, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.140836} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.545300] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-7f7535aa-eede-4389-b9f3-fe528607f4b7 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] Deleted the datastore file {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 839.545492] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-7f7535aa-eede-4389-b9f3-fe528607f4b7 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] [instance: 5b42f744-fdd6-45b1-8563-896869648c23] Deleted contents of the VM from datastore datastore1 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 839.545770] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-7f7535aa-eede-4389-b9f3-fe528607f4b7 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] [instance: 5b42f744-fdd6-45b1-8563-896869648c23] Instance destroyed {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 839.546065] env[63515]: INFO nova.compute.manager [None req-7f7535aa-eede-4389-b9f3-fe528607f4b7 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] [instance: 5b42f744-fdd6-45b1-8563-896869648c23] Took 1.14 seconds to destroy the instance on the hypervisor. [ 839.546283] env[63515]: DEBUG oslo.service.loopingcall [None req-7f7535aa-eede-4389-b9f3-fe528607f4b7 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 839.546506] env[63515]: DEBUG nova.compute.manager [-] [instance: 5b42f744-fdd6-45b1-8563-896869648c23] Deallocating network for instance {{(pid=63515) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 839.546631] env[63515]: DEBUG nova.network.neutron [-] [instance: 5b42f744-fdd6-45b1-8563-896869648c23] deallocate_for_instance() {{(pid=63515) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 839.642904] env[63515]: DEBUG oslo_concurrency.lockutils [None req-fe95cb73-4a2a-4d94-83ac-9bcd96be3709 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Releasing lock "refresh_cache-19475235-dbec-40a5-bd8f-8070616b7ddb" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 839.643414] env[63515]: DEBUG nova.compute.manager [None req-fe95cb73-4a2a-4d94-83ac-9bcd96be3709 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: 19475235-dbec-40a5-bd8f-8070616b7ddb] Instance network_info: |[{"id": "372ea923-ddc7-4b61-92db-c55f8f352b44", "address": "fa:16:3e:fa:c4:2a", "network": {"id": "581d448e-be1b-4d5b-8a6e-c27326aaffc5", "bridge": "br-int", "label": "tempest-ServersTestJSON-1769610073-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "299fdeff647f486390366d5bbf911518", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fb143ef7-8271-4a8a-a4aa-8eba9a89f6a1", "external-id": "nsx-vlan-transportzone-504", "segmentation_id": 504, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap372ea923-dd", "ovs_interfaceid": "372ea923-ddc7-4b61-92db-c55f8f352b44", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 839.643726] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-fe95cb73-4a2a-4d94-83ac-9bcd96be3709 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: 19475235-dbec-40a5-bd8f-8070616b7ddb] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:fa:c4:2a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'fb143ef7-8271-4a8a-a4aa-8eba9a89f6a1', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '372ea923-ddc7-4b61-92db-c55f8f352b44', 'vif_model': 'vmxnet3'}] {{(pid=63515) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 839.651866] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-fe95cb73-4a2a-4d94-83ac-9bcd96be3709 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Creating folder: Project (299fdeff647f486390366d5bbf911518). Parent ref: group-v243370. {{(pid=63515) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 839.654703] env[63515]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-56d4b3ce-8556-4d7d-bafb-345f06ae1e28 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.671144] env[63515]: INFO nova.virt.vmwareapi.vm_util [None req-fe95cb73-4a2a-4d94-83ac-9bcd96be3709 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Created folder: Project (299fdeff647f486390366d5bbf911518) in parent group-v243370. [ 839.671144] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-fe95cb73-4a2a-4d94-83ac-9bcd96be3709 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Creating folder: Instances. Parent ref: group-v243485. {{(pid=63515) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 839.671144] env[63515]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-efe0e85a-3f99-4faf-b0ef-43110ad20caf {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.682026] env[63515]: INFO nova.virt.vmwareapi.vm_util [None req-fe95cb73-4a2a-4d94-83ac-9bcd96be3709 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Created folder: Instances in parent group-v243485. [ 839.682026] env[63515]: DEBUG oslo.service.loopingcall [None req-fe95cb73-4a2a-4d94-83ac-9bcd96be3709 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 839.682352] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 19475235-dbec-40a5-bd8f-8070616b7ddb] Creating VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 839.682352] env[63515]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-afc23021-d681-4b49-aca4-49beacc30623 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.706846] env[63515]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 839.706846] env[63515]: value = "task-1111368" [ 839.706846] env[63515]: _type = "Task" [ 839.706846] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.718339] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1111368, 'name': CreateVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.742072] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9bde418a-36f7-4571-9b00-14d5b733ae70 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.751329] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-430a499e-0644-439a-a50b-2c694906e844 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.789807] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c9f1d0b-bce6-4daa-92f2-d12107ad3270 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.798581] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15b3b713-7034-454e-a9f9-a9cc248ca857 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.815343] env[63515]: DEBUG nova.compute.provider_tree [None req-5a74a6fe-271a-406d-93f5-9b098b35317a tempest-FloatingIPsAssociationNegativeTestJSON-2063029775 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 839.986265] env[63515]: DEBUG nova.compute.manager [req-c6b0ad8a-4ac0-475a-b2cd-585da48f7427 req-0f5ec4fd-133f-48ba-9452-bebf2407d40c service nova] [instance: 19475235-dbec-40a5-bd8f-8070616b7ddb] Received event network-changed-372ea923-ddc7-4b61-92db-c55f8f352b44 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 839.986474] env[63515]: DEBUG nova.compute.manager [req-c6b0ad8a-4ac0-475a-b2cd-585da48f7427 req-0f5ec4fd-133f-48ba-9452-bebf2407d40c service nova] [instance: 19475235-dbec-40a5-bd8f-8070616b7ddb] Refreshing instance network info cache due to event network-changed-372ea923-ddc7-4b61-92db-c55f8f352b44. {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 839.986728] env[63515]: DEBUG oslo_concurrency.lockutils [req-c6b0ad8a-4ac0-475a-b2cd-585da48f7427 req-0f5ec4fd-133f-48ba-9452-bebf2407d40c service nova] Acquiring lock "refresh_cache-19475235-dbec-40a5-bd8f-8070616b7ddb" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 839.988018] env[63515]: DEBUG oslo_concurrency.lockutils [req-c6b0ad8a-4ac0-475a-b2cd-585da48f7427 req-0f5ec4fd-133f-48ba-9452-bebf2407d40c service nova] Acquired lock "refresh_cache-19475235-dbec-40a5-bd8f-8070616b7ddb" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 839.988018] env[63515]: DEBUG nova.network.neutron [req-c6b0ad8a-4ac0-475a-b2cd-585da48f7427 req-0f5ec4fd-133f-48ba-9452-bebf2407d40c service nova] [instance: 19475235-dbec-40a5-bd8f-8070616b7ddb] Refreshing network info cache for port 372ea923-ddc7-4b61-92db-c55f8f352b44 {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 840.217031] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1111368, 'name': CreateVM_Task, 'duration_secs': 0.379622} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 840.217218] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 19475235-dbec-40a5-bd8f-8070616b7ddb] Created VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 840.217914] env[63515]: DEBUG oslo_concurrency.lockutils [None req-fe95cb73-4a2a-4d94-83ac-9bcd96be3709 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 840.218113] env[63515]: DEBUG oslo_concurrency.lockutils [None req-fe95cb73-4a2a-4d94-83ac-9bcd96be3709 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 840.218441] env[63515]: DEBUG oslo_concurrency.lockutils [None req-fe95cb73-4a2a-4d94-83ac-9bcd96be3709 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 840.218705] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e77e7441-a772-4933-840b-47f67e3c5ab7 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.223868] env[63515]: DEBUG oslo_vmware.api [None req-fe95cb73-4a2a-4d94-83ac-9bcd96be3709 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Waiting for the task: (returnval){ [ 840.223868] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]5234e61a-5c3a-b395-1142-239c8e37dc04" [ 840.223868] env[63515]: _type = "Task" [ 840.223868] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 840.232230] env[63515]: DEBUG oslo_concurrency.lockutils [None req-c9f99ed7-857d-4064-a182-30f784598d4b tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Acquiring lock "d2a4f66a-f1a2-4f0d-826d-0b7c4f04bd23" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 840.232480] env[63515]: DEBUG oslo_concurrency.lockutils [None req-c9f99ed7-857d-4064-a182-30f784598d4b tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Lock "d2a4f66a-f1a2-4f0d-826d-0b7c4f04bd23" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 840.232688] env[63515]: DEBUG oslo_concurrency.lockutils [None req-c9f99ed7-857d-4064-a182-30f784598d4b tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Acquiring lock "d2a4f66a-f1a2-4f0d-826d-0b7c4f04bd23-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 840.232877] env[63515]: DEBUG oslo_concurrency.lockutils [None req-c9f99ed7-857d-4064-a182-30f784598d4b tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Lock "d2a4f66a-f1a2-4f0d-826d-0b7c4f04bd23-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 840.233059] env[63515]: DEBUG oslo_concurrency.lockutils [None req-c9f99ed7-857d-4064-a182-30f784598d4b tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Lock "d2a4f66a-f1a2-4f0d-826d-0b7c4f04bd23-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 840.235277] env[63515]: DEBUG oslo_vmware.api [None req-fe95cb73-4a2a-4d94-83ac-9bcd96be3709 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]5234e61a-5c3a-b395-1142-239c8e37dc04, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.237819] env[63515]: INFO nova.compute.manager [None req-c9f99ed7-857d-4064-a182-30f784598d4b tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] [instance: d2a4f66a-f1a2-4f0d-826d-0b7c4f04bd23] Terminating instance [ 840.239846] env[63515]: DEBUG nova.compute.manager [None req-c9f99ed7-857d-4064-a182-30f784598d4b tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] [instance: d2a4f66a-f1a2-4f0d-826d-0b7c4f04bd23] Start destroying the instance on the hypervisor. {{(pid=63515) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 840.240050] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-c9f99ed7-857d-4064-a182-30f784598d4b tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] [instance: d2a4f66a-f1a2-4f0d-826d-0b7c4f04bd23] Destroying instance {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 840.240834] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85698843-bf61-400d-b4f3-03a52d5cdf7b {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.248792] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-c9f99ed7-857d-4064-a182-30f784598d4b tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] [instance: d2a4f66a-f1a2-4f0d-826d-0b7c4f04bd23] Powering off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 840.248996] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-91c39a2b-044a-4540-91e6-3969093a60f9 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.258360] env[63515]: DEBUG oslo_vmware.api [None req-c9f99ed7-857d-4064-a182-30f784598d4b tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Waiting for the task: (returnval){ [ 840.258360] env[63515]: value = "task-1111369" [ 840.258360] env[63515]: _type = "Task" [ 840.258360] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 840.259227] env[63515]: DEBUG nova.compute.manager [req-7f79d2ce-f34d-42fc-9be0-d55a23ac6ea7 req-16d9caef-65ec-4e96-8784-e9aedad484f8 service nova] [instance: 17275472-2921-49c9-b4e0-5485649ebd1c] Received event network-vif-plugged-db47ccfa-8557-4128-9e65-8a8ab57163bb {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 840.259446] env[63515]: DEBUG oslo_concurrency.lockutils [req-7f79d2ce-f34d-42fc-9be0-d55a23ac6ea7 req-16d9caef-65ec-4e96-8784-e9aedad484f8 service nova] Acquiring lock "17275472-2921-49c9-b4e0-5485649ebd1c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 840.259656] env[63515]: DEBUG oslo_concurrency.lockutils [req-7f79d2ce-f34d-42fc-9be0-d55a23ac6ea7 req-16d9caef-65ec-4e96-8784-e9aedad484f8 service nova] Lock "17275472-2921-49c9-b4e0-5485649ebd1c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 840.259825] env[63515]: DEBUG oslo_concurrency.lockutils [req-7f79d2ce-f34d-42fc-9be0-d55a23ac6ea7 req-16d9caef-65ec-4e96-8784-e9aedad484f8 service nova] Lock "17275472-2921-49c9-b4e0-5485649ebd1c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 840.259991] env[63515]: DEBUG nova.compute.manager [req-7f79d2ce-f34d-42fc-9be0-d55a23ac6ea7 req-16d9caef-65ec-4e96-8784-e9aedad484f8 service nova] [instance: 17275472-2921-49c9-b4e0-5485649ebd1c] No waiting events found dispatching network-vif-plugged-db47ccfa-8557-4128-9e65-8a8ab57163bb {{(pid=63515) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 840.260183] env[63515]: WARNING nova.compute.manager [req-7f79d2ce-f34d-42fc-9be0-d55a23ac6ea7 req-16d9caef-65ec-4e96-8784-e9aedad484f8 service nova] [instance: 17275472-2921-49c9-b4e0-5485649ebd1c] Received unexpected event network-vif-plugged-db47ccfa-8557-4128-9e65-8a8ab57163bb for instance with vm_state building and task_state spawning. [ 840.260348] env[63515]: DEBUG nova.compute.manager [req-7f79d2ce-f34d-42fc-9be0-d55a23ac6ea7 req-16d9caef-65ec-4e96-8784-e9aedad484f8 service nova] [instance: 17275472-2921-49c9-b4e0-5485649ebd1c] Received event network-changed-db47ccfa-8557-4128-9e65-8a8ab57163bb {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 840.260503] env[63515]: DEBUG nova.compute.manager [req-7f79d2ce-f34d-42fc-9be0-d55a23ac6ea7 req-16d9caef-65ec-4e96-8784-e9aedad484f8 service nova] [instance: 17275472-2921-49c9-b4e0-5485649ebd1c] Refreshing instance network info cache due to event network-changed-db47ccfa-8557-4128-9e65-8a8ab57163bb. {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 840.260708] env[63515]: DEBUG oslo_concurrency.lockutils [req-7f79d2ce-f34d-42fc-9be0-d55a23ac6ea7 req-16d9caef-65ec-4e96-8784-e9aedad484f8 service nova] Acquiring lock "refresh_cache-17275472-2921-49c9-b4e0-5485649ebd1c" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 840.260827] env[63515]: DEBUG oslo_concurrency.lockutils [req-7f79d2ce-f34d-42fc-9be0-d55a23ac6ea7 req-16d9caef-65ec-4e96-8784-e9aedad484f8 service nova] Acquired lock "refresh_cache-17275472-2921-49c9-b4e0-5485649ebd1c" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 840.260979] env[63515]: DEBUG nova.network.neutron [req-7f79d2ce-f34d-42fc-9be0-d55a23ac6ea7 req-16d9caef-65ec-4e96-8784-e9aedad484f8 service nova] [instance: 17275472-2921-49c9-b4e0-5485649ebd1c] Refreshing network info cache for port db47ccfa-8557-4128-9e65-8a8ab57163bb {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 840.271557] env[63515]: DEBUG oslo_vmware.api [None req-c9f99ed7-857d-4064-a182-30f784598d4b tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Task: {'id': task-1111369, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.318023] env[63515]: DEBUG nova.scheduler.client.report [None req-5a74a6fe-271a-406d-93f5-9b098b35317a tempest-FloatingIPsAssociationNegativeTestJSON-2063029775 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 840.375817] env[63515]: DEBUG nova.network.neutron [-] [instance: 5b42f744-fdd6-45b1-8563-896869648c23] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 840.581606] env[63515]: DEBUG nova.network.neutron [None req-7bca9c72-42f5-4d65-99a2-4374b66c9c04 tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] [instance: 17275472-2921-49c9-b4e0-5485649ebd1c] Successfully updated port: eadc0731-8390-4414-b5cb-0f6a0c660979 {{(pid=63515) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 840.667791] env[63515]: DEBUG nova.compute.manager [None req-ced15047-223c-4e6c-9243-6a752a33316f tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: 70df69e5-687b-44fb-b6fc-cdb08e21dda0] Checking state {{(pid=63515) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 840.668832] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1a9735d-d2ef-4dc5-a7c7-83310e4a0593 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.736961] env[63515]: DEBUG oslo_vmware.api [None req-fe95cb73-4a2a-4d94-83ac-9bcd96be3709 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]5234e61a-5c3a-b395-1142-239c8e37dc04, 'name': SearchDatastore_Task, 'duration_secs': 0.011669} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 840.736961] env[63515]: DEBUG oslo_concurrency.lockutils [None req-fe95cb73-4a2a-4d94-83ac-9bcd96be3709 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 840.737190] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-fe95cb73-4a2a-4d94-83ac-9bcd96be3709 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: 19475235-dbec-40a5-bd8f-8070616b7ddb] Processing image 8a120570-cb06-4099-b262-554ca0ad15c5 {{(pid=63515) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 840.737521] env[63515]: DEBUG oslo_concurrency.lockutils [None req-fe95cb73-4a2a-4d94-83ac-9bcd96be3709 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 840.737629] env[63515]: DEBUG oslo_concurrency.lockutils [None req-fe95cb73-4a2a-4d94-83ac-9bcd96be3709 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 840.737805] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-fe95cb73-4a2a-4d94-83ac-9bcd96be3709 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 840.738123] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6ee4f871-2782-4e20-80b5-11ff90f0727f {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.748027] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-fe95cb73-4a2a-4d94-83ac-9bcd96be3709 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 840.748027] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-fe95cb73-4a2a-4d94-83ac-9bcd96be3709 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63515) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 840.748349] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f56312d3-362d-4422-9808-0714dae5a500 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.754070] env[63515]: DEBUG oslo_vmware.api [None req-fe95cb73-4a2a-4d94-83ac-9bcd96be3709 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Waiting for the task: (returnval){ [ 840.754070] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]5295d833-a61e-0e28-090f-3f0a6b392863" [ 840.754070] env[63515]: _type = "Task" [ 840.754070] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 840.762091] env[63515]: DEBUG oslo_vmware.api [None req-fe95cb73-4a2a-4d94-83ac-9bcd96be3709 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]5295d833-a61e-0e28-090f-3f0a6b392863, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.766044] env[63515]: DEBUG nova.network.neutron [req-c6b0ad8a-4ac0-475a-b2cd-585da48f7427 req-0f5ec4fd-133f-48ba-9452-bebf2407d40c service nova] [instance: 19475235-dbec-40a5-bd8f-8070616b7ddb] Updated VIF entry in instance network info cache for port 372ea923-ddc7-4b61-92db-c55f8f352b44. {{(pid=63515) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 840.766044] env[63515]: DEBUG nova.network.neutron [req-c6b0ad8a-4ac0-475a-b2cd-585da48f7427 req-0f5ec4fd-133f-48ba-9452-bebf2407d40c service nova] [instance: 19475235-dbec-40a5-bd8f-8070616b7ddb] Updating instance_info_cache with network_info: [{"id": "372ea923-ddc7-4b61-92db-c55f8f352b44", "address": "fa:16:3e:fa:c4:2a", "network": {"id": "581d448e-be1b-4d5b-8a6e-c27326aaffc5", "bridge": "br-int", "label": "tempest-ServersTestJSON-1769610073-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "299fdeff647f486390366d5bbf911518", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fb143ef7-8271-4a8a-a4aa-8eba9a89f6a1", "external-id": "nsx-vlan-transportzone-504", "segmentation_id": 504, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap372ea923-dd", "ovs_interfaceid": "372ea923-ddc7-4b61-92db-c55f8f352b44", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 840.774408] env[63515]: DEBUG oslo_vmware.api [None req-c9f99ed7-857d-4064-a182-30f784598d4b tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Task: {'id': task-1111369, 'name': PowerOffVM_Task, 'duration_secs': 0.255131} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 840.774658] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-c9f99ed7-857d-4064-a182-30f784598d4b tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] [instance: d2a4f66a-f1a2-4f0d-826d-0b7c4f04bd23] Powered off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 840.774835] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-c9f99ed7-857d-4064-a182-30f784598d4b tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] [instance: d2a4f66a-f1a2-4f0d-826d-0b7c4f04bd23] Unregistering the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 840.775867] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-133d14a1-a2f5-46cb-9867-c5a3189df2d6 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.805620] env[63515]: DEBUG nova.network.neutron [req-7f79d2ce-f34d-42fc-9be0-d55a23ac6ea7 req-16d9caef-65ec-4e96-8784-e9aedad484f8 service nova] [instance: 17275472-2921-49c9-b4e0-5485649ebd1c] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 840.823313] env[63515]: DEBUG oslo_concurrency.lockutils [None req-5a74a6fe-271a-406d-93f5-9b098b35317a tempest-FloatingIPsAssociationNegativeTestJSON-2063029775 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.880s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 840.826564] env[63515]: DEBUG oslo_concurrency.lockutils [None req-4cdfd447-25bb-42ff-81df-85503f57f660 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 23.046s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 840.842621] env[63515]: INFO nova.scheduler.client.report [None req-5a74a6fe-271a-406d-93f5-9b098b35317a tempest-FloatingIPsAssociationNegativeTestJSON-2063029775 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775-project-member] Deleted allocations for instance 4e453127-1f3e-40ea-819f-6678479826c8 [ 840.878951] env[63515]: INFO nova.compute.manager [-] [instance: 5b42f744-fdd6-45b1-8563-896869648c23] Took 1.33 seconds to deallocate network for instance. [ 840.941491] env[63515]: DEBUG nova.network.neutron [req-7f79d2ce-f34d-42fc-9be0-d55a23ac6ea7 req-16d9caef-65ec-4e96-8784-e9aedad484f8 service nova] [instance: 17275472-2921-49c9-b4e0-5485649ebd1c] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 841.077122] env[63515]: DEBUG oslo_concurrency.lockutils [None req-3db8c3e0-5f0e-4c46-ab36-d6f97ddeb4f1 tempest-ServerActionsV293TestJSON-1193370169 tempest-ServerActionsV293TestJSON-1193370169-project-member] Acquiring lock "2ea99c7a-8b61-4718-bee0-f4ce4fc021af" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 841.077291] env[63515]: DEBUG oslo_concurrency.lockutils [None req-3db8c3e0-5f0e-4c46-ab36-d6f97ddeb4f1 tempest-ServerActionsV293TestJSON-1193370169 tempest-ServerActionsV293TestJSON-1193370169-project-member] Lock "2ea99c7a-8b61-4718-bee0-f4ce4fc021af" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 841.084241] env[63515]: DEBUG oslo_concurrency.lockutils [None req-7bca9c72-42f5-4d65-99a2-4374b66c9c04 tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] Acquiring lock "refresh_cache-17275472-2921-49c9-b4e0-5485649ebd1c" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 841.182901] env[63515]: INFO nova.compute.manager [None req-ced15047-223c-4e6c-9243-6a752a33316f tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: 70df69e5-687b-44fb-b6fc-cdb08e21dda0] instance snapshotting [ 841.186217] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-c9f99ed7-857d-4064-a182-30f784598d4b tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] [instance: d2a4f66a-f1a2-4f0d-826d-0b7c4f04bd23] Unregistered the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 841.186552] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-c9f99ed7-857d-4064-a182-30f784598d4b tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] [instance: d2a4f66a-f1a2-4f0d-826d-0b7c4f04bd23] Deleting contents of the VM from datastore datastore1 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 841.186798] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-c9f99ed7-857d-4064-a182-30f784598d4b tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Deleting the datastore file [datastore1] d2a4f66a-f1a2-4f0d-826d-0b7c4f04bd23 {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 841.187422] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-5349be97-e29c-4fb0-84fa-a0b2bfd1f7ec {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.190068] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-129c2650-9f6e-4c84-a8f9-69f511b33e48 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.212362] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ced9c23d-25ea-4f6b-9240-0f574e94e92b {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.215354] env[63515]: DEBUG oslo_vmware.api [None req-c9f99ed7-857d-4064-a182-30f784598d4b tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Waiting for the task: (returnval){ [ 841.215354] env[63515]: value = "task-1111371" [ 841.215354] env[63515]: _type = "Task" [ 841.215354] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 841.229318] env[63515]: DEBUG oslo_vmware.api [None req-c9f99ed7-857d-4064-a182-30f784598d4b tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Task: {'id': task-1111371, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.266484] env[63515]: DEBUG oslo_vmware.api [None req-fe95cb73-4a2a-4d94-83ac-9bcd96be3709 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]5295d833-a61e-0e28-090f-3f0a6b392863, 'name': SearchDatastore_Task, 'duration_secs': 0.008595} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 841.267390] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e74d65c3-ed8b-4463-b0e9-beed14e72b53 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.270382] env[63515]: DEBUG oslo_concurrency.lockutils [req-c6b0ad8a-4ac0-475a-b2cd-585da48f7427 req-0f5ec4fd-133f-48ba-9452-bebf2407d40c service nova] Releasing lock "refresh_cache-19475235-dbec-40a5-bd8f-8070616b7ddb" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 841.276201] env[63515]: DEBUG oslo_vmware.api [None req-fe95cb73-4a2a-4d94-83ac-9bcd96be3709 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Waiting for the task: (returnval){ [ 841.276201] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]5212b07a-82bc-583f-55e6-6621b0a0ece0" [ 841.276201] env[63515]: _type = "Task" [ 841.276201] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 841.286121] env[63515]: DEBUG oslo_vmware.api [None req-fe95cb73-4a2a-4d94-83ac-9bcd96be3709 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]5212b07a-82bc-583f-55e6-6621b0a0ece0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.331727] env[63515]: INFO nova.compute.claims [None req-4cdfd447-25bb-42ff-81df-85503f57f660 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: d926c699-a64a-4942-9ef4-f0166414661d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 841.352694] env[63515]: DEBUG oslo_concurrency.lockutils [None req-5a74a6fe-271a-406d-93f5-9b098b35317a tempest-FloatingIPsAssociationNegativeTestJSON-2063029775 tempest-FloatingIPsAssociationNegativeTestJSON-2063029775-project-member] Lock "4e453127-1f3e-40ea-819f-6678479826c8" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 27.624s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 841.385435] env[63515]: DEBUG oslo_concurrency.lockutils [None req-7f7535aa-eede-4389-b9f3-fe528607f4b7 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 841.444755] env[63515]: DEBUG oslo_concurrency.lockutils [req-7f79d2ce-f34d-42fc-9be0-d55a23ac6ea7 req-16d9caef-65ec-4e96-8784-e9aedad484f8 service nova] Releasing lock "refresh_cache-17275472-2921-49c9-b4e0-5485649ebd1c" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 841.445154] env[63515]: DEBUG nova.compute.manager [req-7f79d2ce-f34d-42fc-9be0-d55a23ac6ea7 req-16d9caef-65ec-4e96-8784-e9aedad484f8 service nova] [instance: 5b42f744-fdd6-45b1-8563-896869648c23] Received event network-vif-deleted-3a630639-5423-4e62-9beb-ee1e7054f47d {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 841.445417] env[63515]: INFO nova.compute.manager [req-7f79d2ce-f34d-42fc-9be0-d55a23ac6ea7 req-16d9caef-65ec-4e96-8784-e9aedad484f8 service nova] [instance: 5b42f744-fdd6-45b1-8563-896869648c23] Neutron deleted interface 3a630639-5423-4e62-9beb-ee1e7054f47d; detaching it from the instance and deleting it from the info cache [ 841.445733] env[63515]: DEBUG nova.network.neutron [req-7f79d2ce-f34d-42fc-9be0-d55a23ac6ea7 req-16d9caef-65ec-4e96-8784-e9aedad484f8 service nova] [instance: 5b42f744-fdd6-45b1-8563-896869648c23] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 841.447921] env[63515]: DEBUG oslo_concurrency.lockutils [None req-7bca9c72-42f5-4d65-99a2-4374b66c9c04 tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] Acquired lock "refresh_cache-17275472-2921-49c9-b4e0-5485649ebd1c" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 841.448149] env[63515]: DEBUG nova.network.neutron [None req-7bca9c72-42f5-4d65-99a2-4374b66c9c04 tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] [instance: 17275472-2921-49c9-b4e0-5485649ebd1c] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 841.579671] env[63515]: DEBUG nova.compute.manager [None req-3db8c3e0-5f0e-4c46-ab36-d6f97ddeb4f1 tempest-ServerActionsV293TestJSON-1193370169 tempest-ServerActionsV293TestJSON-1193370169-project-member] [instance: 2ea99c7a-8b61-4718-bee0-f4ce4fc021af] Starting instance... {{(pid=63515) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 841.727118] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-ced15047-223c-4e6c-9243-6a752a33316f tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: 70df69e5-687b-44fb-b6fc-cdb08e21dda0] Creating Snapshot of the VM instance {{(pid=63515) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 841.727511] env[63515]: DEBUG oslo_vmware.api [None req-c9f99ed7-857d-4064-a182-30f784598d4b tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Task: {'id': task-1111371, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.286515} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 841.727684] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-e45ed335-7978-4ef0-b81d-055a7bca1834 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.729603] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-c9f99ed7-857d-4064-a182-30f784598d4b tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Deleted the datastore file {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 841.729789] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-c9f99ed7-857d-4064-a182-30f784598d4b tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] [instance: d2a4f66a-f1a2-4f0d-826d-0b7c4f04bd23] Deleted contents of the VM from datastore datastore1 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 841.730809] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-c9f99ed7-857d-4064-a182-30f784598d4b tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] [instance: d2a4f66a-f1a2-4f0d-826d-0b7c4f04bd23] Instance destroyed {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 841.730809] env[63515]: INFO nova.compute.manager [None req-c9f99ed7-857d-4064-a182-30f784598d4b tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] [instance: d2a4f66a-f1a2-4f0d-826d-0b7c4f04bd23] Took 1.49 seconds to destroy the instance on the hypervisor. [ 841.730809] env[63515]: DEBUG oslo.service.loopingcall [None req-c9f99ed7-857d-4064-a182-30f784598d4b tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 841.730809] env[63515]: DEBUG nova.compute.manager [-] [instance: d2a4f66a-f1a2-4f0d-826d-0b7c4f04bd23] Deallocating network for instance {{(pid=63515) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 841.730809] env[63515]: DEBUG nova.network.neutron [-] [instance: d2a4f66a-f1a2-4f0d-826d-0b7c4f04bd23] deallocate_for_instance() {{(pid=63515) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 841.739027] env[63515]: DEBUG oslo_vmware.api [None req-ced15047-223c-4e6c-9243-6a752a33316f tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Waiting for the task: (returnval){ [ 841.739027] env[63515]: value = "task-1111372" [ 841.739027] env[63515]: _type = "Task" [ 841.739027] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 841.750927] env[63515]: DEBUG oslo_vmware.api [None req-ced15047-223c-4e6c-9243-6a752a33316f tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Task: {'id': task-1111372, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.789477] env[63515]: DEBUG oslo_vmware.api [None req-fe95cb73-4a2a-4d94-83ac-9bcd96be3709 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]5212b07a-82bc-583f-55e6-6621b0a0ece0, 'name': SearchDatastore_Task, 'duration_secs': 0.018716} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 841.789730] env[63515]: DEBUG oslo_concurrency.lockutils [None req-fe95cb73-4a2a-4d94-83ac-9bcd96be3709 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 841.790023] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-fe95cb73-4a2a-4d94-83ac-9bcd96be3709 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk to [datastore2] 19475235-dbec-40a5-bd8f-8070616b7ddb/19475235-dbec-40a5-bd8f-8070616b7ddb.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 841.790294] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-354b885f-fe29-49cd-8a50-6bcd10baaa1e {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.801046] env[63515]: DEBUG oslo_vmware.api [None req-fe95cb73-4a2a-4d94-83ac-9bcd96be3709 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Waiting for the task: (returnval){ [ 841.801046] env[63515]: value = "task-1111373" [ 841.801046] env[63515]: _type = "Task" [ 841.801046] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 841.810713] env[63515]: DEBUG oslo_vmware.api [None req-fe95cb73-4a2a-4d94-83ac-9bcd96be3709 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Task: {'id': task-1111373, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.839113] env[63515]: INFO nova.compute.resource_tracker [None req-4cdfd447-25bb-42ff-81df-85503f57f660 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: d926c699-a64a-4942-9ef4-f0166414661d] Updating resource usage from migration ae31ad09-2d25-42de-82b4-75e2937a1149 [ 841.952051] env[63515]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-8abac034-d970-4ed7-bb7a-7ddcd6290b84 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.965722] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61d5e8c7-c7cd-4b20-83e9-09aab5fa34a4 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.009029] env[63515]: DEBUG nova.compute.manager [req-7f79d2ce-f34d-42fc-9be0-d55a23ac6ea7 req-16d9caef-65ec-4e96-8784-e9aedad484f8 service nova] [instance: 5b42f744-fdd6-45b1-8563-896869648c23] Detach interface failed, port_id=3a630639-5423-4e62-9beb-ee1e7054f47d, reason: Instance 5b42f744-fdd6-45b1-8563-896869648c23 could not be found. {{(pid=63515) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 842.009029] env[63515]: DEBUG nova.network.neutron [None req-7bca9c72-42f5-4d65-99a2-4374b66c9c04 tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] [instance: 17275472-2921-49c9-b4e0-5485649ebd1c] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 842.023161] env[63515]: DEBUG nova.compute.manager [req-780e8798-9bba-4dc2-bf7d-941e7442ceaf req-89b6b4da-7399-4b8f-8a4f-c1ee7117eec9 service nova] [instance: 17275472-2921-49c9-b4e0-5485649ebd1c] Received event network-vif-plugged-eadc0731-8390-4414-b5cb-0f6a0c660979 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 842.023461] env[63515]: DEBUG oslo_concurrency.lockutils [req-780e8798-9bba-4dc2-bf7d-941e7442ceaf req-89b6b4da-7399-4b8f-8a4f-c1ee7117eec9 service nova] Acquiring lock "17275472-2921-49c9-b4e0-5485649ebd1c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 842.023647] env[63515]: DEBUG oslo_concurrency.lockutils [req-780e8798-9bba-4dc2-bf7d-941e7442ceaf req-89b6b4da-7399-4b8f-8a4f-c1ee7117eec9 service nova] Lock "17275472-2921-49c9-b4e0-5485649ebd1c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 842.023835] env[63515]: DEBUG oslo_concurrency.lockutils [req-780e8798-9bba-4dc2-bf7d-941e7442ceaf req-89b6b4da-7399-4b8f-8a4f-c1ee7117eec9 service nova] Lock "17275472-2921-49c9-b4e0-5485649ebd1c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 842.024022] env[63515]: DEBUG nova.compute.manager [req-780e8798-9bba-4dc2-bf7d-941e7442ceaf req-89b6b4da-7399-4b8f-8a4f-c1ee7117eec9 service nova] [instance: 17275472-2921-49c9-b4e0-5485649ebd1c] No waiting events found dispatching network-vif-plugged-eadc0731-8390-4414-b5cb-0f6a0c660979 {{(pid=63515) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 842.024198] env[63515]: WARNING nova.compute.manager [req-780e8798-9bba-4dc2-bf7d-941e7442ceaf req-89b6b4da-7399-4b8f-8a4f-c1ee7117eec9 service nova] [instance: 17275472-2921-49c9-b4e0-5485649ebd1c] Received unexpected event network-vif-plugged-eadc0731-8390-4414-b5cb-0f6a0c660979 for instance with vm_state building and task_state spawning. [ 842.024362] env[63515]: DEBUG nova.compute.manager [req-780e8798-9bba-4dc2-bf7d-941e7442ceaf req-89b6b4da-7399-4b8f-8a4f-c1ee7117eec9 service nova] [instance: 17275472-2921-49c9-b4e0-5485649ebd1c] Received event network-changed-eadc0731-8390-4414-b5cb-0f6a0c660979 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 842.024519] env[63515]: DEBUG nova.compute.manager [req-780e8798-9bba-4dc2-bf7d-941e7442ceaf req-89b6b4da-7399-4b8f-8a4f-c1ee7117eec9 service nova] [instance: 17275472-2921-49c9-b4e0-5485649ebd1c] Refreshing instance network info cache due to event network-changed-eadc0731-8390-4414-b5cb-0f6a0c660979. {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 842.024689] env[63515]: DEBUG oslo_concurrency.lockutils [req-780e8798-9bba-4dc2-bf7d-941e7442ceaf req-89b6b4da-7399-4b8f-8a4f-c1ee7117eec9 service nova] Acquiring lock "refresh_cache-17275472-2921-49c9-b4e0-5485649ebd1c" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 842.103311] env[63515]: DEBUG oslo_concurrency.lockutils [None req-3db8c3e0-5f0e-4c46-ab36-d6f97ddeb4f1 tempest-ServerActionsV293TestJSON-1193370169 tempest-ServerActionsV293TestJSON-1193370169-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 842.204516] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0f15a6a-aa0c-4c20-b901-e3d25e09cffd {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.216063] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c1dfff5-0437-41e3-9c21-e6333781883e {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.254880] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01a93c37-7c4c-4486-9649-ce067ea3699d {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.267354] env[63515]: DEBUG oslo_vmware.api [None req-ced15047-223c-4e6c-9243-6a752a33316f tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Task: {'id': task-1111372, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.272065] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7e2a1bf-16fc-407a-87dc-0d0b4b22706b {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.289019] env[63515]: DEBUG nova.compute.provider_tree [None req-4cdfd447-25bb-42ff-81df-85503f57f660 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 842.314368] env[63515]: DEBUG oslo_vmware.api [None req-fe95cb73-4a2a-4d94-83ac-9bcd96be3709 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Task: {'id': task-1111373, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.493959} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 842.314845] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-fe95cb73-4a2a-4d94-83ac-9bcd96be3709 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk to [datastore2] 19475235-dbec-40a5-bd8f-8070616b7ddb/19475235-dbec-40a5-bd8f-8070616b7ddb.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 842.316320] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-fe95cb73-4a2a-4d94-83ac-9bcd96be3709 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: 19475235-dbec-40a5-bd8f-8070616b7ddb] Extending root virtual disk to 1048576 {{(pid=63515) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 842.316320] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b5dca614-d7e1-47a0-a993-eb473d7b1fe5 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.328693] env[63515]: DEBUG oslo_vmware.api [None req-fe95cb73-4a2a-4d94-83ac-9bcd96be3709 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Waiting for the task: (returnval){ [ 842.328693] env[63515]: value = "task-1111374" [ 842.328693] env[63515]: _type = "Task" [ 842.328693] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 842.330378] env[63515]: DEBUG nova.compute.manager [req-33b896c8-05a1-4af7-92a4-41de1a420a38 req-a8f1694a-1278-494e-b980-8201e3c0e524 service nova] [instance: d2a4f66a-f1a2-4f0d-826d-0b7c4f04bd23] Received event network-vif-deleted-3bca52da-5931-4d22-bdf1-33072a3aca76 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 842.331157] env[63515]: INFO nova.compute.manager [req-33b896c8-05a1-4af7-92a4-41de1a420a38 req-a8f1694a-1278-494e-b980-8201e3c0e524 service nova] [instance: d2a4f66a-f1a2-4f0d-826d-0b7c4f04bd23] Neutron deleted interface 3bca52da-5931-4d22-bdf1-33072a3aca76; detaching it from the instance and deleting it from the info cache [ 842.331157] env[63515]: DEBUG nova.network.neutron [req-33b896c8-05a1-4af7-92a4-41de1a420a38 req-a8f1694a-1278-494e-b980-8201e3c0e524 service nova] [instance: d2a4f66a-f1a2-4f0d-826d-0b7c4f04bd23] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 842.346836] env[63515]: DEBUG oslo_vmware.api [None req-fe95cb73-4a2a-4d94-83ac-9bcd96be3709 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Task: {'id': task-1111374, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.666327] env[63515]: DEBUG nova.network.neutron [-] [instance: d2a4f66a-f1a2-4f0d-826d-0b7c4f04bd23] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 842.680728] env[63515]: DEBUG nova.network.neutron [None req-7bca9c72-42f5-4d65-99a2-4374b66c9c04 tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] [instance: 17275472-2921-49c9-b4e0-5485649ebd1c] Updating instance_info_cache with network_info: [{"id": "5514ac51-9696-4c26-9abc-2cc0deef264e", "address": "fa:16:3e:d8:62:e7", "network": {"id": "23240c8d-f699-4df6-a6ba-8bc3b6db56cf", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-236578753", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.205", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "26a52fb1944a4f949f6779829e163eaa", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ab93899c-92b2-4d84-95a6-192234add28c", "external-id": "nsx-vlan-transportzone-697", "segmentation_id": 697, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5514ac51-96", "ovs_interfaceid": "5514ac51-9696-4c26-9abc-2cc0deef264e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "db47ccfa-8557-4128-9e65-8a8ab57163bb", "address": "fa:16:3e:1c:ef:74", "network": {"id": "f38f2c60-a871-4bd1-a969-f5e2e4afb0c1", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1293036143", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.45", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "26a52fb1944a4f949f6779829e163eaa", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "31ac3fea-ebf4-4bed-bf70-1eaecdf71280", "external-id": "nsx-vlan-transportzone-489", "segmentation_id": 489, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdb47ccfa-85", "ovs_interfaceid": "db47ccfa-8557-4128-9e65-8a8ab57163bb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "eadc0731-8390-4414-b5cb-0f6a0c660979", "address": "fa:16:3e:b7:0a:b8", "network": {"id": "23240c8d-f699-4df6-a6ba-8bc3b6db56cf", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-236578753", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.78", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "26a52fb1944a4f949f6779829e163eaa", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ab93899c-92b2-4d84-95a6-192234add28c", "external-id": "nsx-vlan-transportzone-697", "segmentation_id": 697, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapeadc0731-83", "ovs_interfaceid": "eadc0731-8390-4414-b5cb-0f6a0c660979", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 842.760204] env[63515]: DEBUG oslo_vmware.api [None req-ced15047-223c-4e6c-9243-6a752a33316f tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Task: {'id': task-1111372, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.792061] env[63515]: DEBUG nova.scheduler.client.report [None req-4cdfd447-25bb-42ff-81df-85503f57f660 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 842.836377] env[63515]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-820d1198-d77c-4a9c-9d22-8c6fa1fc670a {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.844217] env[63515]: DEBUG oslo_vmware.api [None req-fe95cb73-4a2a-4d94-83ac-9bcd96be3709 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Task: {'id': task-1111374, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.074636} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 842.845711] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-fe95cb73-4a2a-4d94-83ac-9bcd96be3709 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: 19475235-dbec-40a5-bd8f-8070616b7ddb] Extended root virtual disk {{(pid=63515) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 842.846562] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d54ac204-7608-4dcc-8419-b81609d9e2eb {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.851716] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65d4fee1-2525-4b60-a677-1447554debc8 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.886983] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-fe95cb73-4a2a-4d94-83ac-9bcd96be3709 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: 19475235-dbec-40a5-bd8f-8070616b7ddb] Reconfiguring VM instance instance-00000043 to attach disk [datastore2] 19475235-dbec-40a5-bd8f-8070616b7ddb/19475235-dbec-40a5-bd8f-8070616b7ddb.vmdk or device None with type sparse {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 842.894843] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-231c7881-38e6-45cb-b85d-89ddde35cdcb {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.908900] env[63515]: DEBUG nova.compute.manager [req-33b896c8-05a1-4af7-92a4-41de1a420a38 req-a8f1694a-1278-494e-b980-8201e3c0e524 service nova] [instance: d2a4f66a-f1a2-4f0d-826d-0b7c4f04bd23] Detach interface failed, port_id=3bca52da-5931-4d22-bdf1-33072a3aca76, reason: Instance d2a4f66a-f1a2-4f0d-826d-0b7c4f04bd23 could not be found. {{(pid=63515) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 842.916356] env[63515]: DEBUG oslo_vmware.api [None req-fe95cb73-4a2a-4d94-83ac-9bcd96be3709 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Waiting for the task: (returnval){ [ 842.916356] env[63515]: value = "task-1111375" [ 842.916356] env[63515]: _type = "Task" [ 842.916356] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 842.924657] env[63515]: DEBUG oslo_vmware.api [None req-fe95cb73-4a2a-4d94-83ac-9bcd96be3709 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Task: {'id': task-1111375, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.174042] env[63515]: INFO nova.compute.manager [-] [instance: d2a4f66a-f1a2-4f0d-826d-0b7c4f04bd23] Took 1.44 seconds to deallocate network for instance. [ 843.183061] env[63515]: DEBUG oslo_concurrency.lockutils [None req-7bca9c72-42f5-4d65-99a2-4374b66c9c04 tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] Releasing lock "refresh_cache-17275472-2921-49c9-b4e0-5485649ebd1c" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 843.183306] env[63515]: DEBUG nova.compute.manager [None req-7bca9c72-42f5-4d65-99a2-4374b66c9c04 tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] [instance: 17275472-2921-49c9-b4e0-5485649ebd1c] Instance network_info: |[{"id": "5514ac51-9696-4c26-9abc-2cc0deef264e", "address": "fa:16:3e:d8:62:e7", "network": {"id": "23240c8d-f699-4df6-a6ba-8bc3b6db56cf", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-236578753", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.205", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "26a52fb1944a4f949f6779829e163eaa", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ab93899c-92b2-4d84-95a6-192234add28c", "external-id": "nsx-vlan-transportzone-697", "segmentation_id": 697, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5514ac51-96", "ovs_interfaceid": "5514ac51-9696-4c26-9abc-2cc0deef264e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "db47ccfa-8557-4128-9e65-8a8ab57163bb", "address": "fa:16:3e:1c:ef:74", "network": {"id": "f38f2c60-a871-4bd1-a969-f5e2e4afb0c1", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1293036143", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.45", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "26a52fb1944a4f949f6779829e163eaa", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "31ac3fea-ebf4-4bed-bf70-1eaecdf71280", "external-id": "nsx-vlan-transportzone-489", "segmentation_id": 489, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdb47ccfa-85", "ovs_interfaceid": "db47ccfa-8557-4128-9e65-8a8ab57163bb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "eadc0731-8390-4414-b5cb-0f6a0c660979", "address": "fa:16:3e:b7:0a:b8", "network": {"id": "23240c8d-f699-4df6-a6ba-8bc3b6db56cf", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-236578753", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.78", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "26a52fb1944a4f949f6779829e163eaa", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ab93899c-92b2-4d84-95a6-192234add28c", "external-id": "nsx-vlan-transportzone-697", "segmentation_id": 697, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapeadc0731-83", "ovs_interfaceid": "eadc0731-8390-4414-b5cb-0f6a0c660979", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 843.183628] env[63515]: DEBUG oslo_concurrency.lockutils [req-780e8798-9bba-4dc2-bf7d-941e7442ceaf req-89b6b4da-7399-4b8f-8a4f-c1ee7117eec9 service nova] Acquired lock "refresh_cache-17275472-2921-49c9-b4e0-5485649ebd1c" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 843.183808] env[63515]: DEBUG nova.network.neutron [req-780e8798-9bba-4dc2-bf7d-941e7442ceaf req-89b6b4da-7399-4b8f-8a4f-c1ee7117eec9 service nova] [instance: 17275472-2921-49c9-b4e0-5485649ebd1c] Refreshing network info cache for port eadc0731-8390-4414-b5cb-0f6a0c660979 {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 843.184844] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-7bca9c72-42f5-4d65-99a2-4374b66c9c04 tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] [instance: 17275472-2921-49c9-b4e0-5485649ebd1c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:d8:62:e7', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ab93899c-92b2-4d84-95a6-192234add28c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '5514ac51-9696-4c26-9abc-2cc0deef264e', 'vif_model': 'vmxnet3'}, {'network_name': 'br-int', 'mac_address': 'fa:16:3e:1c:ef:74', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '31ac3fea-ebf4-4bed-bf70-1eaecdf71280', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'db47ccfa-8557-4128-9e65-8a8ab57163bb', 'vif_model': 'vmxnet3'}, {'network_name': 'br-int', 'mac_address': 'fa:16:3e:b7:0a:b8', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ab93899c-92b2-4d84-95a6-192234add28c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'eadc0731-8390-4414-b5cb-0f6a0c660979', 'vif_model': 'vmxnet3'}] {{(pid=63515) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 843.198868] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-7bca9c72-42f5-4d65-99a2-4374b66c9c04 tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] Creating folder: Project (26a52fb1944a4f949f6779829e163eaa). Parent ref: group-v243370. {{(pid=63515) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 843.204404] env[63515]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-566e7d8f-b31a-42f3-b793-2d71bc8623c3 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.217731] env[63515]: INFO nova.virt.vmwareapi.vm_util [None req-7bca9c72-42f5-4d65-99a2-4374b66c9c04 tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] Created folder: Project (26a52fb1944a4f949f6779829e163eaa) in parent group-v243370. [ 843.217731] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-7bca9c72-42f5-4d65-99a2-4374b66c9c04 tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] Creating folder: Instances. Parent ref: group-v243489. {{(pid=63515) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 843.217913] env[63515]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-471daf59-ec10-406f-b51f-c56f663b6a80 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.228677] env[63515]: INFO nova.virt.vmwareapi.vm_util [None req-7bca9c72-42f5-4d65-99a2-4374b66c9c04 tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] Created folder: Instances in parent group-v243489. [ 843.228962] env[63515]: DEBUG oslo.service.loopingcall [None req-7bca9c72-42f5-4d65-99a2-4374b66c9c04 tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 843.229180] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 17275472-2921-49c9-b4e0-5485649ebd1c] Creating VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 843.229534] env[63515]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c0dcaa2d-f9d3-4ff0-a168-4949ab0e350a {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.256392] env[63515]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 843.256392] env[63515]: value = "task-1111378" [ 843.256392] env[63515]: _type = "Task" [ 843.256392] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 843.263442] env[63515]: DEBUG oslo_vmware.api [None req-ced15047-223c-4e6c-9243-6a752a33316f tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Task: {'id': task-1111372, 'name': CreateSnapshot_Task, 'duration_secs': 1.146311} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 843.264169] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-ced15047-223c-4e6c-9243-6a752a33316f tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: 70df69e5-687b-44fb-b6fc-cdb08e21dda0] Created Snapshot of the VM instance {{(pid=63515) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 843.264942] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd0b5f39-469f-4daa-b4b1-4e6b47e8b590 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.271156] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1111378, 'name': CreateVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.299372] env[63515]: DEBUG oslo_concurrency.lockutils [None req-4cdfd447-25bb-42ff-81df-85503f57f660 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.473s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 843.299665] env[63515]: INFO nova.compute.manager [None req-4cdfd447-25bb-42ff-81df-85503f57f660 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: d926c699-a64a-4942-9ef4-f0166414661d] Migrating [ 843.308044] env[63515]: DEBUG oslo_concurrency.lockutils [None req-f30dfc84-d4b1-4816-9900-ab0daf6eae41 tempest-ServersTestJSON-324454119 tempest-ServersTestJSON-324454119-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 24.806s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 843.308044] env[63515]: DEBUG nova.objects.instance [None req-f30dfc84-d4b1-4816-9900-ab0daf6eae41 tempest-ServersTestJSON-324454119 tempest-ServersTestJSON-324454119-project-member] Lazy-loading 'resources' on Instance uuid 70ed982f-affd-4dd1-bc90-c64e7c6d49d2 {{(pid=63515) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 843.427479] env[63515]: DEBUG oslo_vmware.api [None req-fe95cb73-4a2a-4d94-83ac-9bcd96be3709 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Task: {'id': task-1111375, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.509385] env[63515]: DEBUG nova.network.neutron [req-780e8798-9bba-4dc2-bf7d-941e7442ceaf req-89b6b4da-7399-4b8f-8a4f-c1ee7117eec9 service nova] [instance: 17275472-2921-49c9-b4e0-5485649ebd1c] Updated VIF entry in instance network info cache for port eadc0731-8390-4414-b5cb-0f6a0c660979. {{(pid=63515) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 843.510310] env[63515]: DEBUG nova.network.neutron [req-780e8798-9bba-4dc2-bf7d-941e7442ceaf req-89b6b4da-7399-4b8f-8a4f-c1ee7117eec9 service nova] [instance: 17275472-2921-49c9-b4e0-5485649ebd1c] Updating instance_info_cache with network_info: [{"id": "5514ac51-9696-4c26-9abc-2cc0deef264e", "address": "fa:16:3e:d8:62:e7", "network": {"id": "23240c8d-f699-4df6-a6ba-8bc3b6db56cf", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-236578753", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.205", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "26a52fb1944a4f949f6779829e163eaa", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ab93899c-92b2-4d84-95a6-192234add28c", "external-id": "nsx-vlan-transportzone-697", "segmentation_id": 697, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5514ac51-96", "ovs_interfaceid": "5514ac51-9696-4c26-9abc-2cc0deef264e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "db47ccfa-8557-4128-9e65-8a8ab57163bb", "address": "fa:16:3e:1c:ef:74", "network": {"id": "f38f2c60-a871-4bd1-a969-f5e2e4afb0c1", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1293036143", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.45", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "26a52fb1944a4f949f6779829e163eaa", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "31ac3fea-ebf4-4bed-bf70-1eaecdf71280", "external-id": "nsx-vlan-transportzone-489", "segmentation_id": 489, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdb47ccfa-85", "ovs_interfaceid": "db47ccfa-8557-4128-9e65-8a8ab57163bb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "eadc0731-8390-4414-b5cb-0f6a0c660979", "address": "fa:16:3e:b7:0a:b8", "network": {"id": "23240c8d-f699-4df6-a6ba-8bc3b6db56cf", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-236578753", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.78", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "26a52fb1944a4f949f6779829e163eaa", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ab93899c-92b2-4d84-95a6-192234add28c", "external-id": "nsx-vlan-transportzone-697", "segmentation_id": 697, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapeadc0731-83", "ovs_interfaceid": "eadc0731-8390-4414-b5cb-0f6a0c660979", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 843.681423] env[63515]: DEBUG oslo_concurrency.lockutils [None req-c9f99ed7-857d-4064-a182-30f784598d4b tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 843.766673] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1111378, 'name': CreateVM_Task} progress is 99%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.787797] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-ced15047-223c-4e6c-9243-6a752a33316f tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: 70df69e5-687b-44fb-b6fc-cdb08e21dda0] Creating linked-clone VM from snapshot {{(pid=63515) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 843.788569] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-4f55e3fe-fcd1-458d-a1e0-8343f4abbc7a {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.798309] env[63515]: DEBUG oslo_vmware.api [None req-ced15047-223c-4e6c-9243-6a752a33316f tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Waiting for the task: (returnval){ [ 843.798309] env[63515]: value = "task-1111379" [ 843.798309] env[63515]: _type = "Task" [ 843.798309] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 843.807704] env[63515]: DEBUG oslo_vmware.api [None req-ced15047-223c-4e6c-9243-6a752a33316f tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Task: {'id': task-1111379, 'name': CloneVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.821860] env[63515]: DEBUG oslo_concurrency.lockutils [None req-4cdfd447-25bb-42ff-81df-85503f57f660 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Acquiring lock "refresh_cache-d926c699-a64a-4942-9ef4-f0166414661d" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 843.821860] env[63515]: DEBUG oslo_concurrency.lockutils [None req-4cdfd447-25bb-42ff-81df-85503f57f660 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Acquired lock "refresh_cache-d926c699-a64a-4942-9ef4-f0166414661d" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 843.821999] env[63515]: DEBUG nova.network.neutron [None req-4cdfd447-25bb-42ff-81df-85503f57f660 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: d926c699-a64a-4942-9ef4-f0166414661d] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 843.936746] env[63515]: DEBUG oslo_vmware.api [None req-fe95cb73-4a2a-4d94-83ac-9bcd96be3709 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Task: {'id': task-1111375, 'name': ReconfigVM_Task, 'duration_secs': 0.656067} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 843.940473] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-fe95cb73-4a2a-4d94-83ac-9bcd96be3709 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: 19475235-dbec-40a5-bd8f-8070616b7ddb] Reconfigured VM instance instance-00000043 to attach disk [datastore2] 19475235-dbec-40a5-bd8f-8070616b7ddb/19475235-dbec-40a5-bd8f-8070616b7ddb.vmdk or device None with type sparse {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 843.941530] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d88c6b98-53e1-4b61-9711-d061c2d63159 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.950779] env[63515]: DEBUG oslo_vmware.api [None req-fe95cb73-4a2a-4d94-83ac-9bcd96be3709 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Waiting for the task: (returnval){ [ 843.950779] env[63515]: value = "task-1111380" [ 843.950779] env[63515]: _type = "Task" [ 843.950779] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 843.967350] env[63515]: DEBUG oslo_vmware.api [None req-fe95cb73-4a2a-4d94-83ac-9bcd96be3709 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Task: {'id': task-1111380, 'name': Rename_Task} progress is 6%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.013210] env[63515]: DEBUG oslo_concurrency.lockutils [req-780e8798-9bba-4dc2-bf7d-941e7442ceaf req-89b6b4da-7399-4b8f-8a4f-c1ee7117eec9 service nova] Releasing lock "refresh_cache-17275472-2921-49c9-b4e0-5485649ebd1c" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 844.114046] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-891ec946-599b-4d6e-b7d9-b828a5f79888 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.122522] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-714cab43-f2b1-4d4e-97ff-13ffc35800c0 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.890027] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1111378, 'name': CreateVM_Task} progress is 99%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.911520] env[63515]: DEBUG oslo_vmware.api [None req-ced15047-223c-4e6c-9243-6a752a33316f tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Task: {'id': task-1111379, 'name': CloneVM_Task} progress is 94%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.912619] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cda4a995-cbc3-4bd6-a680-6a483cfc27cc {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.920739] env[63515]: DEBUG oslo_vmware.api [None req-fe95cb73-4a2a-4d94-83ac-9bcd96be3709 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Task: {'id': task-1111380, 'name': Rename_Task, 'duration_secs': 0.395281} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 844.922932] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-fe95cb73-4a2a-4d94-83ac-9bcd96be3709 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: 19475235-dbec-40a5-bd8f-8070616b7ddb] Powering on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 844.923249] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2c1675d2-fd22-42be-b9bd-f9d33f5129ce {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.925883] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ee0d2e7-7dbc-47ec-bd72-331e60e9d101 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.940466] env[63515]: DEBUG nova.compute.provider_tree [None req-f30dfc84-d4b1-4816-9900-ab0daf6eae41 tempest-ServersTestJSON-324454119 tempest-ServersTestJSON-324454119-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 844.943710] env[63515]: DEBUG oslo_vmware.api [None req-fe95cb73-4a2a-4d94-83ac-9bcd96be3709 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Waiting for the task: (returnval){ [ 844.943710] env[63515]: value = "task-1111381" [ 844.943710] env[63515]: _type = "Task" [ 844.943710] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 844.952703] env[63515]: DEBUG oslo_vmware.api [None req-fe95cb73-4a2a-4d94-83ac-9bcd96be3709 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Task: {'id': task-1111381, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.378881] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1111378, 'name': CreateVM_Task} progress is 99%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.382272] env[63515]: DEBUG oslo_vmware.api [None req-ced15047-223c-4e6c-9243-6a752a33316f tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Task: {'id': task-1111379, 'name': CloneVM_Task, 'duration_secs': 1.557123} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 845.382596] env[63515]: INFO nova.virt.vmwareapi.vmops [None req-ced15047-223c-4e6c-9243-6a752a33316f tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: 70df69e5-687b-44fb-b6fc-cdb08e21dda0] Created linked-clone VM from snapshot [ 845.383400] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89f2e097-98a8-47b1-94b8-0a20d6b4d5c1 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.392836] env[63515]: DEBUG nova.virt.vmwareapi.images [None req-ced15047-223c-4e6c-9243-6a752a33316f tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: 70df69e5-687b-44fb-b6fc-cdb08e21dda0] Uploading image d1918637-9f8a-4d89-9bc2-49e7d0fdaeee {{(pid=63515) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 845.415488] env[63515]: DEBUG oslo_vmware.rw_handles [None req-ced15047-223c-4e6c-9243-6a752a33316f tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 845.415488] env[63515]: value = "vm-243492" [ 845.415488] env[63515]: _type = "VirtualMachine" [ 845.415488] env[63515]: }. {{(pid=63515) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 845.415797] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-0bb56344-c400-4d85-85c3-1dcd2ba332df {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.419422] env[63515]: DEBUG nova.network.neutron [None req-4cdfd447-25bb-42ff-81df-85503f57f660 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: d926c699-a64a-4942-9ef4-f0166414661d] Updating instance_info_cache with network_info: [{"id": "fdec91d2-43e8-4d40-9aa9-885a82fd9b42", "address": "fa:16:3e:5c:b7:75", "network": {"id": "15fdaabe-6793-400f-9737-2cdff8f07238", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1175204866-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f9c98e9f6020475490aaa7e76d907ab7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7d377d75-3add-4a15-8691-74b2eb010924", "external-id": "nsx-vlan-transportzone-71", "segmentation_id": 71, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfdec91d2-43", "ovs_interfaceid": "fdec91d2-43e8-4d40-9aa9-885a82fd9b42", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 845.425276] env[63515]: DEBUG oslo_vmware.rw_handles [None req-ced15047-223c-4e6c-9243-6a752a33316f tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Lease: (returnval){ [ 845.425276] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]52d2e88f-d86a-aa65-fe11-242aa2408e4d" [ 845.425276] env[63515]: _type = "HttpNfcLease" [ 845.425276] env[63515]: } obtained for exporting VM: (result){ [ 845.425276] env[63515]: value = "vm-243492" [ 845.425276] env[63515]: _type = "VirtualMachine" [ 845.425276] env[63515]: }. {{(pid=63515) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 845.425574] env[63515]: DEBUG oslo_vmware.api [None req-ced15047-223c-4e6c-9243-6a752a33316f tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Waiting for the lease: (returnval){ [ 845.425574] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]52d2e88f-d86a-aa65-fe11-242aa2408e4d" [ 845.425574] env[63515]: _type = "HttpNfcLease" [ 845.425574] env[63515]: } to be ready. {{(pid=63515) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 845.433376] env[63515]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 845.433376] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]52d2e88f-d86a-aa65-fe11-242aa2408e4d" [ 845.433376] env[63515]: _type = "HttpNfcLease" [ 845.433376] env[63515]: } is initializing. {{(pid=63515) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 845.444440] env[63515]: DEBUG nova.scheduler.client.report [None req-f30dfc84-d4b1-4816-9900-ab0daf6eae41 tempest-ServersTestJSON-324454119 tempest-ServersTestJSON-324454119-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 845.463391] env[63515]: DEBUG oslo_vmware.api [None req-fe95cb73-4a2a-4d94-83ac-9bcd96be3709 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Task: {'id': task-1111381, 'name': PowerOnVM_Task} progress is 89%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.880912] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1111378, 'name': CreateVM_Task} progress is 99%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.921861] env[63515]: DEBUG oslo_concurrency.lockutils [None req-4cdfd447-25bb-42ff-81df-85503f57f660 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Releasing lock "refresh_cache-d926c699-a64a-4942-9ef4-f0166414661d" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 845.935939] env[63515]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 845.935939] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]52d2e88f-d86a-aa65-fe11-242aa2408e4d" [ 845.935939] env[63515]: _type = "HttpNfcLease" [ 845.935939] env[63515]: } is ready. {{(pid=63515) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 845.940203] env[63515]: DEBUG oslo_vmware.rw_handles [None req-ced15047-223c-4e6c-9243-6a752a33316f tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 845.940203] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]52d2e88f-d86a-aa65-fe11-242aa2408e4d" [ 845.940203] env[63515]: _type = "HttpNfcLease" [ 845.940203] env[63515]: }. {{(pid=63515) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 845.940203] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df38edf3-f1f3-4323-a84b-03704edd5bfb {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.946797] env[63515]: DEBUG oslo_vmware.rw_handles [None req-ced15047-223c-4e6c-9243-6a752a33316f tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5251b1ed-216a-54ea-f1c3-3e298bc1ae33/disk-0.vmdk from lease info. {{(pid=63515) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 845.946993] env[63515]: DEBUG oslo_vmware.rw_handles [None req-ced15047-223c-4e6c-9243-6a752a33316f tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5251b1ed-216a-54ea-f1c3-3e298bc1ae33/disk-0.vmdk for reading. {{(pid=63515) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 845.948581] env[63515]: DEBUG oslo_concurrency.lockutils [None req-f30dfc84-d4b1-4816-9900-ab0daf6eae41 tempest-ServersTestJSON-324454119 tempest-ServersTestJSON-324454119-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.642s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 845.952377] env[63515]: DEBUG oslo_concurrency.lockutils [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 27.250s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 846.020135] env[63515]: INFO nova.scheduler.client.report [None req-f30dfc84-d4b1-4816-9900-ab0daf6eae41 tempest-ServersTestJSON-324454119 tempest-ServersTestJSON-324454119-project-member] Deleted allocations for instance 70ed982f-affd-4dd1-bc90-c64e7c6d49d2 [ 846.029623] env[63515]: DEBUG oslo_vmware.api [None req-fe95cb73-4a2a-4d94-83ac-9bcd96be3709 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Task: {'id': task-1111381, 'name': PowerOnVM_Task, 'duration_secs': 0.638707} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 846.030389] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-fe95cb73-4a2a-4d94-83ac-9bcd96be3709 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: 19475235-dbec-40a5-bd8f-8070616b7ddb] Powered on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 846.030566] env[63515]: INFO nova.compute.manager [None req-fe95cb73-4a2a-4d94-83ac-9bcd96be3709 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: 19475235-dbec-40a5-bd8f-8070616b7ddb] Took 8.89 seconds to spawn the instance on the hypervisor. [ 846.030747] env[63515]: DEBUG nova.compute.manager [None req-fe95cb73-4a2a-4d94-83ac-9bcd96be3709 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: 19475235-dbec-40a5-bd8f-8070616b7ddb] Checking state {{(pid=63515) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 846.032358] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c526704-742b-4ba7-8c03-0869a1800167 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.097532] env[63515]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-3c15a4c5-bde6-4cea-877e-8a8af400c261 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.380497] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1111378, 'name': CreateVM_Task} progress is 99%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.535303] env[63515]: DEBUG oslo_concurrency.lockutils [None req-f30dfc84-d4b1-4816-9900-ab0daf6eae41 tempest-ServersTestJSON-324454119 tempest-ServersTestJSON-324454119-project-member] Lock "70ed982f-affd-4dd1-bc90-c64e7c6d49d2" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 31.462s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 846.554795] env[63515]: INFO nova.compute.manager [None req-fe95cb73-4a2a-4d94-83ac-9bcd96be3709 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: 19475235-dbec-40a5-bd8f-8070616b7ddb] Took 33.90 seconds to build instance. [ 846.884037] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1111378, 'name': CreateVM_Task, 'duration_secs': 3.400906} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 846.884365] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 17275472-2921-49c9-b4e0-5485649ebd1c] Created VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 846.885389] env[63515]: DEBUG oslo_concurrency.lockutils [None req-7bca9c72-42f5-4d65-99a2-4374b66c9c04 tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 846.885645] env[63515]: DEBUG oslo_concurrency.lockutils [None req-7bca9c72-42f5-4d65-99a2-4374b66c9c04 tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 846.886091] env[63515]: DEBUG oslo_concurrency.lockutils [None req-7bca9c72-42f5-4d65-99a2-4374b66c9c04 tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 846.886329] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f1cb38f2-efa5-47fd-b4ac-9b95bb98c83d {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.894677] env[63515]: DEBUG oslo_vmware.api [None req-7bca9c72-42f5-4d65-99a2-4374b66c9c04 tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] Waiting for the task: (returnval){ [ 846.894677] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]52fe7dee-e3de-c852-bfd2-397e64f411cd" [ 846.894677] env[63515]: _type = "Task" [ 846.894677] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 846.904624] env[63515]: DEBUG oslo_vmware.api [None req-7bca9c72-42f5-4d65-99a2-4374b66c9c04 tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52fe7dee-e3de-c852-bfd2-397e64f411cd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.967136] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Applying migration context for instance d926c699-a64a-4942-9ef4-f0166414661d as it has an incoming, in-progress migration ae31ad09-2d25-42de-82b4-75e2937a1149. Migration status is migrating {{(pid=63515) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1016}} [ 846.970498] env[63515]: INFO nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] [instance: d926c699-a64a-4942-9ef4-f0166414661d] Updating resource usage from migration ae31ad09-2d25-42de-82b4-75e2937a1149 [ 847.010829] env[63515]: WARNING nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Instance 5b42f744-fdd6-45b1-8563-896869648c23 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 847.010829] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Instance d7e8dcdb-41cb-46fb-8b61-d251e7c2d372 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63515) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 847.010829] env[63515]: WARNING nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Instance d2a4f66a-f1a2-4f0d-826d-0b7c4f04bd23 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 847.010829] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Instance 5cbce760-0163-4b27-8ae3-e46c926c8916 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63515) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 847.011140] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Instance 63ce797d-7180-4209-ac2c-81978bf7607a actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63515) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 847.011140] env[63515]: WARNING nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Instance 4ef8066a-b68c-457b-9964-b1c34bab0fc3 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 847.057329] env[63515]: DEBUG oslo_concurrency.lockutils [None req-fe95cb73-4a2a-4d94-83ac-9bcd96be3709 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Lock "19475235-dbec-40a5-bd8f-8070616b7ddb" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 35.413s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 847.410068] env[63515]: DEBUG oslo_vmware.api [None req-7bca9c72-42f5-4d65-99a2-4374b66c9c04 tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52fe7dee-e3de-c852-bfd2-397e64f411cd, 'name': SearchDatastore_Task, 'duration_secs': 0.01198} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.410668] env[63515]: DEBUG oslo_concurrency.lockutils [None req-7bca9c72-42f5-4d65-99a2-4374b66c9c04 tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 847.411090] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-7bca9c72-42f5-4d65-99a2-4374b66c9c04 tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] [instance: 17275472-2921-49c9-b4e0-5485649ebd1c] Processing image 8a120570-cb06-4099-b262-554ca0ad15c5 {{(pid=63515) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 847.411748] env[63515]: DEBUG oslo_concurrency.lockutils [None req-7bca9c72-42f5-4d65-99a2-4374b66c9c04 tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 847.412158] env[63515]: DEBUG oslo_concurrency.lockutils [None req-7bca9c72-42f5-4d65-99a2-4374b66c9c04 tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 847.412781] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-7bca9c72-42f5-4d65-99a2-4374b66c9c04 tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 847.413276] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-77ffe222-50e0-4fd1-94c9-aeef2fcf024b {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.428025] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-7bca9c72-42f5-4d65-99a2-4374b66c9c04 tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 847.428025] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-7bca9c72-42f5-4d65-99a2-4374b66c9c04 tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63515) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 847.428025] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9800077c-e398-4885-8f14-539c34b1324a {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.436472] env[63515]: DEBUG oslo_vmware.api [None req-7bca9c72-42f5-4d65-99a2-4374b66c9c04 tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] Waiting for the task: (returnval){ [ 847.436472] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]524fa3b5-0878-9a8f-30b4-0ba7bfe43692" [ 847.436472] env[63515]: _type = "Task" [ 847.436472] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.438101] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5568286d-2e8e-4c40-8e28-c2ffb502263a {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.470695] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-4cdfd447-25bb-42ff-81df-85503f57f660 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: d926c699-a64a-4942-9ef4-f0166414661d] Updating instance 'd926c699-a64a-4942-9ef4-f0166414661d' progress to 0 {{(pid=63515) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 847.474270] env[63515]: DEBUG oslo_vmware.api [None req-7bca9c72-42f5-4d65-99a2-4374b66c9c04 tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]524fa3b5-0878-9a8f-30b4-0ba7bfe43692, 'name': SearchDatastore_Task, 'duration_secs': 0.013157} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.476615] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-126da46e-2d55-4b56-ac90-fc2aefd626f4 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.483342] env[63515]: DEBUG oslo_vmware.api [None req-7bca9c72-42f5-4d65-99a2-4374b66c9c04 tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] Waiting for the task: (returnval){ [ 847.483342] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]521fbea8-40f4-b705-e195-ce73d00018a3" [ 847.483342] env[63515]: _type = "Task" [ 847.483342] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.505481] env[63515]: DEBUG oslo_vmware.api [None req-7bca9c72-42f5-4d65-99a2-4374b66c9c04 tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]521fbea8-40f4-b705-e195-ce73d00018a3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.518407] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Instance 87c468d9-9594-4804-b461-527f01f6118f has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63515) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 847.518407] env[63515]: WARNING nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Instance 9d18c7b6-ce8e-4042-9e70-696bb7f57cb3 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 847.518407] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Instance 70df69e5-687b-44fb-b6fc-cdb08e21dda0 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63515) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 847.518407] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Instance 17275472-2921-49c9-b4e0-5485649ebd1c actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63515) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 847.518407] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Instance 19475235-dbec-40a5-bd8f-8070616b7ddb actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63515) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 847.518407] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Migration ae31ad09-2d25-42de-82b4-75e2937a1149 is active on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63515) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1712}} [ 847.518407] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Instance d926c699-a64a-4942-9ef4-f0166414661d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=63515) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 847.977094] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-4cdfd447-25bb-42ff-81df-85503f57f660 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: d926c699-a64a-4942-9ef4-f0166414661d] Powering off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 847.977489] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a2de2910-041e-47ed-93e8-fcd8a2e2e604 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.990152] env[63515]: DEBUG oslo_vmware.api [None req-4cdfd447-25bb-42ff-81df-85503f57f660 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Waiting for the task: (returnval){ [ 847.990152] env[63515]: value = "task-1111383" [ 847.990152] env[63515]: _type = "Task" [ 847.990152] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.998857] env[63515]: DEBUG oslo_vmware.api [None req-7bca9c72-42f5-4d65-99a2-4374b66c9c04 tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]521fbea8-40f4-b705-e195-ce73d00018a3, 'name': SearchDatastore_Task, 'duration_secs': 0.012089} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.999865] env[63515]: DEBUG oslo_concurrency.lockutils [None req-7bca9c72-42f5-4d65-99a2-4374b66c9c04 tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 848.000240] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-7bca9c72-42f5-4d65-99a2-4374b66c9c04 tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk to [datastore2] 17275472-2921-49c9-b4e0-5485649ebd1c/17275472-2921-49c9-b4e0-5485649ebd1c.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 848.000637] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-341f9859-54ca-48c0-8030-22f08ccf6749 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.008791] env[63515]: DEBUG oslo_vmware.api [None req-4cdfd447-25bb-42ff-81df-85503f57f660 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Task: {'id': task-1111383, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.016223] env[63515]: DEBUG oslo_vmware.api [None req-7bca9c72-42f5-4d65-99a2-4374b66c9c04 tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] Waiting for the task: (returnval){ [ 848.016223] env[63515]: value = "task-1111384" [ 848.016223] env[63515]: _type = "Task" [ 848.016223] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 848.022964] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Instance 705b7e9b-3ab6-40ec-8cf6-d86bf0a1faf0 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63515) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 848.031868] env[63515]: DEBUG oslo_vmware.api [None req-7bca9c72-42f5-4d65-99a2-4374b66c9c04 tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] Task: {'id': task-1111384, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.508216] env[63515]: DEBUG oslo_vmware.api [None req-4cdfd447-25bb-42ff-81df-85503f57f660 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Task: {'id': task-1111383, 'name': PowerOffVM_Task, 'duration_secs': 0.449283} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 848.508538] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-4cdfd447-25bb-42ff-81df-85503f57f660 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: d926c699-a64a-4942-9ef4-f0166414661d] Powered off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 848.508943] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-4cdfd447-25bb-42ff-81df-85503f57f660 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: d926c699-a64a-4942-9ef4-f0166414661d] Updating instance 'd926c699-a64a-4942-9ef4-f0166414661d' progress to 17 {{(pid=63515) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 848.529596] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Instance 2a29f37d-e786-48d3-b126-4854d6e7c67c has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63515) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 848.533320] env[63515]: DEBUG oslo_vmware.api [None req-7bca9c72-42f5-4d65-99a2-4374b66c9c04 tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] Task: {'id': task-1111384, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.761700] env[63515]: DEBUG oslo_concurrency.lockutils [None req-561efd36-c7e8-4824-9224-7669ba1ee132 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Acquiring lock "68766cd4-84be-475b-8494-d7ab43a9e969" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 848.761700] env[63515]: DEBUG oslo_concurrency.lockutils [None req-561efd36-c7e8-4824-9224-7669ba1ee132 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Lock "68766cd4-84be-475b-8494-d7ab43a9e969" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 849.016446] env[63515]: DEBUG nova.virt.hardware [None req-4cdfd447-25bb-42ff-81df-85503f57f660 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T02:52:16Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=8a120570-cb06-4099-b262-554ca0ad15c5,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 849.016866] env[63515]: DEBUG nova.virt.hardware [None req-4cdfd447-25bb-42ff-81df-85503f57f660 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Flavor limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 849.016920] env[63515]: DEBUG nova.virt.hardware [None req-4cdfd447-25bb-42ff-81df-85503f57f660 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Image limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 849.017197] env[63515]: DEBUG nova.virt.hardware [None req-4cdfd447-25bb-42ff-81df-85503f57f660 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Flavor pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 849.017398] env[63515]: DEBUG nova.virt.hardware [None req-4cdfd447-25bb-42ff-81df-85503f57f660 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Image pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 849.017588] env[63515]: DEBUG nova.virt.hardware [None req-4cdfd447-25bb-42ff-81df-85503f57f660 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 849.017958] env[63515]: DEBUG nova.virt.hardware [None req-4cdfd447-25bb-42ff-81df-85503f57f660 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 849.018251] env[63515]: DEBUG nova.virt.hardware [None req-4cdfd447-25bb-42ff-81df-85503f57f660 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 849.018502] env[63515]: DEBUG nova.virt.hardware [None req-4cdfd447-25bb-42ff-81df-85503f57f660 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Got 1 possible topologies {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 849.018722] env[63515]: DEBUG nova.virt.hardware [None req-4cdfd447-25bb-42ff-81df-85503f57f660 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 849.018950] env[63515]: DEBUG nova.virt.hardware [None req-4cdfd447-25bb-42ff-81df-85503f57f660 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 849.026114] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d3b2d560-ecf7-4ee5-8d1c-3bd0f1039fce {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.043162] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Instance 0ee9cce2-94b9-4038-ab38-1b3f96a864bc has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63515) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 849.053320] env[63515]: DEBUG oslo_vmware.api [None req-7bca9c72-42f5-4d65-99a2-4374b66c9c04 tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] Task: {'id': task-1111384, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.565374} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 849.058023] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-7bca9c72-42f5-4d65-99a2-4374b66c9c04 tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk to [datastore2] 17275472-2921-49c9-b4e0-5485649ebd1c/17275472-2921-49c9-b4e0-5485649ebd1c.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 849.058023] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-7bca9c72-42f5-4d65-99a2-4374b66c9c04 tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] [instance: 17275472-2921-49c9-b4e0-5485649ebd1c] Extending root virtual disk to 1048576 {{(pid=63515) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 849.058023] env[63515]: DEBUG oslo_vmware.api [None req-4cdfd447-25bb-42ff-81df-85503f57f660 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Waiting for the task: (returnval){ [ 849.058023] env[63515]: value = "task-1111385" [ 849.058023] env[63515]: _type = "Task" [ 849.058023] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 849.058023] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-38047b37-abc9-4216-bcda-26f6d4b2c811 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.070964] env[63515]: DEBUG oslo_vmware.api [None req-4cdfd447-25bb-42ff-81df-85503f57f660 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Task: {'id': task-1111385, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.072677] env[63515]: DEBUG oslo_vmware.api [None req-7bca9c72-42f5-4d65-99a2-4374b66c9c04 tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] Waiting for the task: (returnval){ [ 849.072677] env[63515]: value = "task-1111386" [ 849.072677] env[63515]: _type = "Task" [ 849.072677] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 849.086050] env[63515]: DEBUG oslo_vmware.api [None req-7bca9c72-42f5-4d65-99a2-4374b66c9c04 tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] Task: {'id': task-1111386, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.256029] env[63515]: DEBUG oslo_concurrency.lockutils [None req-3fd65342-5e9e-46b6-830a-47df91928006 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Acquiring lock "0967283f-8dda-4692-b6b7-601545cabf2f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 849.256310] env[63515]: DEBUG oslo_concurrency.lockutils [None req-3fd65342-5e9e-46b6-830a-47df91928006 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Lock "0967283f-8dda-4692-b6b7-601545cabf2f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 849.263688] env[63515]: DEBUG nova.compute.manager [None req-561efd36-c7e8-4824-9224-7669ba1ee132 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 68766cd4-84be-475b-8494-d7ab43a9e969] Starting instance... {{(pid=63515) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 849.549512] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Instance eab16df9-7bb5-4576-bca0-769a561c5fe9 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63515) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 849.575026] env[63515]: DEBUG oslo_vmware.api [None req-4cdfd447-25bb-42ff-81df-85503f57f660 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Task: {'id': task-1111385, 'name': ReconfigVM_Task, 'duration_secs': 0.348725} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 849.575026] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-4cdfd447-25bb-42ff-81df-85503f57f660 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: d926c699-a64a-4942-9ef4-f0166414661d] Updating instance 'd926c699-a64a-4942-9ef4-f0166414661d' progress to 33 {{(pid=63515) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 849.589050] env[63515]: DEBUG oslo_vmware.api [None req-7bca9c72-42f5-4d65-99a2-4374b66c9c04 tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] Task: {'id': task-1111386, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.069966} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 849.589492] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-7bca9c72-42f5-4d65-99a2-4374b66c9c04 tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] [instance: 17275472-2921-49c9-b4e0-5485649ebd1c] Extended root virtual disk {{(pid=63515) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 849.590404] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13ba8cde-b481-4447-a8d5-3779e98195e9 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.624639] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-7bca9c72-42f5-4d65-99a2-4374b66c9c04 tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] [instance: 17275472-2921-49c9-b4e0-5485649ebd1c] Reconfiguring VM instance instance-00000042 to attach disk [datastore2] 17275472-2921-49c9-b4e0-5485649ebd1c/17275472-2921-49c9-b4e0-5485649ebd1c.vmdk or device None with type sparse {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 849.625840] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e43591c5-e6ac-4135-bfa3-d0f6bfed0634 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.649091] env[63515]: DEBUG oslo_vmware.api [None req-7bca9c72-42f5-4d65-99a2-4374b66c9c04 tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] Waiting for the task: (returnval){ [ 849.649091] env[63515]: value = "task-1111387" [ 849.649091] env[63515]: _type = "Task" [ 849.649091] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 849.659087] env[63515]: DEBUG oslo_vmware.api [None req-7bca9c72-42f5-4d65-99a2-4374b66c9c04 tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] Task: {'id': task-1111387, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.760145] env[63515]: DEBUG nova.compute.manager [None req-3fd65342-5e9e-46b6-830a-47df91928006 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: 0967283f-8dda-4692-b6b7-601545cabf2f] Starting instance... {{(pid=63515) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 849.790926] env[63515]: DEBUG oslo_concurrency.lockutils [None req-561efd36-c7e8-4824-9224-7669ba1ee132 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 850.057745] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Instance 2ea99c7a-8b61-4718-bee0-f4ce4fc021af has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63515) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 850.058079] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Total usable vcpus: 48, total allocated vcpus: 8 {{(pid=63515) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 850.058275] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=2112MB phys_disk=200GB used_disk=8GB total_vcpus=48 used_vcpus=8 pci_stats=[] {{(pid=63515) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 850.082310] env[63515]: DEBUG nova.virt.hardware [None req-4cdfd447-25bb-42ff-81df-85503f57f660 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T02:52:14Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=8a120570-cb06-4099-b262-554ca0ad15c5,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 850.082594] env[63515]: DEBUG nova.virt.hardware [None req-4cdfd447-25bb-42ff-81df-85503f57f660 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Flavor limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 850.082760] env[63515]: DEBUG nova.virt.hardware [None req-4cdfd447-25bb-42ff-81df-85503f57f660 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Image limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 850.082963] env[63515]: DEBUG nova.virt.hardware [None req-4cdfd447-25bb-42ff-81df-85503f57f660 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Flavor pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 850.083321] env[63515]: DEBUG nova.virt.hardware [None req-4cdfd447-25bb-42ff-81df-85503f57f660 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Image pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 850.083502] env[63515]: DEBUG nova.virt.hardware [None req-4cdfd447-25bb-42ff-81df-85503f57f660 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 850.083807] env[63515]: DEBUG nova.virt.hardware [None req-4cdfd447-25bb-42ff-81df-85503f57f660 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 850.083992] env[63515]: DEBUG nova.virt.hardware [None req-4cdfd447-25bb-42ff-81df-85503f57f660 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 850.084190] env[63515]: DEBUG nova.virt.hardware [None req-4cdfd447-25bb-42ff-81df-85503f57f660 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Got 1 possible topologies {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 850.084362] env[63515]: DEBUG nova.virt.hardware [None req-4cdfd447-25bb-42ff-81df-85503f57f660 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 850.084545] env[63515]: DEBUG nova.virt.hardware [None req-4cdfd447-25bb-42ff-81df-85503f57f660 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 850.090293] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-4cdfd447-25bb-42ff-81df-85503f57f660 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: d926c699-a64a-4942-9ef4-f0166414661d] Reconfiguring VM instance instance-0000003f to detach disk 2000 {{(pid=63515) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 850.090870] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9a25bcbb-d2af-46f4-89ba-fc06b3e810d4 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.114588] env[63515]: DEBUG oslo_vmware.api [None req-4cdfd447-25bb-42ff-81df-85503f57f660 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Waiting for the task: (returnval){ [ 850.114588] env[63515]: value = "task-1111388" [ 850.114588] env[63515]: _type = "Task" [ 850.114588] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 850.130012] env[63515]: DEBUG oslo_vmware.api [None req-4cdfd447-25bb-42ff-81df-85503f57f660 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Task: {'id': task-1111388, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.163143] env[63515]: DEBUG oslo_vmware.api [None req-7bca9c72-42f5-4d65-99a2-4374b66c9c04 tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] Task: {'id': task-1111387, 'name': ReconfigVM_Task, 'duration_secs': 0.345536} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 850.163626] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-7bca9c72-42f5-4d65-99a2-4374b66c9c04 tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] [instance: 17275472-2921-49c9-b4e0-5485649ebd1c] Reconfigured VM instance instance-00000042 to attach disk [datastore2] 17275472-2921-49c9-b4e0-5485649ebd1c/17275472-2921-49c9-b4e0-5485649ebd1c.vmdk or device None with type sparse {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 850.164792] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-0869e8af-b09b-424f-90d3-35f4dda697cb {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.176946] env[63515]: DEBUG oslo_vmware.api [None req-7bca9c72-42f5-4d65-99a2-4374b66c9c04 tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] Waiting for the task: (returnval){ [ 850.176946] env[63515]: value = "task-1111389" [ 850.176946] env[63515]: _type = "Task" [ 850.176946] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 850.185497] env[63515]: DEBUG oslo_vmware.api [None req-7bca9c72-42f5-4d65-99a2-4374b66c9c04 tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] Task: {'id': task-1111389, 'name': Rename_Task} progress is 5%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.281401] env[63515]: DEBUG oslo_concurrency.lockutils [None req-3fd65342-5e9e-46b6-830a-47df91928006 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 850.364864] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28ef375b-4117-4b10-8836-81ec33284771 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.373731] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-870f8dc2-701e-475d-89a0-e33713cef2cd {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.405712] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2fe89f52-036b-4a03-974f-73d365ba3b5d {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.414624] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-794c58f3-9481-4539-a3bb-31ddbab0311d {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.430439] env[63515]: DEBUG nova.compute.provider_tree [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 850.630913] env[63515]: DEBUG oslo_vmware.api [None req-4cdfd447-25bb-42ff-81df-85503f57f660 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Task: {'id': task-1111388, 'name': ReconfigVM_Task, 'duration_secs': 0.388578} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 850.631884] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-4cdfd447-25bb-42ff-81df-85503f57f660 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: d926c699-a64a-4942-9ef4-f0166414661d] Reconfigured VM instance instance-0000003f to detach disk 2000 {{(pid=63515) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 850.632370] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2ec83d7-6863-480e-968f-1feba23340d4 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.657669] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-4cdfd447-25bb-42ff-81df-85503f57f660 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: d926c699-a64a-4942-9ef4-f0166414661d] Reconfiguring VM instance instance-0000003f to attach disk [datastore2] d926c699-a64a-4942-9ef4-f0166414661d/d926c699-a64a-4942-9ef4-f0166414661d.vmdk or device None with type thin {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 850.657998] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-36faedca-d496-40ea-b747-b711cfae173f {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.682147] env[63515]: DEBUG oslo_vmware.api [None req-4cdfd447-25bb-42ff-81df-85503f57f660 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Waiting for the task: (returnval){ [ 850.682147] env[63515]: value = "task-1111390" [ 850.682147] env[63515]: _type = "Task" [ 850.682147] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 850.689169] env[63515]: DEBUG oslo_vmware.api [None req-7bca9c72-42f5-4d65-99a2-4374b66c9c04 tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] Task: {'id': task-1111389, 'name': Rename_Task} progress is 14%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.694938] env[63515]: DEBUG oslo_vmware.api [None req-4cdfd447-25bb-42ff-81df-85503f57f660 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Task: {'id': task-1111390, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.933972] env[63515]: DEBUG nova.scheduler.client.report [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 851.192812] env[63515]: DEBUG oslo_vmware.api [None req-7bca9c72-42f5-4d65-99a2-4374b66c9c04 tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] Task: {'id': task-1111389, 'name': Rename_Task, 'duration_secs': 1.004403} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 851.192812] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-7bca9c72-42f5-4d65-99a2-4374b66c9c04 tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] [instance: 17275472-2921-49c9-b4e0-5485649ebd1c] Powering on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 851.192812] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c2456a54-4a95-4b19-bbaa-264a51b82293 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.196031] env[63515]: DEBUG oslo_vmware.api [None req-4cdfd447-25bb-42ff-81df-85503f57f660 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Task: {'id': task-1111390, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.202279] env[63515]: DEBUG oslo_vmware.api [None req-7bca9c72-42f5-4d65-99a2-4374b66c9c04 tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] Waiting for the task: (returnval){ [ 851.202279] env[63515]: value = "task-1111391" [ 851.202279] env[63515]: _type = "Task" [ 851.202279] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 851.211265] env[63515]: DEBUG oslo_vmware.api [None req-7bca9c72-42f5-4d65-99a2-4374b66c9c04 tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] Task: {'id': task-1111391, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.438948] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=63515) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 851.439631] env[63515]: DEBUG oslo_concurrency.lockutils [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 5.488s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 851.440200] env[63515]: DEBUG oslo_concurrency.lockutils [None req-ce131312-5bf5-4677-8923-2d91405ade5d tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 28.985s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 851.440527] env[63515]: DEBUG nova.objects.instance [None req-ce131312-5bf5-4677-8923-2d91405ade5d tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Lazy-loading 'pci_requests' on Instance uuid 87c468d9-9594-4804-b461-527f01f6118f {{(pid=63515) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 851.697048] env[63515]: DEBUG oslo_vmware.api [None req-4cdfd447-25bb-42ff-81df-85503f57f660 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Task: {'id': task-1111390, 'name': ReconfigVM_Task} progress is 99%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.713255] env[63515]: DEBUG oslo_vmware.api [None req-7bca9c72-42f5-4d65-99a2-4374b66c9c04 tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] Task: {'id': task-1111391, 'name': PowerOnVM_Task} progress is 89%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.946038] env[63515]: DEBUG nova.objects.instance [None req-ce131312-5bf5-4677-8923-2d91405ade5d tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Lazy-loading 'numa_topology' on Instance uuid 87c468d9-9594-4804-b461-527f01f6118f {{(pid=63515) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 852.195293] env[63515]: DEBUG oslo_vmware.api [None req-4cdfd447-25bb-42ff-81df-85503f57f660 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Task: {'id': task-1111390, 'name': ReconfigVM_Task} progress is 99%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.213867] env[63515]: DEBUG oslo_vmware.api [None req-7bca9c72-42f5-4d65-99a2-4374b66c9c04 tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] Task: {'id': task-1111391, 'name': PowerOnVM_Task, 'duration_secs': 0.908992} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 852.214335] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-7bca9c72-42f5-4d65-99a2-4374b66c9c04 tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] [instance: 17275472-2921-49c9-b4e0-5485649ebd1c] Powered on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 852.214684] env[63515]: INFO nova.compute.manager [None req-7bca9c72-42f5-4d65-99a2-4374b66c9c04 tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] [instance: 17275472-2921-49c9-b4e0-5485649ebd1c] Took 17.51 seconds to spawn the instance on the hypervisor. [ 852.215078] env[63515]: DEBUG nova.compute.manager [None req-7bca9c72-42f5-4d65-99a2-4374b66c9c04 tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] [instance: 17275472-2921-49c9-b4e0-5485649ebd1c] Checking state {{(pid=63515) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 852.217233] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ab3a28d-70c4-44cf-8e0e-24d3d4d4fca3 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.448745] env[63515]: INFO nova.compute.claims [None req-ce131312-5bf5-4677-8923-2d91405ade5d tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] [instance: 87c468d9-9594-4804-b461-527f01f6118f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 852.694503] env[63515]: DEBUG oslo_vmware.api [None req-4cdfd447-25bb-42ff-81df-85503f57f660 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Task: {'id': task-1111390, 'name': ReconfigVM_Task, 'duration_secs': 1.784992} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 852.694879] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-4cdfd447-25bb-42ff-81df-85503f57f660 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: d926c699-a64a-4942-9ef4-f0166414661d] Reconfigured VM instance instance-0000003f to attach disk [datastore2] d926c699-a64a-4942-9ef4-f0166414661d/d926c699-a64a-4942-9ef4-f0166414661d.vmdk or device None with type thin {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 852.695128] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-4cdfd447-25bb-42ff-81df-85503f57f660 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: d926c699-a64a-4942-9ef4-f0166414661d] Updating instance 'd926c699-a64a-4942-9ef4-f0166414661d' progress to 50 {{(pid=63515) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 852.755929] env[63515]: INFO nova.compute.manager [None req-7bca9c72-42f5-4d65-99a2-4374b66c9c04 tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] [instance: 17275472-2921-49c9-b4e0-5485649ebd1c] Took 42.20 seconds to build instance. [ 853.202533] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de8c3014-ff64-49d6-81af-0bfa3f3b73e1 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.235045] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a969e8f-3de2-42b5-a28c-c1cc3734f33e {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.254732] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-4cdfd447-25bb-42ff-81df-85503f57f660 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: d926c699-a64a-4942-9ef4-f0166414661d] Updating instance 'd926c699-a64a-4942-9ef4-f0166414661d' progress to 67 {{(pid=63515) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 853.258767] env[63515]: DEBUG oslo_concurrency.lockutils [None req-7bca9c72-42f5-4d65-99a2-4374b66c9c04 tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] Lock "17275472-2921-49c9-b4e0-5485649ebd1c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 43.720s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 853.767756] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86caad68-c864-464d-947b-01fc06d13c47 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.779321] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0470acf9-24fb-4c7c-9c29-185658b54c11 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.814104] env[63515]: DEBUG nova.network.neutron [None req-4cdfd447-25bb-42ff-81df-85503f57f660 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: d926c699-a64a-4942-9ef4-f0166414661d] Port fdec91d2-43e8-4d40-9aa9-885a82fd9b42 binding to destination host cpu-1 is already ACTIVE {{(pid=63515) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 853.816077] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-733dbbbf-ee14-4db4-bc51-e79a72846bc8 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.826550] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3aab5e5-7ec4-4e99-9782-cc8a1358d0e9 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.843485] env[63515]: DEBUG nova.compute.provider_tree [None req-ce131312-5bf5-4677-8923-2d91405ade5d tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 853.872459] env[63515]: DEBUG oslo_concurrency.lockutils [None req-593b35b4-5d74-4c58-a93e-9cad00fc0b2d tempest-ServerGroupTestJSON-385159797 tempest-ServerGroupTestJSON-385159797-project-member] Acquiring lock "a2e014a2-9e0f-4250-aa38-3a7154149903" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 853.872720] env[63515]: DEBUG oslo_concurrency.lockutils [None req-593b35b4-5d74-4c58-a93e-9cad00fc0b2d tempest-ServerGroupTestJSON-385159797 tempest-ServerGroupTestJSON-385159797-project-member] Lock "a2e014a2-9e0f-4250-aa38-3a7154149903" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 854.049732] env[63515]: DEBUG oslo_concurrency.lockutils [None req-dd655d0d-92fe-48b9-b34e-d4edd8542788 tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] Acquiring lock "17275472-2921-49c9-b4e0-5485649ebd1c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 854.050026] env[63515]: DEBUG oslo_concurrency.lockutils [None req-dd655d0d-92fe-48b9-b34e-d4edd8542788 tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] Lock "17275472-2921-49c9-b4e0-5485649ebd1c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 854.050266] env[63515]: DEBUG oslo_concurrency.lockutils [None req-dd655d0d-92fe-48b9-b34e-d4edd8542788 tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] Acquiring lock "17275472-2921-49c9-b4e0-5485649ebd1c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 854.050490] env[63515]: DEBUG oslo_concurrency.lockutils [None req-dd655d0d-92fe-48b9-b34e-d4edd8542788 tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] Lock "17275472-2921-49c9-b4e0-5485649ebd1c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 854.050649] env[63515]: DEBUG oslo_concurrency.lockutils [None req-dd655d0d-92fe-48b9-b34e-d4edd8542788 tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] Lock "17275472-2921-49c9-b4e0-5485649ebd1c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 854.052877] env[63515]: INFO nova.compute.manager [None req-dd655d0d-92fe-48b9-b34e-d4edd8542788 tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] [instance: 17275472-2921-49c9-b4e0-5485649ebd1c] Terminating instance [ 854.054813] env[63515]: DEBUG nova.compute.manager [None req-dd655d0d-92fe-48b9-b34e-d4edd8542788 tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] [instance: 17275472-2921-49c9-b4e0-5485649ebd1c] Start destroying the instance on the hypervisor. {{(pid=63515) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 854.055038] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-dd655d0d-92fe-48b9-b34e-d4edd8542788 tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] [instance: 17275472-2921-49c9-b4e0-5485649ebd1c] Destroying instance {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 854.055930] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83a508ee-c7a6-4450-92fe-898075d558d7 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.066105] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-dd655d0d-92fe-48b9-b34e-d4edd8542788 tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] [instance: 17275472-2921-49c9-b4e0-5485649ebd1c] Powering off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 854.066390] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-cbea9432-6255-4550-9b3e-fb25b479d8ad {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.073689] env[63515]: DEBUG oslo_vmware.api [None req-dd655d0d-92fe-48b9-b34e-d4edd8542788 tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] Waiting for the task: (returnval){ [ 854.073689] env[63515]: value = "task-1111392" [ 854.073689] env[63515]: _type = "Task" [ 854.073689] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 854.085736] env[63515]: DEBUG oslo_vmware.api [None req-dd655d0d-92fe-48b9-b34e-d4edd8542788 tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] Task: {'id': task-1111392, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.347482] env[63515]: DEBUG nova.scheduler.client.report [None req-ce131312-5bf5-4677-8923-2d91405ade5d tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 854.375440] env[63515]: DEBUG nova.compute.manager [None req-593b35b4-5d74-4c58-a93e-9cad00fc0b2d tempest-ServerGroupTestJSON-385159797 tempest-ServerGroupTestJSON-385159797-project-member] [instance: a2e014a2-9e0f-4250-aa38-3a7154149903] Starting instance... {{(pid=63515) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 854.584678] env[63515]: DEBUG oslo_vmware.api [None req-dd655d0d-92fe-48b9-b34e-d4edd8542788 tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] Task: {'id': task-1111392, 'name': PowerOffVM_Task, 'duration_secs': 0.238381} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 854.585403] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-dd655d0d-92fe-48b9-b34e-d4edd8542788 tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] [instance: 17275472-2921-49c9-b4e0-5485649ebd1c] Powered off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 854.585403] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-dd655d0d-92fe-48b9-b34e-d4edd8542788 tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] [instance: 17275472-2921-49c9-b4e0-5485649ebd1c] Unregistering the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 854.585630] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3e6a66f6-3346-46ab-b44f-84d66b703324 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.728695] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-dd655d0d-92fe-48b9-b34e-d4edd8542788 tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] [instance: 17275472-2921-49c9-b4e0-5485649ebd1c] Unregistered the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 854.728955] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-dd655d0d-92fe-48b9-b34e-d4edd8542788 tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] [instance: 17275472-2921-49c9-b4e0-5485649ebd1c] Deleting contents of the VM from datastore datastore2 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 854.729168] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-dd655d0d-92fe-48b9-b34e-d4edd8542788 tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] Deleting the datastore file [datastore2] 17275472-2921-49c9-b4e0-5485649ebd1c {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 854.729510] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-12e6dd3b-eebe-4e0a-b483-0b4cebba3b7d {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.738188] env[63515]: DEBUG oslo_vmware.api [None req-dd655d0d-92fe-48b9-b34e-d4edd8542788 tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] Waiting for the task: (returnval){ [ 854.738188] env[63515]: value = "task-1111394" [ 854.738188] env[63515]: _type = "Task" [ 854.738188] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 854.747189] env[63515]: DEBUG oslo_vmware.api [None req-dd655d0d-92fe-48b9-b34e-d4edd8542788 tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] Task: {'id': task-1111394, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.844448] env[63515]: DEBUG oslo_concurrency.lockutils [None req-4cdfd447-25bb-42ff-81df-85503f57f660 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Acquiring lock "d926c699-a64a-4942-9ef4-f0166414661d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 854.844823] env[63515]: DEBUG oslo_concurrency.lockutils [None req-4cdfd447-25bb-42ff-81df-85503f57f660 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Lock "d926c699-a64a-4942-9ef4-f0166414661d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 854.845505] env[63515]: DEBUG oslo_concurrency.lockutils [None req-4cdfd447-25bb-42ff-81df-85503f57f660 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Lock "d926c699-a64a-4942-9ef4-f0166414661d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 854.854161] env[63515]: DEBUG oslo_concurrency.lockutils [None req-ce131312-5bf5-4677-8923-2d91405ade5d tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.412s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 854.854997] env[63515]: DEBUG oslo_concurrency.lockutils [None req-b19ee0f2-2420-400a-b8e9-86908fd35c93 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 31.332s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 854.855517] env[63515]: DEBUG oslo_concurrency.lockutils [None req-b19ee0f2-2420-400a-b8e9-86908fd35c93 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.001s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 854.857669] env[63515]: DEBUG oslo_concurrency.lockutils [None req-957d8582-d135-4677-8fde-d84fe70e2761 tempest-ServerDiagnosticsV248Test-1736022867 tempest-ServerDiagnosticsV248Test-1736022867-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 29.080s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 854.859387] env[63515]: INFO nova.compute.claims [None req-957d8582-d135-4677-8fde-d84fe70e2761 tempest-ServerDiagnosticsV248Test-1736022867 tempest-ServerDiagnosticsV248Test-1736022867-project-member] [instance: 705b7e9b-3ab6-40ec-8cf6-d86bf0a1faf0] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 854.887614] env[63515]: INFO nova.scheduler.client.report [None req-b19ee0f2-2420-400a-b8e9-86908fd35c93 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Deleted allocations for instance 4ef8066a-b68c-457b-9964-b1c34bab0fc3 [ 854.895940] env[63515]: DEBUG oslo_concurrency.lockutils [None req-593b35b4-5d74-4c58-a93e-9cad00fc0b2d tempest-ServerGroupTestJSON-385159797 tempest-ServerGroupTestJSON-385159797-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 854.911960] env[63515]: INFO nova.network.neutron [None req-ce131312-5bf5-4677-8923-2d91405ade5d tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] [instance: 87c468d9-9594-4804-b461-527f01f6118f] Updating port d2098868-0475-4b06-bae3-785c608a861e with attributes {'binding:host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 855.250336] env[63515]: DEBUG oslo_vmware.api [None req-dd655d0d-92fe-48b9-b34e-d4edd8542788 tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] Task: {'id': task-1111394, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.21409} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 855.250744] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-dd655d0d-92fe-48b9-b34e-d4edd8542788 tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] Deleted the datastore file {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 855.250906] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-dd655d0d-92fe-48b9-b34e-d4edd8542788 tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] [instance: 17275472-2921-49c9-b4e0-5485649ebd1c] Deleted contents of the VM from datastore datastore2 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 855.251136] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-dd655d0d-92fe-48b9-b34e-d4edd8542788 tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] [instance: 17275472-2921-49c9-b4e0-5485649ebd1c] Instance destroyed {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 855.251348] env[63515]: INFO nova.compute.manager [None req-dd655d0d-92fe-48b9-b34e-d4edd8542788 tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] [instance: 17275472-2921-49c9-b4e0-5485649ebd1c] Took 1.20 seconds to destroy the instance on the hypervisor. [ 855.251633] env[63515]: DEBUG oslo.service.loopingcall [None req-dd655d0d-92fe-48b9-b34e-d4edd8542788 tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 855.251855] env[63515]: DEBUG nova.compute.manager [-] [instance: 17275472-2921-49c9-b4e0-5485649ebd1c] Deallocating network for instance {{(pid=63515) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 855.251958] env[63515]: DEBUG nova.network.neutron [-] [instance: 17275472-2921-49c9-b4e0-5485649ebd1c] deallocate_for_instance() {{(pid=63515) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 855.399039] env[63515]: DEBUG oslo_concurrency.lockutils [None req-b19ee0f2-2420-400a-b8e9-86908fd35c93 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Lock "4ef8066a-b68c-457b-9964-b1c34bab0fc3" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 36.798s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 855.596979] env[63515]: DEBUG nova.compute.manager [req-2509d33b-e6ba-4db0-9707-1f01517ddef2 req-72d50d79-977f-4d44-9de3-c37f1122a369 service nova] [instance: 17275472-2921-49c9-b4e0-5485649ebd1c] Received event network-vif-deleted-5514ac51-9696-4c26-9abc-2cc0deef264e {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 855.597251] env[63515]: INFO nova.compute.manager [req-2509d33b-e6ba-4db0-9707-1f01517ddef2 req-72d50d79-977f-4d44-9de3-c37f1122a369 service nova] [instance: 17275472-2921-49c9-b4e0-5485649ebd1c] Neutron deleted interface 5514ac51-9696-4c26-9abc-2cc0deef264e; detaching it from the instance and deleting it from the info cache [ 855.597572] env[63515]: DEBUG nova.network.neutron [req-2509d33b-e6ba-4db0-9707-1f01517ddef2 req-72d50d79-977f-4d44-9de3-c37f1122a369 service nova] [instance: 17275472-2921-49c9-b4e0-5485649ebd1c] Updating instance_info_cache with network_info: [{"id": "db47ccfa-8557-4128-9e65-8a8ab57163bb", "address": "fa:16:3e:1c:ef:74", "network": {"id": "f38f2c60-a871-4bd1-a969-f5e2e4afb0c1", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1293036143", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.45", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "26a52fb1944a4f949f6779829e163eaa", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "31ac3fea-ebf4-4bed-bf70-1eaecdf71280", "external-id": "nsx-vlan-transportzone-489", "segmentation_id": 489, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdb47ccfa-85", "ovs_interfaceid": "db47ccfa-8557-4128-9e65-8a8ab57163bb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "eadc0731-8390-4414-b5cb-0f6a0c660979", "address": "fa:16:3e:b7:0a:b8", "network": {"id": "23240c8d-f699-4df6-a6ba-8bc3b6db56cf", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-236578753", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.78", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "26a52fb1944a4f949f6779829e163eaa", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ab93899c-92b2-4d84-95a6-192234add28c", "external-id": "nsx-vlan-transportzone-697", "segmentation_id": 697, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapeadc0731-83", "ovs_interfaceid": "eadc0731-8390-4414-b5cb-0f6a0c660979", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 855.897628] env[63515]: DEBUG oslo_concurrency.lockutils [None req-4cdfd447-25bb-42ff-81df-85503f57f660 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Acquiring lock "refresh_cache-d926c699-a64a-4942-9ef4-f0166414661d" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 855.897846] env[63515]: DEBUG oslo_concurrency.lockutils [None req-4cdfd447-25bb-42ff-81df-85503f57f660 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Acquired lock "refresh_cache-d926c699-a64a-4942-9ef4-f0166414661d" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 855.898047] env[63515]: DEBUG nova.network.neutron [None req-4cdfd447-25bb-42ff-81df-85503f57f660 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: d926c699-a64a-4942-9ef4-f0166414661d] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 856.099948] env[63515]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-12df293c-6c3e-4c78-80f7-d54cd36a5938 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.112489] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68a07046-1c69-482e-8256-a15400bca4ff {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.151420] env[63515]: DEBUG nova.compute.manager [req-2509d33b-e6ba-4db0-9707-1f01517ddef2 req-72d50d79-977f-4d44-9de3-c37f1122a369 service nova] [instance: 17275472-2921-49c9-b4e0-5485649ebd1c] Detach interface failed, port_id=5514ac51-9696-4c26-9abc-2cc0deef264e, reason: Instance 17275472-2921-49c9-b4e0-5485649ebd1c could not be found. {{(pid=63515) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 856.153081] env[63515]: DEBUG nova.network.neutron [-] [instance: 17275472-2921-49c9-b4e0-5485649ebd1c] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 856.219149] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc09d145-034d-4bda-9299-fb82da118ce3 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.229105] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d38c8572-6003-40af-a0c3-c29a55cb0aa0 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.266653] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b2b93de-1a4c-40dc-98f0-7659657ba2f4 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.279134] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb9eec84-124d-4575-893d-98b41470a219 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.297624] env[63515]: DEBUG nova.compute.provider_tree [None req-957d8582-d135-4677-8fde-d84fe70e2761 tempest-ServerDiagnosticsV248Test-1736022867 tempest-ServerDiagnosticsV248Test-1736022867-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 856.327792] env[63515]: DEBUG oslo_vmware.rw_handles [None req-ced15047-223c-4e6c-9243-6a752a33316f tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5251b1ed-216a-54ea-f1c3-3e298bc1ae33/disk-0.vmdk. {{(pid=63515) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 856.328750] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-283d402f-35c4-46c8-8f81-b7bcbb7c41be {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.336462] env[63515]: DEBUG oslo_vmware.rw_handles [None req-ced15047-223c-4e6c-9243-6a752a33316f tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5251b1ed-216a-54ea-f1c3-3e298bc1ae33/disk-0.vmdk is in state: ready. {{(pid=63515) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 856.336642] env[63515]: ERROR oslo_vmware.rw_handles [None req-ced15047-223c-4e6c-9243-6a752a33316f tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5251b1ed-216a-54ea-f1c3-3e298bc1ae33/disk-0.vmdk due to incomplete transfer. [ 856.336912] env[63515]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-43b197e7-51b0-433f-8360-e35bd58b09fb {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.346895] env[63515]: DEBUG oslo_vmware.rw_handles [None req-ced15047-223c-4e6c-9243-6a752a33316f tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5251b1ed-216a-54ea-f1c3-3e298bc1ae33/disk-0.vmdk. {{(pid=63515) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 856.347145] env[63515]: DEBUG nova.virt.vmwareapi.images [None req-ced15047-223c-4e6c-9243-6a752a33316f tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: 70df69e5-687b-44fb-b6fc-cdb08e21dda0] Uploaded image d1918637-9f8a-4d89-9bc2-49e7d0fdaeee to the Glance image server {{(pid=63515) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 856.349329] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-ced15047-223c-4e6c-9243-6a752a33316f tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: 70df69e5-687b-44fb-b6fc-cdb08e21dda0] Destroying the VM {{(pid=63515) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 856.349606] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-464fc506-0502-4755-9f16-4e78fa4cc30f {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.357428] env[63515]: DEBUG oslo_vmware.api [None req-ced15047-223c-4e6c-9243-6a752a33316f tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Waiting for the task: (returnval){ [ 856.357428] env[63515]: value = "task-1111395" [ 856.357428] env[63515]: _type = "Task" [ 856.357428] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 856.367015] env[63515]: DEBUG oslo_vmware.api [None req-ced15047-223c-4e6c-9243-6a752a33316f tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Task: {'id': task-1111395, 'name': Destroy_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.659095] env[63515]: INFO nova.compute.manager [-] [instance: 17275472-2921-49c9-b4e0-5485649ebd1c] Took 1.41 seconds to deallocate network for instance. [ 856.663184] env[63515]: DEBUG nova.compute.manager [req-8ef48bb6-0951-467d-b985-60d8f47d9dfc req-1e5d9810-6b12-48aa-8957-4c97d13226c0 service nova] [instance: 87c468d9-9594-4804-b461-527f01f6118f] Received event network-vif-plugged-d2098868-0475-4b06-bae3-785c608a861e {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 856.663574] env[63515]: DEBUG oslo_concurrency.lockutils [req-8ef48bb6-0951-467d-b985-60d8f47d9dfc req-1e5d9810-6b12-48aa-8957-4c97d13226c0 service nova] Acquiring lock "87c468d9-9594-4804-b461-527f01f6118f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 856.664057] env[63515]: DEBUG oslo_concurrency.lockutils [req-8ef48bb6-0951-467d-b985-60d8f47d9dfc req-1e5d9810-6b12-48aa-8957-4c97d13226c0 service nova] Lock "87c468d9-9594-4804-b461-527f01f6118f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 856.664366] env[63515]: DEBUG oslo_concurrency.lockutils [req-8ef48bb6-0951-467d-b985-60d8f47d9dfc req-1e5d9810-6b12-48aa-8957-4c97d13226c0 service nova] Lock "87c468d9-9594-4804-b461-527f01f6118f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 856.664670] env[63515]: DEBUG nova.compute.manager [req-8ef48bb6-0951-467d-b985-60d8f47d9dfc req-1e5d9810-6b12-48aa-8957-4c97d13226c0 service nova] [instance: 87c468d9-9594-4804-b461-527f01f6118f] No waiting events found dispatching network-vif-plugged-d2098868-0475-4b06-bae3-785c608a861e {{(pid=63515) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 856.664975] env[63515]: WARNING nova.compute.manager [req-8ef48bb6-0951-467d-b985-60d8f47d9dfc req-1e5d9810-6b12-48aa-8957-4c97d13226c0 service nova] [instance: 87c468d9-9594-4804-b461-527f01f6118f] Received unexpected event network-vif-plugged-d2098868-0475-4b06-bae3-785c608a861e for instance with vm_state shelved_offloaded and task_state spawning. [ 856.688568] env[63515]: DEBUG nova.network.neutron [None req-4cdfd447-25bb-42ff-81df-85503f57f660 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: d926c699-a64a-4942-9ef4-f0166414661d] Updating instance_info_cache with network_info: [{"id": "fdec91d2-43e8-4d40-9aa9-885a82fd9b42", "address": "fa:16:3e:5c:b7:75", "network": {"id": "15fdaabe-6793-400f-9737-2cdff8f07238", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1175204866-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f9c98e9f6020475490aaa7e76d907ab7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7d377d75-3add-4a15-8691-74b2eb010924", "external-id": "nsx-vlan-transportzone-71", "segmentation_id": 71, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfdec91d2-43", "ovs_interfaceid": "fdec91d2-43e8-4d40-9aa9-885a82fd9b42", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 856.772039] env[63515]: DEBUG oslo_concurrency.lockutils [None req-ce131312-5bf5-4677-8923-2d91405ade5d tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Acquiring lock "refresh_cache-87c468d9-9594-4804-b461-527f01f6118f" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 856.772261] env[63515]: DEBUG oslo_concurrency.lockutils [None req-ce131312-5bf5-4677-8923-2d91405ade5d tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Acquired lock "refresh_cache-87c468d9-9594-4804-b461-527f01f6118f" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 856.772445] env[63515]: DEBUG nova.network.neutron [None req-ce131312-5bf5-4677-8923-2d91405ade5d tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] [instance: 87c468d9-9594-4804-b461-527f01f6118f] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 856.801210] env[63515]: DEBUG nova.scheduler.client.report [None req-957d8582-d135-4677-8fde-d84fe70e2761 tempest-ServerDiagnosticsV248Test-1736022867 tempest-ServerDiagnosticsV248Test-1736022867-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 856.868631] env[63515]: DEBUG oslo_vmware.api [None req-ced15047-223c-4e6c-9243-6a752a33316f tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Task: {'id': task-1111395, 'name': Destroy_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.174774] env[63515]: DEBUG oslo_concurrency.lockutils [None req-dd655d0d-92fe-48b9-b34e-d4edd8542788 tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 857.191134] env[63515]: DEBUG oslo_concurrency.lockutils [None req-4cdfd447-25bb-42ff-81df-85503f57f660 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Releasing lock "refresh_cache-d926c699-a64a-4942-9ef4-f0166414661d" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 857.306253] env[63515]: DEBUG oslo_concurrency.lockutils [None req-957d8582-d135-4677-8fde-d84fe70e2761 tempest-ServerDiagnosticsV248Test-1736022867 tempest-ServerDiagnosticsV248Test-1736022867-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.448s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 857.306825] env[63515]: DEBUG nova.compute.manager [None req-957d8582-d135-4677-8fde-d84fe70e2761 tempest-ServerDiagnosticsV248Test-1736022867 tempest-ServerDiagnosticsV248Test-1736022867-project-member] [instance: 705b7e9b-3ab6-40ec-8cf6-d86bf0a1faf0] Start building networks asynchronously for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 857.309401] env[63515]: DEBUG oslo_concurrency.lockutils [None req-eae7e17b-7393-4e13-b575-a8d38e798918 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 29.520s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 857.310788] env[63515]: INFO nova.compute.claims [None req-eae7e17b-7393-4e13-b575-a8d38e798918 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] [instance: 2a29f37d-e786-48d3-b126-4854d6e7c67c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 857.370274] env[63515]: DEBUG oslo_vmware.api [None req-ced15047-223c-4e6c-9243-6a752a33316f tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Task: {'id': task-1111395, 'name': Destroy_Task, 'duration_secs': 0.835611} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 857.370611] env[63515]: INFO nova.virt.vmwareapi.vm_util [None req-ced15047-223c-4e6c-9243-6a752a33316f tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: 70df69e5-687b-44fb-b6fc-cdb08e21dda0] Destroyed the VM [ 857.370857] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-ced15047-223c-4e6c-9243-6a752a33316f tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: 70df69e5-687b-44fb-b6fc-cdb08e21dda0] Deleting Snapshot of the VM instance {{(pid=63515) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 857.371153] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-bd8b88ef-a646-4792-a81e-2240dec02b0a {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.378334] env[63515]: DEBUG oslo_vmware.api [None req-ced15047-223c-4e6c-9243-6a752a33316f tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Waiting for the task: (returnval){ [ 857.378334] env[63515]: value = "task-1111396" [ 857.378334] env[63515]: _type = "Task" [ 857.378334] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 857.387827] env[63515]: DEBUG oslo_vmware.api [None req-ced15047-223c-4e6c-9243-6a752a33316f tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Task: {'id': task-1111396, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.528096] env[63515]: DEBUG nova.network.neutron [None req-ce131312-5bf5-4677-8923-2d91405ade5d tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] [instance: 87c468d9-9594-4804-b461-527f01f6118f] Updating instance_info_cache with network_info: [{"id": "d2098868-0475-4b06-bae3-785c608a861e", "address": "fa:16:3e:af:f9:f0", "network": {"id": "09fb20ca-2b17-4d46-b915-6aff85b862d8", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1815529427-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.212", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "19ce3b05eec64475b685faad1fd801f6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ded18042-834c-4792-b3e8-b1c377446432", "external-id": "nsx-vlan-transportzone-293", "segmentation_id": 293, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd2098868-04", "ovs_interfaceid": "d2098868-0475-4b06-bae3-785c608a861e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 857.576143] env[63515]: DEBUG oslo_concurrency.lockutils [None req-d7e3c195-4890-454e-9f6b-af1a7cc4ee6a tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Acquiring lock "ae5d6f34-bac6-4ba0-b0c3-8205982ddc88" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 857.576549] env[63515]: DEBUG oslo_concurrency.lockutils [None req-d7e3c195-4890-454e-9f6b-af1a7cc4ee6a tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Lock "ae5d6f34-bac6-4ba0-b0c3-8205982ddc88" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 857.627318] env[63515]: DEBUG nova.compute.manager [req-ecb05fb3-ff0a-4708-a6cb-112cfb1eb093 req-639691f9-1f51-4920-8957-4f81101ba29f service nova] [instance: 17275472-2921-49c9-b4e0-5485649ebd1c] Received event network-vif-deleted-db47ccfa-8557-4128-9e65-8a8ab57163bb {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 857.627535] env[63515]: DEBUG nova.compute.manager [req-ecb05fb3-ff0a-4708-a6cb-112cfb1eb093 req-639691f9-1f51-4920-8957-4f81101ba29f service nova] [instance: 17275472-2921-49c9-b4e0-5485649ebd1c] Received event network-vif-deleted-eadc0731-8390-4414-b5cb-0f6a0c660979 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 857.721096] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52a5adda-0eb2-4057-9ba3-8f0eb591672a {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.741171] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19d9614f-24bf-49ae-b5e6-d219247f5869 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.749026] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-4cdfd447-25bb-42ff-81df-85503f57f660 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: d926c699-a64a-4942-9ef4-f0166414661d] Updating instance 'd926c699-a64a-4942-9ef4-f0166414661d' progress to 83 {{(pid=63515) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 857.815330] env[63515]: DEBUG nova.compute.utils [None req-957d8582-d135-4677-8fde-d84fe70e2761 tempest-ServerDiagnosticsV248Test-1736022867 tempest-ServerDiagnosticsV248Test-1736022867-project-member] Using /dev/sd instead of None {{(pid=63515) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 857.819645] env[63515]: DEBUG nova.compute.manager [None req-957d8582-d135-4677-8fde-d84fe70e2761 tempest-ServerDiagnosticsV248Test-1736022867 tempest-ServerDiagnosticsV248Test-1736022867-project-member] [instance: 705b7e9b-3ab6-40ec-8cf6-d86bf0a1faf0] Not allocating networking since 'none' was specified. {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 857.888941] env[63515]: DEBUG oslo_vmware.api [None req-ced15047-223c-4e6c-9243-6a752a33316f tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Task: {'id': task-1111396, 'name': RemoveSnapshot_Task} progress is 97%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.034197] env[63515]: DEBUG oslo_concurrency.lockutils [None req-ce131312-5bf5-4677-8923-2d91405ade5d tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Releasing lock "refresh_cache-87c468d9-9594-4804-b461-527f01f6118f" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 858.059576] env[63515]: DEBUG nova.virt.hardware [None req-ce131312-5bf5-4677-8923-2d91405ade5d tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T02:52:14Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='ea0db1064ba9f8c580d546c68498c4aa',container_format='bare',created_at=2024-10-03T02:57:43Z,direct_url=,disk_format='vmdk',id=499b6029-07e9-499c-80d5-a53d8c43ffb7,min_disk=1,min_ram=0,name='tempest-AttachVolumeShelveTestJSON-server-339005594-shelved',owner='19ce3b05eec64475b685faad1fd801f6',properties=ImageMetaProps,protected=,size=31668736,status='active',tags=,updated_at=2024-10-03T02:58:01Z,virtual_size=,visibility=), allow threads: False {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 858.059912] env[63515]: DEBUG nova.virt.hardware [None req-ce131312-5bf5-4677-8923-2d91405ade5d tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Flavor limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 858.060101] env[63515]: DEBUG nova.virt.hardware [None req-ce131312-5bf5-4677-8923-2d91405ade5d tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Image limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 858.060290] env[63515]: DEBUG nova.virt.hardware [None req-ce131312-5bf5-4677-8923-2d91405ade5d tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Flavor pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 858.060441] env[63515]: DEBUG nova.virt.hardware [None req-ce131312-5bf5-4677-8923-2d91405ade5d tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Image pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 858.060589] env[63515]: DEBUG nova.virt.hardware [None req-ce131312-5bf5-4677-8923-2d91405ade5d tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 858.060797] env[63515]: DEBUG nova.virt.hardware [None req-ce131312-5bf5-4677-8923-2d91405ade5d tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 858.060963] env[63515]: DEBUG nova.virt.hardware [None req-ce131312-5bf5-4677-8923-2d91405ade5d tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 858.061193] env[63515]: DEBUG nova.virt.hardware [None req-ce131312-5bf5-4677-8923-2d91405ade5d tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Got 1 possible topologies {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 858.061391] env[63515]: DEBUG nova.virt.hardware [None req-ce131312-5bf5-4677-8923-2d91405ade5d tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 858.061599] env[63515]: DEBUG nova.virt.hardware [None req-ce131312-5bf5-4677-8923-2d91405ade5d tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 858.062466] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd0c61dc-6666-46dd-a496-1343e229a5e1 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.070851] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e521ee7-dfb5-4dc1-b144-f44d1e1d3269 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.085123] env[63515]: DEBUG nova.compute.manager [None req-d7e3c195-4890-454e-9f6b-af1a7cc4ee6a tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: ae5d6f34-bac6-4ba0-b0c3-8205982ddc88] Starting instance... {{(pid=63515) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 858.088816] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-ce131312-5bf5-4677-8923-2d91405ade5d tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] [instance: 87c468d9-9594-4804-b461-527f01f6118f] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:af:f9:f0', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ded18042-834c-4792-b3e8-b1c377446432', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'd2098868-0475-4b06-bae3-785c608a861e', 'vif_model': 'vmxnet3'}] {{(pid=63515) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 858.096414] env[63515]: DEBUG oslo.service.loopingcall [None req-ce131312-5bf5-4677-8923-2d91405ade5d tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 858.096885] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 87c468d9-9594-4804-b461-527f01f6118f] Creating VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 858.097129] env[63515]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-fed2d9f2-0f8e-4a3d-b0b1-4889f8e847fa {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.117412] env[63515]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 858.117412] env[63515]: value = "task-1111397" [ 858.117412] env[63515]: _type = "Task" [ 858.117412] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 858.126103] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1111397, 'name': CreateVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.255596] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-4cdfd447-25bb-42ff-81df-85503f57f660 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: d926c699-a64a-4942-9ef4-f0166414661d] Powering on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 858.255596] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d455a96f-d519-4438-abc0-05d8c01b49de {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.261376] env[63515]: DEBUG oslo_concurrency.lockutils [None req-9a80c171-6c9b-44ae-ab0a-573a8347fcf4 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Acquiring lock "b1bf5a0a-42e7-41a6-90fb-a0c311623fdd" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 858.261607] env[63515]: DEBUG oslo_concurrency.lockutils [None req-9a80c171-6c9b-44ae-ab0a-573a8347fcf4 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Lock "b1bf5a0a-42e7-41a6-90fb-a0c311623fdd" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 858.267593] env[63515]: DEBUG oslo_vmware.api [None req-4cdfd447-25bb-42ff-81df-85503f57f660 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Waiting for the task: (returnval){ [ 858.267593] env[63515]: value = "task-1111398" [ 858.267593] env[63515]: _type = "Task" [ 858.267593] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 858.275831] env[63515]: DEBUG oslo_vmware.api [None req-4cdfd447-25bb-42ff-81df-85503f57f660 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Task: {'id': task-1111398, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.323767] env[63515]: DEBUG nova.compute.manager [None req-957d8582-d135-4677-8fde-d84fe70e2761 tempest-ServerDiagnosticsV248Test-1736022867 tempest-ServerDiagnosticsV248Test-1736022867-project-member] [instance: 705b7e9b-3ab6-40ec-8cf6-d86bf0a1faf0] Start building block device mappings for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 858.389471] env[63515]: DEBUG oslo_vmware.api [None req-ced15047-223c-4e6c-9243-6a752a33316f tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Task: {'id': task-1111396, 'name': RemoveSnapshot_Task, 'duration_secs': 0.600217} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 858.392456] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-ced15047-223c-4e6c-9243-6a752a33316f tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: 70df69e5-687b-44fb-b6fc-cdb08e21dda0] Deleted Snapshot of the VM instance {{(pid=63515) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 858.392950] env[63515]: INFO nova.compute.manager [None req-ced15047-223c-4e6c-9243-6a752a33316f tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: 70df69e5-687b-44fb-b6fc-cdb08e21dda0] Took 17.21 seconds to snapshot the instance on the hypervisor. [ 858.612171] env[63515]: DEBUG oslo_concurrency.lockutils [None req-d7e3c195-4890-454e-9f6b-af1a7cc4ee6a tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 858.613962] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d68b9142-d92c-49aa-adb5-411276460439 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.624740] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65b7cf98-cdd3-4620-83e6-febcf1557697 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.633360] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1111397, 'name': CreateVM_Task, 'duration_secs': 0.356592} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 858.656808] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 87c468d9-9594-4804-b461-527f01f6118f] Created VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 858.657821] env[63515]: DEBUG oslo_concurrency.lockutils [None req-ce131312-5bf5-4677-8923-2d91405ade5d tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/499b6029-07e9-499c-80d5-a53d8c43ffb7" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 858.658014] env[63515]: DEBUG oslo_concurrency.lockutils [None req-ce131312-5bf5-4677-8923-2d91405ade5d tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Acquired lock "[datastore1] devstack-image-cache_base/499b6029-07e9-499c-80d5-a53d8c43ffb7" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 858.658413] env[63515]: DEBUG oslo_concurrency.lockutils [None req-ce131312-5bf5-4677-8923-2d91405ade5d tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/499b6029-07e9-499c-80d5-a53d8c43ffb7" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 858.659164] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d51b6e5a-59f4-4820-be19-757b756f9c18 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.661690] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c8b4811b-6523-40fd-b5a6-06d01b00ae4f {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.666653] env[63515]: DEBUG oslo_vmware.api [None req-ce131312-5bf5-4677-8923-2d91405ade5d tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Waiting for the task: (returnval){ [ 858.666653] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]52f326e9-e7ee-fcc6-dfd0-5e665f24f431" [ 858.666653] env[63515]: _type = "Task" [ 858.666653] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 858.674058] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1aae4ce6-121e-4f94-bf73-60855e956145 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.690305] env[63515]: DEBUG oslo_concurrency.lockutils [None req-ce131312-5bf5-4677-8923-2d91405ade5d tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Releasing lock "[datastore1] devstack-image-cache_base/499b6029-07e9-499c-80d5-a53d8c43ffb7" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 858.690572] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-ce131312-5bf5-4677-8923-2d91405ade5d tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] [instance: 87c468d9-9594-4804-b461-527f01f6118f] Processing image 499b6029-07e9-499c-80d5-a53d8c43ffb7 {{(pid=63515) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 858.690830] env[63515]: DEBUG oslo_concurrency.lockutils [None req-ce131312-5bf5-4677-8923-2d91405ade5d tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/499b6029-07e9-499c-80d5-a53d8c43ffb7/499b6029-07e9-499c-80d5-a53d8c43ffb7.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 858.690964] env[63515]: DEBUG oslo_concurrency.lockutils [None req-ce131312-5bf5-4677-8923-2d91405ade5d tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Acquired lock "[datastore1] devstack-image-cache_base/499b6029-07e9-499c-80d5-a53d8c43ffb7/499b6029-07e9-499c-80d5-a53d8c43ffb7.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 858.691166] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-ce131312-5bf5-4677-8923-2d91405ade5d tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 858.691941] env[63515]: DEBUG nova.compute.provider_tree [None req-eae7e17b-7393-4e13-b575-a8d38e798918 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 858.693898] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c277f6fb-9d40-41a7-b71a-c5c13d97151e {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.696706] env[63515]: DEBUG nova.compute.manager [req-f492aea0-619f-4423-bab1-a088d3be593b req-65c52027-aed7-4767-b8c3-d2d55fb7dda3 service nova] [instance: 87c468d9-9594-4804-b461-527f01f6118f] Received event network-changed-d2098868-0475-4b06-bae3-785c608a861e {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 858.696929] env[63515]: DEBUG nova.compute.manager [req-f492aea0-619f-4423-bab1-a088d3be593b req-65c52027-aed7-4767-b8c3-d2d55fb7dda3 service nova] [instance: 87c468d9-9594-4804-b461-527f01f6118f] Refreshing instance network info cache due to event network-changed-d2098868-0475-4b06-bae3-785c608a861e. {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 858.697154] env[63515]: DEBUG oslo_concurrency.lockutils [req-f492aea0-619f-4423-bab1-a088d3be593b req-65c52027-aed7-4767-b8c3-d2d55fb7dda3 service nova] Acquiring lock "refresh_cache-87c468d9-9594-4804-b461-527f01f6118f" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 858.697369] env[63515]: DEBUG oslo_concurrency.lockutils [req-f492aea0-619f-4423-bab1-a088d3be593b req-65c52027-aed7-4767-b8c3-d2d55fb7dda3 service nova] Acquired lock "refresh_cache-87c468d9-9594-4804-b461-527f01f6118f" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 858.697462] env[63515]: DEBUG nova.network.neutron [req-f492aea0-619f-4423-bab1-a088d3be593b req-65c52027-aed7-4767-b8c3-d2d55fb7dda3 service nova] [instance: 87c468d9-9594-4804-b461-527f01f6118f] Refreshing network info cache for port d2098868-0475-4b06-bae3-785c608a861e {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 858.708825] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-ce131312-5bf5-4677-8923-2d91405ade5d tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 858.709040] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-ce131312-5bf5-4677-8923-2d91405ade5d tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63515) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 858.709838] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4050b332-0adb-4d94-9000-a956556a6bf4 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.716387] env[63515]: DEBUG oslo_vmware.api [None req-ce131312-5bf5-4677-8923-2d91405ade5d tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Waiting for the task: (returnval){ [ 858.716387] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]52760b44-17f1-5cdb-3ac7-e216b479d4a2" [ 858.716387] env[63515]: _type = "Task" [ 858.716387] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 858.725444] env[63515]: DEBUG oslo_vmware.api [None req-ce131312-5bf5-4677-8923-2d91405ade5d tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52760b44-17f1-5cdb-3ac7-e216b479d4a2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.764214] env[63515]: DEBUG nova.compute.manager [None req-9a80c171-6c9b-44ae-ab0a-573a8347fcf4 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: b1bf5a0a-42e7-41a6-90fb-a0c311623fdd] Starting instance... {{(pid=63515) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 858.777441] env[63515]: DEBUG oslo_vmware.api [None req-4cdfd447-25bb-42ff-81df-85503f57f660 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Task: {'id': task-1111398, 'name': PowerOnVM_Task, 'duration_secs': 0.395638} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 858.777822] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-4cdfd447-25bb-42ff-81df-85503f57f660 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: d926c699-a64a-4942-9ef4-f0166414661d] Powered on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 858.777999] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-4cdfd447-25bb-42ff-81df-85503f57f660 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: d926c699-a64a-4942-9ef4-f0166414661d] Updating instance 'd926c699-a64a-4942-9ef4-f0166414661d' progress to 100 {{(pid=63515) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 859.200370] env[63515]: DEBUG nova.scheduler.client.report [None req-eae7e17b-7393-4e13-b575-a8d38e798918 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 859.228800] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-ce131312-5bf5-4677-8923-2d91405ade5d tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] [instance: 87c468d9-9594-4804-b461-527f01f6118f] Preparing fetch location {{(pid=63515) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 859.229184] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-ce131312-5bf5-4677-8923-2d91405ade5d tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] [instance: 87c468d9-9594-4804-b461-527f01f6118f] Fetch image to [datastore1] OSTACK_IMG_d97e85a2-748d-4e52-8257-3fd60729ce2e/OSTACK_IMG_d97e85a2-748d-4e52-8257-3fd60729ce2e.vmdk {{(pid=63515) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 859.229293] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-ce131312-5bf5-4677-8923-2d91405ade5d tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] [instance: 87c468d9-9594-4804-b461-527f01f6118f] Downloading stream optimized image 499b6029-07e9-499c-80d5-a53d8c43ffb7 to [datastore1] OSTACK_IMG_d97e85a2-748d-4e52-8257-3fd60729ce2e/OSTACK_IMG_d97e85a2-748d-4e52-8257-3fd60729ce2e.vmdk on the data store datastore1 as vApp {{(pid=63515) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 859.229571] env[63515]: DEBUG nova.virt.vmwareapi.images [None req-ce131312-5bf5-4677-8923-2d91405ade5d tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] [instance: 87c468d9-9594-4804-b461-527f01f6118f] Downloading image file data 499b6029-07e9-499c-80d5-a53d8c43ffb7 to the ESX as VM named 'OSTACK_IMG_d97e85a2-748d-4e52-8257-3fd60729ce2e' {{(pid=63515) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 859.289345] env[63515]: DEBUG oslo_concurrency.lockutils [None req-9a80c171-6c9b-44ae-ab0a-573a8347fcf4 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 859.315688] env[63515]: DEBUG oslo_vmware.rw_handles [None req-ce131312-5bf5-4677-8923-2d91405ade5d tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 859.315688] env[63515]: value = "resgroup-9" [ 859.315688] env[63515]: _type = "ResourcePool" [ 859.315688] env[63515]: }. {{(pid=63515) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 859.315922] env[63515]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-4d5c855e-954f-4150-a410-0d4e59b1143d {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.335270] env[63515]: DEBUG nova.compute.manager [None req-957d8582-d135-4677-8fde-d84fe70e2761 tempest-ServerDiagnosticsV248Test-1736022867 tempest-ServerDiagnosticsV248Test-1736022867-project-member] [instance: 705b7e9b-3ab6-40ec-8cf6-d86bf0a1faf0] Start spawning the instance on the hypervisor. {{(pid=63515) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 859.339380] env[63515]: DEBUG oslo_vmware.rw_handles [None req-ce131312-5bf5-4677-8923-2d91405ade5d tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Lease: (returnval){ [ 859.339380] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]52560a7e-9cc3-33f9-9767-cc1bc213e35c" [ 859.339380] env[63515]: _type = "HttpNfcLease" [ 859.339380] env[63515]: } obtained for vApp import into resource pool (val){ [ 859.339380] env[63515]: value = "resgroup-9" [ 859.339380] env[63515]: _type = "ResourcePool" [ 859.339380] env[63515]: }. {{(pid=63515) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 859.339622] env[63515]: DEBUG oslo_vmware.api [None req-ce131312-5bf5-4677-8923-2d91405ade5d tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Waiting for the lease: (returnval){ [ 859.339622] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]52560a7e-9cc3-33f9-9767-cc1bc213e35c" [ 859.339622] env[63515]: _type = "HttpNfcLease" [ 859.339622] env[63515]: } to be ready. {{(pid=63515) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 859.348374] env[63515]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 859.348374] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]52560a7e-9cc3-33f9-9767-cc1bc213e35c" [ 859.348374] env[63515]: _type = "HttpNfcLease" [ 859.348374] env[63515]: } is initializing. {{(pid=63515) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 859.357461] env[63515]: DEBUG nova.virt.hardware [None req-957d8582-d135-4677-8fde-d84fe70e2761 tempest-ServerDiagnosticsV248Test-1736022867 tempest-ServerDiagnosticsV248Test-1736022867-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T02:52:14Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T02:51:56Z,direct_url=,disk_format='vmdk',id=8a120570-cb06-4099-b262-554ca0ad15c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b569255cc43e42e7ae3f2b4ad37c6ef4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T02:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 859.357706] env[63515]: DEBUG nova.virt.hardware [None req-957d8582-d135-4677-8fde-d84fe70e2761 tempest-ServerDiagnosticsV248Test-1736022867 tempest-ServerDiagnosticsV248Test-1736022867-project-member] Flavor limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 859.357869] env[63515]: DEBUG nova.virt.hardware [None req-957d8582-d135-4677-8fde-d84fe70e2761 tempest-ServerDiagnosticsV248Test-1736022867 tempest-ServerDiagnosticsV248Test-1736022867-project-member] Image limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 859.358097] env[63515]: DEBUG nova.virt.hardware [None req-957d8582-d135-4677-8fde-d84fe70e2761 tempest-ServerDiagnosticsV248Test-1736022867 tempest-ServerDiagnosticsV248Test-1736022867-project-member] Flavor pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 859.358298] env[63515]: DEBUG nova.virt.hardware [None req-957d8582-d135-4677-8fde-d84fe70e2761 tempest-ServerDiagnosticsV248Test-1736022867 tempest-ServerDiagnosticsV248Test-1736022867-project-member] Image pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 859.358493] env[63515]: DEBUG nova.virt.hardware [None req-957d8582-d135-4677-8fde-d84fe70e2761 tempest-ServerDiagnosticsV248Test-1736022867 tempest-ServerDiagnosticsV248Test-1736022867-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 859.358709] env[63515]: DEBUG nova.virt.hardware [None req-957d8582-d135-4677-8fde-d84fe70e2761 tempest-ServerDiagnosticsV248Test-1736022867 tempest-ServerDiagnosticsV248Test-1736022867-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 859.358875] env[63515]: DEBUG nova.virt.hardware [None req-957d8582-d135-4677-8fde-d84fe70e2761 tempest-ServerDiagnosticsV248Test-1736022867 tempest-ServerDiagnosticsV248Test-1736022867-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 859.359061] env[63515]: DEBUG nova.virt.hardware [None req-957d8582-d135-4677-8fde-d84fe70e2761 tempest-ServerDiagnosticsV248Test-1736022867 tempest-ServerDiagnosticsV248Test-1736022867-project-member] Got 1 possible topologies {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 859.359233] env[63515]: DEBUG nova.virt.hardware [None req-957d8582-d135-4677-8fde-d84fe70e2761 tempest-ServerDiagnosticsV248Test-1736022867 tempest-ServerDiagnosticsV248Test-1736022867-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 859.359406] env[63515]: DEBUG nova.virt.hardware [None req-957d8582-d135-4677-8fde-d84fe70e2761 tempest-ServerDiagnosticsV248Test-1736022867 tempest-ServerDiagnosticsV248Test-1736022867-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 859.360469] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bfe98590-c6a5-477b-904e-48139b682c49 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.368344] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b96b8dd-e98f-4335-b86a-ec3f2257d495 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.382247] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-957d8582-d135-4677-8fde-d84fe70e2761 tempest-ServerDiagnosticsV248Test-1736022867 tempest-ServerDiagnosticsV248Test-1736022867-project-member] [instance: 705b7e9b-3ab6-40ec-8cf6-d86bf0a1faf0] Instance VIF info [] {{(pid=63515) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 859.387641] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-957d8582-d135-4677-8fde-d84fe70e2761 tempest-ServerDiagnosticsV248Test-1736022867 tempest-ServerDiagnosticsV248Test-1736022867-project-member] Creating folder: Project (5344912cbf724502b77ef4f58fb790b1). Parent ref: group-v243370. {{(pid=63515) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 859.390495] env[63515]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-7e104d35-22fc-49eb-8086-995fb921c446 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.400178] env[63515]: INFO nova.virt.vmwareapi.vm_util [None req-957d8582-d135-4677-8fde-d84fe70e2761 tempest-ServerDiagnosticsV248Test-1736022867 tempest-ServerDiagnosticsV248Test-1736022867-project-member] Created folder: Project (5344912cbf724502b77ef4f58fb790b1) in parent group-v243370. [ 859.400396] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-957d8582-d135-4677-8fde-d84fe70e2761 tempest-ServerDiagnosticsV248Test-1736022867 tempest-ServerDiagnosticsV248Test-1736022867-project-member] Creating folder: Instances. Parent ref: group-v243495. {{(pid=63515) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 859.400870] env[63515]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d0074815-f898-4721-9bf9-549ad3ee71fd {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.411024] env[63515]: INFO nova.virt.vmwareapi.vm_util [None req-957d8582-d135-4677-8fde-d84fe70e2761 tempest-ServerDiagnosticsV248Test-1736022867 tempest-ServerDiagnosticsV248Test-1736022867-project-member] Created folder: Instances in parent group-v243495. [ 859.411285] env[63515]: DEBUG oslo.service.loopingcall [None req-957d8582-d135-4677-8fde-d84fe70e2761 tempest-ServerDiagnosticsV248Test-1736022867 tempest-ServerDiagnosticsV248Test-1736022867-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 859.411487] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 705b7e9b-3ab6-40ec-8cf6-d86bf0a1faf0] Creating VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 859.411696] env[63515]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-416a0185-bd28-4f20-bfc0-dd966c2ab3f7 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.427676] env[63515]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 859.427676] env[63515]: value = "task-1111402" [ 859.427676] env[63515]: _type = "Task" [ 859.427676] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.435421] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1111402, 'name': CreateVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.475539] env[63515]: DEBUG nova.network.neutron [req-f492aea0-619f-4423-bab1-a088d3be593b req-65c52027-aed7-4767-b8c3-d2d55fb7dda3 service nova] [instance: 87c468d9-9594-4804-b461-527f01f6118f] Updated VIF entry in instance network info cache for port d2098868-0475-4b06-bae3-785c608a861e. {{(pid=63515) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 859.476015] env[63515]: DEBUG nova.network.neutron [req-f492aea0-619f-4423-bab1-a088d3be593b req-65c52027-aed7-4767-b8c3-d2d55fb7dda3 service nova] [instance: 87c468d9-9594-4804-b461-527f01f6118f] Updating instance_info_cache with network_info: [{"id": "d2098868-0475-4b06-bae3-785c608a861e", "address": "fa:16:3e:af:f9:f0", "network": {"id": "09fb20ca-2b17-4d46-b915-6aff85b862d8", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1815529427-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.212", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "19ce3b05eec64475b685faad1fd801f6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ded18042-834c-4792-b3e8-b1c377446432", "external-id": "nsx-vlan-transportzone-293", "segmentation_id": 293, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd2098868-04", "ovs_interfaceid": "d2098868-0475-4b06-bae3-785c608a861e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 859.707389] env[63515]: DEBUG oslo_concurrency.lockutils [None req-eae7e17b-7393-4e13-b575-a8d38e798918 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.397s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 859.707877] env[63515]: DEBUG nova.compute.manager [None req-eae7e17b-7393-4e13-b575-a8d38e798918 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] [instance: 2a29f37d-e786-48d3-b126-4854d6e7c67c] Start building networks asynchronously for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 859.711290] env[63515]: DEBUG oslo_concurrency.lockutils [None req-baaf7dbf-037b-4d17-a0eb-d252e746ead6 tempest-InstanceActionsTestJSON-247065106 tempest-InstanceActionsTestJSON-247065106-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 31.466s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 859.713524] env[63515]: INFO nova.compute.claims [None req-baaf7dbf-037b-4d17-a0eb-d252e746ead6 tempest-InstanceActionsTestJSON-247065106 tempest-InstanceActionsTestJSON-247065106-project-member] [instance: 0ee9cce2-94b9-4038-ab38-1b3f96a864bc] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 859.848734] env[63515]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 859.848734] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]52560a7e-9cc3-33f9-9767-cc1bc213e35c" [ 859.848734] env[63515]: _type = "HttpNfcLease" [ 859.848734] env[63515]: } is ready. {{(pid=63515) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 859.849125] env[63515]: DEBUG oslo_vmware.rw_handles [None req-ce131312-5bf5-4677-8923-2d91405ade5d tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 859.849125] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]52560a7e-9cc3-33f9-9767-cc1bc213e35c" [ 859.849125] env[63515]: _type = "HttpNfcLease" [ 859.849125] env[63515]: }. {{(pid=63515) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 859.849791] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65cfe054-0a1f-4afe-98e1-28d475b775e2 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.857733] env[63515]: DEBUG oslo_vmware.rw_handles [None req-ce131312-5bf5-4677-8923-2d91405ade5d tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/520d7a2f-331f-4a3e-2aae-6efa230e5397/disk-0.vmdk from lease info. {{(pid=63515) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 859.857733] env[63515]: DEBUG oslo_vmware.rw_handles [None req-ce131312-5bf5-4677-8923-2d91405ade5d tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Creating HTTP connection to write to file with size = 31668736 and URL = https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/520d7a2f-331f-4a3e-2aae-6efa230e5397/disk-0.vmdk. {{(pid=63515) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 859.922209] env[63515]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-4d58b533-ebea-40db-a5ab-f9de857c6579 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.936666] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1111402, 'name': CreateVM_Task, 'duration_secs': 0.255459} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 859.936905] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 705b7e9b-3ab6-40ec-8cf6-d86bf0a1faf0] Created VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 859.937361] env[63515]: DEBUG oslo_concurrency.lockutils [None req-957d8582-d135-4677-8fde-d84fe70e2761 tempest-ServerDiagnosticsV248Test-1736022867 tempest-ServerDiagnosticsV248Test-1736022867-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 859.937539] env[63515]: DEBUG oslo_concurrency.lockutils [None req-957d8582-d135-4677-8fde-d84fe70e2761 tempest-ServerDiagnosticsV248Test-1736022867 tempest-ServerDiagnosticsV248Test-1736022867-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 859.937865] env[63515]: DEBUG oslo_concurrency.lockutils [None req-957d8582-d135-4677-8fde-d84fe70e2761 tempest-ServerDiagnosticsV248Test-1736022867 tempest-ServerDiagnosticsV248Test-1736022867-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 859.938135] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cbf19cae-abd6-4a3d-b44b-8fcf82459211 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.942156] env[63515]: DEBUG oslo_vmware.api [None req-957d8582-d135-4677-8fde-d84fe70e2761 tempest-ServerDiagnosticsV248Test-1736022867 tempest-ServerDiagnosticsV248Test-1736022867-project-member] Waiting for the task: (returnval){ [ 859.942156] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]52016d32-f1c6-aa6c-4644-3341064866a7" [ 859.942156] env[63515]: _type = "Task" [ 859.942156] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.949736] env[63515]: DEBUG oslo_vmware.api [None req-957d8582-d135-4677-8fde-d84fe70e2761 tempest-ServerDiagnosticsV248Test-1736022867 tempest-ServerDiagnosticsV248Test-1736022867-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52016d32-f1c6-aa6c-4644-3341064866a7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.978706] env[63515]: DEBUG oslo_concurrency.lockutils [req-f492aea0-619f-4423-bab1-a088d3be593b req-65c52027-aed7-4767-b8c3-d2d55fb7dda3 service nova] Releasing lock "refresh_cache-87c468d9-9594-4804-b461-527f01f6118f" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 860.218095] env[63515]: DEBUG nova.compute.utils [None req-eae7e17b-7393-4e13-b575-a8d38e798918 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Using /dev/sd instead of None {{(pid=63515) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 860.224589] env[63515]: DEBUG nova.compute.manager [None req-eae7e17b-7393-4e13-b575-a8d38e798918 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] [instance: 2a29f37d-e786-48d3-b126-4854d6e7c67c] Allocating IP information in the background. {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 860.224589] env[63515]: DEBUG nova.network.neutron [None req-eae7e17b-7393-4e13-b575-a8d38e798918 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] [instance: 2a29f37d-e786-48d3-b126-4854d6e7c67c] allocate_for_instance() {{(pid=63515) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 860.287132] env[63515]: DEBUG nova.policy [None req-eae7e17b-7393-4e13-b575-a8d38e798918 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2b86a44d6b814d2baeb7e7eb87f12361', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '08c9aa51b1864410809f8877c7cd49c4', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63515) authorize /opt/stack/nova/nova/policy.py:201}} [ 860.453137] env[63515]: DEBUG oslo_vmware.api [None req-957d8582-d135-4677-8fde-d84fe70e2761 tempest-ServerDiagnosticsV248Test-1736022867 tempest-ServerDiagnosticsV248Test-1736022867-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52016d32-f1c6-aa6c-4644-3341064866a7, 'name': SearchDatastore_Task, 'duration_secs': 0.008357} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.453543] env[63515]: DEBUG oslo_concurrency.lockutils [None req-957d8582-d135-4677-8fde-d84fe70e2761 tempest-ServerDiagnosticsV248Test-1736022867 tempest-ServerDiagnosticsV248Test-1736022867-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 860.454218] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-957d8582-d135-4677-8fde-d84fe70e2761 tempest-ServerDiagnosticsV248Test-1736022867 tempest-ServerDiagnosticsV248Test-1736022867-project-member] [instance: 705b7e9b-3ab6-40ec-8cf6-d86bf0a1faf0] Processing image 8a120570-cb06-4099-b262-554ca0ad15c5 {{(pid=63515) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 860.454218] env[63515]: DEBUG oslo_concurrency.lockutils [None req-957d8582-d135-4677-8fde-d84fe70e2761 tempest-ServerDiagnosticsV248Test-1736022867 tempest-ServerDiagnosticsV248Test-1736022867-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 860.454319] env[63515]: DEBUG oslo_concurrency.lockutils [None req-957d8582-d135-4677-8fde-d84fe70e2761 tempest-ServerDiagnosticsV248Test-1736022867 tempest-ServerDiagnosticsV248Test-1736022867-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 860.454524] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-957d8582-d135-4677-8fde-d84fe70e2761 tempest-ServerDiagnosticsV248Test-1736022867 tempest-ServerDiagnosticsV248Test-1736022867-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 860.454885] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a110a33c-1ea9-475a-b6d1-c2cccc966857 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.465956] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-957d8582-d135-4677-8fde-d84fe70e2761 tempest-ServerDiagnosticsV248Test-1736022867 tempest-ServerDiagnosticsV248Test-1736022867-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 860.466276] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-957d8582-d135-4677-8fde-d84fe70e2761 tempest-ServerDiagnosticsV248Test-1736022867 tempest-ServerDiagnosticsV248Test-1736022867-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63515) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 860.467055] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cfd57aac-8621-4f21-8a93-790553c09251 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.481199] env[63515]: DEBUG oslo_vmware.api [None req-957d8582-d135-4677-8fde-d84fe70e2761 tempest-ServerDiagnosticsV248Test-1736022867 tempest-ServerDiagnosticsV248Test-1736022867-project-member] Waiting for the task: (returnval){ [ 860.481199] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]5235f2b9-2721-350a-2abb-60083d763f23" [ 860.481199] env[63515]: _type = "Task" [ 860.481199] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.498039] env[63515]: DEBUG oslo_vmware.api [None req-957d8582-d135-4677-8fde-d84fe70e2761 tempest-ServerDiagnosticsV248Test-1736022867 tempest-ServerDiagnosticsV248Test-1736022867-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]5235f2b9-2721-350a-2abb-60083d763f23, 'name': SearchDatastore_Task, 'duration_secs': 0.014353} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.500440] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-230e1c3d-95ac-4d0e-baf3-f2d2f1422088 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.507239] env[63515]: DEBUG oslo_vmware.api [None req-957d8582-d135-4677-8fde-d84fe70e2761 tempest-ServerDiagnosticsV248Test-1736022867 tempest-ServerDiagnosticsV248Test-1736022867-project-member] Waiting for the task: (returnval){ [ 860.507239] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]52019cdf-a3fb-d5b0-f9e1-75ab9e0d3657" [ 860.507239] env[63515]: _type = "Task" [ 860.507239] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.517444] env[63515]: DEBUG oslo_vmware.api [None req-957d8582-d135-4677-8fde-d84fe70e2761 tempest-ServerDiagnosticsV248Test-1736022867 tempest-ServerDiagnosticsV248Test-1736022867-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52019cdf-a3fb-d5b0-f9e1-75ab9e0d3657, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.606924] env[63515]: DEBUG nova.network.neutron [None req-eae7e17b-7393-4e13-b575-a8d38e798918 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] [instance: 2a29f37d-e786-48d3-b126-4854d6e7c67c] Successfully created port: 956cd361-0d42-4914-85d3-5f75a72a68a1 {{(pid=63515) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 860.731798] env[63515]: DEBUG nova.compute.manager [None req-eae7e17b-7393-4e13-b575-a8d38e798918 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] [instance: 2a29f37d-e786-48d3-b126-4854d6e7c67c] Start building block device mappings for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 861.024032] env[63515]: DEBUG oslo_vmware.api [None req-957d8582-d135-4677-8fde-d84fe70e2761 tempest-ServerDiagnosticsV248Test-1736022867 tempest-ServerDiagnosticsV248Test-1736022867-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52019cdf-a3fb-d5b0-f9e1-75ab9e0d3657, 'name': SearchDatastore_Task, 'duration_secs': 0.012549} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.028575] env[63515]: DEBUG oslo_concurrency.lockutils [None req-957d8582-d135-4677-8fde-d84fe70e2761 tempest-ServerDiagnosticsV248Test-1736022867 tempest-ServerDiagnosticsV248Test-1736022867-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 861.028780] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-957d8582-d135-4677-8fde-d84fe70e2761 tempest-ServerDiagnosticsV248Test-1736022867 tempest-ServerDiagnosticsV248Test-1736022867-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk to [datastore1] 705b7e9b-3ab6-40ec-8cf6-d86bf0a1faf0/705b7e9b-3ab6-40ec-8cf6-d86bf0a1faf0.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 861.029363] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-555b01ea-4da8-4e4c-bb00-9a7e2c97c45b {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.044670] env[63515]: DEBUG oslo_vmware.api [None req-957d8582-d135-4677-8fde-d84fe70e2761 tempest-ServerDiagnosticsV248Test-1736022867 tempest-ServerDiagnosticsV248Test-1736022867-project-member] Waiting for the task: (returnval){ [ 861.044670] env[63515]: value = "task-1111403" [ 861.044670] env[63515]: _type = "Task" [ 861.044670] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 861.059634] env[63515]: DEBUG oslo_vmware.api [None req-957d8582-d135-4677-8fde-d84fe70e2761 tempest-ServerDiagnosticsV248Test-1736022867 tempest-ServerDiagnosticsV248Test-1736022867-project-member] Task: {'id': task-1111403, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.093145] env[63515]: DEBUG oslo_vmware.rw_handles [None req-ce131312-5bf5-4677-8923-2d91405ade5d tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Completed reading data from the image iterator. {{(pid=63515) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 861.093343] env[63515]: DEBUG oslo_vmware.rw_handles [None req-ce131312-5bf5-4677-8923-2d91405ade5d tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/520d7a2f-331f-4a3e-2aae-6efa230e5397/disk-0.vmdk. {{(pid=63515) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 861.094656] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4c9064c-21a9-4594-9a96-7fb4a1587adc {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.108598] env[63515]: DEBUG oslo_vmware.rw_handles [None req-ce131312-5bf5-4677-8923-2d91405ade5d tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/520d7a2f-331f-4a3e-2aae-6efa230e5397/disk-0.vmdk is in state: ready. {{(pid=63515) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 861.108598] env[63515]: DEBUG oslo_vmware.rw_handles [None req-ce131312-5bf5-4677-8923-2d91405ade5d tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Releasing lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/520d7a2f-331f-4a3e-2aae-6efa230e5397/disk-0.vmdk. {{(pid=63515) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 861.108790] env[63515]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-de5ab7f8-1c81-48b7-9e41-b875491eaa4a {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.149050] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1884c31-a17c-4fbf-a4b6-57cae62e52ce {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.157227] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a75b19c-6802-4ce9-bcfd-e870ed2c9ccf {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.190572] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c71a8ea-8521-42ff-aac2-fa16b2d51f3c {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.198469] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-205d91dd-d0c3-4149-aeb8-ee5a4ef55d02 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.212404] env[63515]: DEBUG nova.compute.provider_tree [None req-baaf7dbf-037b-4d17-a0eb-d252e746ead6 tempest-InstanceActionsTestJSON-247065106 tempest-InstanceActionsTestJSON-247065106-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 861.510615] env[63515]: DEBUG oslo_vmware.rw_handles [None req-ce131312-5bf5-4677-8923-2d91405ade5d tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Closed VMDK write handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/520d7a2f-331f-4a3e-2aae-6efa230e5397/disk-0.vmdk. {{(pid=63515) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 861.510615] env[63515]: INFO nova.virt.vmwareapi.images [None req-ce131312-5bf5-4677-8923-2d91405ade5d tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] [instance: 87c468d9-9594-4804-b461-527f01f6118f] Downloaded image file data 499b6029-07e9-499c-80d5-a53d8c43ffb7 [ 861.511339] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6cab119e-5904-46a7-884b-e74ffc44380b {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.532342] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f9beaf30-a2ff-4235-8022-e79f49fb9933 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.547216] env[63515]: DEBUG oslo_concurrency.lockutils [None req-bfac7ffd-4aa9-41c4-9bda-3f289bbda0a2 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Acquiring lock "d926c699-a64a-4942-9ef4-f0166414661d" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 861.547547] env[63515]: DEBUG oslo_concurrency.lockutils [None req-bfac7ffd-4aa9-41c4-9bda-3f289bbda0a2 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Lock "d926c699-a64a-4942-9ef4-f0166414661d" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.001s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 861.547748] env[63515]: DEBUG nova.compute.manager [None req-bfac7ffd-4aa9-41c4-9bda-3f289bbda0a2 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: d926c699-a64a-4942-9ef4-f0166414661d] Going to confirm migration 2 {{(pid=63515) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 861.558295] env[63515]: DEBUG oslo_vmware.api [None req-957d8582-d135-4677-8fde-d84fe70e2761 tempest-ServerDiagnosticsV248Test-1736022867 tempest-ServerDiagnosticsV248Test-1736022867-project-member] Task: {'id': task-1111403, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.617913] env[63515]: INFO nova.virt.vmwareapi.images [None req-ce131312-5bf5-4677-8923-2d91405ade5d tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] [instance: 87c468d9-9594-4804-b461-527f01f6118f] The imported VM was unregistered [ 861.620279] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-ce131312-5bf5-4677-8923-2d91405ade5d tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] [instance: 87c468d9-9594-4804-b461-527f01f6118f] Caching image {{(pid=63515) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 861.620403] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-ce131312-5bf5-4677-8923-2d91405ade5d tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Creating directory with path [datastore1] devstack-image-cache_base/499b6029-07e9-499c-80d5-a53d8c43ffb7 {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 861.620752] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-4cb44d21-b558-4d48-8c4f-7f52a96efd08 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.631025] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-ce131312-5bf5-4677-8923-2d91405ade5d tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Created directory with path [datastore1] devstack-image-cache_base/499b6029-07e9-499c-80d5-a53d8c43ffb7 {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 861.631025] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-ce131312-5bf5-4677-8923-2d91405ade5d tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Moving virtual disk from [datastore1] OSTACK_IMG_d97e85a2-748d-4e52-8257-3fd60729ce2e/OSTACK_IMG_d97e85a2-748d-4e52-8257-3fd60729ce2e.vmdk to [datastore1] devstack-image-cache_base/499b6029-07e9-499c-80d5-a53d8c43ffb7/499b6029-07e9-499c-80d5-a53d8c43ffb7.vmdk. {{(pid=63515) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 861.631194] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-d029e6a1-103d-43b3-b712-6123c6ca5f61 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.638165] env[63515]: DEBUG oslo_vmware.api [None req-ce131312-5bf5-4677-8923-2d91405ade5d tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Waiting for the task: (returnval){ [ 861.638165] env[63515]: value = "task-1111405" [ 861.638165] env[63515]: _type = "Task" [ 861.638165] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 861.645852] env[63515]: DEBUG oslo_vmware.api [None req-ce131312-5bf5-4677-8923-2d91405ade5d tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Task: {'id': task-1111405, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.715586] env[63515]: DEBUG nova.scheduler.client.report [None req-baaf7dbf-037b-4d17-a0eb-d252e746ead6 tempest-InstanceActionsTestJSON-247065106 tempest-InstanceActionsTestJSON-247065106-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 861.747249] env[63515]: DEBUG nova.compute.manager [None req-eae7e17b-7393-4e13-b575-a8d38e798918 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] [instance: 2a29f37d-e786-48d3-b126-4854d6e7c67c] Start spawning the instance on the hypervisor. {{(pid=63515) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 861.772127] env[63515]: DEBUG nova.virt.hardware [None req-eae7e17b-7393-4e13-b575-a8d38e798918 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T02:52:14Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T02:51:56Z,direct_url=,disk_format='vmdk',id=8a120570-cb06-4099-b262-554ca0ad15c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b569255cc43e42e7ae3f2b4ad37c6ef4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T02:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 861.772411] env[63515]: DEBUG nova.virt.hardware [None req-eae7e17b-7393-4e13-b575-a8d38e798918 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Flavor limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 861.772571] env[63515]: DEBUG nova.virt.hardware [None req-eae7e17b-7393-4e13-b575-a8d38e798918 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Image limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 861.772758] env[63515]: DEBUG nova.virt.hardware [None req-eae7e17b-7393-4e13-b575-a8d38e798918 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Flavor pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 861.772911] env[63515]: DEBUG nova.virt.hardware [None req-eae7e17b-7393-4e13-b575-a8d38e798918 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Image pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 861.773081] env[63515]: DEBUG nova.virt.hardware [None req-eae7e17b-7393-4e13-b575-a8d38e798918 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 861.773301] env[63515]: DEBUG nova.virt.hardware [None req-eae7e17b-7393-4e13-b575-a8d38e798918 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 861.773467] env[63515]: DEBUG nova.virt.hardware [None req-eae7e17b-7393-4e13-b575-a8d38e798918 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 861.773667] env[63515]: DEBUG nova.virt.hardware [None req-eae7e17b-7393-4e13-b575-a8d38e798918 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Got 1 possible topologies {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 861.773898] env[63515]: DEBUG nova.virt.hardware [None req-eae7e17b-7393-4e13-b575-a8d38e798918 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 861.774102] env[63515]: DEBUG nova.virt.hardware [None req-eae7e17b-7393-4e13-b575-a8d38e798918 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 861.774954] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce362639-b57f-46dd-be4d-434a0d966b10 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.782920] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2fc0c9b0-d44d-4d24-a672-56d1b44b13f9 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.064852] env[63515]: DEBUG oslo_vmware.api [None req-957d8582-d135-4677-8fde-d84fe70e2761 tempest-ServerDiagnosticsV248Test-1736022867 tempest-ServerDiagnosticsV248Test-1736022867-project-member] Task: {'id': task-1111403, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.548764} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 862.065120] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-957d8582-d135-4677-8fde-d84fe70e2761 tempest-ServerDiagnosticsV248Test-1736022867 tempest-ServerDiagnosticsV248Test-1736022867-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk to [datastore1] 705b7e9b-3ab6-40ec-8cf6-d86bf0a1faf0/705b7e9b-3ab6-40ec-8cf6-d86bf0a1faf0.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 862.065396] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-957d8582-d135-4677-8fde-d84fe70e2761 tempest-ServerDiagnosticsV248Test-1736022867 tempest-ServerDiagnosticsV248Test-1736022867-project-member] [instance: 705b7e9b-3ab6-40ec-8cf6-d86bf0a1faf0] Extending root virtual disk to 1048576 {{(pid=63515) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 862.065660] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-80976e9d-3e55-4007-a88f-efd2119b8ba0 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.076057] env[63515]: DEBUG oslo_vmware.api [None req-957d8582-d135-4677-8fde-d84fe70e2761 tempest-ServerDiagnosticsV248Test-1736022867 tempest-ServerDiagnosticsV248Test-1736022867-project-member] Waiting for the task: (returnval){ [ 862.076057] env[63515]: value = "task-1111406" [ 862.076057] env[63515]: _type = "Task" [ 862.076057] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.085803] env[63515]: DEBUG oslo_vmware.api [None req-957d8582-d135-4677-8fde-d84fe70e2761 tempest-ServerDiagnosticsV248Test-1736022867 tempest-ServerDiagnosticsV248Test-1736022867-project-member] Task: {'id': task-1111406, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.121109] env[63515]: DEBUG oslo_concurrency.lockutils [None req-bfac7ffd-4aa9-41c4-9bda-3f289bbda0a2 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Acquiring lock "refresh_cache-d926c699-a64a-4942-9ef4-f0166414661d" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 862.121109] env[63515]: DEBUG oslo_concurrency.lockutils [None req-bfac7ffd-4aa9-41c4-9bda-3f289bbda0a2 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Acquired lock "refresh_cache-d926c699-a64a-4942-9ef4-f0166414661d" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 862.121109] env[63515]: DEBUG nova.network.neutron [None req-bfac7ffd-4aa9-41c4-9bda-3f289bbda0a2 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: d926c699-a64a-4942-9ef4-f0166414661d] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 862.121109] env[63515]: DEBUG nova.objects.instance [None req-bfac7ffd-4aa9-41c4-9bda-3f289bbda0a2 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Lazy-loading 'info_cache' on Instance uuid d926c699-a64a-4942-9ef4-f0166414661d {{(pid=63515) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 862.153071] env[63515]: DEBUG oslo_vmware.api [None req-ce131312-5bf5-4677-8923-2d91405ade5d tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Task: {'id': task-1111405, 'name': MoveVirtualDisk_Task} progress is 21%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.181068] env[63515]: DEBUG nova.compute.manager [req-571dc88f-b7ae-4228-93d6-7f2ea49fc901 req-9d39f94c-633e-4cb9-93eb-4b363e1395e6 service nova] [instance: 2a29f37d-e786-48d3-b126-4854d6e7c67c] Received event network-vif-plugged-956cd361-0d42-4914-85d3-5f75a72a68a1 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 862.181470] env[63515]: DEBUG oslo_concurrency.lockutils [req-571dc88f-b7ae-4228-93d6-7f2ea49fc901 req-9d39f94c-633e-4cb9-93eb-4b363e1395e6 service nova] Acquiring lock "2a29f37d-e786-48d3-b126-4854d6e7c67c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 862.181680] env[63515]: DEBUG oslo_concurrency.lockutils [req-571dc88f-b7ae-4228-93d6-7f2ea49fc901 req-9d39f94c-633e-4cb9-93eb-4b363e1395e6 service nova] Lock "2a29f37d-e786-48d3-b126-4854d6e7c67c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 862.181866] env[63515]: DEBUG oslo_concurrency.lockutils [req-571dc88f-b7ae-4228-93d6-7f2ea49fc901 req-9d39f94c-633e-4cb9-93eb-4b363e1395e6 service nova] Lock "2a29f37d-e786-48d3-b126-4854d6e7c67c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 862.182052] env[63515]: DEBUG nova.compute.manager [req-571dc88f-b7ae-4228-93d6-7f2ea49fc901 req-9d39f94c-633e-4cb9-93eb-4b363e1395e6 service nova] [instance: 2a29f37d-e786-48d3-b126-4854d6e7c67c] No waiting events found dispatching network-vif-plugged-956cd361-0d42-4914-85d3-5f75a72a68a1 {{(pid=63515) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 862.182238] env[63515]: WARNING nova.compute.manager [req-571dc88f-b7ae-4228-93d6-7f2ea49fc901 req-9d39f94c-633e-4cb9-93eb-4b363e1395e6 service nova] [instance: 2a29f37d-e786-48d3-b126-4854d6e7c67c] Received unexpected event network-vif-plugged-956cd361-0d42-4914-85d3-5f75a72a68a1 for instance with vm_state building and task_state spawning. [ 862.221279] env[63515]: DEBUG oslo_concurrency.lockutils [None req-baaf7dbf-037b-4d17-a0eb-d252e746ead6 tempest-InstanceActionsTestJSON-247065106 tempest-InstanceActionsTestJSON-247065106-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.510s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 862.221980] env[63515]: DEBUG nova.compute.manager [None req-baaf7dbf-037b-4d17-a0eb-d252e746ead6 tempest-InstanceActionsTestJSON-247065106 tempest-InstanceActionsTestJSON-247065106-project-member] [instance: 0ee9cce2-94b9-4038-ab38-1b3f96a864bc] Start building networks asynchronously for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 862.225126] env[63515]: DEBUG oslo_concurrency.lockutils [None req-789266fe-c183-402c-aef4-222438c1eee4 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 32.228s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 862.228163] env[63515]: INFO nova.compute.claims [None req-789266fe-c183-402c-aef4-222438c1eee4 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: eab16df9-7bb5-4576-bca0-769a561c5fe9] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 862.306725] env[63515]: DEBUG nova.network.neutron [None req-eae7e17b-7393-4e13-b575-a8d38e798918 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] [instance: 2a29f37d-e786-48d3-b126-4854d6e7c67c] Successfully updated port: 956cd361-0d42-4914-85d3-5f75a72a68a1 {{(pid=63515) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 862.587602] env[63515]: DEBUG oslo_vmware.api [None req-957d8582-d135-4677-8fde-d84fe70e2761 tempest-ServerDiagnosticsV248Test-1736022867 tempest-ServerDiagnosticsV248Test-1736022867-project-member] Task: {'id': task-1111406, 'name': ExtendVirtualDisk_Task} progress is 50%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.649255] env[63515]: DEBUG oslo_vmware.api [None req-ce131312-5bf5-4677-8923-2d91405ade5d tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Task: {'id': task-1111405, 'name': MoveVirtualDisk_Task} progress is 43%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.736261] env[63515]: DEBUG nova.compute.utils [None req-baaf7dbf-037b-4d17-a0eb-d252e746ead6 tempest-InstanceActionsTestJSON-247065106 tempest-InstanceActionsTestJSON-247065106-project-member] Using /dev/sd instead of None {{(pid=63515) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 862.738313] env[63515]: DEBUG nova.compute.manager [None req-baaf7dbf-037b-4d17-a0eb-d252e746ead6 tempest-InstanceActionsTestJSON-247065106 tempest-InstanceActionsTestJSON-247065106-project-member] [instance: 0ee9cce2-94b9-4038-ab38-1b3f96a864bc] Allocating IP information in the background. {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 862.738484] env[63515]: DEBUG nova.network.neutron [None req-baaf7dbf-037b-4d17-a0eb-d252e746ead6 tempest-InstanceActionsTestJSON-247065106 tempest-InstanceActionsTestJSON-247065106-project-member] [instance: 0ee9cce2-94b9-4038-ab38-1b3f96a864bc] allocate_for_instance() {{(pid=63515) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 862.787898] env[63515]: DEBUG nova.policy [None req-baaf7dbf-037b-4d17-a0eb-d252e746ead6 tempest-InstanceActionsTestJSON-247065106 tempest-InstanceActionsTestJSON-247065106-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2213043e34f4452eaef2e5893798cc1a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '486afecc6edc4a94b18a1e34e4f92fe5', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63515) authorize /opt/stack/nova/nova/policy.py:201}} [ 862.813539] env[63515]: DEBUG oslo_concurrency.lockutils [None req-eae7e17b-7393-4e13-b575-a8d38e798918 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Acquiring lock "refresh_cache-2a29f37d-e786-48d3-b126-4854d6e7c67c" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 862.813629] env[63515]: DEBUG oslo_concurrency.lockutils [None req-eae7e17b-7393-4e13-b575-a8d38e798918 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Acquired lock "refresh_cache-2a29f37d-e786-48d3-b126-4854d6e7c67c" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 862.813827] env[63515]: DEBUG nova.network.neutron [None req-eae7e17b-7393-4e13-b575-a8d38e798918 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] [instance: 2a29f37d-e786-48d3-b126-4854d6e7c67c] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 863.058636] env[63515]: DEBUG nova.network.neutron [None req-baaf7dbf-037b-4d17-a0eb-d252e746ead6 tempest-InstanceActionsTestJSON-247065106 tempest-InstanceActionsTestJSON-247065106-project-member] [instance: 0ee9cce2-94b9-4038-ab38-1b3f96a864bc] Successfully created port: 2fc6eecd-e0cd-41f2-8707-6eb57c2fef5a {{(pid=63515) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 863.088960] env[63515]: DEBUG oslo_vmware.api [None req-957d8582-d135-4677-8fde-d84fe70e2761 tempest-ServerDiagnosticsV248Test-1736022867 tempest-ServerDiagnosticsV248Test-1736022867-project-member] Task: {'id': task-1111406, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.553211} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 863.089354] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-957d8582-d135-4677-8fde-d84fe70e2761 tempest-ServerDiagnosticsV248Test-1736022867 tempest-ServerDiagnosticsV248Test-1736022867-project-member] [instance: 705b7e9b-3ab6-40ec-8cf6-d86bf0a1faf0] Extended root virtual disk {{(pid=63515) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 863.090285] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb53f0b5-65ce-43d3-b126-212641f34255 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.114604] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-957d8582-d135-4677-8fde-d84fe70e2761 tempest-ServerDiagnosticsV248Test-1736022867 tempest-ServerDiagnosticsV248Test-1736022867-project-member] [instance: 705b7e9b-3ab6-40ec-8cf6-d86bf0a1faf0] Reconfiguring VM instance instance-00000044 to attach disk [datastore1] 705b7e9b-3ab6-40ec-8cf6-d86bf0a1faf0/705b7e9b-3ab6-40ec-8cf6-d86bf0a1faf0.vmdk or device None with type sparse {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 863.114791] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0ccd2582-c054-4ee2-b6ac-0933bc430010 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.146433] env[63515]: DEBUG oslo_vmware.api [None req-957d8582-d135-4677-8fde-d84fe70e2761 tempest-ServerDiagnosticsV248Test-1736022867 tempest-ServerDiagnosticsV248Test-1736022867-project-member] Waiting for the task: (returnval){ [ 863.146433] env[63515]: value = "task-1111407" [ 863.146433] env[63515]: _type = "Task" [ 863.146433] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.152698] env[63515]: DEBUG oslo_vmware.api [None req-ce131312-5bf5-4677-8923-2d91405ade5d tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Task: {'id': task-1111405, 'name': MoveVirtualDisk_Task} progress is 66%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.161270] env[63515]: DEBUG oslo_vmware.api [None req-957d8582-d135-4677-8fde-d84fe70e2761 tempest-ServerDiagnosticsV248Test-1736022867 tempest-ServerDiagnosticsV248Test-1736022867-project-member] Task: {'id': task-1111407, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.241745] env[63515]: DEBUG nova.compute.manager [None req-baaf7dbf-037b-4d17-a0eb-d252e746ead6 tempest-InstanceActionsTestJSON-247065106 tempest-InstanceActionsTestJSON-247065106-project-member] [instance: 0ee9cce2-94b9-4038-ab38-1b3f96a864bc] Start building block device mappings for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 863.365436] env[63515]: DEBUG nova.network.neutron [None req-eae7e17b-7393-4e13-b575-a8d38e798918 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] [instance: 2a29f37d-e786-48d3-b126-4854d6e7c67c] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 863.403046] env[63515]: DEBUG nova.network.neutron [None req-bfac7ffd-4aa9-41c4-9bda-3f289bbda0a2 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: d926c699-a64a-4942-9ef4-f0166414661d] Updating instance_info_cache with network_info: [{"id": "fdec91d2-43e8-4d40-9aa9-885a82fd9b42", "address": "fa:16:3e:5c:b7:75", "network": {"id": "15fdaabe-6793-400f-9737-2cdff8f07238", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1175204866-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f9c98e9f6020475490aaa7e76d907ab7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7d377d75-3add-4a15-8691-74b2eb010924", "external-id": "nsx-vlan-transportzone-71", "segmentation_id": 71, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfdec91d2-43", "ovs_interfaceid": "fdec91d2-43e8-4d40-9aa9-885a82fd9b42", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 863.594447] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c522655-4cd5-4caf-a70d-d39fbf7cdd1a {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.606027] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20d0ad16-cc19-45ae-8645-29c51ddbc27e {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.640382] env[63515]: DEBUG nova.network.neutron [None req-eae7e17b-7393-4e13-b575-a8d38e798918 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] [instance: 2a29f37d-e786-48d3-b126-4854d6e7c67c] Updating instance_info_cache with network_info: [{"id": "956cd361-0d42-4914-85d3-5f75a72a68a1", "address": "fa:16:3e:a3:72:bb", "network": {"id": "850bfa95-3df6-4019-8260-ade20ec58799", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-832000501-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "08c9aa51b1864410809f8877c7cd49c4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3cf748a8-7ae0-4dca-817d-e727c30d72f4", "external-id": "nsx-vlan-transportzone-853", "segmentation_id": 853, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap956cd361-0d", "ovs_interfaceid": "956cd361-0d42-4914-85d3-5f75a72a68a1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 863.642413] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e0d719f-063d-4c57-a446-60830d70d384 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.665673] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc06643e-6cdf-4afa-a7a7-ee4a34e41f97 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.673026] env[63515]: DEBUG oslo_vmware.api [None req-ce131312-5bf5-4677-8923-2d91405ade5d tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Task: {'id': task-1111405, 'name': MoveVirtualDisk_Task} progress is 85%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.674042] env[63515]: DEBUG oslo_vmware.api [None req-957d8582-d135-4677-8fde-d84fe70e2761 tempest-ServerDiagnosticsV248Test-1736022867 tempest-ServerDiagnosticsV248Test-1736022867-project-member] Task: {'id': task-1111407, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.686374] env[63515]: DEBUG nova.compute.provider_tree [None req-789266fe-c183-402c-aef4-222438c1eee4 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 863.906503] env[63515]: DEBUG oslo_concurrency.lockutils [None req-bfac7ffd-4aa9-41c4-9bda-3f289bbda0a2 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Releasing lock "refresh_cache-d926c699-a64a-4942-9ef4-f0166414661d" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 863.906930] env[63515]: DEBUG nova.objects.instance [None req-bfac7ffd-4aa9-41c4-9bda-3f289bbda0a2 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Lazy-loading 'migration_context' on Instance uuid d926c699-a64a-4942-9ef4-f0166414661d {{(pid=63515) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 864.146763] env[63515]: DEBUG oslo_concurrency.lockutils [None req-eae7e17b-7393-4e13-b575-a8d38e798918 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Releasing lock "refresh_cache-2a29f37d-e786-48d3-b126-4854d6e7c67c" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 864.146763] env[63515]: DEBUG nova.compute.manager [None req-eae7e17b-7393-4e13-b575-a8d38e798918 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] [instance: 2a29f37d-e786-48d3-b126-4854d6e7c67c] Instance network_info: |[{"id": "956cd361-0d42-4914-85d3-5f75a72a68a1", "address": "fa:16:3e:a3:72:bb", "network": {"id": "850bfa95-3df6-4019-8260-ade20ec58799", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-832000501-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "08c9aa51b1864410809f8877c7cd49c4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3cf748a8-7ae0-4dca-817d-e727c30d72f4", "external-id": "nsx-vlan-transportzone-853", "segmentation_id": 853, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap956cd361-0d", "ovs_interfaceid": "956cd361-0d42-4914-85d3-5f75a72a68a1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 864.150436] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-eae7e17b-7393-4e13-b575-a8d38e798918 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] [instance: 2a29f37d-e786-48d3-b126-4854d6e7c67c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:a3:72:bb', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '3cf748a8-7ae0-4dca-817d-e727c30d72f4', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '956cd361-0d42-4914-85d3-5f75a72a68a1', 'vif_model': 'vmxnet3'}] {{(pid=63515) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 864.157979] env[63515]: DEBUG oslo.service.loopingcall [None req-eae7e17b-7393-4e13-b575-a8d38e798918 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 864.158334] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2a29f37d-e786-48d3-b126-4854d6e7c67c] Creating VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 864.162131] env[63515]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-8b97057a-b374-4de0-b443-114641a9b5e4 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.181031] env[63515]: DEBUG oslo_vmware.api [None req-ce131312-5bf5-4677-8923-2d91405ade5d tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Task: {'id': task-1111405, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.40288} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.181927] env[63515]: INFO nova.virt.vmwareapi.ds_util [None req-ce131312-5bf5-4677-8923-2d91405ade5d tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Moved virtual disk from [datastore1] OSTACK_IMG_d97e85a2-748d-4e52-8257-3fd60729ce2e/OSTACK_IMG_d97e85a2-748d-4e52-8257-3fd60729ce2e.vmdk to [datastore1] devstack-image-cache_base/499b6029-07e9-499c-80d5-a53d8c43ffb7/499b6029-07e9-499c-80d5-a53d8c43ffb7.vmdk. [ 864.182179] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-ce131312-5bf5-4677-8923-2d91405ade5d tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] [instance: 87c468d9-9594-4804-b461-527f01f6118f] Cleaning up location [datastore1] OSTACK_IMG_d97e85a2-748d-4e52-8257-3fd60729ce2e {{(pid=63515) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 864.182369] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-ce131312-5bf5-4677-8923-2d91405ade5d tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Deleting the datastore file [datastore1] OSTACK_IMG_d97e85a2-748d-4e52-8257-3fd60729ce2e {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 864.182625] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-62327ba5-56d5-4e03-bedc-e6659eda62a5 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.187301] env[63515]: DEBUG oslo_vmware.api [None req-957d8582-d135-4677-8fde-d84fe70e2761 tempest-ServerDiagnosticsV248Test-1736022867 tempest-ServerDiagnosticsV248Test-1736022867-project-member] Task: {'id': task-1111407, 'name': ReconfigVM_Task, 'duration_secs': 1.045613} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.188781] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-957d8582-d135-4677-8fde-d84fe70e2761 tempest-ServerDiagnosticsV248Test-1736022867 tempest-ServerDiagnosticsV248Test-1736022867-project-member] [instance: 705b7e9b-3ab6-40ec-8cf6-d86bf0a1faf0] Reconfigured VM instance instance-00000044 to attach disk [datastore1] 705b7e9b-3ab6-40ec-8cf6-d86bf0a1faf0/705b7e9b-3ab6-40ec-8cf6-d86bf0a1faf0.vmdk or device None with type sparse {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 864.189945] env[63515]: DEBUG nova.scheduler.client.report [None req-789266fe-c183-402c-aef4-222438c1eee4 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 864.193164] env[63515]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 864.193164] env[63515]: value = "task-1111408" [ 864.193164] env[63515]: _type = "Task" [ 864.193164] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.194671] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-091f1ffe-a438-47c8-9105-199bd5ac27f6 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.196364] env[63515]: DEBUG oslo_vmware.api [None req-ce131312-5bf5-4677-8923-2d91405ade5d tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Waiting for the task: (returnval){ [ 864.196364] env[63515]: value = "task-1111409" [ 864.196364] env[63515]: _type = "Task" [ 864.196364] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.205370] env[63515]: DEBUG oslo_vmware.api [None req-957d8582-d135-4677-8fde-d84fe70e2761 tempest-ServerDiagnosticsV248Test-1736022867 tempest-ServerDiagnosticsV248Test-1736022867-project-member] Waiting for the task: (returnval){ [ 864.205370] env[63515]: value = "task-1111410" [ 864.205370] env[63515]: _type = "Task" [ 864.205370] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.215265] env[63515]: DEBUG oslo_vmware.api [None req-ce131312-5bf5-4677-8923-2d91405ade5d tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Task: {'id': task-1111409, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.215464] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1111408, 'name': CreateVM_Task} progress is 6%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.216581] env[63515]: DEBUG nova.compute.manager [req-748a8323-3ea7-462b-86a8-d9fceb8238d7 req-20f90ec6-b2a2-40e8-8aa1-970ee8f9840a service nova] [instance: 2a29f37d-e786-48d3-b126-4854d6e7c67c] Received event network-changed-956cd361-0d42-4914-85d3-5f75a72a68a1 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 864.216786] env[63515]: DEBUG nova.compute.manager [req-748a8323-3ea7-462b-86a8-d9fceb8238d7 req-20f90ec6-b2a2-40e8-8aa1-970ee8f9840a service nova] [instance: 2a29f37d-e786-48d3-b126-4854d6e7c67c] Refreshing instance network info cache due to event network-changed-956cd361-0d42-4914-85d3-5f75a72a68a1. {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 864.217027] env[63515]: DEBUG oslo_concurrency.lockutils [req-748a8323-3ea7-462b-86a8-d9fceb8238d7 req-20f90ec6-b2a2-40e8-8aa1-970ee8f9840a service nova] Acquiring lock "refresh_cache-2a29f37d-e786-48d3-b126-4854d6e7c67c" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 864.217182] env[63515]: DEBUG oslo_concurrency.lockutils [req-748a8323-3ea7-462b-86a8-d9fceb8238d7 req-20f90ec6-b2a2-40e8-8aa1-970ee8f9840a service nova] Acquired lock "refresh_cache-2a29f37d-e786-48d3-b126-4854d6e7c67c" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 864.217348] env[63515]: DEBUG nova.network.neutron [req-748a8323-3ea7-462b-86a8-d9fceb8238d7 req-20f90ec6-b2a2-40e8-8aa1-970ee8f9840a service nova] [instance: 2a29f37d-e786-48d3-b126-4854d6e7c67c] Refreshing network info cache for port 956cd361-0d42-4914-85d3-5f75a72a68a1 {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 864.226918] env[63515]: DEBUG oslo_vmware.api [None req-957d8582-d135-4677-8fde-d84fe70e2761 tempest-ServerDiagnosticsV248Test-1736022867 tempest-ServerDiagnosticsV248Test-1736022867-project-member] Task: {'id': task-1111410, 'name': Rename_Task} progress is 14%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.251576] env[63515]: DEBUG nova.compute.manager [None req-baaf7dbf-037b-4d17-a0eb-d252e746ead6 tempest-InstanceActionsTestJSON-247065106 tempest-InstanceActionsTestJSON-247065106-project-member] [instance: 0ee9cce2-94b9-4038-ab38-1b3f96a864bc] Start spawning the instance on the hypervisor. {{(pid=63515) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 864.277245] env[63515]: DEBUG nova.virt.hardware [None req-baaf7dbf-037b-4d17-a0eb-d252e746ead6 tempest-InstanceActionsTestJSON-247065106 tempest-InstanceActionsTestJSON-247065106-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T02:52:14Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T02:51:56Z,direct_url=,disk_format='vmdk',id=8a120570-cb06-4099-b262-554ca0ad15c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b569255cc43e42e7ae3f2b4ad37c6ef4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T02:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 864.277529] env[63515]: DEBUG nova.virt.hardware [None req-baaf7dbf-037b-4d17-a0eb-d252e746ead6 tempest-InstanceActionsTestJSON-247065106 tempest-InstanceActionsTestJSON-247065106-project-member] Flavor limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 864.277699] env[63515]: DEBUG nova.virt.hardware [None req-baaf7dbf-037b-4d17-a0eb-d252e746ead6 tempest-InstanceActionsTestJSON-247065106 tempest-InstanceActionsTestJSON-247065106-project-member] Image limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 864.277889] env[63515]: DEBUG nova.virt.hardware [None req-baaf7dbf-037b-4d17-a0eb-d252e746ead6 tempest-InstanceActionsTestJSON-247065106 tempest-InstanceActionsTestJSON-247065106-project-member] Flavor pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 864.278074] env[63515]: DEBUG nova.virt.hardware [None req-baaf7dbf-037b-4d17-a0eb-d252e746ead6 tempest-InstanceActionsTestJSON-247065106 tempest-InstanceActionsTestJSON-247065106-project-member] Image pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 864.278261] env[63515]: DEBUG nova.virt.hardware [None req-baaf7dbf-037b-4d17-a0eb-d252e746ead6 tempest-InstanceActionsTestJSON-247065106 tempest-InstanceActionsTestJSON-247065106-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 864.278510] env[63515]: DEBUG nova.virt.hardware [None req-baaf7dbf-037b-4d17-a0eb-d252e746ead6 tempest-InstanceActionsTestJSON-247065106 tempest-InstanceActionsTestJSON-247065106-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 864.278640] env[63515]: DEBUG nova.virt.hardware [None req-baaf7dbf-037b-4d17-a0eb-d252e746ead6 tempest-InstanceActionsTestJSON-247065106 tempest-InstanceActionsTestJSON-247065106-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 864.278811] env[63515]: DEBUG nova.virt.hardware [None req-baaf7dbf-037b-4d17-a0eb-d252e746ead6 tempest-InstanceActionsTestJSON-247065106 tempest-InstanceActionsTestJSON-247065106-project-member] Got 1 possible topologies {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 864.278977] env[63515]: DEBUG nova.virt.hardware [None req-baaf7dbf-037b-4d17-a0eb-d252e746ead6 tempest-InstanceActionsTestJSON-247065106 tempest-InstanceActionsTestJSON-247065106-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 864.279212] env[63515]: DEBUG nova.virt.hardware [None req-baaf7dbf-037b-4d17-a0eb-d252e746ead6 tempest-InstanceActionsTestJSON-247065106 tempest-InstanceActionsTestJSON-247065106-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 864.280045] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b40431e3-2ae6-40f3-831e-469d842abb96 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.288151] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e0a6628-ae4b-42e5-91b6-03b5d8051d1b {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.410060] env[63515]: DEBUG nova.objects.base [None req-bfac7ffd-4aa9-41c4-9bda-3f289bbda0a2 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Object Instance lazy-loaded attributes: info_cache,migration_context {{(pid=63515) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 864.411067] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe6881d2-57ea-496f-995b-165e7ab046db {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.432222] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4596e419-e469-4d96-91f2-c3bb13148149 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.437267] env[63515]: DEBUG oslo_vmware.api [None req-bfac7ffd-4aa9-41c4-9bda-3f289bbda0a2 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Waiting for the task: (returnval){ [ 864.437267] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]52c202d9-5a68-5ea4-4c35-acc9e922042a" [ 864.437267] env[63515]: _type = "Task" [ 864.437267] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.444812] env[63515]: DEBUG oslo_vmware.api [None req-bfac7ffd-4aa9-41c4-9bda-3f289bbda0a2 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52c202d9-5a68-5ea4-4c35-acc9e922042a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.700579] env[63515]: DEBUG oslo_concurrency.lockutils [None req-789266fe-c183-402c-aef4-222438c1eee4 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.475s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 864.701101] env[63515]: DEBUG nova.compute.manager [None req-789266fe-c183-402c-aef4-222438c1eee4 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: eab16df9-7bb5-4576-bca0-769a561c5fe9] Start building networks asynchronously for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 864.703957] env[63515]: DEBUG oslo_concurrency.lockutils [None req-ed70636f-053b-400c-8dba-ba2b774acd5f tempest-ServerMetadataNegativeTestJSON-502053314 tempest-ServerMetadataNegativeTestJSON-502053314-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 29.217s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 864.704204] env[63515]: DEBUG oslo_concurrency.lockutils [None req-ed70636f-053b-400c-8dba-ba2b774acd5f tempest-ServerMetadataNegativeTestJSON-502053314 tempest-ServerMetadataNegativeTestJSON-502053314-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 864.706434] env[63515]: DEBUG oslo_concurrency.lockutils [None req-7f7535aa-eede-4389-b9f3-fe528607f4b7 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 23.321s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 864.706805] env[63515]: DEBUG oslo_concurrency.lockutils [None req-7f7535aa-eede-4389-b9f3-fe528607f4b7 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 864.708235] env[63515]: DEBUG oslo_concurrency.lockutils [None req-3db8c3e0-5f0e-4c46-ab36-d6f97ddeb4f1 tempest-ServerActionsV293TestJSON-1193370169 tempest-ServerActionsV293TestJSON-1193370169-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 22.605s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 864.712321] env[63515]: INFO nova.compute.claims [None req-3db8c3e0-5f0e-4c46-ab36-d6f97ddeb4f1 tempest-ServerActionsV293TestJSON-1193370169 tempest-ServerActionsV293TestJSON-1193370169-project-member] [instance: 2ea99c7a-8b61-4718-bee0-f4ce4fc021af] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 864.724709] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1111408, 'name': CreateVM_Task} progress is 25%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.729101] env[63515]: DEBUG oslo_vmware.api [None req-ce131312-5bf5-4677-8923-2d91405ade5d tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Task: {'id': task-1111409, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.0477} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.729101] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-ce131312-5bf5-4677-8923-2d91405ade5d tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Deleted the datastore file {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 864.729101] env[63515]: DEBUG oslo_concurrency.lockutils [None req-ce131312-5bf5-4677-8923-2d91405ade5d tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Releasing lock "[datastore1] devstack-image-cache_base/499b6029-07e9-499c-80d5-a53d8c43ffb7/499b6029-07e9-499c-80d5-a53d8c43ffb7.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 864.729789] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-ce131312-5bf5-4677-8923-2d91405ade5d tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/499b6029-07e9-499c-80d5-a53d8c43ffb7/499b6029-07e9-499c-80d5-a53d8c43ffb7.vmdk to [datastore1] 87c468d9-9594-4804-b461-527f01f6118f/87c468d9-9594-4804-b461-527f01f6118f.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 864.732193] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3e1322af-bac3-4251-9017-588b1e6d8a20 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.734430] env[63515]: DEBUG oslo_vmware.api [None req-957d8582-d135-4677-8fde-d84fe70e2761 tempest-ServerDiagnosticsV248Test-1736022867 tempest-ServerDiagnosticsV248Test-1736022867-project-member] Task: {'id': task-1111410, 'name': Rename_Task, 'duration_secs': 0.162812} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.734678] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-957d8582-d135-4677-8fde-d84fe70e2761 tempest-ServerDiagnosticsV248Test-1736022867 tempest-ServerDiagnosticsV248Test-1736022867-project-member] [instance: 705b7e9b-3ab6-40ec-8cf6-d86bf0a1faf0] Powering on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 864.735264] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c954b71d-8659-46a5-81b2-2d2ffb6ce18e {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.739850] env[63515]: DEBUG oslo_vmware.api [None req-ce131312-5bf5-4677-8923-2d91405ade5d tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Waiting for the task: (returnval){ [ 864.739850] env[63515]: value = "task-1111411" [ 864.739850] env[63515]: _type = "Task" [ 864.739850] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.745160] env[63515]: DEBUG oslo_vmware.api [None req-957d8582-d135-4677-8fde-d84fe70e2761 tempest-ServerDiagnosticsV248Test-1736022867 tempest-ServerDiagnosticsV248Test-1736022867-project-member] Waiting for the task: (returnval){ [ 864.745160] env[63515]: value = "task-1111412" [ 864.745160] env[63515]: _type = "Task" [ 864.745160] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.749719] env[63515]: INFO nova.scheduler.client.report [None req-ed70636f-053b-400c-8dba-ba2b774acd5f tempest-ServerMetadataNegativeTestJSON-502053314 tempest-ServerMetadataNegativeTestJSON-502053314-project-member] Deleted allocations for instance 9d18c7b6-ce8e-4042-9e70-696bb7f57cb3 [ 864.753918] env[63515]: INFO nova.scheduler.client.report [None req-7f7535aa-eede-4389-b9f3-fe528607f4b7 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] Deleted allocations for instance 5b42f744-fdd6-45b1-8563-896869648c23 [ 864.755119] env[63515]: DEBUG oslo_vmware.api [None req-ce131312-5bf5-4677-8923-2d91405ade5d tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Task: {'id': task-1111411, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.765126] env[63515]: DEBUG oslo_vmware.api [None req-957d8582-d135-4677-8fde-d84fe70e2761 tempest-ServerDiagnosticsV248Test-1736022867 tempest-ServerDiagnosticsV248Test-1736022867-project-member] Task: {'id': task-1111412, 'name': PowerOnVM_Task} progress is 33%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.790060] env[63515]: DEBUG nova.network.neutron [None req-baaf7dbf-037b-4d17-a0eb-d252e746ead6 tempest-InstanceActionsTestJSON-247065106 tempest-InstanceActionsTestJSON-247065106-project-member] [instance: 0ee9cce2-94b9-4038-ab38-1b3f96a864bc] Successfully updated port: 2fc6eecd-e0cd-41f2-8707-6eb57c2fef5a {{(pid=63515) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 864.947364] env[63515]: DEBUG oslo_vmware.api [None req-bfac7ffd-4aa9-41c4-9bda-3f289bbda0a2 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52c202d9-5a68-5ea4-4c35-acc9e922042a, 'name': SearchDatastore_Task, 'duration_secs': 0.006123} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.948057] env[63515]: DEBUG oslo_concurrency.lockutils [None req-bfac7ffd-4aa9-41c4-9bda-3f289bbda0a2 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 865.152899] env[63515]: DEBUG nova.network.neutron [req-748a8323-3ea7-462b-86a8-d9fceb8238d7 req-20f90ec6-b2a2-40e8-8aa1-970ee8f9840a service nova] [instance: 2a29f37d-e786-48d3-b126-4854d6e7c67c] Updated VIF entry in instance network info cache for port 956cd361-0d42-4914-85d3-5f75a72a68a1. {{(pid=63515) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 865.153343] env[63515]: DEBUG nova.network.neutron [req-748a8323-3ea7-462b-86a8-d9fceb8238d7 req-20f90ec6-b2a2-40e8-8aa1-970ee8f9840a service nova] [instance: 2a29f37d-e786-48d3-b126-4854d6e7c67c] Updating instance_info_cache with network_info: [{"id": "956cd361-0d42-4914-85d3-5f75a72a68a1", "address": "fa:16:3e:a3:72:bb", "network": {"id": "850bfa95-3df6-4019-8260-ade20ec58799", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-832000501-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "08c9aa51b1864410809f8877c7cd49c4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3cf748a8-7ae0-4dca-817d-e727c30d72f4", "external-id": "nsx-vlan-transportzone-853", "segmentation_id": 853, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap956cd361-0d", "ovs_interfaceid": "956cd361-0d42-4914-85d3-5f75a72a68a1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 865.208234] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1111408, 'name': CreateVM_Task, 'duration_secs': 0.875134} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 865.208386] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2a29f37d-e786-48d3-b126-4854d6e7c67c] Created VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 865.209040] env[63515]: DEBUG oslo_concurrency.lockutils [None req-eae7e17b-7393-4e13-b575-a8d38e798918 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 865.209210] env[63515]: DEBUG oslo_concurrency.lockutils [None req-eae7e17b-7393-4e13-b575-a8d38e798918 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 865.209535] env[63515]: DEBUG oslo_concurrency.lockutils [None req-eae7e17b-7393-4e13-b575-a8d38e798918 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 865.209805] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d54a5fda-4a88-4e4a-a619-2743590a40d9 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.220186] env[63515]: DEBUG nova.compute.utils [None req-789266fe-c183-402c-aef4-222438c1eee4 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Using /dev/sd instead of None {{(pid=63515) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 865.223594] env[63515]: DEBUG oslo_vmware.api [None req-eae7e17b-7393-4e13-b575-a8d38e798918 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Waiting for the task: (returnval){ [ 865.223594] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]52ad103b-4a2a-1884-3a9e-3196b0de40cf" [ 865.223594] env[63515]: _type = "Task" [ 865.223594] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 865.224183] env[63515]: DEBUG nova.compute.manager [None req-789266fe-c183-402c-aef4-222438c1eee4 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: eab16df9-7bb5-4576-bca0-769a561c5fe9] Allocating IP information in the background. {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 865.224346] env[63515]: DEBUG nova.network.neutron [None req-789266fe-c183-402c-aef4-222438c1eee4 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: eab16df9-7bb5-4576-bca0-769a561c5fe9] allocate_for_instance() {{(pid=63515) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 865.243124] env[63515]: DEBUG oslo_vmware.api [None req-eae7e17b-7393-4e13-b575-a8d38e798918 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52ad103b-4a2a-1884-3a9e-3196b0de40cf, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.270041] env[63515]: DEBUG oslo_vmware.api [None req-ce131312-5bf5-4677-8923-2d91405ade5d tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Task: {'id': task-1111411, 'name': CopyVirtualDisk_Task} progress is 15%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.270300] env[63515]: DEBUG oslo_vmware.api [None req-957d8582-d135-4677-8fde-d84fe70e2761 tempest-ServerDiagnosticsV248Test-1736022867 tempest-ServerDiagnosticsV248Test-1736022867-project-member] Task: {'id': task-1111412, 'name': PowerOnVM_Task} progress is 88%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.270736] env[63515]: DEBUG oslo_concurrency.lockutils [None req-7f7535aa-eede-4389-b9f3-fe528607f4b7 tempest-ServersAdminNegativeTestJSON-1962214329 tempest-ServersAdminNegativeTestJSON-1962214329-project-member] Lock "5b42f744-fdd6-45b1-8563-896869648c23" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 26.866s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 865.271726] env[63515]: DEBUG oslo_concurrency.lockutils [None req-ed70636f-053b-400c-8dba-ba2b774acd5f tempest-ServerMetadataNegativeTestJSON-502053314 tempest-ServerMetadataNegativeTestJSON-502053314-project-member] Lock "9d18c7b6-ce8e-4042-9e70-696bb7f57cb3" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 32.836s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 865.295054] env[63515]: DEBUG oslo_concurrency.lockutils [None req-baaf7dbf-037b-4d17-a0eb-d252e746ead6 tempest-InstanceActionsTestJSON-247065106 tempest-InstanceActionsTestJSON-247065106-project-member] Acquiring lock "refresh_cache-0ee9cce2-94b9-4038-ab38-1b3f96a864bc" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 865.295280] env[63515]: DEBUG oslo_concurrency.lockutils [None req-baaf7dbf-037b-4d17-a0eb-d252e746ead6 tempest-InstanceActionsTestJSON-247065106 tempest-InstanceActionsTestJSON-247065106-project-member] Acquired lock "refresh_cache-0ee9cce2-94b9-4038-ab38-1b3f96a864bc" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 865.295641] env[63515]: DEBUG nova.network.neutron [None req-baaf7dbf-037b-4d17-a0eb-d252e746ead6 tempest-InstanceActionsTestJSON-247065106 tempest-InstanceActionsTestJSON-247065106-project-member] [instance: 0ee9cce2-94b9-4038-ab38-1b3f96a864bc] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 865.299100] env[63515]: DEBUG nova.policy [None req-789266fe-c183-402c-aef4-222438c1eee4 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '16880abf1c5b4341800c94ada2c756c7', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7d4ada9ec35f42b19c6480a9101d21a4', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63515) authorize /opt/stack/nova/nova/policy.py:201}} [ 865.656749] env[63515]: DEBUG oslo_concurrency.lockutils [req-748a8323-3ea7-462b-86a8-d9fceb8238d7 req-20f90ec6-b2a2-40e8-8aa1-970ee8f9840a service nova] Releasing lock "refresh_cache-2a29f37d-e786-48d3-b126-4854d6e7c67c" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 865.727218] env[63515]: DEBUG nova.compute.manager [None req-789266fe-c183-402c-aef4-222438c1eee4 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: eab16df9-7bb5-4576-bca0-769a561c5fe9] Start building block device mappings for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 865.746472] env[63515]: DEBUG oslo_vmware.api [None req-eae7e17b-7393-4e13-b575-a8d38e798918 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52ad103b-4a2a-1884-3a9e-3196b0de40cf, 'name': SearchDatastore_Task, 'duration_secs': 0.070909} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 865.751541] env[63515]: DEBUG oslo_concurrency.lockutils [None req-eae7e17b-7393-4e13-b575-a8d38e798918 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 865.752278] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-eae7e17b-7393-4e13-b575-a8d38e798918 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] [instance: 2a29f37d-e786-48d3-b126-4854d6e7c67c] Processing image 8a120570-cb06-4099-b262-554ca0ad15c5 {{(pid=63515) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 865.752724] env[63515]: DEBUG oslo_concurrency.lockutils [None req-eae7e17b-7393-4e13-b575-a8d38e798918 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 865.753453] env[63515]: DEBUG oslo_concurrency.lockutils [None req-eae7e17b-7393-4e13-b575-a8d38e798918 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 865.755017] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-eae7e17b-7393-4e13-b575-a8d38e798918 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 865.755017] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0be6a44f-c994-486e-bc40-953a8afc59ba {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.772069] env[63515]: DEBUG oslo_vmware.api [None req-ce131312-5bf5-4677-8923-2d91405ade5d tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Task: {'id': task-1111411, 'name': CopyVirtualDisk_Task} progress is 38%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.776489] env[63515]: DEBUG oslo_vmware.api [None req-957d8582-d135-4677-8fde-d84fe70e2761 tempest-ServerDiagnosticsV248Test-1736022867 tempest-ServerDiagnosticsV248Test-1736022867-project-member] Task: {'id': task-1111412, 'name': PowerOnVM_Task} progress is 89%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.781393] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-eae7e17b-7393-4e13-b575-a8d38e798918 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 865.781765] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-eae7e17b-7393-4e13-b575-a8d38e798918 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63515) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 865.785588] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ad440b81-620f-4dae-bc9f-7c7d567406e5 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.795192] env[63515]: DEBUG oslo_vmware.api [None req-eae7e17b-7393-4e13-b575-a8d38e798918 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Waiting for the task: (returnval){ [ 865.795192] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]52d51f66-34e3-2379-4d5c-91a8b0abef03" [ 865.795192] env[63515]: _type = "Task" [ 865.795192] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 865.811527] env[63515]: DEBUG oslo_vmware.api [None req-eae7e17b-7393-4e13-b575-a8d38e798918 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52d51f66-34e3-2379-4d5c-91a8b0abef03, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.069200] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-853a7258-a127-4dc0-b66c-284d3612f5c8 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.079725] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8487a03-1454-453d-bf86-25ec4377f169 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.118859] env[63515]: DEBUG nova.network.neutron [None req-baaf7dbf-037b-4d17-a0eb-d252e746ead6 tempest-InstanceActionsTestJSON-247065106 tempest-InstanceActionsTestJSON-247065106-project-member] [instance: 0ee9cce2-94b9-4038-ab38-1b3f96a864bc] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 866.122198] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa7434d3-fd32-4780-8e55-aaa2ddc3261a {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.132977] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72cdd218-1249-4b7a-8429-891c01f66572 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.152230] env[63515]: DEBUG nova.compute.provider_tree [None req-3db8c3e0-5f0e-4c46-ab36-d6f97ddeb4f1 tempest-ServerActionsV293TestJSON-1193370169 tempest-ServerActionsV293TestJSON-1193370169-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 866.258561] env[63515]: DEBUG oslo_vmware.api [None req-ce131312-5bf5-4677-8923-2d91405ade5d tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Task: {'id': task-1111411, 'name': CopyVirtualDisk_Task} progress is 57%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.275064] env[63515]: DEBUG oslo_vmware.api [None req-957d8582-d135-4677-8fde-d84fe70e2761 tempest-ServerDiagnosticsV248Test-1736022867 tempest-ServerDiagnosticsV248Test-1736022867-project-member] Task: {'id': task-1111412, 'name': PowerOnVM_Task, 'duration_secs': 1.442745} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 866.275064] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-957d8582-d135-4677-8fde-d84fe70e2761 tempest-ServerDiagnosticsV248Test-1736022867 tempest-ServerDiagnosticsV248Test-1736022867-project-member] [instance: 705b7e9b-3ab6-40ec-8cf6-d86bf0a1faf0] Powered on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 866.275064] env[63515]: INFO nova.compute.manager [None req-957d8582-d135-4677-8fde-d84fe70e2761 tempest-ServerDiagnosticsV248Test-1736022867 tempest-ServerDiagnosticsV248Test-1736022867-project-member] [instance: 705b7e9b-3ab6-40ec-8cf6-d86bf0a1faf0] Took 6.94 seconds to spawn the instance on the hypervisor. [ 866.275064] env[63515]: DEBUG nova.compute.manager [None req-957d8582-d135-4677-8fde-d84fe70e2761 tempest-ServerDiagnosticsV248Test-1736022867 tempest-ServerDiagnosticsV248Test-1736022867-project-member] [instance: 705b7e9b-3ab6-40ec-8cf6-d86bf0a1faf0] Checking state {{(pid=63515) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 866.275064] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-367aa927-d607-4c7f-93ff-41100bba6212 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.306462] env[63515]: DEBUG oslo_vmware.api [None req-eae7e17b-7393-4e13-b575-a8d38e798918 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52d51f66-34e3-2379-4d5c-91a8b0abef03, 'name': SearchDatastore_Task, 'duration_secs': 0.073204} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 866.307663] env[63515]: DEBUG nova.network.neutron [None req-789266fe-c183-402c-aef4-222438c1eee4 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: eab16df9-7bb5-4576-bca0-769a561c5fe9] Successfully created port: 4168f5a5-c282-4f13-9c97-0b0236bd659b {{(pid=63515) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 866.311164] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-81ad89c7-e910-4d5c-a66f-7f97c57d34c8 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.324227] env[63515]: DEBUG oslo_vmware.api [None req-eae7e17b-7393-4e13-b575-a8d38e798918 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Waiting for the task: (returnval){ [ 866.324227] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]52cf7d62-f820-1b4d-5765-42b567876084" [ 866.324227] env[63515]: _type = "Task" [ 866.324227] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.331767] env[63515]: DEBUG nova.compute.manager [req-acfbfd38-5e9b-4768-8fd0-cf6c44fb7b64 req-c9384dab-48d7-4b86-a97b-4ac56ab4ff99 service nova] [instance: 0ee9cce2-94b9-4038-ab38-1b3f96a864bc] Received event network-vif-plugged-2fc6eecd-e0cd-41f2-8707-6eb57c2fef5a {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 866.331767] env[63515]: DEBUG oslo_concurrency.lockutils [req-acfbfd38-5e9b-4768-8fd0-cf6c44fb7b64 req-c9384dab-48d7-4b86-a97b-4ac56ab4ff99 service nova] Acquiring lock "0ee9cce2-94b9-4038-ab38-1b3f96a864bc-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 866.331951] env[63515]: DEBUG oslo_concurrency.lockutils [req-acfbfd38-5e9b-4768-8fd0-cf6c44fb7b64 req-c9384dab-48d7-4b86-a97b-4ac56ab4ff99 service nova] Lock "0ee9cce2-94b9-4038-ab38-1b3f96a864bc-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 866.332158] env[63515]: DEBUG oslo_concurrency.lockutils [req-acfbfd38-5e9b-4768-8fd0-cf6c44fb7b64 req-c9384dab-48d7-4b86-a97b-4ac56ab4ff99 service nova] Lock "0ee9cce2-94b9-4038-ab38-1b3f96a864bc-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 866.332336] env[63515]: DEBUG nova.compute.manager [req-acfbfd38-5e9b-4768-8fd0-cf6c44fb7b64 req-c9384dab-48d7-4b86-a97b-4ac56ab4ff99 service nova] [instance: 0ee9cce2-94b9-4038-ab38-1b3f96a864bc] No waiting events found dispatching network-vif-plugged-2fc6eecd-e0cd-41f2-8707-6eb57c2fef5a {{(pid=63515) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 866.332514] env[63515]: WARNING nova.compute.manager [req-acfbfd38-5e9b-4768-8fd0-cf6c44fb7b64 req-c9384dab-48d7-4b86-a97b-4ac56ab4ff99 service nova] [instance: 0ee9cce2-94b9-4038-ab38-1b3f96a864bc] Received unexpected event network-vif-plugged-2fc6eecd-e0cd-41f2-8707-6eb57c2fef5a for instance with vm_state building and task_state spawning. [ 866.334286] env[63515]: DEBUG nova.compute.manager [req-acfbfd38-5e9b-4768-8fd0-cf6c44fb7b64 req-c9384dab-48d7-4b86-a97b-4ac56ab4ff99 service nova] [instance: 0ee9cce2-94b9-4038-ab38-1b3f96a864bc] Received event network-changed-2fc6eecd-e0cd-41f2-8707-6eb57c2fef5a {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 866.334286] env[63515]: DEBUG nova.compute.manager [req-acfbfd38-5e9b-4768-8fd0-cf6c44fb7b64 req-c9384dab-48d7-4b86-a97b-4ac56ab4ff99 service nova] [instance: 0ee9cce2-94b9-4038-ab38-1b3f96a864bc] Refreshing instance network info cache due to event network-changed-2fc6eecd-e0cd-41f2-8707-6eb57c2fef5a. {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 866.334286] env[63515]: DEBUG oslo_concurrency.lockutils [req-acfbfd38-5e9b-4768-8fd0-cf6c44fb7b64 req-c9384dab-48d7-4b86-a97b-4ac56ab4ff99 service nova] Acquiring lock "refresh_cache-0ee9cce2-94b9-4038-ab38-1b3f96a864bc" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 866.341592] env[63515]: DEBUG oslo_vmware.api [None req-eae7e17b-7393-4e13-b575-a8d38e798918 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52cf7d62-f820-1b4d-5765-42b567876084, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.595277] env[63515]: DEBUG nova.network.neutron [None req-baaf7dbf-037b-4d17-a0eb-d252e746ead6 tempest-InstanceActionsTestJSON-247065106 tempest-InstanceActionsTestJSON-247065106-project-member] [instance: 0ee9cce2-94b9-4038-ab38-1b3f96a864bc] Updating instance_info_cache with network_info: [{"id": "2fc6eecd-e0cd-41f2-8707-6eb57c2fef5a", "address": "fa:16:3e:e1:7e:62", "network": {"id": "9411d1db-b1ec-4a75-ad97-8942272c078e", "bridge": "br-int", "label": "tempest-InstanceActionsTestJSON-1741013332-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "486afecc6edc4a94b18a1e34e4f92fe5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fc48e29b-113c-4849-850c-35435eab4052", "external-id": "nsx-vlan-transportzone-958", "segmentation_id": 958, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2fc6eecd-e0", "ovs_interfaceid": "2fc6eecd-e0cd-41f2-8707-6eb57c2fef5a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 866.656158] env[63515]: DEBUG nova.scheduler.client.report [None req-3db8c3e0-5f0e-4c46-ab36-d6f97ddeb4f1 tempest-ServerActionsV293TestJSON-1193370169 tempest-ServerActionsV293TestJSON-1193370169-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 866.743028] env[63515]: DEBUG nova.compute.manager [None req-789266fe-c183-402c-aef4-222438c1eee4 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: eab16df9-7bb5-4576-bca0-769a561c5fe9] Start spawning the instance on the hypervisor. {{(pid=63515) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 866.760511] env[63515]: DEBUG oslo_vmware.api [None req-ce131312-5bf5-4677-8923-2d91405ade5d tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Task: {'id': task-1111411, 'name': CopyVirtualDisk_Task} progress is 80%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.775041] env[63515]: DEBUG nova.virt.hardware [None req-789266fe-c183-402c-aef4-222438c1eee4 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T02:52:14Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T02:51:56Z,direct_url=,disk_format='vmdk',id=8a120570-cb06-4099-b262-554ca0ad15c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b569255cc43e42e7ae3f2b4ad37c6ef4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T02:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 866.775361] env[63515]: DEBUG nova.virt.hardware [None req-789266fe-c183-402c-aef4-222438c1eee4 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Flavor limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 866.775560] env[63515]: DEBUG nova.virt.hardware [None req-789266fe-c183-402c-aef4-222438c1eee4 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Image limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 866.775702] env[63515]: DEBUG nova.virt.hardware [None req-789266fe-c183-402c-aef4-222438c1eee4 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Flavor pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 866.775898] env[63515]: DEBUG nova.virt.hardware [None req-789266fe-c183-402c-aef4-222438c1eee4 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Image pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 866.776409] env[63515]: DEBUG nova.virt.hardware [None req-789266fe-c183-402c-aef4-222438c1eee4 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 866.777212] env[63515]: DEBUG nova.virt.hardware [None req-789266fe-c183-402c-aef4-222438c1eee4 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 866.777495] env[63515]: DEBUG nova.virt.hardware [None req-789266fe-c183-402c-aef4-222438c1eee4 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 866.778018] env[63515]: DEBUG nova.virt.hardware [None req-789266fe-c183-402c-aef4-222438c1eee4 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Got 1 possible topologies {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 866.778253] env[63515]: DEBUG nova.virt.hardware [None req-789266fe-c183-402c-aef4-222438c1eee4 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 866.778447] env[63515]: DEBUG nova.virt.hardware [None req-789266fe-c183-402c-aef4-222438c1eee4 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 866.779644] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-621e7663-7674-4e3c-98a3-5cc6e8d3e601 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.792069] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab4997a9-4b8e-4c47-abde-507d3a1a1b09 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.807359] env[63515]: INFO nova.compute.manager [None req-957d8582-d135-4677-8fde-d84fe70e2761 tempest-ServerDiagnosticsV248Test-1736022867 tempest-ServerDiagnosticsV248Test-1736022867-project-member] [instance: 705b7e9b-3ab6-40ec-8cf6-d86bf0a1faf0] Took 41.05 seconds to build instance. [ 866.837224] env[63515]: DEBUG oslo_vmware.api [None req-eae7e17b-7393-4e13-b575-a8d38e798918 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52cf7d62-f820-1b4d-5765-42b567876084, 'name': SearchDatastore_Task, 'duration_secs': 0.073404} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 866.839064] env[63515]: DEBUG oslo_concurrency.lockutils [None req-eae7e17b-7393-4e13-b575-a8d38e798918 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 866.839064] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-eae7e17b-7393-4e13-b575-a8d38e798918 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk to [datastore1] 2a29f37d-e786-48d3-b126-4854d6e7c67c/2a29f37d-e786-48d3-b126-4854d6e7c67c.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 866.839064] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-9fc3d299-36e9-4704-8649-efcdf1f292fa {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.847624] env[63515]: DEBUG oslo_vmware.api [None req-eae7e17b-7393-4e13-b575-a8d38e798918 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Waiting for the task: (returnval){ [ 866.847624] env[63515]: value = "task-1111413" [ 866.847624] env[63515]: _type = "Task" [ 866.847624] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.859526] env[63515]: DEBUG oslo_vmware.api [None req-eae7e17b-7393-4e13-b575-a8d38e798918 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Task: {'id': task-1111413, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.094596] env[63515]: DEBUG nova.compute.manager [None req-1859dc3b-8bb8-4a4f-b3f4-9974d10d8e4c tempest-ServerDiagnosticsV248Test-1898563194 tempest-ServerDiagnosticsV248Test-1898563194-project-admin] [instance: 705b7e9b-3ab6-40ec-8cf6-d86bf0a1faf0] Checking state {{(pid=63515) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 867.095914] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96cb27d0-b83b-44f5-87ed-40cc2da0f147 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.099242] env[63515]: DEBUG oslo_concurrency.lockutils [None req-baaf7dbf-037b-4d17-a0eb-d252e746ead6 tempest-InstanceActionsTestJSON-247065106 tempest-InstanceActionsTestJSON-247065106-project-member] Releasing lock "refresh_cache-0ee9cce2-94b9-4038-ab38-1b3f96a864bc" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 867.099751] env[63515]: DEBUG nova.compute.manager [None req-baaf7dbf-037b-4d17-a0eb-d252e746ead6 tempest-InstanceActionsTestJSON-247065106 tempest-InstanceActionsTestJSON-247065106-project-member] [instance: 0ee9cce2-94b9-4038-ab38-1b3f96a864bc] Instance network_info: |[{"id": "2fc6eecd-e0cd-41f2-8707-6eb57c2fef5a", "address": "fa:16:3e:e1:7e:62", "network": {"id": "9411d1db-b1ec-4a75-ad97-8942272c078e", "bridge": "br-int", "label": "tempest-InstanceActionsTestJSON-1741013332-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "486afecc6edc4a94b18a1e34e4f92fe5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fc48e29b-113c-4849-850c-35435eab4052", "external-id": "nsx-vlan-transportzone-958", "segmentation_id": 958, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2fc6eecd-e0", "ovs_interfaceid": "2fc6eecd-e0cd-41f2-8707-6eb57c2fef5a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 867.100416] env[63515]: DEBUG oslo_concurrency.lockutils [req-acfbfd38-5e9b-4768-8fd0-cf6c44fb7b64 req-c9384dab-48d7-4b86-a97b-4ac56ab4ff99 service nova] Acquired lock "refresh_cache-0ee9cce2-94b9-4038-ab38-1b3f96a864bc" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 867.100617] env[63515]: DEBUG nova.network.neutron [req-acfbfd38-5e9b-4768-8fd0-cf6c44fb7b64 req-c9384dab-48d7-4b86-a97b-4ac56ab4ff99 service nova] [instance: 0ee9cce2-94b9-4038-ab38-1b3f96a864bc] Refreshing network info cache for port 2fc6eecd-e0cd-41f2-8707-6eb57c2fef5a {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 867.102017] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-baaf7dbf-037b-4d17-a0eb-d252e746ead6 tempest-InstanceActionsTestJSON-247065106 tempest-InstanceActionsTestJSON-247065106-project-member] [instance: 0ee9cce2-94b9-4038-ab38-1b3f96a864bc] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:e1:7e:62', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'fc48e29b-113c-4849-850c-35435eab4052', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '2fc6eecd-e0cd-41f2-8707-6eb57c2fef5a', 'vif_model': 'vmxnet3'}] {{(pid=63515) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 867.109590] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-baaf7dbf-037b-4d17-a0eb-d252e746ead6 tempest-InstanceActionsTestJSON-247065106 tempest-InstanceActionsTestJSON-247065106-project-member] Creating folder: Project (486afecc6edc4a94b18a1e34e4f92fe5). Parent ref: group-v243370. {{(pid=63515) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 867.111361] env[63515]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-895681e7-d2b6-4b51-967b-109a830919d8 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.115800] env[63515]: INFO nova.compute.manager [None req-1859dc3b-8bb8-4a4f-b3f4-9974d10d8e4c tempest-ServerDiagnosticsV248Test-1898563194 tempest-ServerDiagnosticsV248Test-1898563194-project-admin] [instance: 705b7e9b-3ab6-40ec-8cf6-d86bf0a1faf0] Retrieving diagnostics [ 867.116661] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dfc87461-f67a-4d70-a023-4aaa036e1d8d {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.123560] env[63515]: INFO nova.virt.vmwareapi.vm_util [None req-baaf7dbf-037b-4d17-a0eb-d252e746ead6 tempest-InstanceActionsTestJSON-247065106 tempest-InstanceActionsTestJSON-247065106-project-member] Created folder: Project (486afecc6edc4a94b18a1e34e4f92fe5) in parent group-v243370. [ 867.123772] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-baaf7dbf-037b-4d17-a0eb-d252e746ead6 tempest-InstanceActionsTestJSON-247065106 tempest-InstanceActionsTestJSON-247065106-project-member] Creating folder: Instances. Parent ref: group-v243499. {{(pid=63515) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 867.124773] env[63515]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-6a2c9576-d0d5-4874-be16-75cd0a14163d {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.160821] env[63515]: DEBUG oslo_concurrency.lockutils [None req-3db8c3e0-5f0e-4c46-ab36-d6f97ddeb4f1 tempest-ServerActionsV293TestJSON-1193370169 tempest-ServerActionsV293TestJSON-1193370169-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.452s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 867.161228] env[63515]: DEBUG nova.compute.manager [None req-3db8c3e0-5f0e-4c46-ab36-d6f97ddeb4f1 tempest-ServerActionsV293TestJSON-1193370169 tempest-ServerActionsV293TestJSON-1193370169-project-member] [instance: 2ea99c7a-8b61-4718-bee0-f4ce4fc021af] Start building networks asynchronously for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 867.166071] env[63515]: DEBUG oslo_concurrency.lockutils [None req-c9f99ed7-857d-4064-a182-30f784598d4b tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 23.485s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 867.166334] env[63515]: DEBUG oslo_concurrency.lockutils [None req-c9f99ed7-857d-4064-a182-30f784598d4b tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 867.168434] env[63515]: DEBUG oslo_concurrency.lockutils [None req-561efd36-c7e8-4824-9224-7669ba1ee132 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.378s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 867.170036] env[63515]: INFO nova.compute.claims [None req-561efd36-c7e8-4824-9224-7669ba1ee132 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 68766cd4-84be-475b-8494-d7ab43a9e969] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 867.172732] env[63515]: INFO nova.virt.vmwareapi.vm_util [None req-baaf7dbf-037b-4d17-a0eb-d252e746ead6 tempest-InstanceActionsTestJSON-247065106 tempest-InstanceActionsTestJSON-247065106-project-member] Created folder: Instances in parent group-v243499. [ 867.172980] env[63515]: DEBUG oslo.service.loopingcall [None req-baaf7dbf-037b-4d17-a0eb-d252e746ead6 tempest-InstanceActionsTestJSON-247065106 tempest-InstanceActionsTestJSON-247065106-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 867.174148] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0ee9cce2-94b9-4038-ab38-1b3f96a864bc] Creating VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 867.174594] env[63515]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-21f8a167-d8ff-46fb-a07c-a7a7399d8984 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.195785] env[63515]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 867.195785] env[63515]: value = "task-1111416" [ 867.195785] env[63515]: _type = "Task" [ 867.195785] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 867.204355] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1111416, 'name': CreateVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.218448] env[63515]: INFO nova.scheduler.client.report [None req-c9f99ed7-857d-4064-a182-30f784598d4b tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Deleted allocations for instance d2a4f66a-f1a2-4f0d-826d-0b7c4f04bd23 [ 867.262488] env[63515]: DEBUG oslo_vmware.api [None req-ce131312-5bf5-4677-8923-2d91405ade5d tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Task: {'id': task-1111411, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.318762] env[63515]: DEBUG oslo_concurrency.lockutils [None req-957d8582-d135-4677-8fde-d84fe70e2761 tempest-ServerDiagnosticsV248Test-1736022867 tempest-ServerDiagnosticsV248Test-1736022867-project-member] Lock "705b7e9b-3ab6-40ec-8cf6-d86bf0a1faf0" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 42.569s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 867.358041] env[63515]: DEBUG oslo_vmware.api [None req-eae7e17b-7393-4e13-b575-a8d38e798918 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Task: {'id': task-1111413, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.675408] env[63515]: DEBUG nova.compute.utils [None req-3db8c3e0-5f0e-4c46-ab36-d6f97ddeb4f1 tempest-ServerActionsV293TestJSON-1193370169 tempest-ServerActionsV293TestJSON-1193370169-project-member] Using /dev/sd instead of None {{(pid=63515) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 867.678383] env[63515]: DEBUG nova.compute.manager [None req-3db8c3e0-5f0e-4c46-ab36-d6f97ddeb4f1 tempest-ServerActionsV293TestJSON-1193370169 tempest-ServerActionsV293TestJSON-1193370169-project-member] [instance: 2ea99c7a-8b61-4718-bee0-f4ce4fc021af] Allocating IP information in the background. {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 867.678600] env[63515]: DEBUG nova.network.neutron [None req-3db8c3e0-5f0e-4c46-ab36-d6f97ddeb4f1 tempest-ServerActionsV293TestJSON-1193370169 tempest-ServerActionsV293TestJSON-1193370169-project-member] [instance: 2ea99c7a-8b61-4718-bee0-f4ce4fc021af] allocate_for_instance() {{(pid=63515) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 867.709444] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1111416, 'name': CreateVM_Task} progress is 25%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.729438] env[63515]: DEBUG oslo_concurrency.lockutils [None req-c9f99ed7-857d-4064-a182-30f784598d4b tempest-ListServerFiltersTestJSON-707093826 tempest-ListServerFiltersTestJSON-707093826-project-member] Lock "d2a4f66a-f1a2-4f0d-826d-0b7c4f04bd23" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 27.497s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 867.757034] env[63515]: DEBUG oslo_vmware.api [None req-ce131312-5bf5-4677-8923-2d91405ade5d tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Task: {'id': task-1111411, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.544029} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.758148] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-ce131312-5bf5-4677-8923-2d91405ade5d tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/499b6029-07e9-499c-80d5-a53d8c43ffb7/499b6029-07e9-499c-80d5-a53d8c43ffb7.vmdk to [datastore1] 87c468d9-9594-4804-b461-527f01f6118f/87c468d9-9594-4804-b461-527f01f6118f.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 867.758148] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63ec5662-96ee-42a6-813f-4af279f48d76 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.782816] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-ce131312-5bf5-4677-8923-2d91405ade5d tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] [instance: 87c468d9-9594-4804-b461-527f01f6118f] Reconfiguring VM instance instance-00000026 to attach disk [datastore1] 87c468d9-9594-4804-b461-527f01f6118f/87c468d9-9594-4804-b461-527f01f6118f.vmdk or device None with type streamOptimized {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 867.786823] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6bffe774-36eb-40ee-ad97-307fab36ebf3 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.806414] env[63515]: DEBUG oslo_vmware.api [None req-ce131312-5bf5-4677-8923-2d91405ade5d tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Waiting for the task: (returnval){ [ 867.806414] env[63515]: value = "task-1111417" [ 867.806414] env[63515]: _type = "Task" [ 867.806414] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 867.811357] env[63515]: DEBUG nova.policy [None req-3db8c3e0-5f0e-4c46-ab36-d6f97ddeb4f1 tempest-ServerActionsV293TestJSON-1193370169 tempest-ServerActionsV293TestJSON-1193370169-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'eb9a461be0ca4072be071326f32801bd', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ae6fb15713db493793797f015fe6fcf9', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63515) authorize /opt/stack/nova/nova/policy.py:201}} [ 867.819584] env[63515]: DEBUG oslo_vmware.api [None req-ce131312-5bf5-4677-8923-2d91405ade5d tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Task: {'id': task-1111417, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.858332] env[63515]: DEBUG oslo_vmware.api [None req-eae7e17b-7393-4e13-b575-a8d38e798918 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Task: {'id': task-1111413, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.184042] env[63515]: DEBUG nova.compute.manager [None req-3db8c3e0-5f0e-4c46-ab36-d6f97ddeb4f1 tempest-ServerActionsV293TestJSON-1193370169 tempest-ServerActionsV293TestJSON-1193370169-project-member] [instance: 2ea99c7a-8b61-4718-bee0-f4ce4fc021af] Start building block device mappings for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 868.210174] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1111416, 'name': CreateVM_Task} progress is 99%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.221879] env[63515]: DEBUG nova.network.neutron [req-acfbfd38-5e9b-4768-8fd0-cf6c44fb7b64 req-c9384dab-48d7-4b86-a97b-4ac56ab4ff99 service nova] [instance: 0ee9cce2-94b9-4038-ab38-1b3f96a864bc] Updated VIF entry in instance network info cache for port 2fc6eecd-e0cd-41f2-8707-6eb57c2fef5a. {{(pid=63515) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 868.222351] env[63515]: DEBUG nova.network.neutron [req-acfbfd38-5e9b-4768-8fd0-cf6c44fb7b64 req-c9384dab-48d7-4b86-a97b-4ac56ab4ff99 service nova] [instance: 0ee9cce2-94b9-4038-ab38-1b3f96a864bc] Updating instance_info_cache with network_info: [{"id": "2fc6eecd-e0cd-41f2-8707-6eb57c2fef5a", "address": "fa:16:3e:e1:7e:62", "network": {"id": "9411d1db-b1ec-4a75-ad97-8942272c078e", "bridge": "br-int", "label": "tempest-InstanceActionsTestJSON-1741013332-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "486afecc6edc4a94b18a1e34e4f92fe5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fc48e29b-113c-4849-850c-35435eab4052", "external-id": "nsx-vlan-transportzone-958", "segmentation_id": 958, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2fc6eecd-e0", "ovs_interfaceid": "2fc6eecd-e0cd-41f2-8707-6eb57c2fef5a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 868.281228] env[63515]: DEBUG nova.network.neutron [None req-3db8c3e0-5f0e-4c46-ab36-d6f97ddeb4f1 tempest-ServerActionsV293TestJSON-1193370169 tempest-ServerActionsV293TestJSON-1193370169-project-member] [instance: 2ea99c7a-8b61-4718-bee0-f4ce4fc021af] Successfully created port: 1ddac51a-f040-4ab9-b3bd-3f75ed5e1f5b {{(pid=63515) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 868.320485] env[63515]: DEBUG oslo_vmware.api [None req-ce131312-5bf5-4677-8923-2d91405ade5d tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Task: {'id': task-1111417, 'name': ReconfigVM_Task, 'duration_secs': 0.391739} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 868.323350] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-ce131312-5bf5-4677-8923-2d91405ade5d tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] [instance: 87c468d9-9594-4804-b461-527f01f6118f] Reconfigured VM instance instance-00000026 to attach disk [datastore1] 87c468d9-9594-4804-b461-527f01f6118f/87c468d9-9594-4804-b461-527f01f6118f.vmdk or device None with type streamOptimized {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 868.324799] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-ce131312-5bf5-4677-8923-2d91405ade5d tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] [instance: 87c468d9-9594-4804-b461-527f01f6118f] Block device information present: {'root_device_name': '/dev/sda', 'image': [{'disk_bus': None, 'encrypted': False, 'encryption_secret_uuid': None, 'encryption_format': None, 'size': 0, 'guest_format': None, 'device_name': '/dev/sda', 'device_type': 'disk', 'boot_index': 0, 'encryption_options': None, 'image_id': '8a120570-cb06-4099-b262-554ca0ad15c5'}], 'ephemerals': [], 'block_device_mapping': [{'connection_info': {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-243476', 'volume_id': '67893a17-820c-4c63-b387-06761ba8e0d5', 'name': 'volume-67893a17-820c-4c63-b387-06761ba8e0d5', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attached', 'instance': '87c468d9-9594-4804-b461-527f01f6118f', 'attached_at': '', 'detached_at': '', 'volume_id': '67893a17-820c-4c63-b387-06761ba8e0d5', 'serial': '67893a17-820c-4c63-b387-06761ba8e0d5'}, 'disk_bus': None, 'attachment_id': '67842379-9552-4358-b269-226527568f79', 'guest_format': None, 'mount_device': '/dev/sdb', 'delete_on_termination': False, 'device_type': None, 'boot_index': None, 'volume_type': None}], 'swap': None} {{(pid=63515) spawn /opt/stack/nova/nova/virt/vmwareapi/vmops.py:799}} [ 868.325066] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-ce131312-5bf5-4677-8923-2d91405ade5d tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] [instance: 87c468d9-9594-4804-b461-527f01f6118f] Volume attach. Driver type: vmdk {{(pid=63515) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 868.325264] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-ce131312-5bf5-4677-8923-2d91405ade5d tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] [instance: 87c468d9-9594-4804-b461-527f01f6118f] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-243476', 'volume_id': '67893a17-820c-4c63-b387-06761ba8e0d5', 'name': 'volume-67893a17-820c-4c63-b387-06761ba8e0d5', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attached', 'instance': '87c468d9-9594-4804-b461-527f01f6118f', 'attached_at': '', 'detached_at': '', 'volume_id': '67893a17-820c-4c63-b387-06761ba8e0d5', 'serial': '67893a17-820c-4c63-b387-06761ba8e0d5'} {{(pid=63515) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 868.326989] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d13d13be-390d-4e33-9d9c-6842065364c7 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.349103] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7dd79952-a1ed-499c-acf2-3a96f419259b {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.362284] env[63515]: DEBUG oslo_vmware.api [None req-eae7e17b-7393-4e13-b575-a8d38e798918 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Task: {'id': task-1111413, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.396324} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 868.378371] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-eae7e17b-7393-4e13-b575-a8d38e798918 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk to [datastore1] 2a29f37d-e786-48d3-b126-4854d6e7c67c/2a29f37d-e786-48d3-b126-4854d6e7c67c.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 868.378371] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-eae7e17b-7393-4e13-b575-a8d38e798918 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] [instance: 2a29f37d-e786-48d3-b126-4854d6e7c67c] Extending root virtual disk to 1048576 {{(pid=63515) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 868.387018] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-ce131312-5bf5-4677-8923-2d91405ade5d tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] [instance: 87c468d9-9594-4804-b461-527f01f6118f] Reconfiguring VM instance instance-00000026 to attach disk [datastore2] volume-67893a17-820c-4c63-b387-06761ba8e0d5/volume-67893a17-820c-4c63-b387-06761ba8e0d5.vmdk or device None with type thin {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 868.391558] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-3dfaa072-0bce-4cd7-b504-878c59dad7de {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.393992] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ae96c7cf-6197-4973-af37-a77e1301ab58 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.408683] env[63515]: DEBUG nova.compute.manager [req-fb9124da-55c7-4a03-bddf-d1b56386e4b0 req-2135148c-49b7-40d3-8d91-9fcd0bd74303 service nova] [instance: eab16df9-7bb5-4576-bca0-769a561c5fe9] Received event network-vif-plugged-4168f5a5-c282-4f13-9c97-0b0236bd659b {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 868.409139] env[63515]: DEBUG oslo_concurrency.lockutils [req-fb9124da-55c7-4a03-bddf-d1b56386e4b0 req-2135148c-49b7-40d3-8d91-9fcd0bd74303 service nova] Acquiring lock "eab16df9-7bb5-4576-bca0-769a561c5fe9-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 868.409376] env[63515]: DEBUG oslo_concurrency.lockutils [req-fb9124da-55c7-4a03-bddf-d1b56386e4b0 req-2135148c-49b7-40d3-8d91-9fcd0bd74303 service nova] Lock "eab16df9-7bb5-4576-bca0-769a561c5fe9-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 868.409551] env[63515]: DEBUG oslo_concurrency.lockutils [req-fb9124da-55c7-4a03-bddf-d1b56386e4b0 req-2135148c-49b7-40d3-8d91-9fcd0bd74303 service nova] Lock "eab16df9-7bb5-4576-bca0-769a561c5fe9-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 868.409852] env[63515]: DEBUG nova.compute.manager [req-fb9124da-55c7-4a03-bddf-d1b56386e4b0 req-2135148c-49b7-40d3-8d91-9fcd0bd74303 service nova] [instance: eab16df9-7bb5-4576-bca0-769a561c5fe9] No waiting events found dispatching network-vif-plugged-4168f5a5-c282-4f13-9c97-0b0236bd659b {{(pid=63515) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 868.409949] env[63515]: WARNING nova.compute.manager [req-fb9124da-55c7-4a03-bddf-d1b56386e4b0 req-2135148c-49b7-40d3-8d91-9fcd0bd74303 service nova] [instance: eab16df9-7bb5-4576-bca0-769a561c5fe9] Received unexpected event network-vif-plugged-4168f5a5-c282-4f13-9c97-0b0236bd659b for instance with vm_state building and task_state spawning. [ 868.416960] env[63515]: DEBUG oslo_vmware.api [None req-eae7e17b-7393-4e13-b575-a8d38e798918 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Waiting for the task: (returnval){ [ 868.416960] env[63515]: value = "task-1111418" [ 868.416960] env[63515]: _type = "Task" [ 868.416960] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 868.419329] env[63515]: DEBUG oslo_vmware.api [None req-ce131312-5bf5-4677-8923-2d91405ade5d tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Waiting for the task: (returnval){ [ 868.419329] env[63515]: value = "task-1111419" [ 868.419329] env[63515]: _type = "Task" [ 868.419329] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 868.431140] env[63515]: DEBUG oslo_vmware.api [None req-eae7e17b-7393-4e13-b575-a8d38e798918 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Task: {'id': task-1111418, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.436854] env[63515]: DEBUG oslo_vmware.api [None req-ce131312-5bf5-4677-8923-2d91405ade5d tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Task: {'id': task-1111419, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.580011] env[63515]: DEBUG nova.network.neutron [None req-789266fe-c183-402c-aef4-222438c1eee4 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: eab16df9-7bb5-4576-bca0-769a561c5fe9] Successfully updated port: 4168f5a5-c282-4f13-9c97-0b0236bd659b {{(pid=63515) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 868.611358] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1337092e-f076-450c-ab1a-912da33e6a06 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.620302] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-771fdcd4-7990-471d-8b00-df1d9daa7bbd {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.653303] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15cafac7-33c9-42b3-9c35-98bafb69d3a9 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.661776] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e22a8645-b450-414b-a94d-06d47a9dc869 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.678867] env[63515]: DEBUG nova.compute.provider_tree [None req-561efd36-c7e8-4824-9224-7669ba1ee132 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 868.692772] env[63515]: INFO nova.virt.block_device [None req-3db8c3e0-5f0e-4c46-ab36-d6f97ddeb4f1 tempest-ServerActionsV293TestJSON-1193370169 tempest-ServerActionsV293TestJSON-1193370169-project-member] [instance: 2ea99c7a-8b61-4718-bee0-f4ce4fc021af] Booting with volume 211b1a87-a01e-4d13-aa27-91391bf9cde1 at /dev/sda [ 868.712539] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1111416, 'name': CreateVM_Task, 'duration_secs': 1.111512} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 868.712736] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0ee9cce2-94b9-4038-ab38-1b3f96a864bc] Created VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 868.713440] env[63515]: DEBUG oslo_concurrency.lockutils [None req-baaf7dbf-037b-4d17-a0eb-d252e746ead6 tempest-InstanceActionsTestJSON-247065106 tempest-InstanceActionsTestJSON-247065106-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 868.713612] env[63515]: DEBUG oslo_concurrency.lockutils [None req-baaf7dbf-037b-4d17-a0eb-d252e746ead6 tempest-InstanceActionsTestJSON-247065106 tempest-InstanceActionsTestJSON-247065106-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 868.714050] env[63515]: DEBUG oslo_concurrency.lockutils [None req-baaf7dbf-037b-4d17-a0eb-d252e746ead6 tempest-InstanceActionsTestJSON-247065106 tempest-InstanceActionsTestJSON-247065106-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 868.714201] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-299e86f8-bf3d-4d68-844a-c946ac6ff923 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.719617] env[63515]: DEBUG oslo_vmware.api [None req-baaf7dbf-037b-4d17-a0eb-d252e746ead6 tempest-InstanceActionsTestJSON-247065106 tempest-InstanceActionsTestJSON-247065106-project-member] Waiting for the task: (returnval){ [ 868.719617] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]528923f5-835a-325c-236c-2e8eda90a994" [ 868.719617] env[63515]: _type = "Task" [ 868.719617] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 868.731569] env[63515]: DEBUG oslo_concurrency.lockutils [req-acfbfd38-5e9b-4768-8fd0-cf6c44fb7b64 req-c9384dab-48d7-4b86-a97b-4ac56ab4ff99 service nova] Releasing lock "refresh_cache-0ee9cce2-94b9-4038-ab38-1b3f96a864bc" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 868.732078] env[63515]: DEBUG oslo_vmware.api [None req-baaf7dbf-037b-4d17-a0eb-d252e746ead6 tempest-InstanceActionsTestJSON-247065106 tempest-InstanceActionsTestJSON-247065106-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]528923f5-835a-325c-236c-2e8eda90a994, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.741167] env[63515]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-aa6118ed-4e04-4c13-8aa1-edc4a28d6abb {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.749524] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13fd61c3-425b-459c-9d72-141b634c0baf {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.782937] env[63515]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-db238637-48d7-45c7-9a6a-2281ec733d8b {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.791096] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70d59157-f1c1-403e-b442-37d6ead4789c {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.821467] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11cd216a-8424-49da-aaec-a27110470b85 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.828621] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20fbc609-351e-4667-beac-ec9f0e781dac {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.842109] env[63515]: DEBUG nova.virt.block_device [None req-3db8c3e0-5f0e-4c46-ab36-d6f97ddeb4f1 tempest-ServerActionsV293TestJSON-1193370169 tempest-ServerActionsV293TestJSON-1193370169-project-member] [instance: 2ea99c7a-8b61-4718-bee0-f4ce4fc021af] Updating existing volume attachment record: 899e81f3-63dd-4e06-9a5e-d90a6c020f76 {{(pid=63515) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 868.928585] env[63515]: DEBUG oslo_vmware.api [None req-eae7e17b-7393-4e13-b575-a8d38e798918 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Task: {'id': task-1111418, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.099556} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 868.928997] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-eae7e17b-7393-4e13-b575-a8d38e798918 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] [instance: 2a29f37d-e786-48d3-b126-4854d6e7c67c] Extended root virtual disk {{(pid=63515) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 868.929931] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-161f4ec0-3a43-4470-86f3-90fbc751829a {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.935466] env[63515]: DEBUG oslo_vmware.api [None req-ce131312-5bf5-4677-8923-2d91405ade5d tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Task: {'id': task-1111419, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.957056] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-eae7e17b-7393-4e13-b575-a8d38e798918 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] [instance: 2a29f37d-e786-48d3-b126-4854d6e7c67c] Reconfiguring VM instance instance-00000045 to attach disk [datastore1] 2a29f37d-e786-48d3-b126-4854d6e7c67c/2a29f37d-e786-48d3-b126-4854d6e7c67c.vmdk or device None with type sparse {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 868.957698] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-416ffc81-87b7-4786-99b7-8d04a02e27ad {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.978136] env[63515]: DEBUG oslo_vmware.api [None req-eae7e17b-7393-4e13-b575-a8d38e798918 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Waiting for the task: (returnval){ [ 868.978136] env[63515]: value = "task-1111420" [ 868.978136] env[63515]: _type = "Task" [ 868.978136] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 868.985142] env[63515]: DEBUG oslo_vmware.api [None req-eae7e17b-7393-4e13-b575-a8d38e798918 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Task: {'id': task-1111420, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.082520] env[63515]: DEBUG oslo_concurrency.lockutils [None req-789266fe-c183-402c-aef4-222438c1eee4 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Acquiring lock "refresh_cache-eab16df9-7bb5-4576-bca0-769a561c5fe9" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 869.082520] env[63515]: DEBUG oslo_concurrency.lockutils [None req-789266fe-c183-402c-aef4-222438c1eee4 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Acquired lock "refresh_cache-eab16df9-7bb5-4576-bca0-769a561c5fe9" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 869.082520] env[63515]: DEBUG nova.network.neutron [None req-789266fe-c183-402c-aef4-222438c1eee4 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: eab16df9-7bb5-4576-bca0-769a561c5fe9] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 869.183738] env[63515]: DEBUG nova.scheduler.client.report [None req-561efd36-c7e8-4824-9224-7669ba1ee132 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 869.230331] env[63515]: DEBUG oslo_vmware.api [None req-baaf7dbf-037b-4d17-a0eb-d252e746ead6 tempest-InstanceActionsTestJSON-247065106 tempest-InstanceActionsTestJSON-247065106-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]528923f5-835a-325c-236c-2e8eda90a994, 'name': SearchDatastore_Task, 'duration_secs': 0.018545} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 869.230484] env[63515]: DEBUG oslo_concurrency.lockutils [None req-baaf7dbf-037b-4d17-a0eb-d252e746ead6 tempest-InstanceActionsTestJSON-247065106 tempest-InstanceActionsTestJSON-247065106-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 869.230733] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-baaf7dbf-037b-4d17-a0eb-d252e746ead6 tempest-InstanceActionsTestJSON-247065106 tempest-InstanceActionsTestJSON-247065106-project-member] [instance: 0ee9cce2-94b9-4038-ab38-1b3f96a864bc] Processing image 8a120570-cb06-4099-b262-554ca0ad15c5 {{(pid=63515) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 869.230988] env[63515]: DEBUG oslo_concurrency.lockutils [None req-baaf7dbf-037b-4d17-a0eb-d252e746ead6 tempest-InstanceActionsTestJSON-247065106 tempest-InstanceActionsTestJSON-247065106-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 869.231157] env[63515]: DEBUG oslo_concurrency.lockutils [None req-baaf7dbf-037b-4d17-a0eb-d252e746ead6 tempest-InstanceActionsTestJSON-247065106 tempest-InstanceActionsTestJSON-247065106-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 869.231360] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-baaf7dbf-037b-4d17-a0eb-d252e746ead6 tempest-InstanceActionsTestJSON-247065106 tempest-InstanceActionsTestJSON-247065106-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 869.231648] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-fc7ec200-fed3-44bd-ad06-fae1ef7b335b {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.241486] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-baaf7dbf-037b-4d17-a0eb-d252e746ead6 tempest-InstanceActionsTestJSON-247065106 tempest-InstanceActionsTestJSON-247065106-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 869.241705] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-baaf7dbf-037b-4d17-a0eb-d252e746ead6 tempest-InstanceActionsTestJSON-247065106 tempest-InstanceActionsTestJSON-247065106-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63515) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 869.242625] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e8cbc60d-4287-4b20-bfd1-5fe9fdff6f2a {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.248330] env[63515]: DEBUG oslo_vmware.api [None req-baaf7dbf-037b-4d17-a0eb-d252e746ead6 tempest-InstanceActionsTestJSON-247065106 tempest-InstanceActionsTestJSON-247065106-project-member] Waiting for the task: (returnval){ [ 869.248330] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]52dfa565-12a6-a514-da69-3de6009f2831" [ 869.248330] env[63515]: _type = "Task" [ 869.248330] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 869.256432] env[63515]: DEBUG oslo_vmware.api [None req-baaf7dbf-037b-4d17-a0eb-d252e746ead6 tempest-InstanceActionsTestJSON-247065106 tempest-InstanceActionsTestJSON-247065106-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52dfa565-12a6-a514-da69-3de6009f2831, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.431948] env[63515]: DEBUG oslo_vmware.api [None req-ce131312-5bf5-4677-8923-2d91405ade5d tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Task: {'id': task-1111419, 'name': ReconfigVM_Task} progress is 99%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.487674] env[63515]: DEBUG oslo_vmware.api [None req-eae7e17b-7393-4e13-b575-a8d38e798918 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Task: {'id': task-1111420, 'name': ReconfigVM_Task, 'duration_secs': 0.311878} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 869.487984] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-eae7e17b-7393-4e13-b575-a8d38e798918 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] [instance: 2a29f37d-e786-48d3-b126-4854d6e7c67c] Reconfigured VM instance instance-00000045 to attach disk [datastore1] 2a29f37d-e786-48d3-b126-4854d6e7c67c/2a29f37d-e786-48d3-b126-4854d6e7c67c.vmdk or device None with type sparse {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 869.488695] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-8061ff99-1321-424d-87b3-b7fc6bdd9c56 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.497210] env[63515]: DEBUG oslo_vmware.api [None req-eae7e17b-7393-4e13-b575-a8d38e798918 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Waiting for the task: (returnval){ [ 869.497210] env[63515]: value = "task-1111421" [ 869.497210] env[63515]: _type = "Task" [ 869.497210] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 869.503740] env[63515]: DEBUG oslo_vmware.api [None req-eae7e17b-7393-4e13-b575-a8d38e798918 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Task: {'id': task-1111421, 'name': Rename_Task} progress is 5%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.627402] env[63515]: DEBUG nova.network.neutron [None req-789266fe-c183-402c-aef4-222438c1eee4 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: eab16df9-7bb5-4576-bca0-769a561c5fe9] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 869.689403] env[63515]: DEBUG oslo_concurrency.lockutils [None req-561efd36-c7e8-4824-9224-7669ba1ee132 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.521s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 869.689927] env[63515]: DEBUG nova.compute.manager [None req-561efd36-c7e8-4824-9224-7669ba1ee132 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 68766cd4-84be-475b-8494-d7ab43a9e969] Start building networks asynchronously for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 869.693424] env[63515]: DEBUG oslo_concurrency.lockutils [None req-3fd65342-5e9e-46b6-830a-47df91928006 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.412s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 869.694861] env[63515]: INFO nova.compute.claims [None req-3fd65342-5e9e-46b6-830a-47df91928006 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: 0967283f-8dda-4692-b6b7-601545cabf2f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 869.759601] env[63515]: DEBUG oslo_vmware.api [None req-baaf7dbf-037b-4d17-a0eb-d252e746ead6 tempest-InstanceActionsTestJSON-247065106 tempest-InstanceActionsTestJSON-247065106-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52dfa565-12a6-a514-da69-3de6009f2831, 'name': SearchDatastore_Task, 'duration_secs': 0.018666} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 869.760593] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-661cd5fe-0091-4831-877d-2092eb323fa1 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.770395] env[63515]: DEBUG oslo_vmware.api [None req-baaf7dbf-037b-4d17-a0eb-d252e746ead6 tempest-InstanceActionsTestJSON-247065106 tempest-InstanceActionsTestJSON-247065106-project-member] Waiting for the task: (returnval){ [ 869.770395] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]522b794c-a387-c211-ba0a-f9cfdabd3b07" [ 869.770395] env[63515]: _type = "Task" [ 869.770395] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 869.784244] env[63515]: DEBUG oslo_vmware.api [None req-baaf7dbf-037b-4d17-a0eb-d252e746ead6 tempest-InstanceActionsTestJSON-247065106 tempest-InstanceActionsTestJSON-247065106-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]522b794c-a387-c211-ba0a-f9cfdabd3b07, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.934121] env[63515]: DEBUG oslo_vmware.api [None req-ce131312-5bf5-4677-8923-2d91405ade5d tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Task: {'id': task-1111419, 'name': ReconfigVM_Task, 'duration_secs': 1.052263} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 869.935078] env[63515]: DEBUG nova.network.neutron [None req-789266fe-c183-402c-aef4-222438c1eee4 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: eab16df9-7bb5-4576-bca0-769a561c5fe9] Updating instance_info_cache with network_info: [{"id": "4168f5a5-c282-4f13-9c97-0b0236bd659b", "address": "fa:16:3e:a0:6a:50", "network": {"id": "88dfb29d-41e0-4df7-8f70-e52cda05620c", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1355578400-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7d4ada9ec35f42b19c6480a9101d21a4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "357d2811-e990-4985-9f9e-b158d10d3699", "external-id": "nsx-vlan-transportzone-641", "segmentation_id": 641, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4168f5a5-c2", "ovs_interfaceid": "4168f5a5-c282-4f13-9c97-0b0236bd659b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 869.936463] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-ce131312-5bf5-4677-8923-2d91405ade5d tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] [instance: 87c468d9-9594-4804-b461-527f01f6118f] Reconfigured VM instance instance-00000026 to attach disk [datastore2] volume-67893a17-820c-4c63-b387-06761ba8e0d5/volume-67893a17-820c-4c63-b387-06761ba8e0d5.vmdk or device None with type thin {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 869.943561] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7bb599c3-975b-435e-8d83-d356300f6c8c {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.962341] env[63515]: DEBUG oslo_vmware.api [None req-ce131312-5bf5-4677-8923-2d91405ade5d tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Waiting for the task: (returnval){ [ 869.962341] env[63515]: value = "task-1111422" [ 869.962341] env[63515]: _type = "Task" [ 869.962341] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 869.975566] env[63515]: DEBUG oslo_vmware.api [None req-ce131312-5bf5-4677-8923-2d91405ade5d tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Task: {'id': task-1111422, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.004119] env[63515]: DEBUG oslo_vmware.api [None req-eae7e17b-7393-4e13-b575-a8d38e798918 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Task: {'id': task-1111421, 'name': Rename_Task, 'duration_secs': 0.301149} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 870.006122] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-eae7e17b-7393-4e13-b575-a8d38e798918 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] [instance: 2a29f37d-e786-48d3-b126-4854d6e7c67c] Powering on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 870.006122] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ce636d47-b599-42ea-9369-5a47cf21f03f {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.011476] env[63515]: DEBUG oslo_vmware.api [None req-eae7e17b-7393-4e13-b575-a8d38e798918 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Waiting for the task: (returnval){ [ 870.011476] env[63515]: value = "task-1111423" [ 870.011476] env[63515]: _type = "Task" [ 870.011476] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 870.019846] env[63515]: DEBUG oslo_vmware.api [None req-eae7e17b-7393-4e13-b575-a8d38e798918 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Task: {'id': task-1111423, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.035511] env[63515]: DEBUG nova.compute.manager [req-0a8f42ce-bb10-4891-9d39-efc6a88674e0 req-6ae66b6e-0dde-4ad9-b02b-362c22cd2768 service nova] [instance: 2ea99c7a-8b61-4718-bee0-f4ce4fc021af] Received event network-vif-plugged-1ddac51a-f040-4ab9-b3bd-3f75ed5e1f5b {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 870.035771] env[63515]: DEBUG oslo_concurrency.lockutils [req-0a8f42ce-bb10-4891-9d39-efc6a88674e0 req-6ae66b6e-0dde-4ad9-b02b-362c22cd2768 service nova] Acquiring lock "2ea99c7a-8b61-4718-bee0-f4ce4fc021af-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 870.035996] env[63515]: DEBUG oslo_concurrency.lockutils [req-0a8f42ce-bb10-4891-9d39-efc6a88674e0 req-6ae66b6e-0dde-4ad9-b02b-362c22cd2768 service nova] Lock "2ea99c7a-8b61-4718-bee0-f4ce4fc021af-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 870.036191] env[63515]: DEBUG oslo_concurrency.lockutils [req-0a8f42ce-bb10-4891-9d39-efc6a88674e0 req-6ae66b6e-0dde-4ad9-b02b-362c22cd2768 service nova] Lock "2ea99c7a-8b61-4718-bee0-f4ce4fc021af-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 870.036367] env[63515]: DEBUG nova.compute.manager [req-0a8f42ce-bb10-4891-9d39-efc6a88674e0 req-6ae66b6e-0dde-4ad9-b02b-362c22cd2768 service nova] [instance: 2ea99c7a-8b61-4718-bee0-f4ce4fc021af] No waiting events found dispatching network-vif-plugged-1ddac51a-f040-4ab9-b3bd-3f75ed5e1f5b {{(pid=63515) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 870.036556] env[63515]: WARNING nova.compute.manager [req-0a8f42ce-bb10-4891-9d39-efc6a88674e0 req-6ae66b6e-0dde-4ad9-b02b-362c22cd2768 service nova] [instance: 2ea99c7a-8b61-4718-bee0-f4ce4fc021af] Received unexpected event network-vif-plugged-1ddac51a-f040-4ab9-b3bd-3f75ed5e1f5b for instance with vm_state building and task_state block_device_mapping. [ 870.204676] env[63515]: DEBUG nova.compute.utils [None req-561efd36-c7e8-4824-9224-7669ba1ee132 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Using /dev/sd instead of None {{(pid=63515) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 870.212830] env[63515]: DEBUG nova.compute.manager [None req-561efd36-c7e8-4824-9224-7669ba1ee132 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 68766cd4-84be-475b-8494-d7ab43a9e969] Allocating IP information in the background. {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 870.213061] env[63515]: DEBUG nova.network.neutron [None req-561efd36-c7e8-4824-9224-7669ba1ee132 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 68766cd4-84be-475b-8494-d7ab43a9e969] allocate_for_instance() {{(pid=63515) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 870.223515] env[63515]: DEBUG nova.network.neutron [None req-3db8c3e0-5f0e-4c46-ab36-d6f97ddeb4f1 tempest-ServerActionsV293TestJSON-1193370169 tempest-ServerActionsV293TestJSON-1193370169-project-member] [instance: 2ea99c7a-8b61-4718-bee0-f4ce4fc021af] Successfully updated port: 1ddac51a-f040-4ab9-b3bd-3f75ed5e1f5b {{(pid=63515) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 870.281616] env[63515]: DEBUG oslo_vmware.api [None req-baaf7dbf-037b-4d17-a0eb-d252e746ead6 tempest-InstanceActionsTestJSON-247065106 tempest-InstanceActionsTestJSON-247065106-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]522b794c-a387-c211-ba0a-f9cfdabd3b07, 'name': SearchDatastore_Task, 'duration_secs': 0.024134} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 870.281897] env[63515]: DEBUG oslo_concurrency.lockutils [None req-baaf7dbf-037b-4d17-a0eb-d252e746ead6 tempest-InstanceActionsTestJSON-247065106 tempest-InstanceActionsTestJSON-247065106-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 870.282509] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-baaf7dbf-037b-4d17-a0eb-d252e746ead6 tempest-InstanceActionsTestJSON-247065106 tempest-InstanceActionsTestJSON-247065106-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk to [datastore1] 0ee9cce2-94b9-4038-ab38-1b3f96a864bc/0ee9cce2-94b9-4038-ab38-1b3f96a864bc.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 870.282815] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ecfda338-95c0-48d8-8ab9-848a2a3e208c {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.290109] env[63515]: DEBUG oslo_vmware.api [None req-baaf7dbf-037b-4d17-a0eb-d252e746ead6 tempest-InstanceActionsTestJSON-247065106 tempest-InstanceActionsTestJSON-247065106-project-member] Waiting for the task: (returnval){ [ 870.290109] env[63515]: value = "task-1111424" [ 870.290109] env[63515]: _type = "Task" [ 870.290109] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 870.299968] env[63515]: DEBUG oslo_vmware.api [None req-baaf7dbf-037b-4d17-a0eb-d252e746ead6 tempest-InstanceActionsTestJSON-247065106 tempest-InstanceActionsTestJSON-247065106-project-member] Task: {'id': task-1111424, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.442126] env[63515]: DEBUG nova.policy [None req-561efd36-c7e8-4824-9224-7669ba1ee132 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '82fe830c309f41a4ab512a25d8481742', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b18d4daacbc84e758f9f9ae4f3f09d28', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63515) authorize /opt/stack/nova/nova/policy.py:201}} [ 870.443977] env[63515]: DEBUG oslo_concurrency.lockutils [None req-789266fe-c183-402c-aef4-222438c1eee4 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Releasing lock "refresh_cache-eab16df9-7bb5-4576-bca0-769a561c5fe9" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 870.444643] env[63515]: DEBUG nova.compute.manager [None req-789266fe-c183-402c-aef4-222438c1eee4 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: eab16df9-7bb5-4576-bca0-769a561c5fe9] Instance network_info: |[{"id": "4168f5a5-c282-4f13-9c97-0b0236bd659b", "address": "fa:16:3e:a0:6a:50", "network": {"id": "88dfb29d-41e0-4df7-8f70-e52cda05620c", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1355578400-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7d4ada9ec35f42b19c6480a9101d21a4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "357d2811-e990-4985-9f9e-b158d10d3699", "external-id": "nsx-vlan-transportzone-641", "segmentation_id": 641, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4168f5a5-c2", "ovs_interfaceid": "4168f5a5-c282-4f13-9c97-0b0236bd659b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 870.445376] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-789266fe-c183-402c-aef4-222438c1eee4 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: eab16df9-7bb5-4576-bca0-769a561c5fe9] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:a0:6a:50', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '357d2811-e990-4985-9f9e-b158d10d3699', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '4168f5a5-c282-4f13-9c97-0b0236bd659b', 'vif_model': 'vmxnet3'}] {{(pid=63515) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 870.460504] env[63515]: DEBUG oslo.service.loopingcall [None req-789266fe-c183-402c-aef4-222438c1eee4 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 870.474635] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: eab16df9-7bb5-4576-bca0-769a561c5fe9] Creating VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 870.475157] env[63515]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-017cbf9a-c6a0-4f16-9e26-9a6085cd221e {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.507794] env[63515]: DEBUG oslo_vmware.api [None req-ce131312-5bf5-4677-8923-2d91405ade5d tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Task: {'id': task-1111422, 'name': ReconfigVM_Task, 'duration_secs': 0.195343} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 870.510812] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-ce131312-5bf5-4677-8923-2d91405ade5d tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] [instance: 87c468d9-9594-4804-b461-527f01f6118f] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-243476', 'volume_id': '67893a17-820c-4c63-b387-06761ba8e0d5', 'name': 'volume-67893a17-820c-4c63-b387-06761ba8e0d5', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attached', 'instance': '87c468d9-9594-4804-b461-527f01f6118f', 'attached_at': '', 'detached_at': '', 'volume_id': '67893a17-820c-4c63-b387-06761ba8e0d5', 'serial': '67893a17-820c-4c63-b387-06761ba8e0d5'} {{(pid=63515) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 870.510812] env[63515]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 870.510812] env[63515]: value = "task-1111425" [ 870.510812] env[63515]: _type = "Task" [ 870.510812] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 870.510812] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-bf2eac8b-146d-40ae-b5ab-d1dfccc42022 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.531307] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1111425, 'name': CreateVM_Task} progress is 6%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.537289] env[63515]: DEBUG oslo_vmware.api [None req-eae7e17b-7393-4e13-b575-a8d38e798918 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Task: {'id': task-1111423, 'name': PowerOnVM_Task, 'duration_secs': 0.493346} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 870.537810] env[63515]: DEBUG oslo_vmware.api [None req-ce131312-5bf5-4677-8923-2d91405ade5d tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Waiting for the task: (returnval){ [ 870.537810] env[63515]: value = "task-1111426" [ 870.537810] env[63515]: _type = "Task" [ 870.537810] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 870.538108] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-eae7e17b-7393-4e13-b575-a8d38e798918 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] [instance: 2a29f37d-e786-48d3-b126-4854d6e7c67c] Powered on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 870.538356] env[63515]: INFO nova.compute.manager [None req-eae7e17b-7393-4e13-b575-a8d38e798918 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] [instance: 2a29f37d-e786-48d3-b126-4854d6e7c67c] Took 8.79 seconds to spawn the instance on the hypervisor. [ 870.538572] env[63515]: DEBUG nova.compute.manager [None req-eae7e17b-7393-4e13-b575-a8d38e798918 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] [instance: 2a29f37d-e786-48d3-b126-4854d6e7c67c] Checking state {{(pid=63515) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 870.539586] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bacdf63e-8133-4992-94ca-5012faff685c {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.563213] env[63515]: DEBUG nova.compute.manager [req-0800dd56-10d9-41d3-a726-2a49d9e20d0a req-7fc15a16-9815-40ed-8122-54f0278b135c service nova] [instance: eab16df9-7bb5-4576-bca0-769a561c5fe9] Received event network-changed-4168f5a5-c282-4f13-9c97-0b0236bd659b {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 870.563213] env[63515]: DEBUG nova.compute.manager [req-0800dd56-10d9-41d3-a726-2a49d9e20d0a req-7fc15a16-9815-40ed-8122-54f0278b135c service nova] [instance: eab16df9-7bb5-4576-bca0-769a561c5fe9] Refreshing instance network info cache due to event network-changed-4168f5a5-c282-4f13-9c97-0b0236bd659b. {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 870.563213] env[63515]: DEBUG oslo_concurrency.lockutils [req-0800dd56-10d9-41d3-a726-2a49d9e20d0a req-7fc15a16-9815-40ed-8122-54f0278b135c service nova] Acquiring lock "refresh_cache-eab16df9-7bb5-4576-bca0-769a561c5fe9" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 870.563213] env[63515]: DEBUG oslo_concurrency.lockutils [req-0800dd56-10d9-41d3-a726-2a49d9e20d0a req-7fc15a16-9815-40ed-8122-54f0278b135c service nova] Acquired lock "refresh_cache-eab16df9-7bb5-4576-bca0-769a561c5fe9" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 870.563213] env[63515]: DEBUG nova.network.neutron [req-0800dd56-10d9-41d3-a726-2a49d9e20d0a req-7fc15a16-9815-40ed-8122-54f0278b135c service nova] [instance: eab16df9-7bb5-4576-bca0-769a561c5fe9] Refreshing network info cache for port 4168f5a5-c282-4f13-9c97-0b0236bd659b {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 870.713698] env[63515]: DEBUG nova.compute.manager [None req-561efd36-c7e8-4824-9224-7669ba1ee132 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 68766cd4-84be-475b-8494-d7ab43a9e969] Start building block device mappings for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 870.731954] env[63515]: DEBUG oslo_concurrency.lockutils [None req-3db8c3e0-5f0e-4c46-ab36-d6f97ddeb4f1 tempest-ServerActionsV293TestJSON-1193370169 tempest-ServerActionsV293TestJSON-1193370169-project-member] Acquiring lock "refresh_cache-2ea99c7a-8b61-4718-bee0-f4ce4fc021af" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 870.732186] env[63515]: DEBUG oslo_concurrency.lockutils [None req-3db8c3e0-5f0e-4c46-ab36-d6f97ddeb4f1 tempest-ServerActionsV293TestJSON-1193370169 tempest-ServerActionsV293TestJSON-1193370169-project-member] Acquired lock "refresh_cache-2ea99c7a-8b61-4718-bee0-f4ce4fc021af" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 870.735176] env[63515]: DEBUG nova.network.neutron [None req-3db8c3e0-5f0e-4c46-ab36-d6f97ddeb4f1 tempest-ServerActionsV293TestJSON-1193370169 tempest-ServerActionsV293TestJSON-1193370169-project-member] [instance: 2ea99c7a-8b61-4718-bee0-f4ce4fc021af] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 870.807622] env[63515]: DEBUG oslo_vmware.api [None req-baaf7dbf-037b-4d17-a0eb-d252e746ead6 tempest-InstanceActionsTestJSON-247065106 tempest-InstanceActionsTestJSON-247065106-project-member] Task: {'id': task-1111424, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.975204] env[63515]: DEBUG nova.compute.manager [None req-3db8c3e0-5f0e-4c46-ab36-d6f97ddeb4f1 tempest-ServerActionsV293TestJSON-1193370169 tempest-ServerActionsV293TestJSON-1193370169-project-member] [instance: 2ea99c7a-8b61-4718-bee0-f4ce4fc021af] Start spawning the instance on the hypervisor. {{(pid=63515) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 870.975870] env[63515]: DEBUG nova.virt.hardware [None req-3db8c3e0-5f0e-4c46-ab36-d6f97ddeb4f1 tempest-ServerActionsV293TestJSON-1193370169 tempest-ServerActionsV293TestJSON-1193370169-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T02:52:14Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=0,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=1073741824,status='active',tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 870.976054] env[63515]: DEBUG nova.virt.hardware [None req-3db8c3e0-5f0e-4c46-ab36-d6f97ddeb4f1 tempest-ServerActionsV293TestJSON-1193370169 tempest-ServerActionsV293TestJSON-1193370169-project-member] Flavor limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 870.976222] env[63515]: DEBUG nova.virt.hardware [None req-3db8c3e0-5f0e-4c46-ab36-d6f97ddeb4f1 tempest-ServerActionsV293TestJSON-1193370169 tempest-ServerActionsV293TestJSON-1193370169-project-member] Image limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 870.976452] env[63515]: DEBUG nova.virt.hardware [None req-3db8c3e0-5f0e-4c46-ab36-d6f97ddeb4f1 tempest-ServerActionsV293TestJSON-1193370169 tempest-ServerActionsV293TestJSON-1193370169-project-member] Flavor pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 870.976561] env[63515]: DEBUG nova.virt.hardware [None req-3db8c3e0-5f0e-4c46-ab36-d6f97ddeb4f1 tempest-ServerActionsV293TestJSON-1193370169 tempest-ServerActionsV293TestJSON-1193370169-project-member] Image pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 870.976740] env[63515]: DEBUG nova.virt.hardware [None req-3db8c3e0-5f0e-4c46-ab36-d6f97ddeb4f1 tempest-ServerActionsV293TestJSON-1193370169 tempest-ServerActionsV293TestJSON-1193370169-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 870.976994] env[63515]: DEBUG nova.virt.hardware [None req-3db8c3e0-5f0e-4c46-ab36-d6f97ddeb4f1 tempest-ServerActionsV293TestJSON-1193370169 tempest-ServerActionsV293TestJSON-1193370169-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 870.984763] env[63515]: DEBUG nova.virt.hardware [None req-3db8c3e0-5f0e-4c46-ab36-d6f97ddeb4f1 tempest-ServerActionsV293TestJSON-1193370169 tempest-ServerActionsV293TestJSON-1193370169-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 870.984763] env[63515]: DEBUG nova.virt.hardware [None req-3db8c3e0-5f0e-4c46-ab36-d6f97ddeb4f1 tempest-ServerActionsV293TestJSON-1193370169 tempest-ServerActionsV293TestJSON-1193370169-project-member] Got 1 possible topologies {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 870.984763] env[63515]: DEBUG nova.virt.hardware [None req-3db8c3e0-5f0e-4c46-ab36-d6f97ddeb4f1 tempest-ServerActionsV293TestJSON-1193370169 tempest-ServerActionsV293TestJSON-1193370169-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 870.984763] env[63515]: DEBUG nova.virt.hardware [None req-3db8c3e0-5f0e-4c46-ab36-d6f97ddeb4f1 tempest-ServerActionsV293TestJSON-1193370169 tempest-ServerActionsV293TestJSON-1193370169-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 870.987273] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bca3b56b-94be-4c67-9673-84613f7b225e {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.997106] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4828af26-947d-4eed-aa67-d066ed9fa1d0 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.026234] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1111425, 'name': CreateVM_Task, 'duration_secs': 0.435652} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 871.026409] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: eab16df9-7bb5-4576-bca0-769a561c5fe9] Created VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 871.027142] env[63515]: DEBUG oslo_concurrency.lockutils [None req-789266fe-c183-402c-aef4-222438c1eee4 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 871.027360] env[63515]: DEBUG oslo_concurrency.lockutils [None req-789266fe-c183-402c-aef4-222438c1eee4 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 871.027648] env[63515]: DEBUG oslo_concurrency.lockutils [None req-789266fe-c183-402c-aef4-222438c1eee4 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 871.027911] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ed6be664-855c-4ad5-bc0e-c2bad2096002 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.033032] env[63515]: DEBUG oslo_vmware.api [None req-789266fe-c183-402c-aef4-222438c1eee4 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Waiting for the task: (returnval){ [ 871.033032] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]526cf637-eecd-d8cc-f0ae-ad2ef1f57f6e" [ 871.033032] env[63515]: _type = "Task" [ 871.033032] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 871.045386] env[63515]: DEBUG oslo_vmware.api [None req-789266fe-c183-402c-aef4-222438c1eee4 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]526cf637-eecd-d8cc-f0ae-ad2ef1f57f6e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.051074] env[63515]: DEBUG oslo_vmware.api [None req-ce131312-5bf5-4677-8923-2d91405ade5d tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Task: {'id': task-1111426, 'name': Rename_Task, 'duration_secs': 0.231533} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 871.051074] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-ce131312-5bf5-4677-8923-2d91405ade5d tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] [instance: 87c468d9-9594-4804-b461-527f01f6118f] Powering on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 871.051564] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f0740884-228a-4be6-9061-544069574c1f {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.061829] env[63515]: DEBUG oslo_vmware.api [None req-ce131312-5bf5-4677-8923-2d91405ade5d tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Waiting for the task: (returnval){ [ 871.061829] env[63515]: value = "task-1111427" [ 871.061829] env[63515]: _type = "Task" [ 871.061829] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 871.080451] env[63515]: INFO nova.compute.manager [None req-eae7e17b-7393-4e13-b575-a8d38e798918 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] [instance: 2a29f37d-e786-48d3-b126-4854d6e7c67c] Took 43.31 seconds to build instance. [ 871.089205] env[63515]: DEBUG oslo_vmware.api [None req-ce131312-5bf5-4677-8923-2d91405ade5d tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Task: {'id': task-1111427, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.119581] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3226405b-89d2-457c-b37d-2eb06e78401f {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.127659] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6dd5c89c-51e8-4a82-a8e1-eed769ba4aea {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.137911] env[63515]: DEBUG nova.network.neutron [None req-561efd36-c7e8-4824-9224-7669ba1ee132 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 68766cd4-84be-475b-8494-d7ab43a9e969] Successfully created port: 28b59076-140f-46b5-b4a9-0bf530a1d43d {{(pid=63515) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 871.172209] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed212a28-c7d9-4d4d-8ed1-54d9ed47f96f {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.185874] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9397ffb-bf37-41f2-978e-e979488537d9 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.201379] env[63515]: DEBUG nova.compute.provider_tree [None req-3fd65342-5e9e-46b6-830a-47df91928006 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 871.303396] env[63515]: DEBUG oslo_vmware.api [None req-baaf7dbf-037b-4d17-a0eb-d252e746ead6 tempest-InstanceActionsTestJSON-247065106 tempest-InstanceActionsTestJSON-247065106-project-member] Task: {'id': task-1111424, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.586428} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 871.303861] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-baaf7dbf-037b-4d17-a0eb-d252e746ead6 tempest-InstanceActionsTestJSON-247065106 tempest-InstanceActionsTestJSON-247065106-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk to [datastore1] 0ee9cce2-94b9-4038-ab38-1b3f96a864bc/0ee9cce2-94b9-4038-ab38-1b3f96a864bc.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 871.304060] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-baaf7dbf-037b-4d17-a0eb-d252e746ead6 tempest-InstanceActionsTestJSON-247065106 tempest-InstanceActionsTestJSON-247065106-project-member] [instance: 0ee9cce2-94b9-4038-ab38-1b3f96a864bc] Extending root virtual disk to 1048576 {{(pid=63515) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 871.304361] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-6c169339-b288-4191-b930-629609587972 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.313809] env[63515]: DEBUG oslo_vmware.api [None req-baaf7dbf-037b-4d17-a0eb-d252e746ead6 tempest-InstanceActionsTestJSON-247065106 tempest-InstanceActionsTestJSON-247065106-project-member] Waiting for the task: (returnval){ [ 871.313809] env[63515]: value = "task-1111428" [ 871.313809] env[63515]: _type = "Task" [ 871.313809] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 871.324833] env[63515]: DEBUG oslo_vmware.api [None req-baaf7dbf-037b-4d17-a0eb-d252e746ead6 tempest-InstanceActionsTestJSON-247065106 tempest-InstanceActionsTestJSON-247065106-project-member] Task: {'id': task-1111428, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.339045] env[63515]: DEBUG nova.network.neutron [None req-3db8c3e0-5f0e-4c46-ab36-d6f97ddeb4f1 tempest-ServerActionsV293TestJSON-1193370169 tempest-ServerActionsV293TestJSON-1193370169-project-member] [instance: 2ea99c7a-8b61-4718-bee0-f4ce4fc021af] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 871.545122] env[63515]: DEBUG oslo_vmware.api [None req-789266fe-c183-402c-aef4-222438c1eee4 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]526cf637-eecd-d8cc-f0ae-ad2ef1f57f6e, 'name': SearchDatastore_Task, 'duration_secs': 0.024664} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 871.545696] env[63515]: DEBUG oslo_concurrency.lockutils [None req-789266fe-c183-402c-aef4-222438c1eee4 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 871.546075] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-789266fe-c183-402c-aef4-222438c1eee4 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: eab16df9-7bb5-4576-bca0-769a561c5fe9] Processing image 8a120570-cb06-4099-b262-554ca0ad15c5 {{(pid=63515) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 871.546337] env[63515]: DEBUG oslo_concurrency.lockutils [None req-789266fe-c183-402c-aef4-222438c1eee4 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 871.546490] env[63515]: DEBUG oslo_concurrency.lockutils [None req-789266fe-c183-402c-aef4-222438c1eee4 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 871.546674] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-789266fe-c183-402c-aef4-222438c1eee4 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 871.546997] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-12a4108a-d4d9-4dc7-b528-8946f7bf2bd6 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.556395] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-789266fe-c183-402c-aef4-222438c1eee4 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 871.556530] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-789266fe-c183-402c-aef4-222438c1eee4 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63515) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 871.557559] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-db666e42-978b-44f1-9886-b7dc9e6d099f {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.562283] env[63515]: DEBUG oslo_vmware.api [None req-789266fe-c183-402c-aef4-222438c1eee4 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Waiting for the task: (returnval){ [ 871.562283] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]52fda87f-0481-d061-0280-1394e6c76912" [ 871.562283] env[63515]: _type = "Task" [ 871.562283] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 871.576197] env[63515]: DEBUG oslo_vmware.api [None req-789266fe-c183-402c-aef4-222438c1eee4 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52fda87f-0481-d061-0280-1394e6c76912, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.581843] env[63515]: DEBUG oslo_concurrency.lockutils [None req-eae7e17b-7393-4e13-b575-a8d38e798918 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Lock "2a29f37d-e786-48d3-b126-4854d6e7c67c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 44.826s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 871.581929] env[63515]: DEBUG oslo_vmware.api [None req-ce131312-5bf5-4677-8923-2d91405ade5d tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Task: {'id': task-1111427, 'name': PowerOnVM_Task} progress is 89%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.705531] env[63515]: DEBUG nova.scheduler.client.report [None req-3fd65342-5e9e-46b6-830a-47df91928006 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 871.735515] env[63515]: DEBUG nova.compute.manager [None req-561efd36-c7e8-4824-9224-7669ba1ee132 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 68766cd4-84be-475b-8494-d7ab43a9e969] Start spawning the instance on the hypervisor. {{(pid=63515) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 871.765896] env[63515]: DEBUG nova.virt.hardware [None req-561efd36-c7e8-4824-9224-7669ba1ee132 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T02:52:14Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T02:51:56Z,direct_url=,disk_format='vmdk',id=8a120570-cb06-4099-b262-554ca0ad15c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b569255cc43e42e7ae3f2b4ad37c6ef4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T02:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 871.766280] env[63515]: DEBUG nova.virt.hardware [None req-561efd36-c7e8-4824-9224-7669ba1ee132 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Flavor limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 871.766402] env[63515]: DEBUG nova.virt.hardware [None req-561efd36-c7e8-4824-9224-7669ba1ee132 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Image limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 871.766588] env[63515]: DEBUG nova.virt.hardware [None req-561efd36-c7e8-4824-9224-7669ba1ee132 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Flavor pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 871.766744] env[63515]: DEBUG nova.virt.hardware [None req-561efd36-c7e8-4824-9224-7669ba1ee132 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Image pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 871.766942] env[63515]: DEBUG nova.virt.hardware [None req-561efd36-c7e8-4824-9224-7669ba1ee132 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 871.767182] env[63515]: DEBUG nova.virt.hardware [None req-561efd36-c7e8-4824-9224-7669ba1ee132 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 871.767354] env[63515]: DEBUG nova.virt.hardware [None req-561efd36-c7e8-4824-9224-7669ba1ee132 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 871.767525] env[63515]: DEBUG nova.virt.hardware [None req-561efd36-c7e8-4824-9224-7669ba1ee132 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Got 1 possible topologies {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 871.767694] env[63515]: DEBUG nova.virt.hardware [None req-561efd36-c7e8-4824-9224-7669ba1ee132 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 871.767872] env[63515]: DEBUG nova.virt.hardware [None req-561efd36-c7e8-4824-9224-7669ba1ee132 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 871.768955] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce0cff3e-b974-42c2-ac78-274f4c455123 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.778608] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2a7575b-b04f-44db-a0f2-f017cb4da8d8 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.819644] env[63515]: DEBUG nova.network.neutron [req-0800dd56-10d9-41d3-a726-2a49d9e20d0a req-7fc15a16-9815-40ed-8122-54f0278b135c service nova] [instance: eab16df9-7bb5-4576-bca0-769a561c5fe9] Updated VIF entry in instance network info cache for port 4168f5a5-c282-4f13-9c97-0b0236bd659b. {{(pid=63515) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 871.820198] env[63515]: DEBUG nova.network.neutron [req-0800dd56-10d9-41d3-a726-2a49d9e20d0a req-7fc15a16-9815-40ed-8122-54f0278b135c service nova] [instance: eab16df9-7bb5-4576-bca0-769a561c5fe9] Updating instance_info_cache with network_info: [{"id": "4168f5a5-c282-4f13-9c97-0b0236bd659b", "address": "fa:16:3e:a0:6a:50", "network": {"id": "88dfb29d-41e0-4df7-8f70-e52cda05620c", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1355578400-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7d4ada9ec35f42b19c6480a9101d21a4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "357d2811-e990-4985-9f9e-b158d10d3699", "external-id": "nsx-vlan-transportzone-641", "segmentation_id": 641, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4168f5a5-c2", "ovs_interfaceid": "4168f5a5-c282-4f13-9c97-0b0236bd659b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 871.827663] env[63515]: DEBUG oslo_vmware.api [None req-baaf7dbf-037b-4d17-a0eb-d252e746ead6 tempest-InstanceActionsTestJSON-247065106 tempest-InstanceActionsTestJSON-247065106-project-member] Task: {'id': task-1111428, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.073511} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 871.827663] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-baaf7dbf-037b-4d17-a0eb-d252e746ead6 tempest-InstanceActionsTestJSON-247065106 tempest-InstanceActionsTestJSON-247065106-project-member] [instance: 0ee9cce2-94b9-4038-ab38-1b3f96a864bc] Extended root virtual disk {{(pid=63515) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 871.828150] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-262d03d3-2a70-414a-b677-643ccd5309fe {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.858023] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-baaf7dbf-037b-4d17-a0eb-d252e746ead6 tempest-InstanceActionsTestJSON-247065106 tempest-InstanceActionsTestJSON-247065106-project-member] [instance: 0ee9cce2-94b9-4038-ab38-1b3f96a864bc] Reconfiguring VM instance instance-00000046 to attach disk [datastore1] 0ee9cce2-94b9-4038-ab38-1b3f96a864bc/0ee9cce2-94b9-4038-ab38-1b3f96a864bc.vmdk or device None with type sparse {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 871.858023] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1c3acb60-0c67-4190-865a-37734d725c73 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.876368] env[63515]: DEBUG oslo_vmware.api [None req-baaf7dbf-037b-4d17-a0eb-d252e746ead6 tempest-InstanceActionsTestJSON-247065106 tempest-InstanceActionsTestJSON-247065106-project-member] Waiting for the task: (returnval){ [ 871.876368] env[63515]: value = "task-1111429" [ 871.876368] env[63515]: _type = "Task" [ 871.876368] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 871.881491] env[63515]: DEBUG nova.network.neutron [None req-3db8c3e0-5f0e-4c46-ab36-d6f97ddeb4f1 tempest-ServerActionsV293TestJSON-1193370169 tempest-ServerActionsV293TestJSON-1193370169-project-member] [instance: 2ea99c7a-8b61-4718-bee0-f4ce4fc021af] Updating instance_info_cache with network_info: [{"id": "1ddac51a-f040-4ab9-b3bd-3f75ed5e1f5b", "address": "fa:16:3e:f0:73:5e", "network": {"id": "4803e9b2-07bf-4e7d-b71a-5ea8fd8cde6c", "bridge": "br-int", "label": "tempest-ServerActionsV293TestJSON-439688245-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ae6fb15713db493793797f015fe6fcf9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "48512b02-ad5c-4105-ba7d-fd4775acf8e1", "external-id": "nsx-vlan-transportzone-516", "segmentation_id": 516, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1ddac51a-f0", "ovs_interfaceid": "1ddac51a-f040-4ab9-b3bd-3f75ed5e1f5b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 871.887805] env[63515]: DEBUG oslo_vmware.api [None req-baaf7dbf-037b-4d17-a0eb-d252e746ead6 tempest-InstanceActionsTestJSON-247065106 tempest-InstanceActionsTestJSON-247065106-project-member] Task: {'id': task-1111429, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.075914] env[63515]: DEBUG oslo_vmware.api [None req-789266fe-c183-402c-aef4-222438c1eee4 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52fda87f-0481-d061-0280-1394e6c76912, 'name': SearchDatastore_Task, 'duration_secs': 0.041284} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 872.078360] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cfa3b7dd-1fa4-4c7e-9513-d36981d4d53f {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.086044] env[63515]: DEBUG oslo_vmware.api [None req-ce131312-5bf5-4677-8923-2d91405ade5d tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Task: {'id': task-1111427, 'name': PowerOnVM_Task, 'duration_secs': 0.624609} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 872.088473] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-ce131312-5bf5-4677-8923-2d91405ade5d tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] [instance: 87c468d9-9594-4804-b461-527f01f6118f] Powered on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 872.090176] env[63515]: DEBUG oslo_vmware.api [None req-789266fe-c183-402c-aef4-222438c1eee4 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Waiting for the task: (returnval){ [ 872.090176] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]52366aba-8dfc-c723-8160-45edbd8fd122" [ 872.090176] env[63515]: _type = "Task" [ 872.090176] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 872.098667] env[63515]: DEBUG oslo_vmware.api [None req-789266fe-c183-402c-aef4-222438c1eee4 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52366aba-8dfc-c723-8160-45edbd8fd122, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.195546] env[63515]: DEBUG nova.compute.manager [req-9fe21552-0dc2-42a8-a20e-41b15f707aef req-68187fb5-dd1f-4fd8-aeee-3e95959174d3 service nova] [instance: 2ea99c7a-8b61-4718-bee0-f4ce4fc021af] Received event network-changed-1ddac51a-f040-4ab9-b3bd-3f75ed5e1f5b {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 872.195752] env[63515]: DEBUG nova.compute.manager [req-9fe21552-0dc2-42a8-a20e-41b15f707aef req-68187fb5-dd1f-4fd8-aeee-3e95959174d3 service nova] [instance: 2ea99c7a-8b61-4718-bee0-f4ce4fc021af] Refreshing instance network info cache due to event network-changed-1ddac51a-f040-4ab9-b3bd-3f75ed5e1f5b. {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 872.196016] env[63515]: DEBUG oslo_concurrency.lockutils [req-9fe21552-0dc2-42a8-a20e-41b15f707aef req-68187fb5-dd1f-4fd8-aeee-3e95959174d3 service nova] Acquiring lock "refresh_cache-2ea99c7a-8b61-4718-bee0-f4ce4fc021af" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 872.211436] env[63515]: DEBUG oslo_concurrency.lockutils [None req-3fd65342-5e9e-46b6-830a-47df91928006 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.518s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 872.211868] env[63515]: DEBUG nova.compute.manager [None req-3fd65342-5e9e-46b6-830a-47df91928006 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: 0967283f-8dda-4692-b6b7-601545cabf2f] Start building networks asynchronously for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 872.215251] env[63515]: DEBUG oslo_concurrency.lockutils [None req-593b35b4-5d74-4c58-a93e-9cad00fc0b2d tempest-ServerGroupTestJSON-385159797 tempest-ServerGroupTestJSON-385159797-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.319s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 872.216785] env[63515]: INFO nova.compute.claims [None req-593b35b4-5d74-4c58-a93e-9cad00fc0b2d tempest-ServerGroupTestJSON-385159797 tempest-ServerGroupTestJSON-385159797-project-member] [instance: a2e014a2-9e0f-4250-aa38-3a7154149903] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 872.227121] env[63515]: DEBUG nova.compute.manager [None req-ce131312-5bf5-4677-8923-2d91405ade5d tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] [instance: 87c468d9-9594-4804-b461-527f01f6118f] Checking state {{(pid=63515) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 872.230344] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f592c3c-1bb1-4233-9249-cf992f26f47c {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.325177] env[63515]: DEBUG oslo_concurrency.lockutils [req-0800dd56-10d9-41d3-a726-2a49d9e20d0a req-7fc15a16-9815-40ed-8122-54f0278b135c service nova] Releasing lock "refresh_cache-eab16df9-7bb5-4576-bca0-769a561c5fe9" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 872.383488] env[63515]: DEBUG oslo_concurrency.lockutils [None req-3db8c3e0-5f0e-4c46-ab36-d6f97ddeb4f1 tempest-ServerActionsV293TestJSON-1193370169 tempest-ServerActionsV293TestJSON-1193370169-project-member] Releasing lock "refresh_cache-2ea99c7a-8b61-4718-bee0-f4ce4fc021af" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 872.384212] env[63515]: DEBUG nova.compute.manager [None req-3db8c3e0-5f0e-4c46-ab36-d6f97ddeb4f1 tempest-ServerActionsV293TestJSON-1193370169 tempest-ServerActionsV293TestJSON-1193370169-project-member] [instance: 2ea99c7a-8b61-4718-bee0-f4ce4fc021af] Instance network_info: |[{"id": "1ddac51a-f040-4ab9-b3bd-3f75ed5e1f5b", "address": "fa:16:3e:f0:73:5e", "network": {"id": "4803e9b2-07bf-4e7d-b71a-5ea8fd8cde6c", "bridge": "br-int", "label": "tempest-ServerActionsV293TestJSON-439688245-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ae6fb15713db493793797f015fe6fcf9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "48512b02-ad5c-4105-ba7d-fd4775acf8e1", "external-id": "nsx-vlan-transportzone-516", "segmentation_id": 516, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1ddac51a-f0", "ovs_interfaceid": "1ddac51a-f040-4ab9-b3bd-3f75ed5e1f5b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 872.384706] env[63515]: DEBUG oslo_concurrency.lockutils [req-9fe21552-0dc2-42a8-a20e-41b15f707aef req-68187fb5-dd1f-4fd8-aeee-3e95959174d3 service nova] Acquired lock "refresh_cache-2ea99c7a-8b61-4718-bee0-f4ce4fc021af" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 872.385030] env[63515]: DEBUG nova.network.neutron [req-9fe21552-0dc2-42a8-a20e-41b15f707aef req-68187fb5-dd1f-4fd8-aeee-3e95959174d3 service nova] [instance: 2ea99c7a-8b61-4718-bee0-f4ce4fc021af] Refreshing network info cache for port 1ddac51a-f040-4ab9-b3bd-3f75ed5e1f5b {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 872.386517] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-3db8c3e0-5f0e-4c46-ab36-d6f97ddeb4f1 tempest-ServerActionsV293TestJSON-1193370169 tempest-ServerActionsV293TestJSON-1193370169-project-member] [instance: 2ea99c7a-8b61-4718-bee0-f4ce4fc021af] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:f0:73:5e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '48512b02-ad5c-4105-ba7d-fd4775acf8e1', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '1ddac51a-f040-4ab9-b3bd-3f75ed5e1f5b', 'vif_model': 'vmxnet3'}] {{(pid=63515) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 872.398878] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-3db8c3e0-5f0e-4c46-ab36-d6f97ddeb4f1 tempest-ServerActionsV293TestJSON-1193370169 tempest-ServerActionsV293TestJSON-1193370169-project-member] Creating folder: Project (ae6fb15713db493793797f015fe6fcf9). Parent ref: group-v243370. {{(pid=63515) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 872.400536] env[63515]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-2e4e09ff-f6b1-4ec3-81fe-472a0a64d595 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.408132] env[63515]: DEBUG oslo_vmware.api [None req-baaf7dbf-037b-4d17-a0eb-d252e746ead6 tempest-InstanceActionsTestJSON-247065106 tempest-InstanceActionsTestJSON-247065106-project-member] Task: {'id': task-1111429, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.422364] env[63515]: WARNING suds.client [-] Web service reported a SOAP processing fault using an unexpected HTTP status code 200. Reporting as an internal server error. [ 872.422364] env[63515]: DEBUG oslo_vmware.api [-] Fault list: [DuplicateName] {{(pid=63515) _invoke_api /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:337}} [ 872.422364] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-3db8c3e0-5f0e-4c46-ab36-d6f97ddeb4f1 tempest-ServerActionsV293TestJSON-1193370169 tempest-ServerActionsV293TestJSON-1193370169-project-member] Folder already exists: Project (ae6fb15713db493793797f015fe6fcf9). Parent ref: group-v243370. {{(pid=63515) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1609}} [ 872.422364] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-3db8c3e0-5f0e-4c46-ab36-d6f97ddeb4f1 tempest-ServerActionsV293TestJSON-1193370169 tempest-ServerActionsV293TestJSON-1193370169-project-member] Creating folder: Instances. Parent ref: group-v243477. {{(pid=63515) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 872.422364] env[63515]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-4d156e62-53a2-4850-bad0-70fa8d1bc4ab {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.434675] env[63515]: INFO nova.virt.vmwareapi.vm_util [None req-3db8c3e0-5f0e-4c46-ab36-d6f97ddeb4f1 tempest-ServerActionsV293TestJSON-1193370169 tempest-ServerActionsV293TestJSON-1193370169-project-member] Created folder: Instances in parent group-v243477. [ 872.434675] env[63515]: DEBUG oslo.service.loopingcall [None req-3db8c3e0-5f0e-4c46-ab36-d6f97ddeb4f1 tempest-ServerActionsV293TestJSON-1193370169 tempest-ServerActionsV293TestJSON-1193370169-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 872.434675] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2ea99c7a-8b61-4718-bee0-f4ce4fc021af] Creating VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 872.434675] env[63515]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-66aca2bb-1aec-4b8a-97b8-813524654ac1 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.454231] env[63515]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 872.454231] env[63515]: value = "task-1111432" [ 872.454231] env[63515]: _type = "Task" [ 872.454231] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 872.468297] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1111432, 'name': CreateVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.600947] env[63515]: DEBUG oslo_vmware.api [None req-789266fe-c183-402c-aef4-222438c1eee4 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52366aba-8dfc-c723-8160-45edbd8fd122, 'name': SearchDatastore_Task, 'duration_secs': 0.073362} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 872.601260] env[63515]: DEBUG oslo_concurrency.lockutils [None req-789266fe-c183-402c-aef4-222438c1eee4 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 872.601522] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-789266fe-c183-402c-aef4-222438c1eee4 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk to [datastore1] eab16df9-7bb5-4576-bca0-769a561c5fe9/eab16df9-7bb5-4576-bca0-769a561c5fe9.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 872.601796] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-17f2ecea-842d-442d-91af-4255a7bf418c {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.608637] env[63515]: DEBUG oslo_vmware.api [None req-789266fe-c183-402c-aef4-222438c1eee4 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Waiting for the task: (returnval){ [ 872.608637] env[63515]: value = "task-1111433" [ 872.608637] env[63515]: _type = "Task" [ 872.608637] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 872.617683] env[63515]: DEBUG oslo_vmware.api [None req-789266fe-c183-402c-aef4-222438c1eee4 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Task: {'id': task-1111433, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.717129] env[63515]: DEBUG nova.compute.utils [None req-3fd65342-5e9e-46b6-830a-47df91928006 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Using /dev/sd instead of None {{(pid=63515) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 872.718627] env[63515]: DEBUG nova.compute.manager [None req-3fd65342-5e9e-46b6-830a-47df91928006 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: 0967283f-8dda-4692-b6b7-601545cabf2f] Allocating IP information in the background. {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 872.718872] env[63515]: DEBUG nova.network.neutron [None req-3fd65342-5e9e-46b6-830a-47df91928006 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: 0967283f-8dda-4692-b6b7-601545cabf2f] allocate_for_instance() {{(pid=63515) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 872.752475] env[63515]: DEBUG oslo_concurrency.lockutils [None req-ce131312-5bf5-4677-8923-2d91405ade5d tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Lock "87c468d9-9594-4804-b461-527f01f6118f" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 57.706s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 872.802702] env[63515]: DEBUG nova.policy [None req-3fd65342-5e9e-46b6-830a-47df91928006 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9b19bb2d32d84c019541c3b2e711a202', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '299fdeff647f486390366d5bbf911518', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63515) authorize /opt/stack/nova/nova/policy.py:201}} [ 872.892779] env[63515]: DEBUG oslo_vmware.api [None req-baaf7dbf-037b-4d17-a0eb-d252e746ead6 tempest-InstanceActionsTestJSON-247065106 tempest-InstanceActionsTestJSON-247065106-project-member] Task: {'id': task-1111429, 'name': ReconfigVM_Task, 'duration_secs': 0.712017} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 872.893301] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-baaf7dbf-037b-4d17-a0eb-d252e746ead6 tempest-InstanceActionsTestJSON-247065106 tempest-InstanceActionsTestJSON-247065106-project-member] [instance: 0ee9cce2-94b9-4038-ab38-1b3f96a864bc] Reconfigured VM instance instance-00000046 to attach disk [datastore1] 0ee9cce2-94b9-4038-ab38-1b3f96a864bc/0ee9cce2-94b9-4038-ab38-1b3f96a864bc.vmdk or device None with type sparse {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 872.894214] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6961bc3b-c9f1-40c6-b950-3c54545c4eb1 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.904111] env[63515]: DEBUG oslo_vmware.api [None req-baaf7dbf-037b-4d17-a0eb-d252e746ead6 tempest-InstanceActionsTestJSON-247065106 tempest-InstanceActionsTestJSON-247065106-project-member] Waiting for the task: (returnval){ [ 872.904111] env[63515]: value = "task-1111434" [ 872.904111] env[63515]: _type = "Task" [ 872.904111] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 872.914430] env[63515]: DEBUG oslo_vmware.api [None req-baaf7dbf-037b-4d17-a0eb-d252e746ead6 tempest-InstanceActionsTestJSON-247065106 tempest-InstanceActionsTestJSON-247065106-project-member] Task: {'id': task-1111434, 'name': Rename_Task} progress is 5%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.970426] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1111432, 'name': CreateVM_Task, 'duration_secs': 0.321926} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 872.974024] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2ea99c7a-8b61-4718-bee0-f4ce4fc021af] Created VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 872.976387] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-3db8c3e0-5f0e-4c46-ab36-d6f97ddeb4f1 tempest-ServerActionsV293TestJSON-1193370169 tempest-ServerActionsV293TestJSON-1193370169-project-member] [instance: 2ea99c7a-8b61-4718-bee0-f4ce4fc021af] Block device information present: {'root_device_name': '/dev/sda', 'image': [], 'ephemerals': [], 'block_device_mapping': [{'connection_info': {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-243483', 'volume_id': '211b1a87-a01e-4d13-aa27-91391bf9cde1', 'name': 'volume-211b1a87-a01e-4d13-aa27-91391bf9cde1', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '2ea99c7a-8b61-4718-bee0-f4ce4fc021af', 'attached_at': '', 'detached_at': '', 'volume_id': '211b1a87-a01e-4d13-aa27-91391bf9cde1', 'serial': '211b1a87-a01e-4d13-aa27-91391bf9cde1'}, 'disk_bus': None, 'attachment_id': '899e81f3-63dd-4e06-9a5e-d90a6c020f76', 'guest_format': None, 'mount_device': '/dev/sda', 'delete_on_termination': True, 'device_type': None, 'boot_index': 0, 'volume_type': None}], 'swap': None} {{(pid=63515) spawn /opt/stack/nova/nova/virt/vmwareapi/vmops.py:799}} [ 872.976387] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-3db8c3e0-5f0e-4c46-ab36-d6f97ddeb4f1 tempest-ServerActionsV293TestJSON-1193370169 tempest-ServerActionsV293TestJSON-1193370169-project-member] [instance: 2ea99c7a-8b61-4718-bee0-f4ce4fc021af] Root volume attach. Driver type: vmdk {{(pid=63515) attach_root_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:661}} [ 872.976387] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc34d6f6-93f4-4427-87dc-38c2ea16d1eb {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.992590] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a10c7c6-113b-459f-88ea-3f206c3ac010 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.011750] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab067a45-89c1-414d-b2a6-62eeb9d4c8d9 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.021479] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.RelocateVM_Task with opID=oslo.vmware-4764da37-6c22-4911-9e63-3fd7d83a1d72 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.031716] env[63515]: DEBUG oslo_vmware.api [None req-3db8c3e0-5f0e-4c46-ab36-d6f97ddeb4f1 tempest-ServerActionsV293TestJSON-1193370169 tempest-ServerActionsV293TestJSON-1193370169-project-member] Waiting for the task: (returnval){ [ 873.031716] env[63515]: value = "task-1111435" [ 873.031716] env[63515]: _type = "Task" [ 873.031716] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 873.041535] env[63515]: DEBUG oslo_vmware.api [None req-3db8c3e0-5f0e-4c46-ab36-d6f97ddeb4f1 tempest-ServerActionsV293TestJSON-1193370169 tempest-ServerActionsV293TestJSON-1193370169-project-member] Task: {'id': task-1111435, 'name': RelocateVM_Task} progress is 5%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.121535] env[63515]: DEBUG oslo_vmware.api [None req-789266fe-c183-402c-aef4-222438c1eee4 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Task: {'id': task-1111433, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.211456] env[63515]: DEBUG nova.network.neutron [req-9fe21552-0dc2-42a8-a20e-41b15f707aef req-68187fb5-dd1f-4fd8-aeee-3e95959174d3 service nova] [instance: 2ea99c7a-8b61-4718-bee0-f4ce4fc021af] Updated VIF entry in instance network info cache for port 1ddac51a-f040-4ab9-b3bd-3f75ed5e1f5b. {{(pid=63515) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 873.211958] env[63515]: DEBUG nova.network.neutron [req-9fe21552-0dc2-42a8-a20e-41b15f707aef req-68187fb5-dd1f-4fd8-aeee-3e95959174d3 service nova] [instance: 2ea99c7a-8b61-4718-bee0-f4ce4fc021af] Updating instance_info_cache with network_info: [{"id": "1ddac51a-f040-4ab9-b3bd-3f75ed5e1f5b", "address": "fa:16:3e:f0:73:5e", "network": {"id": "4803e9b2-07bf-4e7d-b71a-5ea8fd8cde6c", "bridge": "br-int", "label": "tempest-ServerActionsV293TestJSON-439688245-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ae6fb15713db493793797f015fe6fcf9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "48512b02-ad5c-4105-ba7d-fd4775acf8e1", "external-id": "nsx-vlan-transportzone-516", "segmentation_id": 516, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1ddac51a-f0", "ovs_interfaceid": "1ddac51a-f040-4ab9-b3bd-3f75ed5e1f5b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 873.222533] env[63515]: DEBUG nova.compute.manager [None req-3fd65342-5e9e-46b6-830a-47df91928006 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: 0967283f-8dda-4692-b6b7-601545cabf2f] Start building block device mappings for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 873.411296] env[63515]: DEBUG nova.network.neutron [None req-3fd65342-5e9e-46b6-830a-47df91928006 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: 0967283f-8dda-4692-b6b7-601545cabf2f] Successfully created port: 66571248-83b1-46c5-b892-f84ab3882407 {{(pid=63515) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 873.417443] env[63515]: DEBUG oslo_vmware.api [None req-baaf7dbf-037b-4d17-a0eb-d252e746ead6 tempest-InstanceActionsTestJSON-247065106 tempest-InstanceActionsTestJSON-247065106-project-member] Task: {'id': task-1111434, 'name': Rename_Task, 'duration_secs': 0.354004} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 873.417755] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-baaf7dbf-037b-4d17-a0eb-d252e746ead6 tempest-InstanceActionsTestJSON-247065106 tempest-InstanceActionsTestJSON-247065106-project-member] [instance: 0ee9cce2-94b9-4038-ab38-1b3f96a864bc] Powering on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 873.418047] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-4e35e2b9-3857-4308-b807-cc0869aa87d7 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.425330] env[63515]: DEBUG oslo_vmware.api [None req-baaf7dbf-037b-4d17-a0eb-d252e746ead6 tempest-InstanceActionsTestJSON-247065106 tempest-InstanceActionsTestJSON-247065106-project-member] Waiting for the task: (returnval){ [ 873.425330] env[63515]: value = "task-1111436" [ 873.425330] env[63515]: _type = "Task" [ 873.425330] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 873.435738] env[63515]: DEBUG oslo_vmware.api [None req-baaf7dbf-037b-4d17-a0eb-d252e746ead6 tempest-InstanceActionsTestJSON-247065106 tempest-InstanceActionsTestJSON-247065106-project-member] Task: {'id': task-1111436, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.548178] env[63515]: DEBUG oslo_vmware.api [None req-3db8c3e0-5f0e-4c46-ab36-d6f97ddeb4f1 tempest-ServerActionsV293TestJSON-1193370169 tempest-ServerActionsV293TestJSON-1193370169-project-member] Task: {'id': task-1111435, 'name': RelocateVM_Task, 'duration_secs': 0.46076} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 873.548339] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-3db8c3e0-5f0e-4c46-ab36-d6f97ddeb4f1 tempest-ServerActionsV293TestJSON-1193370169 tempest-ServerActionsV293TestJSON-1193370169-project-member] [instance: 2ea99c7a-8b61-4718-bee0-f4ce4fc021af] Volume attach. Driver type: vmdk {{(pid=63515) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 873.548651] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-3db8c3e0-5f0e-4c46-ab36-d6f97ddeb4f1 tempest-ServerActionsV293TestJSON-1193370169 tempest-ServerActionsV293TestJSON-1193370169-project-member] [instance: 2ea99c7a-8b61-4718-bee0-f4ce4fc021af] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-243483', 'volume_id': '211b1a87-a01e-4d13-aa27-91391bf9cde1', 'name': 'volume-211b1a87-a01e-4d13-aa27-91391bf9cde1', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '2ea99c7a-8b61-4718-bee0-f4ce4fc021af', 'attached_at': '', 'detached_at': '', 'volume_id': '211b1a87-a01e-4d13-aa27-91391bf9cde1', 'serial': '211b1a87-a01e-4d13-aa27-91391bf9cde1'} {{(pid=63515) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 873.549879] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e8ccc6e-65d1-4fdc-ab08-53e2f78d0468 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.582848] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0adc3ea9-144f-4fc1-a5ea-73d9d6a1a3fc {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.587759] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23ea8cb9-b3bc-42f1-b6b3-2778d44bca4c {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.614012] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-3db8c3e0-5f0e-4c46-ab36-d6f97ddeb4f1 tempest-ServerActionsV293TestJSON-1193370169 tempest-ServerActionsV293TestJSON-1193370169-project-member] [instance: 2ea99c7a-8b61-4718-bee0-f4ce4fc021af] Reconfiguring VM instance instance-00000048 to attach disk [datastore2] volume-211b1a87-a01e-4d13-aa27-91391bf9cde1/volume-211b1a87-a01e-4d13-aa27-91391bf9cde1.vmdk or device None with type thin {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 873.616505] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3e80b67a-f7ae-4b59-9f6a-5123c164aebd {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.635890] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c725fd36-05ea-455d-8887-33ce9438e097 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.644818] env[63515]: DEBUG oslo_vmware.api [None req-789266fe-c183-402c-aef4-222438c1eee4 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Task: {'id': task-1111433, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.557879} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 873.670913] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-789266fe-c183-402c-aef4-222438c1eee4 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk to [datastore1] eab16df9-7bb5-4576-bca0-769a561c5fe9/eab16df9-7bb5-4576-bca0-769a561c5fe9.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 873.671345] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-789266fe-c183-402c-aef4-222438c1eee4 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: eab16df9-7bb5-4576-bca0-769a561c5fe9] Extending root virtual disk to 1048576 {{(pid=63515) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 873.671450] env[63515]: DEBUG oslo_vmware.api [None req-3db8c3e0-5f0e-4c46-ab36-d6f97ddeb4f1 tempest-ServerActionsV293TestJSON-1193370169 tempest-ServerActionsV293TestJSON-1193370169-project-member] Waiting for the task: (returnval){ [ 873.671450] env[63515]: value = "task-1111437" [ 873.671450] env[63515]: _type = "Task" [ 873.671450] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 873.672464] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-408ba72a-cfb1-43e2-9e12-ee37ee5df2a4 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.675715] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-edb9d5a1-69a6-465e-8f4c-17d69e81116b {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.695199] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49ded08c-f51e-4af3-afca-50ec29f7626b {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.699830] env[63515]: DEBUG oslo_vmware.api [None req-789266fe-c183-402c-aef4-222438c1eee4 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Waiting for the task: (returnval){ [ 873.699830] env[63515]: value = "task-1111438" [ 873.699830] env[63515]: _type = "Task" [ 873.699830] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 873.700390] env[63515]: DEBUG oslo_vmware.api [None req-3db8c3e0-5f0e-4c46-ab36-d6f97ddeb4f1 tempest-ServerActionsV293TestJSON-1193370169 tempest-ServerActionsV293TestJSON-1193370169-project-member] Task: {'id': task-1111437, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.713871] env[63515]: DEBUG nova.compute.provider_tree [None req-593b35b4-5d74-4c58-a93e-9cad00fc0b2d tempest-ServerGroupTestJSON-385159797 tempest-ServerGroupTestJSON-385159797-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 873.716523] env[63515]: DEBUG oslo_concurrency.lockutils [req-9fe21552-0dc2-42a8-a20e-41b15f707aef req-68187fb5-dd1f-4fd8-aeee-3e95959174d3 service nova] Releasing lock "refresh_cache-2ea99c7a-8b61-4718-bee0-f4ce4fc021af" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 873.720422] env[63515]: DEBUG oslo_vmware.api [None req-789266fe-c183-402c-aef4-222438c1eee4 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Task: {'id': task-1111438, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.747798] env[63515]: DEBUG nova.network.neutron [None req-561efd36-c7e8-4824-9224-7669ba1ee132 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 68766cd4-84be-475b-8494-d7ab43a9e969] Successfully updated port: 28b59076-140f-46b5-b4a9-0bf530a1d43d {{(pid=63515) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 873.934271] env[63515]: DEBUG nova.compute.manager [req-da73337e-1271-4376-bd6d-1cc5a23bbae6 req-31146ea0-3577-4f4c-9eb2-d51d347d1194 service nova] [instance: 68766cd4-84be-475b-8494-d7ab43a9e969] Received event network-vif-plugged-28b59076-140f-46b5-b4a9-0bf530a1d43d {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 873.934271] env[63515]: DEBUG oslo_concurrency.lockutils [req-da73337e-1271-4376-bd6d-1cc5a23bbae6 req-31146ea0-3577-4f4c-9eb2-d51d347d1194 service nova] Acquiring lock "68766cd4-84be-475b-8494-d7ab43a9e969-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 873.934271] env[63515]: DEBUG oslo_concurrency.lockutils [req-da73337e-1271-4376-bd6d-1cc5a23bbae6 req-31146ea0-3577-4f4c-9eb2-d51d347d1194 service nova] Lock "68766cd4-84be-475b-8494-d7ab43a9e969-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 873.934271] env[63515]: DEBUG oslo_concurrency.lockutils [req-da73337e-1271-4376-bd6d-1cc5a23bbae6 req-31146ea0-3577-4f4c-9eb2-d51d347d1194 service nova] Lock "68766cd4-84be-475b-8494-d7ab43a9e969-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 873.934271] env[63515]: DEBUG nova.compute.manager [req-da73337e-1271-4376-bd6d-1cc5a23bbae6 req-31146ea0-3577-4f4c-9eb2-d51d347d1194 service nova] [instance: 68766cd4-84be-475b-8494-d7ab43a9e969] No waiting events found dispatching network-vif-plugged-28b59076-140f-46b5-b4a9-0bf530a1d43d {{(pid=63515) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 873.934271] env[63515]: WARNING nova.compute.manager [req-da73337e-1271-4376-bd6d-1cc5a23bbae6 req-31146ea0-3577-4f4c-9eb2-d51d347d1194 service nova] [instance: 68766cd4-84be-475b-8494-d7ab43a9e969] Received unexpected event network-vif-plugged-28b59076-140f-46b5-b4a9-0bf530a1d43d for instance with vm_state building and task_state spawning. [ 873.938095] env[63515]: DEBUG oslo_vmware.api [None req-baaf7dbf-037b-4d17-a0eb-d252e746ead6 tempest-InstanceActionsTestJSON-247065106 tempest-InstanceActionsTestJSON-247065106-project-member] Task: {'id': task-1111436, 'name': PowerOnVM_Task} progress is 88%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.185728] env[63515]: DEBUG oslo_vmware.api [None req-3db8c3e0-5f0e-4c46-ab36-d6f97ddeb4f1 tempest-ServerActionsV293TestJSON-1193370169 tempest-ServerActionsV293TestJSON-1193370169-project-member] Task: {'id': task-1111437, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.210317] env[63515]: DEBUG oslo_vmware.api [None req-789266fe-c183-402c-aef4-222438c1eee4 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Task: {'id': task-1111438, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.073391} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 874.210648] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-789266fe-c183-402c-aef4-222438c1eee4 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: eab16df9-7bb5-4576-bca0-769a561c5fe9] Extended root virtual disk {{(pid=63515) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 874.211532] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a9ac24f-49a1-40ed-bffd-6199121958dd {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.218284] env[63515]: DEBUG nova.scheduler.client.report [None req-593b35b4-5d74-4c58-a93e-9cad00fc0b2d tempest-ServerGroupTestJSON-385159797 tempest-ServerGroupTestJSON-385159797-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 874.244044] env[63515]: DEBUG nova.compute.manager [None req-3fd65342-5e9e-46b6-830a-47df91928006 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: 0967283f-8dda-4692-b6b7-601545cabf2f] Start spawning the instance on the hypervisor. {{(pid=63515) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 874.255807] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-789266fe-c183-402c-aef4-222438c1eee4 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: eab16df9-7bb5-4576-bca0-769a561c5fe9] Reconfiguring VM instance instance-00000047 to attach disk [datastore1] eab16df9-7bb5-4576-bca0-769a561c5fe9/eab16df9-7bb5-4576-bca0-769a561c5fe9.vmdk or device None with type sparse {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 874.258131] env[63515]: DEBUG oslo_concurrency.lockutils [None req-87ca4373-5e35-4af3-b945-64d29205298c tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Acquiring lock "d7e8dcdb-41cb-46fb-8b61-d251e7c2d372" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 874.258373] env[63515]: DEBUG oslo_concurrency.lockutils [None req-87ca4373-5e35-4af3-b945-64d29205298c tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Lock "d7e8dcdb-41cb-46fb-8b61-d251e7c2d372" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 874.258612] env[63515]: DEBUG oslo_concurrency.lockutils [None req-87ca4373-5e35-4af3-b945-64d29205298c tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Acquiring lock "d7e8dcdb-41cb-46fb-8b61-d251e7c2d372-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 874.258913] env[63515]: DEBUG oslo_concurrency.lockutils [None req-87ca4373-5e35-4af3-b945-64d29205298c tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Lock "d7e8dcdb-41cb-46fb-8b61-d251e7c2d372-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 874.259372] env[63515]: DEBUG oslo_concurrency.lockutils [None req-87ca4373-5e35-4af3-b945-64d29205298c tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Lock "d7e8dcdb-41cb-46fb-8b61-d251e7c2d372-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 874.261363] env[63515]: DEBUG oslo_concurrency.lockutils [None req-561efd36-c7e8-4824-9224-7669ba1ee132 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Acquiring lock "refresh_cache-68766cd4-84be-475b-8494-d7ab43a9e969" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 874.261501] env[63515]: DEBUG oslo_concurrency.lockutils [None req-561efd36-c7e8-4824-9224-7669ba1ee132 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Acquired lock "refresh_cache-68766cd4-84be-475b-8494-d7ab43a9e969" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 874.261998] env[63515]: DEBUG nova.network.neutron [None req-561efd36-c7e8-4824-9224-7669ba1ee132 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 68766cd4-84be-475b-8494-d7ab43a9e969] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 874.263184] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c6d96068-4837-4cd5-acd0-fa2485b4bf18 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.283060] env[63515]: INFO nova.compute.manager [None req-87ca4373-5e35-4af3-b945-64d29205298c tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] [instance: d7e8dcdb-41cb-46fb-8b61-d251e7c2d372] Terminating instance [ 874.287688] env[63515]: DEBUG nova.compute.manager [None req-87ca4373-5e35-4af3-b945-64d29205298c tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] [instance: d7e8dcdb-41cb-46fb-8b61-d251e7c2d372] Start destroying the instance on the hypervisor. {{(pid=63515) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 874.287974] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-87ca4373-5e35-4af3-b945-64d29205298c tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] [instance: d7e8dcdb-41cb-46fb-8b61-d251e7c2d372] Destroying instance {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 874.291349] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f019508-f58c-4772-8c8f-25d91f6b05b2 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.296454] env[63515]: DEBUG oslo_vmware.api [None req-789266fe-c183-402c-aef4-222438c1eee4 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Waiting for the task: (returnval){ [ 874.296454] env[63515]: value = "task-1111439" [ 874.296454] env[63515]: _type = "Task" [ 874.296454] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 874.302959] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-87ca4373-5e35-4af3-b945-64d29205298c tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] [instance: d7e8dcdb-41cb-46fb-8b61-d251e7c2d372] Powering off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 874.305254] env[63515]: DEBUG nova.virt.hardware [None req-3fd65342-5e9e-46b6-830a-47df91928006 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T02:52:14Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T02:51:56Z,direct_url=,disk_format='vmdk',id=8a120570-cb06-4099-b262-554ca0ad15c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b569255cc43e42e7ae3f2b4ad37c6ef4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T02:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 874.305628] env[63515]: DEBUG nova.virt.hardware [None req-3fd65342-5e9e-46b6-830a-47df91928006 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Flavor limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 874.305662] env[63515]: DEBUG nova.virt.hardware [None req-3fd65342-5e9e-46b6-830a-47df91928006 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Image limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 874.305916] env[63515]: DEBUG nova.virt.hardware [None req-3fd65342-5e9e-46b6-830a-47df91928006 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Flavor pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 874.306108] env[63515]: DEBUG nova.virt.hardware [None req-3fd65342-5e9e-46b6-830a-47df91928006 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Image pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 874.306284] env[63515]: DEBUG nova.virt.hardware [None req-3fd65342-5e9e-46b6-830a-47df91928006 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 874.306504] env[63515]: DEBUG nova.virt.hardware [None req-3fd65342-5e9e-46b6-830a-47df91928006 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 874.306672] env[63515]: DEBUG nova.virt.hardware [None req-3fd65342-5e9e-46b6-830a-47df91928006 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 874.306846] env[63515]: DEBUG nova.virt.hardware [None req-3fd65342-5e9e-46b6-830a-47df91928006 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Got 1 possible topologies {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 874.307035] env[63515]: DEBUG nova.virt.hardware [None req-3fd65342-5e9e-46b6-830a-47df91928006 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 874.307228] env[63515]: DEBUG nova.virt.hardware [None req-3fd65342-5e9e-46b6-830a-47df91928006 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 874.307865] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4af7b523-2b19-40d1-9c0c-31ea6c3efbc5 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.310168] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d506fdfd-884d-4845-acf7-ce5202647f42 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.319197] env[63515]: DEBUG oslo_vmware.api [None req-789266fe-c183-402c-aef4-222438c1eee4 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Task: {'id': task-1111439, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.325442] env[63515]: DEBUG nova.network.neutron [None req-561efd36-c7e8-4824-9224-7669ba1ee132 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 68766cd4-84be-475b-8494-d7ab43a9e969] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 874.328619] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-791d1936-4685-4736-88d0-ff4bbdf70714 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.332801] env[63515]: DEBUG oslo_vmware.api [None req-87ca4373-5e35-4af3-b945-64d29205298c tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Waiting for the task: (returnval){ [ 874.332801] env[63515]: value = "task-1111440" [ 874.332801] env[63515]: _type = "Task" [ 874.332801] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 874.350896] env[63515]: DEBUG oslo_vmware.api [None req-87ca4373-5e35-4af3-b945-64d29205298c tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Task: {'id': task-1111440, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.411854] env[63515]: DEBUG nova.compute.manager [req-b035ecee-444a-4056-9738-992e26b2fa53 req-9da2e22f-2ea5-497e-9b7e-de87dab780f3 service nova] [instance: 2a29f37d-e786-48d3-b126-4854d6e7c67c] Received event network-changed-956cd361-0d42-4914-85d3-5f75a72a68a1 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 874.411854] env[63515]: DEBUG nova.compute.manager [req-b035ecee-444a-4056-9738-992e26b2fa53 req-9da2e22f-2ea5-497e-9b7e-de87dab780f3 service nova] [instance: 2a29f37d-e786-48d3-b126-4854d6e7c67c] Refreshing instance network info cache due to event network-changed-956cd361-0d42-4914-85d3-5f75a72a68a1. {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 874.411993] env[63515]: DEBUG oslo_concurrency.lockutils [req-b035ecee-444a-4056-9738-992e26b2fa53 req-9da2e22f-2ea5-497e-9b7e-de87dab780f3 service nova] Acquiring lock "refresh_cache-2a29f37d-e786-48d3-b126-4854d6e7c67c" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 874.412113] env[63515]: DEBUG oslo_concurrency.lockutils [req-b035ecee-444a-4056-9738-992e26b2fa53 req-9da2e22f-2ea5-497e-9b7e-de87dab780f3 service nova] Acquired lock "refresh_cache-2a29f37d-e786-48d3-b126-4854d6e7c67c" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 874.412262] env[63515]: DEBUG nova.network.neutron [req-b035ecee-444a-4056-9738-992e26b2fa53 req-9da2e22f-2ea5-497e-9b7e-de87dab780f3 service nova] [instance: 2a29f37d-e786-48d3-b126-4854d6e7c67c] Refreshing network info cache for port 956cd361-0d42-4914-85d3-5f75a72a68a1 {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 874.436669] env[63515]: DEBUG oslo_vmware.api [None req-baaf7dbf-037b-4d17-a0eb-d252e746ead6 tempest-InstanceActionsTestJSON-247065106 tempest-InstanceActionsTestJSON-247065106-project-member] Task: {'id': task-1111436, 'name': PowerOnVM_Task, 'duration_secs': 0.824101} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 874.437229] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-baaf7dbf-037b-4d17-a0eb-d252e746ead6 tempest-InstanceActionsTestJSON-247065106 tempest-InstanceActionsTestJSON-247065106-project-member] [instance: 0ee9cce2-94b9-4038-ab38-1b3f96a864bc] Powered on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 874.437477] env[63515]: INFO nova.compute.manager [None req-baaf7dbf-037b-4d17-a0eb-d252e746ead6 tempest-InstanceActionsTestJSON-247065106 tempest-InstanceActionsTestJSON-247065106-project-member] [instance: 0ee9cce2-94b9-4038-ab38-1b3f96a864bc] Took 10.19 seconds to spawn the instance on the hypervisor. [ 874.437671] env[63515]: DEBUG nova.compute.manager [None req-baaf7dbf-037b-4d17-a0eb-d252e746ead6 tempest-InstanceActionsTestJSON-247065106 tempest-InstanceActionsTestJSON-247065106-project-member] [instance: 0ee9cce2-94b9-4038-ab38-1b3f96a864bc] Checking state {{(pid=63515) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 874.438599] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-609f9b3b-0dbc-4a91-8ecf-08c9f0499e7f {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.563813] env[63515]: DEBUG nova.network.neutron [None req-561efd36-c7e8-4824-9224-7669ba1ee132 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 68766cd4-84be-475b-8494-d7ab43a9e969] Updating instance_info_cache with network_info: [{"id": "28b59076-140f-46b5-b4a9-0bf530a1d43d", "address": "fa:16:3e:a3:55:0c", "network": {"id": "5758c885-e1c9-45c8-86aa-27a802101f52", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-734458138-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b18d4daacbc84e758f9f9ae4f3f09d28", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69351262-8d39-441a-85ba-3a78df436d17", "external-id": "nsx-vlan-transportzone-205", "segmentation_id": 205, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap28b59076-14", "ovs_interfaceid": "28b59076-140f-46b5-b4a9-0bf530a1d43d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 874.690155] env[63515]: DEBUG oslo_vmware.api [None req-3db8c3e0-5f0e-4c46-ab36-d6f97ddeb4f1 tempest-ServerActionsV293TestJSON-1193370169 tempest-ServerActionsV293TestJSON-1193370169-project-member] Task: {'id': task-1111437, 'name': ReconfigVM_Task, 'duration_secs': 0.762155} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 874.690444] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-3db8c3e0-5f0e-4c46-ab36-d6f97ddeb4f1 tempest-ServerActionsV293TestJSON-1193370169 tempest-ServerActionsV293TestJSON-1193370169-project-member] [instance: 2ea99c7a-8b61-4718-bee0-f4ce4fc021af] Reconfigured VM instance instance-00000048 to attach disk [datastore2] volume-211b1a87-a01e-4d13-aa27-91391bf9cde1/volume-211b1a87-a01e-4d13-aa27-91391bf9cde1.vmdk or device None with type thin {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 874.698722] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-38c9799e-66d2-4191-98f2-6fc001d3154a {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.720032] env[63515]: DEBUG oslo_vmware.api [None req-3db8c3e0-5f0e-4c46-ab36-d6f97ddeb4f1 tempest-ServerActionsV293TestJSON-1193370169 tempest-ServerActionsV293TestJSON-1193370169-project-member] Waiting for the task: (returnval){ [ 874.720032] env[63515]: value = "task-1111441" [ 874.720032] env[63515]: _type = "Task" [ 874.720032] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 874.728979] env[63515]: DEBUG oslo_vmware.api [None req-3db8c3e0-5f0e-4c46-ab36-d6f97ddeb4f1 tempest-ServerActionsV293TestJSON-1193370169 tempest-ServerActionsV293TestJSON-1193370169-project-member] Task: {'id': task-1111441, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.742492] env[63515]: DEBUG oslo_concurrency.lockutils [None req-593b35b4-5d74-4c58-a93e-9cad00fc0b2d tempest-ServerGroupTestJSON-385159797 tempest-ServerGroupTestJSON-385159797-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.527s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 874.743684] env[63515]: DEBUG oslo_concurrency.lockutils [None req-dd655d0d-92fe-48b9-b34e-d4edd8542788 tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 17.569s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 874.744052] env[63515]: DEBUG nova.objects.instance [None req-dd655d0d-92fe-48b9-b34e-d4edd8542788 tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] Lazy-loading 'resources' on Instance uuid 17275472-2921-49c9-b4e0-5485649ebd1c {{(pid=63515) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 874.807290] env[63515]: DEBUG oslo_vmware.api [None req-789266fe-c183-402c-aef4-222438c1eee4 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Task: {'id': task-1111439, 'name': ReconfigVM_Task, 'duration_secs': 0.471761} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 874.807648] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-789266fe-c183-402c-aef4-222438c1eee4 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: eab16df9-7bb5-4576-bca0-769a561c5fe9] Reconfigured VM instance instance-00000047 to attach disk [datastore1] eab16df9-7bb5-4576-bca0-769a561c5fe9/eab16df9-7bb5-4576-bca0-769a561c5fe9.vmdk or device None with type sparse {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 874.808393] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-5a34884e-dffa-4c95-a0b4-4c92a9a6ffd7 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.817020] env[63515]: DEBUG oslo_vmware.api [None req-789266fe-c183-402c-aef4-222438c1eee4 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Waiting for the task: (returnval){ [ 874.817020] env[63515]: value = "task-1111442" [ 874.817020] env[63515]: _type = "Task" [ 874.817020] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 874.825602] env[63515]: DEBUG oslo_vmware.api [None req-789266fe-c183-402c-aef4-222438c1eee4 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Task: {'id': task-1111442, 'name': Rename_Task} progress is 5%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.843171] env[63515]: DEBUG oslo_vmware.api [None req-87ca4373-5e35-4af3-b945-64d29205298c tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Task: {'id': task-1111440, 'name': PowerOffVM_Task, 'duration_secs': 0.422316} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 874.843743] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-87ca4373-5e35-4af3-b945-64d29205298c tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] [instance: d7e8dcdb-41cb-46fb-8b61-d251e7c2d372] Powered off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 874.843743] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-87ca4373-5e35-4af3-b945-64d29205298c tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] [instance: d7e8dcdb-41cb-46fb-8b61-d251e7c2d372] Unregistering the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 874.843955] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-464450c6-0a73-4bd5-9349-588c64f6fa9c {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.922167] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-87ca4373-5e35-4af3-b945-64d29205298c tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] [instance: d7e8dcdb-41cb-46fb-8b61-d251e7c2d372] Unregistered the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 874.922421] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-87ca4373-5e35-4af3-b945-64d29205298c tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] [instance: d7e8dcdb-41cb-46fb-8b61-d251e7c2d372] Deleting contents of the VM from datastore datastore1 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 874.922601] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-87ca4373-5e35-4af3-b945-64d29205298c tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Deleting the datastore file [datastore1] d7e8dcdb-41cb-46fb-8b61-d251e7c2d372 {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 874.923298] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-2e068ce6-b9f9-49bd-9ee3-9ad986f39526 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.930697] env[63515]: DEBUG oslo_vmware.api [None req-87ca4373-5e35-4af3-b945-64d29205298c tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Waiting for the task: (returnval){ [ 874.930697] env[63515]: value = "task-1111444" [ 874.930697] env[63515]: _type = "Task" [ 874.930697] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 874.944107] env[63515]: DEBUG oslo_vmware.api [None req-87ca4373-5e35-4af3-b945-64d29205298c tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Task: {'id': task-1111444, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.964188] env[63515]: INFO nova.compute.manager [None req-baaf7dbf-037b-4d17-a0eb-d252e746ead6 tempest-InstanceActionsTestJSON-247065106 tempest-InstanceActionsTestJSON-247065106-project-member] [instance: 0ee9cce2-94b9-4038-ab38-1b3f96a864bc] Took 46.73 seconds to build instance. [ 875.070084] env[63515]: DEBUG oslo_concurrency.lockutils [None req-561efd36-c7e8-4824-9224-7669ba1ee132 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Releasing lock "refresh_cache-68766cd4-84be-475b-8494-d7ab43a9e969" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 875.072099] env[63515]: DEBUG nova.compute.manager [None req-561efd36-c7e8-4824-9224-7669ba1ee132 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 68766cd4-84be-475b-8494-d7ab43a9e969] Instance network_info: |[{"id": "28b59076-140f-46b5-b4a9-0bf530a1d43d", "address": "fa:16:3e:a3:55:0c", "network": {"id": "5758c885-e1c9-45c8-86aa-27a802101f52", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-734458138-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b18d4daacbc84e758f9f9ae4f3f09d28", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69351262-8d39-441a-85ba-3a78df436d17", "external-id": "nsx-vlan-transportzone-205", "segmentation_id": 205, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap28b59076-14", "ovs_interfaceid": "28b59076-140f-46b5-b4a9-0bf530a1d43d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 875.072099] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-561efd36-c7e8-4824-9224-7669ba1ee132 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 68766cd4-84be-475b-8494-d7ab43a9e969] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:a3:55:0c', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '69351262-8d39-441a-85ba-3a78df436d17', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '28b59076-140f-46b5-b4a9-0bf530a1d43d', 'vif_model': 'vmxnet3'}] {{(pid=63515) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 875.080200] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-561efd36-c7e8-4824-9224-7669ba1ee132 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Creating folder: Project (b18d4daacbc84e758f9f9ae4f3f09d28). Parent ref: group-v243370. {{(pid=63515) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 875.080495] env[63515]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-9d469e36-4c38-4197-9ab7-5248341cbf19 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.092025] env[63515]: INFO nova.virt.vmwareapi.vm_util [None req-561efd36-c7e8-4824-9224-7669ba1ee132 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Created folder: Project (b18d4daacbc84e758f9f9ae4f3f09d28) in parent group-v243370. [ 875.092025] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-561efd36-c7e8-4824-9224-7669ba1ee132 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Creating folder: Instances. Parent ref: group-v243505. {{(pid=63515) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 875.092025] env[63515]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-839062ad-544d-4300-af0f-bdc3f0c15ff4 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.102101] env[63515]: INFO nova.virt.vmwareapi.vm_util [None req-561efd36-c7e8-4824-9224-7669ba1ee132 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Created folder: Instances in parent group-v243505. [ 875.102101] env[63515]: DEBUG oslo.service.loopingcall [None req-561efd36-c7e8-4824-9224-7669ba1ee132 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 875.102331] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 68766cd4-84be-475b-8494-d7ab43a9e969] Creating VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 875.102588] env[63515]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-5a7010fe-4d9e-4fe4-8f25-af2ff8837a4b {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.121865] env[63515]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 875.121865] env[63515]: value = "task-1111447" [ 875.121865] env[63515]: _type = "Task" [ 875.121865] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 875.130152] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1111447, 'name': CreateVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.236238] env[63515]: DEBUG oslo_vmware.api [None req-3db8c3e0-5f0e-4c46-ab36-d6f97ddeb4f1 tempest-ServerActionsV293TestJSON-1193370169 tempest-ServerActionsV293TestJSON-1193370169-project-member] Task: {'id': task-1111441, 'name': ReconfigVM_Task, 'duration_secs': 0.238884} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 875.236631] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-3db8c3e0-5f0e-4c46-ab36-d6f97ddeb4f1 tempest-ServerActionsV293TestJSON-1193370169 tempest-ServerActionsV293TestJSON-1193370169-project-member] [instance: 2ea99c7a-8b61-4718-bee0-f4ce4fc021af] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-243483', 'volume_id': '211b1a87-a01e-4d13-aa27-91391bf9cde1', 'name': 'volume-211b1a87-a01e-4d13-aa27-91391bf9cde1', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '2ea99c7a-8b61-4718-bee0-f4ce4fc021af', 'attached_at': '', 'detached_at': '', 'volume_id': '211b1a87-a01e-4d13-aa27-91391bf9cde1', 'serial': '211b1a87-a01e-4d13-aa27-91391bf9cde1'} {{(pid=63515) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 875.237327] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-9333b19a-ed93-4e46-86ef-e5e46ea375b0 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.244363] env[63515]: DEBUG oslo_vmware.api [None req-3db8c3e0-5f0e-4c46-ab36-d6f97ddeb4f1 tempest-ServerActionsV293TestJSON-1193370169 tempest-ServerActionsV293TestJSON-1193370169-project-member] Waiting for the task: (returnval){ [ 875.244363] env[63515]: value = "task-1111448" [ 875.244363] env[63515]: _type = "Task" [ 875.244363] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 875.250449] env[63515]: DEBUG oslo_concurrency.lockutils [None req-593b35b4-5d74-4c58-a93e-9cad00fc0b2d tempest-ServerGroupTestJSON-385159797 tempest-ServerGroupTestJSON-385159797-project-member] Acquiring lock "0dbb3077-8407-49be-b0ce-345e1b63383c" by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 875.251030] env[63515]: DEBUG oslo_concurrency.lockutils [None req-593b35b4-5d74-4c58-a93e-9cad00fc0b2d tempest-ServerGroupTestJSON-385159797 tempest-ServerGroupTestJSON-385159797-project-member] Lock "0dbb3077-8407-49be-b0ce-345e1b63383c" acquired by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" :: waited 0.001s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 875.266111] env[63515]: DEBUG oslo_vmware.api [None req-3db8c3e0-5f0e-4c46-ab36-d6f97ddeb4f1 tempest-ServerActionsV293TestJSON-1193370169 tempest-ServerActionsV293TestJSON-1193370169-project-member] Task: {'id': task-1111448, 'name': Rename_Task} progress is 10%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.328945] env[63515]: DEBUG oslo_vmware.api [None req-789266fe-c183-402c-aef4-222438c1eee4 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Task: {'id': task-1111442, 'name': Rename_Task, 'duration_secs': 0.231676} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 875.329459] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-789266fe-c183-402c-aef4-222438c1eee4 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: eab16df9-7bb5-4576-bca0-769a561c5fe9] Powering on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 875.332553] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-237cf1e9-ec37-4c11-8a86-ee202c778126 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.340084] env[63515]: DEBUG oslo_vmware.api [None req-789266fe-c183-402c-aef4-222438c1eee4 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Waiting for the task: (returnval){ [ 875.340084] env[63515]: value = "task-1111449" [ 875.340084] env[63515]: _type = "Task" [ 875.340084] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 875.355665] env[63515]: DEBUG oslo_vmware.api [None req-789266fe-c183-402c-aef4-222438c1eee4 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Task: {'id': task-1111449, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.446504] env[63515]: DEBUG oslo_vmware.api [None req-87ca4373-5e35-4af3-b945-64d29205298c tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Task: {'id': task-1111444, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.45511} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 875.450086] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-87ca4373-5e35-4af3-b945-64d29205298c tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Deleted the datastore file {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 875.450660] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-87ca4373-5e35-4af3-b945-64d29205298c tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] [instance: d7e8dcdb-41cb-46fb-8b61-d251e7c2d372] Deleted contents of the VM from datastore datastore1 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 875.451105] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-87ca4373-5e35-4af3-b945-64d29205298c tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] [instance: d7e8dcdb-41cb-46fb-8b61-d251e7c2d372] Instance destroyed {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 875.451588] env[63515]: INFO nova.compute.manager [None req-87ca4373-5e35-4af3-b945-64d29205298c tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] [instance: d7e8dcdb-41cb-46fb-8b61-d251e7c2d372] Took 1.16 seconds to destroy the instance on the hypervisor. [ 875.453039] env[63515]: DEBUG oslo.service.loopingcall [None req-87ca4373-5e35-4af3-b945-64d29205298c tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 875.453039] env[63515]: DEBUG nova.compute.manager [-] [instance: d7e8dcdb-41cb-46fb-8b61-d251e7c2d372] Deallocating network for instance {{(pid=63515) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 875.453039] env[63515]: DEBUG nova.network.neutron [-] [instance: d7e8dcdb-41cb-46fb-8b61-d251e7c2d372] deallocate_for_instance() {{(pid=63515) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 875.465352] env[63515]: DEBUG oslo_concurrency.lockutils [None req-baaf7dbf-037b-4d17-a0eb-d252e746ead6 tempest-InstanceActionsTestJSON-247065106 tempest-InstanceActionsTestJSON-247065106-project-member] Lock "0ee9cce2-94b9-4038-ab38-1b3f96a864bc" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 48.244s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 875.596358] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c836179-af9f-4a13-984b-8c8c38270047 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.605436] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da2a4b24-4f6b-47c9-a95c-a6bdf0c3abc2 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.649262] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d44729c6-a299-43cf-b44e-c8da910152fe {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.660582] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee373a4d-7213-45ad-bb8d-8e4d8f62e8b4 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.664970] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1111447, 'name': CreateVM_Task, 'duration_secs': 0.507659} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 875.665167] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 68766cd4-84be-475b-8494-d7ab43a9e969] Created VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 875.666244] env[63515]: DEBUG oslo_concurrency.lockutils [None req-561efd36-c7e8-4824-9224-7669ba1ee132 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 875.666762] env[63515]: DEBUG oslo_concurrency.lockutils [None req-561efd36-c7e8-4824-9224-7669ba1ee132 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 875.667149] env[63515]: DEBUG oslo_concurrency.lockutils [None req-561efd36-c7e8-4824-9224-7669ba1ee132 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 875.677806] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-32976efd-31af-4d56-9f09-bd4bc2dc01c5 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.680085] env[63515]: DEBUG nova.compute.provider_tree [None req-dd655d0d-92fe-48b9-b34e-d4edd8542788 tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 875.686151] env[63515]: DEBUG oslo_concurrency.lockutils [None req-dbe24c17-88c2-40e6-987c-d70326fd4bd3 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Acquiring lock "e13da90c-28e6-43d2-99b7-19c5095954ca" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 875.686393] env[63515]: DEBUG oslo_concurrency.lockutils [None req-dbe24c17-88c2-40e6-987c-d70326fd4bd3 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Lock "e13da90c-28e6-43d2-99b7-19c5095954ca" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 875.689701] env[63515]: DEBUG oslo_vmware.api [None req-561efd36-c7e8-4824-9224-7669ba1ee132 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Waiting for the task: (returnval){ [ 875.689701] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]529f20a0-98d8-da29-63d1-633f6d0e69b8" [ 875.689701] env[63515]: _type = "Task" [ 875.689701] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 875.700776] env[63515]: DEBUG oslo_vmware.api [None req-561efd36-c7e8-4824-9224-7669ba1ee132 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]529f20a0-98d8-da29-63d1-633f6d0e69b8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.718482] env[63515]: DEBUG nova.network.neutron [req-b035ecee-444a-4056-9738-992e26b2fa53 req-9da2e22f-2ea5-497e-9b7e-de87dab780f3 service nova] [instance: 2a29f37d-e786-48d3-b126-4854d6e7c67c] Updated VIF entry in instance network info cache for port 956cd361-0d42-4914-85d3-5f75a72a68a1. {{(pid=63515) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 875.718861] env[63515]: DEBUG nova.network.neutron [req-b035ecee-444a-4056-9738-992e26b2fa53 req-9da2e22f-2ea5-497e-9b7e-de87dab780f3 service nova] [instance: 2a29f37d-e786-48d3-b126-4854d6e7c67c] Updating instance_info_cache with network_info: [{"id": "956cd361-0d42-4914-85d3-5f75a72a68a1", "address": "fa:16:3e:a3:72:bb", "network": {"id": "850bfa95-3df6-4019-8260-ade20ec58799", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-832000501-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.153", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "08c9aa51b1864410809f8877c7cd49c4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3cf748a8-7ae0-4dca-817d-e727c30d72f4", "external-id": "nsx-vlan-transportzone-853", "segmentation_id": 853, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap956cd361-0d", "ovs_interfaceid": "956cd361-0d42-4914-85d3-5f75a72a68a1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 875.755089] env[63515]: DEBUG oslo_vmware.api [None req-3db8c3e0-5f0e-4c46-ab36-d6f97ddeb4f1 tempest-ServerActionsV293TestJSON-1193370169 tempest-ServerActionsV293TestJSON-1193370169-project-member] Task: {'id': task-1111448, 'name': Rename_Task, 'duration_secs': 0.178473} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 875.755393] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-3db8c3e0-5f0e-4c46-ab36-d6f97ddeb4f1 tempest-ServerActionsV293TestJSON-1193370169 tempest-ServerActionsV293TestJSON-1193370169-project-member] [instance: 2ea99c7a-8b61-4718-bee0-f4ce4fc021af] Powering on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 875.755649] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-13db9e62-2e15-4f6e-b0ca-b75c53ad8e07 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.757480] env[63515]: DEBUG oslo_concurrency.lockutils [None req-593b35b4-5d74-4c58-a93e-9cad00fc0b2d tempest-ServerGroupTestJSON-385159797 tempest-ServerGroupTestJSON-385159797-project-member] Lock "0dbb3077-8407-49be-b0ce-345e1b63383c" "released" by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" :: held 0.507s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 875.757971] env[63515]: DEBUG nova.compute.manager [None req-593b35b4-5d74-4c58-a93e-9cad00fc0b2d tempest-ServerGroupTestJSON-385159797 tempest-ServerGroupTestJSON-385159797-project-member] [instance: a2e014a2-9e0f-4250-aa38-3a7154149903] Start building networks asynchronously for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 875.766620] env[63515]: DEBUG oslo_vmware.api [None req-3db8c3e0-5f0e-4c46-ab36-d6f97ddeb4f1 tempest-ServerActionsV293TestJSON-1193370169 tempest-ServerActionsV293TestJSON-1193370169-project-member] Waiting for the task: (returnval){ [ 875.766620] env[63515]: value = "task-1111450" [ 875.766620] env[63515]: _type = "Task" [ 875.766620] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 875.775215] env[63515]: DEBUG oslo_vmware.api [None req-3db8c3e0-5f0e-4c46-ab36-d6f97ddeb4f1 tempest-ServerActionsV293TestJSON-1193370169 tempest-ServerActionsV293TestJSON-1193370169-project-member] Task: {'id': task-1111450, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.788786] env[63515]: DEBUG nova.network.neutron [None req-3fd65342-5e9e-46b6-830a-47df91928006 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: 0967283f-8dda-4692-b6b7-601545cabf2f] Successfully updated port: 66571248-83b1-46c5-b892-f84ab3882407 {{(pid=63515) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 875.850166] env[63515]: DEBUG oslo_vmware.api [None req-789266fe-c183-402c-aef4-222438c1eee4 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Task: {'id': task-1111449, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.979950] env[63515]: DEBUG nova.compute.manager [req-bb6e93bc-46d8-42a6-a5f2-32d744515850 req-344797c7-c665-442c-81a3-1cd0c366000e service nova] [instance: 68766cd4-84be-475b-8494-d7ab43a9e969] Received event network-changed-28b59076-140f-46b5-b4a9-0bf530a1d43d {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 875.981931] env[63515]: DEBUG nova.compute.manager [req-bb6e93bc-46d8-42a6-a5f2-32d744515850 req-344797c7-c665-442c-81a3-1cd0c366000e service nova] [instance: 68766cd4-84be-475b-8494-d7ab43a9e969] Refreshing instance network info cache due to event network-changed-28b59076-140f-46b5-b4a9-0bf530a1d43d. {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 875.981931] env[63515]: DEBUG oslo_concurrency.lockutils [req-bb6e93bc-46d8-42a6-a5f2-32d744515850 req-344797c7-c665-442c-81a3-1cd0c366000e service nova] Acquiring lock "refresh_cache-68766cd4-84be-475b-8494-d7ab43a9e969" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 875.981931] env[63515]: DEBUG oslo_concurrency.lockutils [req-bb6e93bc-46d8-42a6-a5f2-32d744515850 req-344797c7-c665-442c-81a3-1cd0c366000e service nova] Acquired lock "refresh_cache-68766cd4-84be-475b-8494-d7ab43a9e969" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 875.981931] env[63515]: DEBUG nova.network.neutron [req-bb6e93bc-46d8-42a6-a5f2-32d744515850 req-344797c7-c665-442c-81a3-1cd0c366000e service nova] [instance: 68766cd4-84be-475b-8494-d7ab43a9e969] Refreshing network info cache for port 28b59076-140f-46b5-b4a9-0bf530a1d43d {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 876.187041] env[63515]: DEBUG nova.scheduler.client.report [None req-dd655d0d-92fe-48b9-b34e-d4edd8542788 tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 876.193382] env[63515]: DEBUG nova.compute.manager [None req-dbe24c17-88c2-40e6-987c-d70326fd4bd3 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: e13da90c-28e6-43d2-99b7-19c5095954ca] Starting instance... {{(pid=63515) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 876.215300] env[63515]: DEBUG oslo_vmware.api [None req-561efd36-c7e8-4824-9224-7669ba1ee132 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]529f20a0-98d8-da29-63d1-633f6d0e69b8, 'name': SearchDatastore_Task, 'duration_secs': 0.026382} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 876.215748] env[63515]: DEBUG oslo_concurrency.lockutils [None req-561efd36-c7e8-4824-9224-7669ba1ee132 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 876.217189] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-561efd36-c7e8-4824-9224-7669ba1ee132 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 68766cd4-84be-475b-8494-d7ab43a9e969] Processing image 8a120570-cb06-4099-b262-554ca0ad15c5 {{(pid=63515) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 876.219841] env[63515]: DEBUG oslo_concurrency.lockutils [None req-561efd36-c7e8-4824-9224-7669ba1ee132 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 876.219841] env[63515]: DEBUG oslo_concurrency.lockutils [None req-561efd36-c7e8-4824-9224-7669ba1ee132 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 876.219841] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-561efd36-c7e8-4824-9224-7669ba1ee132 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 876.219841] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c12be1b8-1f84-4ad0-95f2-2ea7b29f1222 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.222123] env[63515]: DEBUG oslo_concurrency.lockutils [req-b035ecee-444a-4056-9738-992e26b2fa53 req-9da2e22f-2ea5-497e-9b7e-de87dab780f3 service nova] Releasing lock "refresh_cache-2a29f37d-e786-48d3-b126-4854d6e7c67c" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 876.235055] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-561efd36-c7e8-4824-9224-7669ba1ee132 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 876.235055] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-561efd36-c7e8-4824-9224-7669ba1ee132 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63515) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 876.235055] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5b57f013-c933-4c6c-95ef-c5bb9dda7b00 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.242538] env[63515]: DEBUG oslo_vmware.api [None req-561efd36-c7e8-4824-9224-7669ba1ee132 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Waiting for the task: (returnval){ [ 876.242538] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]5270e7a0-11d1-b603-c5eb-93adff44d533" [ 876.242538] env[63515]: _type = "Task" [ 876.242538] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.251281] env[63515]: DEBUG oslo_vmware.api [None req-561efd36-c7e8-4824-9224-7669ba1ee132 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]5270e7a0-11d1-b603-c5eb-93adff44d533, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.262133] env[63515]: DEBUG nova.compute.utils [None req-593b35b4-5d74-4c58-a93e-9cad00fc0b2d tempest-ServerGroupTestJSON-385159797 tempest-ServerGroupTestJSON-385159797-project-member] Using /dev/sd instead of None {{(pid=63515) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 876.264339] env[63515]: DEBUG nova.compute.manager [None req-593b35b4-5d74-4c58-a93e-9cad00fc0b2d tempest-ServerGroupTestJSON-385159797 tempest-ServerGroupTestJSON-385159797-project-member] [instance: a2e014a2-9e0f-4250-aa38-3a7154149903] Allocating IP information in the background. {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 876.264339] env[63515]: DEBUG nova.network.neutron [None req-593b35b4-5d74-4c58-a93e-9cad00fc0b2d tempest-ServerGroupTestJSON-385159797 tempest-ServerGroupTestJSON-385159797-project-member] [instance: a2e014a2-9e0f-4250-aa38-3a7154149903] allocate_for_instance() {{(pid=63515) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 876.280019] env[63515]: DEBUG oslo_vmware.api [None req-3db8c3e0-5f0e-4c46-ab36-d6f97ddeb4f1 tempest-ServerActionsV293TestJSON-1193370169 tempest-ServerActionsV293TestJSON-1193370169-project-member] Task: {'id': task-1111450, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.299074] env[63515]: DEBUG oslo_concurrency.lockutils [None req-3fd65342-5e9e-46b6-830a-47df91928006 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Acquiring lock "refresh_cache-0967283f-8dda-4692-b6b7-601545cabf2f" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 876.299074] env[63515]: DEBUG oslo_concurrency.lockutils [None req-3fd65342-5e9e-46b6-830a-47df91928006 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Acquired lock "refresh_cache-0967283f-8dda-4692-b6b7-601545cabf2f" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 876.299074] env[63515]: DEBUG nova.network.neutron [None req-3fd65342-5e9e-46b6-830a-47df91928006 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: 0967283f-8dda-4692-b6b7-601545cabf2f] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 876.341231] env[63515]: DEBUG nova.policy [None req-593b35b4-5d74-4c58-a93e-9cad00fc0b2d tempest-ServerGroupTestJSON-385159797 tempest-ServerGroupTestJSON-385159797-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'cd993ea10064445a83a38acd908bebb6', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '07921f3cdb3140b9b5703c418cb6a90e', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63515) authorize /opt/stack/nova/nova/policy.py:201}} [ 876.353903] env[63515]: DEBUG oslo_vmware.api [None req-789266fe-c183-402c-aef4-222438c1eee4 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Task: {'id': task-1111449, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.536352] env[63515]: DEBUG oslo_concurrency.lockutils [None req-82894ab5-858a-446e-a869-69b8a9c84386 tempest-InstanceActionsTestJSON-247065106 tempest-InstanceActionsTestJSON-247065106-project-member] Acquiring lock "0ee9cce2-94b9-4038-ab38-1b3f96a864bc" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 876.536613] env[63515]: DEBUG oslo_concurrency.lockutils [None req-82894ab5-858a-446e-a869-69b8a9c84386 tempest-InstanceActionsTestJSON-247065106 tempest-InstanceActionsTestJSON-247065106-project-member] Lock "0ee9cce2-94b9-4038-ab38-1b3f96a864bc" acquired by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 876.536932] env[63515]: INFO nova.compute.manager [None req-82894ab5-858a-446e-a869-69b8a9c84386 tempest-InstanceActionsTestJSON-247065106 tempest-InstanceActionsTestJSON-247065106-project-member] [instance: 0ee9cce2-94b9-4038-ab38-1b3f96a864bc] Rebooting instance [ 876.541632] env[63515]: DEBUG nova.network.neutron [-] [instance: d7e8dcdb-41cb-46fb-8b61-d251e7c2d372] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 876.700591] env[63515]: DEBUG oslo_concurrency.lockutils [None req-dd655d0d-92fe-48b9-b34e-d4edd8542788 tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.957s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 876.704201] env[63515]: DEBUG oslo_concurrency.lockutils [None req-d7e3c195-4890-454e-9f6b-af1a7cc4ee6a tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.091s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 876.708116] env[63515]: INFO nova.compute.claims [None req-d7e3c195-4890-454e-9f6b-af1a7cc4ee6a tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: ae5d6f34-bac6-4ba0-b0c3-8205982ddc88] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 876.735140] env[63515]: INFO nova.scheduler.client.report [None req-dd655d0d-92fe-48b9-b34e-d4edd8542788 tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] Deleted allocations for instance 17275472-2921-49c9-b4e0-5485649ebd1c [ 876.744733] env[63515]: DEBUG oslo_concurrency.lockutils [None req-dbe24c17-88c2-40e6-987c-d70326fd4bd3 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 876.763840] env[63515]: DEBUG oslo_vmware.api [None req-561efd36-c7e8-4824-9224-7669ba1ee132 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]5270e7a0-11d1-b603-c5eb-93adff44d533, 'name': SearchDatastore_Task, 'duration_secs': 0.01217} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 876.766723] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0b8310eb-f55c-4d1d-8cb4-5a2e2dd2efcb {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.768925] env[63515]: DEBUG nova.compute.manager [None req-593b35b4-5d74-4c58-a93e-9cad00fc0b2d tempest-ServerGroupTestJSON-385159797 tempest-ServerGroupTestJSON-385159797-project-member] [instance: a2e014a2-9e0f-4250-aa38-3a7154149903] Start building block device mappings for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 876.777404] env[63515]: DEBUG oslo_vmware.api [None req-561efd36-c7e8-4824-9224-7669ba1ee132 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Waiting for the task: (returnval){ [ 876.777404] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]52e39198-0f77-eafc-5dab-04534e3e710c" [ 876.777404] env[63515]: _type = "Task" [ 876.777404] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.785310] env[63515]: DEBUG oslo_vmware.api [None req-3db8c3e0-5f0e-4c46-ab36-d6f97ddeb4f1 tempest-ServerActionsV293TestJSON-1193370169 tempest-ServerActionsV293TestJSON-1193370169-project-member] Task: {'id': task-1111450, 'name': PowerOnVM_Task, 'duration_secs': 0.871605} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 876.786256] env[63515]: DEBUG nova.network.neutron [None req-593b35b4-5d74-4c58-a93e-9cad00fc0b2d tempest-ServerGroupTestJSON-385159797 tempest-ServerGroupTestJSON-385159797-project-member] [instance: a2e014a2-9e0f-4250-aa38-3a7154149903] Successfully created port: a453f43a-c524-4eb7-8dcc-9edc8aabdd2c {{(pid=63515) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 876.789980] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-3db8c3e0-5f0e-4c46-ab36-d6f97ddeb4f1 tempest-ServerActionsV293TestJSON-1193370169 tempest-ServerActionsV293TestJSON-1193370169-project-member] [instance: 2ea99c7a-8b61-4718-bee0-f4ce4fc021af] Powered on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 876.790795] env[63515]: INFO nova.compute.manager [None req-3db8c3e0-5f0e-4c46-ab36-d6f97ddeb4f1 tempest-ServerActionsV293TestJSON-1193370169 tempest-ServerActionsV293TestJSON-1193370169-project-member] [instance: 2ea99c7a-8b61-4718-bee0-f4ce4fc021af] Took 5.81 seconds to spawn the instance on the hypervisor. [ 876.790795] env[63515]: DEBUG nova.compute.manager [None req-3db8c3e0-5f0e-4c46-ab36-d6f97ddeb4f1 tempest-ServerActionsV293TestJSON-1193370169 tempest-ServerActionsV293TestJSON-1193370169-project-member] [instance: 2ea99c7a-8b61-4718-bee0-f4ce4fc021af] Checking state {{(pid=63515) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 876.791637] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64c28463-7f25-429e-978a-f0b915712b5b {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.801154] env[63515]: DEBUG nova.compute.manager [req-55e8c293-3776-4f1e-b309-01d5c7852ff5 req-d60377d0-4931-4f83-b645-09b1ff7a11c7 service nova] [instance: d7e8dcdb-41cb-46fb-8b61-d251e7c2d372] Received event network-vif-deleted-9378f59d-f9bc-4c2e-8336-2c5710f987cf {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 876.801937] env[63515]: DEBUG oslo_vmware.api [None req-561efd36-c7e8-4824-9224-7669ba1ee132 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52e39198-0f77-eafc-5dab-04534e3e710c, 'name': SearchDatastore_Task, 'duration_secs': 0.01246} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 876.803496] env[63515]: DEBUG oslo_concurrency.lockutils [None req-561efd36-c7e8-4824-9224-7669ba1ee132 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 876.804073] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-561efd36-c7e8-4824-9224-7669ba1ee132 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk to [datastore2] 68766cd4-84be-475b-8494-d7ab43a9e969/68766cd4-84be-475b-8494-d7ab43a9e969.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 876.807945] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f9deb610-a564-4ba1-b87d-ed65a09bcc4c {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.823528] env[63515]: DEBUG oslo_vmware.api [None req-561efd36-c7e8-4824-9224-7669ba1ee132 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Waiting for the task: (returnval){ [ 876.823528] env[63515]: value = "task-1111451" [ 876.823528] env[63515]: _type = "Task" [ 876.823528] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.833613] env[63515]: DEBUG oslo_vmware.api [None req-561efd36-c7e8-4824-9224-7669ba1ee132 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Task: {'id': task-1111451, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.840122] env[63515]: DEBUG nova.network.neutron [req-bb6e93bc-46d8-42a6-a5f2-32d744515850 req-344797c7-c665-442c-81a3-1cd0c366000e service nova] [instance: 68766cd4-84be-475b-8494-d7ab43a9e969] Updated VIF entry in instance network info cache for port 28b59076-140f-46b5-b4a9-0bf530a1d43d. {{(pid=63515) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 876.840272] env[63515]: DEBUG nova.network.neutron [req-bb6e93bc-46d8-42a6-a5f2-32d744515850 req-344797c7-c665-442c-81a3-1cd0c366000e service nova] [instance: 68766cd4-84be-475b-8494-d7ab43a9e969] Updating instance_info_cache with network_info: [{"id": "28b59076-140f-46b5-b4a9-0bf530a1d43d", "address": "fa:16:3e:a3:55:0c", "network": {"id": "5758c885-e1c9-45c8-86aa-27a802101f52", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-734458138-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b18d4daacbc84e758f9f9ae4f3f09d28", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69351262-8d39-441a-85ba-3a78df436d17", "external-id": "nsx-vlan-transportzone-205", "segmentation_id": 205, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap28b59076-14", "ovs_interfaceid": "28b59076-140f-46b5-b4a9-0bf530a1d43d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 876.854640] env[63515]: DEBUG oslo_vmware.api [None req-789266fe-c183-402c-aef4-222438c1eee4 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Task: {'id': task-1111449, 'name': PowerOnVM_Task} progress is 78%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.856945] env[63515]: DEBUG nova.network.neutron [None req-3fd65342-5e9e-46b6-830a-47df91928006 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: 0967283f-8dda-4692-b6b7-601545cabf2f] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 877.043633] env[63515]: INFO nova.compute.manager [-] [instance: d7e8dcdb-41cb-46fb-8b61-d251e7c2d372] Took 1.59 seconds to deallocate network for instance. [ 877.069031] env[63515]: DEBUG oslo_concurrency.lockutils [None req-82894ab5-858a-446e-a869-69b8a9c84386 tempest-InstanceActionsTestJSON-247065106 tempest-InstanceActionsTestJSON-247065106-project-member] Acquiring lock "refresh_cache-0ee9cce2-94b9-4038-ab38-1b3f96a864bc" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 877.069031] env[63515]: DEBUG oslo_concurrency.lockutils [None req-82894ab5-858a-446e-a869-69b8a9c84386 tempest-InstanceActionsTestJSON-247065106 tempest-InstanceActionsTestJSON-247065106-project-member] Acquired lock "refresh_cache-0ee9cce2-94b9-4038-ab38-1b3f96a864bc" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 877.069031] env[63515]: DEBUG nova.network.neutron [None req-82894ab5-858a-446e-a869-69b8a9c84386 tempest-InstanceActionsTestJSON-247065106 tempest-InstanceActionsTestJSON-247065106-project-member] [instance: 0ee9cce2-94b9-4038-ab38-1b3f96a864bc] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 877.085223] env[63515]: DEBUG nova.network.neutron [None req-3fd65342-5e9e-46b6-830a-47df91928006 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: 0967283f-8dda-4692-b6b7-601545cabf2f] Updating instance_info_cache with network_info: [{"id": "66571248-83b1-46c5-b892-f84ab3882407", "address": "fa:16:3e:7e:e0:aa", "network": {"id": "581d448e-be1b-4d5b-8a6e-c27326aaffc5", "bridge": "br-int", "label": "tempest-ServersTestJSON-1769610073-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "299fdeff647f486390366d5bbf911518", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fb143ef7-8271-4a8a-a4aa-8eba9a89f6a1", "external-id": "nsx-vlan-transportzone-504", "segmentation_id": 504, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap66571248-83", "ovs_interfaceid": "66571248-83b1-46c5-b892-f84ab3882407", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 877.260411] env[63515]: DEBUG oslo_concurrency.lockutils [None req-dd655d0d-92fe-48b9-b34e-d4edd8542788 tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] Lock "17275472-2921-49c9-b4e0-5485649ebd1c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 23.210s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 877.324524] env[63515]: INFO nova.compute.manager [None req-3db8c3e0-5f0e-4c46-ab36-d6f97ddeb4f1 tempest-ServerActionsV293TestJSON-1193370169 tempest-ServerActionsV293TestJSON-1193370169-project-member] [instance: 2ea99c7a-8b61-4718-bee0-f4ce4fc021af] Took 35.24 seconds to build instance. [ 877.334741] env[63515]: DEBUG oslo_vmware.api [None req-561efd36-c7e8-4824-9224-7669ba1ee132 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Task: {'id': task-1111451, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.342394] env[63515]: DEBUG oslo_concurrency.lockutils [req-bb6e93bc-46d8-42a6-a5f2-32d744515850 req-344797c7-c665-442c-81a3-1cd0c366000e service nova] Releasing lock "refresh_cache-68766cd4-84be-475b-8494-d7ab43a9e969" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 877.342648] env[63515]: DEBUG nova.compute.manager [req-bb6e93bc-46d8-42a6-a5f2-32d744515850 req-344797c7-c665-442c-81a3-1cd0c366000e service nova] [instance: 0967283f-8dda-4692-b6b7-601545cabf2f] Received event network-vif-plugged-66571248-83b1-46c5-b892-f84ab3882407 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 877.342844] env[63515]: DEBUG oslo_concurrency.lockutils [req-bb6e93bc-46d8-42a6-a5f2-32d744515850 req-344797c7-c665-442c-81a3-1cd0c366000e service nova] Acquiring lock "0967283f-8dda-4692-b6b7-601545cabf2f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 877.343064] env[63515]: DEBUG oslo_concurrency.lockutils [req-bb6e93bc-46d8-42a6-a5f2-32d744515850 req-344797c7-c665-442c-81a3-1cd0c366000e service nova] Lock "0967283f-8dda-4692-b6b7-601545cabf2f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 877.343237] env[63515]: DEBUG oslo_concurrency.lockutils [req-bb6e93bc-46d8-42a6-a5f2-32d744515850 req-344797c7-c665-442c-81a3-1cd0c366000e service nova] Lock "0967283f-8dda-4692-b6b7-601545cabf2f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 877.343409] env[63515]: DEBUG nova.compute.manager [req-bb6e93bc-46d8-42a6-a5f2-32d744515850 req-344797c7-c665-442c-81a3-1cd0c366000e service nova] [instance: 0967283f-8dda-4692-b6b7-601545cabf2f] No waiting events found dispatching network-vif-plugged-66571248-83b1-46c5-b892-f84ab3882407 {{(pid=63515) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 877.343582] env[63515]: WARNING nova.compute.manager [req-bb6e93bc-46d8-42a6-a5f2-32d744515850 req-344797c7-c665-442c-81a3-1cd0c366000e service nova] [instance: 0967283f-8dda-4692-b6b7-601545cabf2f] Received unexpected event network-vif-plugged-66571248-83b1-46c5-b892-f84ab3882407 for instance with vm_state building and task_state spawning. [ 877.343754] env[63515]: DEBUG nova.compute.manager [req-bb6e93bc-46d8-42a6-a5f2-32d744515850 req-344797c7-c665-442c-81a3-1cd0c366000e service nova] [instance: 0967283f-8dda-4692-b6b7-601545cabf2f] Received event network-changed-66571248-83b1-46c5-b892-f84ab3882407 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 877.343915] env[63515]: DEBUG nova.compute.manager [req-bb6e93bc-46d8-42a6-a5f2-32d744515850 req-344797c7-c665-442c-81a3-1cd0c366000e service nova] [instance: 0967283f-8dda-4692-b6b7-601545cabf2f] Refreshing instance network info cache due to event network-changed-66571248-83b1-46c5-b892-f84ab3882407. {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 877.344101] env[63515]: DEBUG oslo_concurrency.lockutils [req-bb6e93bc-46d8-42a6-a5f2-32d744515850 req-344797c7-c665-442c-81a3-1cd0c366000e service nova] Acquiring lock "refresh_cache-0967283f-8dda-4692-b6b7-601545cabf2f" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 877.353719] env[63515]: DEBUG oslo_vmware.api [None req-789266fe-c183-402c-aef4-222438c1eee4 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Task: {'id': task-1111449, 'name': PowerOnVM_Task, 'duration_secs': 1.665652} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 877.353982] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-789266fe-c183-402c-aef4-222438c1eee4 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: eab16df9-7bb5-4576-bca0-769a561c5fe9] Powered on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 877.354289] env[63515]: INFO nova.compute.manager [None req-789266fe-c183-402c-aef4-222438c1eee4 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: eab16df9-7bb5-4576-bca0-769a561c5fe9] Took 10.61 seconds to spawn the instance on the hypervisor. [ 877.354719] env[63515]: DEBUG nova.compute.manager [None req-789266fe-c183-402c-aef4-222438c1eee4 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: eab16df9-7bb5-4576-bca0-769a561c5fe9] Checking state {{(pid=63515) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 877.355969] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2271a729-e3d2-4e04-ae1d-af4f62e803a7 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.553954] env[63515]: DEBUG oslo_concurrency.lockutils [None req-87ca4373-5e35-4af3-b945-64d29205298c tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 877.590134] env[63515]: DEBUG oslo_concurrency.lockutils [None req-3fd65342-5e9e-46b6-830a-47df91928006 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Releasing lock "refresh_cache-0967283f-8dda-4692-b6b7-601545cabf2f" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 877.590134] env[63515]: DEBUG nova.compute.manager [None req-3fd65342-5e9e-46b6-830a-47df91928006 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: 0967283f-8dda-4692-b6b7-601545cabf2f] Instance network_info: |[{"id": "66571248-83b1-46c5-b892-f84ab3882407", "address": "fa:16:3e:7e:e0:aa", "network": {"id": "581d448e-be1b-4d5b-8a6e-c27326aaffc5", "bridge": "br-int", "label": "tempest-ServersTestJSON-1769610073-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "299fdeff647f486390366d5bbf911518", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fb143ef7-8271-4a8a-a4aa-8eba9a89f6a1", "external-id": "nsx-vlan-transportzone-504", "segmentation_id": 504, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap66571248-83", "ovs_interfaceid": "66571248-83b1-46c5-b892-f84ab3882407", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 877.590134] env[63515]: DEBUG oslo_concurrency.lockutils [req-bb6e93bc-46d8-42a6-a5f2-32d744515850 req-344797c7-c665-442c-81a3-1cd0c366000e service nova] Acquired lock "refresh_cache-0967283f-8dda-4692-b6b7-601545cabf2f" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 877.590134] env[63515]: DEBUG nova.network.neutron [req-bb6e93bc-46d8-42a6-a5f2-32d744515850 req-344797c7-c665-442c-81a3-1cd0c366000e service nova] [instance: 0967283f-8dda-4692-b6b7-601545cabf2f] Refreshing network info cache for port 66571248-83b1-46c5-b892-f84ab3882407 {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 877.590134] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-3fd65342-5e9e-46b6-830a-47df91928006 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: 0967283f-8dda-4692-b6b7-601545cabf2f] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:7e:e0:aa', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'fb143ef7-8271-4a8a-a4aa-8eba9a89f6a1', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '66571248-83b1-46c5-b892-f84ab3882407', 'vif_model': 'vmxnet3'}] {{(pid=63515) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 877.597293] env[63515]: DEBUG oslo.service.loopingcall [None req-3fd65342-5e9e-46b6-830a-47df91928006 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 877.600469] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0967283f-8dda-4692-b6b7-601545cabf2f] Creating VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 877.601129] env[63515]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-5328ba91-a705-4cf1-b29b-4c309d436b0a {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.625066] env[63515]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 877.625066] env[63515]: value = "task-1111452" [ 877.625066] env[63515]: _type = "Task" [ 877.625066] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 877.632544] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1111452, 'name': CreateVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.649020] env[63515]: DEBUG nova.compute.manager [None req-1eac0833-270f-4d52-8c20-847d896aff86 tempest-ServerDiagnosticsV248Test-1898563194 tempest-ServerDiagnosticsV248Test-1898563194-project-admin] [instance: 705b7e9b-3ab6-40ec-8cf6-d86bf0a1faf0] Checking state {{(pid=63515) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 877.649020] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10900557-8790-44a8-aec2-25b8ba75566f {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.661021] env[63515]: INFO nova.compute.manager [None req-1eac0833-270f-4d52-8c20-847d896aff86 tempest-ServerDiagnosticsV248Test-1898563194 tempest-ServerDiagnosticsV248Test-1898563194-project-admin] [instance: 705b7e9b-3ab6-40ec-8cf6-d86bf0a1faf0] Retrieving diagnostics [ 877.661021] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-192d5f5b-bc25-4cdd-a1f4-6e8151024859 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.782077] env[63515]: DEBUG nova.compute.manager [None req-593b35b4-5d74-4c58-a93e-9cad00fc0b2d tempest-ServerGroupTestJSON-385159797 tempest-ServerGroupTestJSON-385159797-project-member] [instance: a2e014a2-9e0f-4250-aa38-3a7154149903] Start spawning the instance on the hypervisor. {{(pid=63515) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 877.809713] env[63515]: DEBUG nova.virt.hardware [None req-593b35b4-5d74-4c58-a93e-9cad00fc0b2d tempest-ServerGroupTestJSON-385159797 tempest-ServerGroupTestJSON-385159797-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T02:52:14Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T02:51:56Z,direct_url=,disk_format='vmdk',id=8a120570-cb06-4099-b262-554ca0ad15c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b569255cc43e42e7ae3f2b4ad37c6ef4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T02:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 877.810035] env[63515]: DEBUG nova.virt.hardware [None req-593b35b4-5d74-4c58-a93e-9cad00fc0b2d tempest-ServerGroupTestJSON-385159797 tempest-ServerGroupTestJSON-385159797-project-member] Flavor limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 877.810250] env[63515]: DEBUG nova.virt.hardware [None req-593b35b4-5d74-4c58-a93e-9cad00fc0b2d tempest-ServerGroupTestJSON-385159797 tempest-ServerGroupTestJSON-385159797-project-member] Image limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 877.810744] env[63515]: DEBUG nova.virt.hardware [None req-593b35b4-5d74-4c58-a93e-9cad00fc0b2d tempest-ServerGroupTestJSON-385159797 tempest-ServerGroupTestJSON-385159797-project-member] Flavor pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 877.811034] env[63515]: DEBUG nova.virt.hardware [None req-593b35b4-5d74-4c58-a93e-9cad00fc0b2d tempest-ServerGroupTestJSON-385159797 tempest-ServerGroupTestJSON-385159797-project-member] Image pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 877.811382] env[63515]: DEBUG nova.virt.hardware [None req-593b35b4-5d74-4c58-a93e-9cad00fc0b2d tempest-ServerGroupTestJSON-385159797 tempest-ServerGroupTestJSON-385159797-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 877.811673] env[63515]: DEBUG nova.virt.hardware [None req-593b35b4-5d74-4c58-a93e-9cad00fc0b2d tempest-ServerGroupTestJSON-385159797 tempest-ServerGroupTestJSON-385159797-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 877.811889] env[63515]: DEBUG nova.virt.hardware [None req-593b35b4-5d74-4c58-a93e-9cad00fc0b2d tempest-ServerGroupTestJSON-385159797 tempest-ServerGroupTestJSON-385159797-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 877.812106] env[63515]: DEBUG nova.virt.hardware [None req-593b35b4-5d74-4c58-a93e-9cad00fc0b2d tempest-ServerGroupTestJSON-385159797 tempest-ServerGroupTestJSON-385159797-project-member] Got 1 possible topologies {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 877.812287] env[63515]: DEBUG nova.virt.hardware [None req-593b35b4-5d74-4c58-a93e-9cad00fc0b2d tempest-ServerGroupTestJSON-385159797 tempest-ServerGroupTestJSON-385159797-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 877.812467] env[63515]: DEBUG nova.virt.hardware [None req-593b35b4-5d74-4c58-a93e-9cad00fc0b2d tempest-ServerGroupTestJSON-385159797 tempest-ServerGroupTestJSON-385159797-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 877.813639] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3f20034-1883-4ffd-987e-96a55f07d649 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.832172] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5de8dd03-1135-45fb-a790-a68f82c88bb6 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.838976] env[63515]: DEBUG oslo_concurrency.lockutils [None req-3db8c3e0-5f0e-4c46-ab36-d6f97ddeb4f1 tempest-ServerActionsV293TestJSON-1193370169 tempest-ServerActionsV293TestJSON-1193370169-project-member] Lock "2ea99c7a-8b61-4718-bee0-f4ce4fc021af" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 36.762s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 877.844652] env[63515]: DEBUG oslo_vmware.api [None req-561efd36-c7e8-4824-9224-7669ba1ee132 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Task: {'id': task-1111451, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.560926} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 877.853455] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-561efd36-c7e8-4824-9224-7669ba1ee132 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk to [datastore2] 68766cd4-84be-475b-8494-d7ab43a9e969/68766cd4-84be-475b-8494-d7ab43a9e969.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 877.853770] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-561efd36-c7e8-4824-9224-7669ba1ee132 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 68766cd4-84be-475b-8494-d7ab43a9e969] Extending root virtual disk to 1048576 {{(pid=63515) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 877.858750] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-8b4502c8-5692-4fbb-8f08-397506a125e7 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.875317] env[63515]: DEBUG oslo_vmware.api [None req-561efd36-c7e8-4824-9224-7669ba1ee132 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Waiting for the task: (returnval){ [ 877.875317] env[63515]: value = "task-1111453" [ 877.875317] env[63515]: _type = "Task" [ 877.875317] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 877.880940] env[63515]: INFO nova.compute.manager [None req-789266fe-c183-402c-aef4-222438c1eee4 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: eab16df9-7bb5-4576-bca0-769a561c5fe9] Took 47.90 seconds to build instance. [ 877.886935] env[63515]: DEBUG oslo_vmware.api [None req-561efd36-c7e8-4824-9224-7669ba1ee132 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Task: {'id': task-1111453, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.018337] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-088e727f-0709-4167-9e1d-649db8f0df25 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.028509] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7281ac9f-7e96-4e49-9c49-99851e5af3ed {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.063211] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3c32703-c5cd-447c-aa12-5ce91db3036c {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.071093] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4daa6771-5b81-4520-8048-5ae023906c72 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.085086] env[63515]: DEBUG nova.compute.provider_tree [None req-d7e3c195-4890-454e-9f6b-af1a7cc4ee6a tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 878.133838] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1111452, 'name': CreateVM_Task} progress is 25%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.204338] env[63515]: DEBUG nova.network.neutron [req-bb6e93bc-46d8-42a6-a5f2-32d744515850 req-344797c7-c665-442c-81a3-1cd0c366000e service nova] [instance: 0967283f-8dda-4692-b6b7-601545cabf2f] Updated VIF entry in instance network info cache for port 66571248-83b1-46c5-b892-f84ab3882407. {{(pid=63515) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 878.204709] env[63515]: DEBUG nova.network.neutron [req-bb6e93bc-46d8-42a6-a5f2-32d744515850 req-344797c7-c665-442c-81a3-1cd0c366000e service nova] [instance: 0967283f-8dda-4692-b6b7-601545cabf2f] Updating instance_info_cache with network_info: [{"id": "66571248-83b1-46c5-b892-f84ab3882407", "address": "fa:16:3e:7e:e0:aa", "network": {"id": "581d448e-be1b-4d5b-8a6e-c27326aaffc5", "bridge": "br-int", "label": "tempest-ServersTestJSON-1769610073-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "299fdeff647f486390366d5bbf911518", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fb143ef7-8271-4a8a-a4aa-8eba9a89f6a1", "external-id": "nsx-vlan-transportzone-504", "segmentation_id": 504, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap66571248-83", "ovs_interfaceid": "66571248-83b1-46c5-b892-f84ab3882407", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 878.219401] env[63515]: DEBUG nova.network.neutron [None req-82894ab5-858a-446e-a869-69b8a9c84386 tempest-InstanceActionsTestJSON-247065106 tempest-InstanceActionsTestJSON-247065106-project-member] [instance: 0ee9cce2-94b9-4038-ab38-1b3f96a864bc] Updating instance_info_cache with network_info: [{"id": "2fc6eecd-e0cd-41f2-8707-6eb57c2fef5a", "address": "fa:16:3e:e1:7e:62", "network": {"id": "9411d1db-b1ec-4a75-ad97-8942272c078e", "bridge": "br-int", "label": "tempest-InstanceActionsTestJSON-1741013332-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "486afecc6edc4a94b18a1e34e4f92fe5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fc48e29b-113c-4849-850c-35435eab4052", "external-id": "nsx-vlan-transportzone-958", "segmentation_id": 958, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2fc6eecd-e0", "ovs_interfaceid": "2fc6eecd-e0cd-41f2-8707-6eb57c2fef5a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 878.389284] env[63515]: DEBUG oslo_concurrency.lockutils [None req-789266fe-c183-402c-aef4-222438c1eee4 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Lock "eab16df9-7bb5-4576-bca0-769a561c5fe9" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 49.439s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 878.389608] env[63515]: DEBUG oslo_vmware.api [None req-561efd36-c7e8-4824-9224-7669ba1ee132 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Task: {'id': task-1111453, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.121135} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 878.389867] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-561efd36-c7e8-4824-9224-7669ba1ee132 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 68766cd4-84be-475b-8494-d7ab43a9e969] Extended root virtual disk {{(pid=63515) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 878.392102] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32b3bdf5-4ceb-4e68-8f1f-77d007e9a524 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.417348] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-561efd36-c7e8-4824-9224-7669ba1ee132 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 68766cd4-84be-475b-8494-d7ab43a9e969] Reconfiguring VM instance instance-00000049 to attach disk [datastore2] 68766cd4-84be-475b-8494-d7ab43a9e969/68766cd4-84be-475b-8494-d7ab43a9e969.vmdk or device None with type sparse {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 878.417691] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0d127cab-7884-4df3-809a-a1c67c304ca5 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.438804] env[63515]: DEBUG oslo_vmware.api [None req-561efd36-c7e8-4824-9224-7669ba1ee132 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Waiting for the task: (returnval){ [ 878.438804] env[63515]: value = "task-1111454" [ 878.438804] env[63515]: _type = "Task" [ 878.438804] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 878.448950] env[63515]: DEBUG oslo_vmware.api [None req-561efd36-c7e8-4824-9224-7669ba1ee132 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Task: {'id': task-1111454, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.591462] env[63515]: DEBUG nova.scheduler.client.report [None req-d7e3c195-4890-454e-9f6b-af1a7cc4ee6a tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 878.606951] env[63515]: DEBUG nova.network.neutron [None req-593b35b4-5d74-4c58-a93e-9cad00fc0b2d tempest-ServerGroupTestJSON-385159797 tempest-ServerGroupTestJSON-385159797-project-member] [instance: a2e014a2-9e0f-4250-aa38-3a7154149903] Successfully updated port: a453f43a-c524-4eb7-8dcc-9edc8aabdd2c {{(pid=63515) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 878.636633] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1111452, 'name': CreateVM_Task, 'duration_secs': 0.786824} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 878.637039] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0967283f-8dda-4692-b6b7-601545cabf2f] Created VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 878.638147] env[63515]: DEBUG oslo_concurrency.lockutils [None req-3fd65342-5e9e-46b6-830a-47df91928006 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 878.638654] env[63515]: DEBUG oslo_concurrency.lockutils [None req-3fd65342-5e9e-46b6-830a-47df91928006 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 878.639118] env[63515]: DEBUG oslo_concurrency.lockutils [None req-3fd65342-5e9e-46b6-830a-47df91928006 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 878.639510] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ce002d1e-8192-44a1-b3f8-6d03290c6792 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.647025] env[63515]: DEBUG oslo_vmware.api [None req-3fd65342-5e9e-46b6-830a-47df91928006 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Waiting for the task: (returnval){ [ 878.647025] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]52e941ac-2436-56e8-fb44-d2ec3c3d8724" [ 878.647025] env[63515]: _type = "Task" [ 878.647025] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 878.659453] env[63515]: DEBUG oslo_vmware.api [None req-3fd65342-5e9e-46b6-830a-47df91928006 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52e941ac-2436-56e8-fb44-d2ec3c3d8724, 'name': SearchDatastore_Task, 'duration_secs': 0.011138} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 878.660106] env[63515]: DEBUG oslo_concurrency.lockutils [None req-3fd65342-5e9e-46b6-830a-47df91928006 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 878.660501] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-3fd65342-5e9e-46b6-830a-47df91928006 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: 0967283f-8dda-4692-b6b7-601545cabf2f] Processing image 8a120570-cb06-4099-b262-554ca0ad15c5 {{(pid=63515) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 878.660869] env[63515]: DEBUG oslo_concurrency.lockutils [None req-3fd65342-5e9e-46b6-830a-47df91928006 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 878.661354] env[63515]: DEBUG oslo_concurrency.lockutils [None req-3fd65342-5e9e-46b6-830a-47df91928006 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 878.661889] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-3fd65342-5e9e-46b6-830a-47df91928006 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 878.662286] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1806524f-475f-45a2-ab9c-a3121e60ff11 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.672276] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-3fd65342-5e9e-46b6-830a-47df91928006 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 878.672643] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-3fd65342-5e9e-46b6-830a-47df91928006 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63515) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 878.674150] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b9f55674-4a03-458e-8cb1-647d415915dd {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.682111] env[63515]: DEBUG oslo_vmware.api [None req-3fd65342-5e9e-46b6-830a-47df91928006 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Waiting for the task: (returnval){ [ 878.682111] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]52ba37e7-e29c-705a-d1eb-55607e2c32b9" [ 878.682111] env[63515]: _type = "Task" [ 878.682111] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 878.691654] env[63515]: DEBUG oslo_vmware.api [None req-3fd65342-5e9e-46b6-830a-47df91928006 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52ba37e7-e29c-705a-d1eb-55607e2c32b9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.707825] env[63515]: DEBUG oslo_concurrency.lockutils [req-bb6e93bc-46d8-42a6-a5f2-32d744515850 req-344797c7-c665-442c-81a3-1cd0c366000e service nova] Releasing lock "refresh_cache-0967283f-8dda-4692-b6b7-601545cabf2f" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 878.727183] env[63515]: DEBUG oslo_concurrency.lockutils [None req-82894ab5-858a-446e-a869-69b8a9c84386 tempest-InstanceActionsTestJSON-247065106 tempest-InstanceActionsTestJSON-247065106-project-member] Releasing lock "refresh_cache-0ee9cce2-94b9-4038-ab38-1b3f96a864bc" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 878.731023] env[63515]: DEBUG nova.compute.manager [None req-82894ab5-858a-446e-a869-69b8a9c84386 tempest-InstanceActionsTestJSON-247065106 tempest-InstanceActionsTestJSON-247065106-project-member] [instance: 0ee9cce2-94b9-4038-ab38-1b3f96a864bc] Checking state {{(pid=63515) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 878.731814] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-457298b3-635d-4aeb-a0f9-c39fcb461320 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.952211] env[63515]: DEBUG oslo_vmware.api [None req-561efd36-c7e8-4824-9224-7669ba1ee132 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Task: {'id': task-1111454, 'name': ReconfigVM_Task} progress is 99%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.055540] env[63515]: DEBUG oslo_concurrency.lockutils [None req-cf46911e-9c80-4bd3-a167-7e0168ef6827 tempest-ServerDiagnosticsV248Test-1736022867 tempest-ServerDiagnosticsV248Test-1736022867-project-member] Acquiring lock "705b7e9b-3ab6-40ec-8cf6-d86bf0a1faf0" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 879.055870] env[63515]: DEBUG oslo_concurrency.lockutils [None req-cf46911e-9c80-4bd3-a167-7e0168ef6827 tempest-ServerDiagnosticsV248Test-1736022867 tempest-ServerDiagnosticsV248Test-1736022867-project-member] Lock "705b7e9b-3ab6-40ec-8cf6-d86bf0a1faf0" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 879.056116] env[63515]: DEBUG oslo_concurrency.lockutils [None req-cf46911e-9c80-4bd3-a167-7e0168ef6827 tempest-ServerDiagnosticsV248Test-1736022867 tempest-ServerDiagnosticsV248Test-1736022867-project-member] Acquiring lock "705b7e9b-3ab6-40ec-8cf6-d86bf0a1faf0-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 879.056347] env[63515]: DEBUG oslo_concurrency.lockutils [None req-cf46911e-9c80-4bd3-a167-7e0168ef6827 tempest-ServerDiagnosticsV248Test-1736022867 tempest-ServerDiagnosticsV248Test-1736022867-project-member] Lock "705b7e9b-3ab6-40ec-8cf6-d86bf0a1faf0-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 879.056559] env[63515]: DEBUG oslo_concurrency.lockutils [None req-cf46911e-9c80-4bd3-a167-7e0168ef6827 tempest-ServerDiagnosticsV248Test-1736022867 tempest-ServerDiagnosticsV248Test-1736022867-project-member] Lock "705b7e9b-3ab6-40ec-8cf6-d86bf0a1faf0-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 879.060183] env[63515]: INFO nova.compute.manager [None req-cf46911e-9c80-4bd3-a167-7e0168ef6827 tempest-ServerDiagnosticsV248Test-1736022867 tempest-ServerDiagnosticsV248Test-1736022867-project-member] [instance: 705b7e9b-3ab6-40ec-8cf6-d86bf0a1faf0] Terminating instance [ 879.062768] env[63515]: DEBUG oslo_concurrency.lockutils [None req-cf46911e-9c80-4bd3-a167-7e0168ef6827 tempest-ServerDiagnosticsV248Test-1736022867 tempest-ServerDiagnosticsV248Test-1736022867-project-member] Acquiring lock "refresh_cache-705b7e9b-3ab6-40ec-8cf6-d86bf0a1faf0" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 879.062984] env[63515]: DEBUG oslo_concurrency.lockutils [None req-cf46911e-9c80-4bd3-a167-7e0168ef6827 tempest-ServerDiagnosticsV248Test-1736022867 tempest-ServerDiagnosticsV248Test-1736022867-project-member] Acquired lock "refresh_cache-705b7e9b-3ab6-40ec-8cf6-d86bf0a1faf0" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 879.063229] env[63515]: DEBUG nova.network.neutron [None req-cf46911e-9c80-4bd3-a167-7e0168ef6827 tempest-ServerDiagnosticsV248Test-1736022867 tempest-ServerDiagnosticsV248Test-1736022867-project-member] [instance: 705b7e9b-3ab6-40ec-8cf6-d86bf0a1faf0] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 879.095066] env[63515]: DEBUG oslo_concurrency.lockutils [None req-d7e3c195-4890-454e-9f6b-af1a7cc4ee6a tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.391s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 879.095671] env[63515]: DEBUG nova.compute.manager [None req-d7e3c195-4890-454e-9f6b-af1a7cc4ee6a tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: ae5d6f34-bac6-4ba0-b0c3-8205982ddc88] Start building networks asynchronously for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 879.099062] env[63515]: DEBUG oslo_concurrency.lockutils [None req-9a80c171-6c9b-44ae-ab0a-573a8347fcf4 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.810s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 879.101387] env[63515]: INFO nova.compute.claims [None req-9a80c171-6c9b-44ae-ab0a-573a8347fcf4 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: b1bf5a0a-42e7-41a6-90fb-a0c311623fdd] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 879.114015] env[63515]: DEBUG oslo_concurrency.lockutils [None req-593b35b4-5d74-4c58-a93e-9cad00fc0b2d tempest-ServerGroupTestJSON-385159797 tempest-ServerGroupTestJSON-385159797-project-member] Acquiring lock "refresh_cache-a2e014a2-9e0f-4250-aa38-3a7154149903" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 879.114015] env[63515]: DEBUG oslo_concurrency.lockutils [None req-593b35b4-5d74-4c58-a93e-9cad00fc0b2d tempest-ServerGroupTestJSON-385159797 tempest-ServerGroupTestJSON-385159797-project-member] Acquired lock "refresh_cache-a2e014a2-9e0f-4250-aa38-3a7154149903" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 879.114015] env[63515]: DEBUG nova.network.neutron [None req-593b35b4-5d74-4c58-a93e-9cad00fc0b2d tempest-ServerGroupTestJSON-385159797 tempest-ServerGroupTestJSON-385159797-project-member] [instance: a2e014a2-9e0f-4250-aa38-3a7154149903] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 879.190874] env[63515]: DEBUG oslo_vmware.api [None req-3fd65342-5e9e-46b6-830a-47df91928006 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52ba37e7-e29c-705a-d1eb-55607e2c32b9, 'name': SearchDatastore_Task, 'duration_secs': 0.012124} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 879.191753] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5239e89a-9fdc-47d9-8eb5-d4ee79441b47 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.197608] env[63515]: DEBUG oslo_vmware.api [None req-3fd65342-5e9e-46b6-830a-47df91928006 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Waiting for the task: (returnval){ [ 879.197608] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]527e9ad7-7391-968c-7d64-0e91abcf5f61" [ 879.197608] env[63515]: _type = "Task" [ 879.197608] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.206324] env[63515]: DEBUG oslo_vmware.api [None req-3fd65342-5e9e-46b6-830a-47df91928006 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]527e9ad7-7391-968c-7d64-0e91abcf5f61, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.361050] env[63515]: DEBUG nova.compute.manager [req-85bff543-95b6-4e87-8673-1c77db611942 req-bda2bce6-201c-4892-a6c3-f55478cee521 service nova] [instance: a2e014a2-9e0f-4250-aa38-3a7154149903] Received event network-vif-plugged-a453f43a-c524-4eb7-8dcc-9edc8aabdd2c {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 879.361284] env[63515]: DEBUG oslo_concurrency.lockutils [req-85bff543-95b6-4e87-8673-1c77db611942 req-bda2bce6-201c-4892-a6c3-f55478cee521 service nova] Acquiring lock "a2e014a2-9e0f-4250-aa38-3a7154149903-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 879.361499] env[63515]: DEBUG oslo_concurrency.lockutils [req-85bff543-95b6-4e87-8673-1c77db611942 req-bda2bce6-201c-4892-a6c3-f55478cee521 service nova] Lock "a2e014a2-9e0f-4250-aa38-3a7154149903-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 879.361671] env[63515]: DEBUG oslo_concurrency.lockutils [req-85bff543-95b6-4e87-8673-1c77db611942 req-bda2bce6-201c-4892-a6c3-f55478cee521 service nova] Lock "a2e014a2-9e0f-4250-aa38-3a7154149903-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 879.361841] env[63515]: DEBUG nova.compute.manager [req-85bff543-95b6-4e87-8673-1c77db611942 req-bda2bce6-201c-4892-a6c3-f55478cee521 service nova] [instance: a2e014a2-9e0f-4250-aa38-3a7154149903] No waiting events found dispatching network-vif-plugged-a453f43a-c524-4eb7-8dcc-9edc8aabdd2c {{(pid=63515) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 879.362017] env[63515]: WARNING nova.compute.manager [req-85bff543-95b6-4e87-8673-1c77db611942 req-bda2bce6-201c-4892-a6c3-f55478cee521 service nova] [instance: a2e014a2-9e0f-4250-aa38-3a7154149903] Received unexpected event network-vif-plugged-a453f43a-c524-4eb7-8dcc-9edc8aabdd2c for instance with vm_state building and task_state spawning. [ 879.362711] env[63515]: DEBUG nova.compute.manager [req-85bff543-95b6-4e87-8673-1c77db611942 req-bda2bce6-201c-4892-a6c3-f55478cee521 service nova] [instance: a2e014a2-9e0f-4250-aa38-3a7154149903] Received event network-changed-a453f43a-c524-4eb7-8dcc-9edc8aabdd2c {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 879.362947] env[63515]: DEBUG nova.compute.manager [req-85bff543-95b6-4e87-8673-1c77db611942 req-bda2bce6-201c-4892-a6c3-f55478cee521 service nova] [instance: a2e014a2-9e0f-4250-aa38-3a7154149903] Refreshing instance network info cache due to event network-changed-a453f43a-c524-4eb7-8dcc-9edc8aabdd2c. {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 879.363206] env[63515]: DEBUG oslo_concurrency.lockutils [req-85bff543-95b6-4e87-8673-1c77db611942 req-bda2bce6-201c-4892-a6c3-f55478cee521 service nova] Acquiring lock "refresh_cache-a2e014a2-9e0f-4250-aa38-3a7154149903" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 879.462801] env[63515]: DEBUG oslo_vmware.api [None req-561efd36-c7e8-4824-9224-7669ba1ee132 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Task: {'id': task-1111454, 'name': ReconfigVM_Task, 'duration_secs': 0.53136} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 879.463223] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-561efd36-c7e8-4824-9224-7669ba1ee132 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 68766cd4-84be-475b-8494-d7ab43a9e969] Reconfigured VM instance instance-00000049 to attach disk [datastore2] 68766cd4-84be-475b-8494-d7ab43a9e969/68766cd4-84be-475b-8494-d7ab43a9e969.vmdk or device None with type sparse {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 879.464660] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-2838a5f9-9df1-4e00-82e7-69b13352742d {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.476055] env[63515]: DEBUG oslo_vmware.api [None req-561efd36-c7e8-4824-9224-7669ba1ee132 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Waiting for the task: (returnval){ [ 879.476055] env[63515]: value = "task-1111455" [ 879.476055] env[63515]: _type = "Task" [ 879.476055] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.487668] env[63515]: DEBUG oslo_vmware.api [None req-561efd36-c7e8-4824-9224-7669ba1ee132 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Task: {'id': task-1111455, 'name': Rename_Task} progress is 10%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.608978] env[63515]: DEBUG nova.compute.utils [None req-d7e3c195-4890-454e-9f6b-af1a7cc4ee6a tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Using /dev/sd instead of None {{(pid=63515) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 879.619597] env[63515]: DEBUG nova.network.neutron [None req-cf46911e-9c80-4bd3-a167-7e0168ef6827 tempest-ServerDiagnosticsV248Test-1736022867 tempest-ServerDiagnosticsV248Test-1736022867-project-member] [instance: 705b7e9b-3ab6-40ec-8cf6-d86bf0a1faf0] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 879.621327] env[63515]: DEBUG nova.compute.manager [None req-d7e3c195-4890-454e-9f6b-af1a7cc4ee6a tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: ae5d6f34-bac6-4ba0-b0c3-8205982ddc88] Allocating IP information in the background. {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 879.621514] env[63515]: DEBUG nova.network.neutron [None req-d7e3c195-4890-454e-9f6b-af1a7cc4ee6a tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: ae5d6f34-bac6-4ba0-b0c3-8205982ddc88] allocate_for_instance() {{(pid=63515) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 879.715180] env[63515]: DEBUG nova.network.neutron [None req-593b35b4-5d74-4c58-a93e-9cad00fc0b2d tempest-ServerGroupTestJSON-385159797 tempest-ServerGroupTestJSON-385159797-project-member] [instance: a2e014a2-9e0f-4250-aa38-3a7154149903] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 879.728654] env[63515]: DEBUG oslo_vmware.api [None req-3fd65342-5e9e-46b6-830a-47df91928006 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]527e9ad7-7391-968c-7d64-0e91abcf5f61, 'name': SearchDatastore_Task, 'duration_secs': 0.012204} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 879.728912] env[63515]: DEBUG oslo_concurrency.lockutils [None req-3fd65342-5e9e-46b6-830a-47df91928006 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 879.729384] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-3fd65342-5e9e-46b6-830a-47df91928006 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk to [datastore2] 0967283f-8dda-4692-b6b7-601545cabf2f/0967283f-8dda-4692-b6b7-601545cabf2f.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 879.729948] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-cfce2e98-36cc-400a-b4dd-4ed9bc3ea03d {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.745221] env[63515]: DEBUG oslo_vmware.api [None req-3fd65342-5e9e-46b6-830a-47df91928006 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Waiting for the task: (returnval){ [ 879.745221] env[63515]: value = "task-1111456" [ 879.745221] env[63515]: _type = "Task" [ 879.745221] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.758403] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5481873-9c48-4063-826d-38c4adcc6a4d {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.763588] env[63515]: DEBUG oslo_vmware.api [None req-3fd65342-5e9e-46b6-830a-47df91928006 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Task: {'id': task-1111456, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.773130] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-82894ab5-858a-446e-a869-69b8a9c84386 tempest-InstanceActionsTestJSON-247065106 tempest-InstanceActionsTestJSON-247065106-project-member] [instance: 0ee9cce2-94b9-4038-ab38-1b3f96a864bc] Doing hard reboot of VM {{(pid=63515) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1063}} [ 879.773417] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ResetVM_Task with opID=oslo.vmware-0d91d52d-2ea7-4ac8-b5ca-a122dd5a7ba6 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.779944] env[63515]: DEBUG oslo_vmware.api [None req-82894ab5-858a-446e-a869-69b8a9c84386 tempest-InstanceActionsTestJSON-247065106 tempest-InstanceActionsTestJSON-247065106-project-member] Waiting for the task: (returnval){ [ 879.779944] env[63515]: value = "task-1111457" [ 879.779944] env[63515]: _type = "Task" [ 879.779944] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.796217] env[63515]: DEBUG oslo_vmware.api [None req-82894ab5-858a-446e-a869-69b8a9c84386 tempest-InstanceActionsTestJSON-247065106 tempest-InstanceActionsTestJSON-247065106-project-member] Task: {'id': task-1111457, 'name': ResetVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.801050] env[63515]: DEBUG nova.policy [None req-d7e3c195-4890-454e-9f6b-af1a7cc4ee6a tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f58ac812970845dc825202b842feb5d0', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ea099ce07bfb4a8da014d9303cf552eb', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63515) authorize /opt/stack/nova/nova/policy.py:201}} [ 879.804095] env[63515]: DEBUG nova.network.neutron [None req-cf46911e-9c80-4bd3-a167-7e0168ef6827 tempest-ServerDiagnosticsV248Test-1736022867 tempest-ServerDiagnosticsV248Test-1736022867-project-member] [instance: 705b7e9b-3ab6-40ec-8cf6-d86bf0a1faf0] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 879.956528] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52d38a5c-4784-4bc7-8464-d7efa256c2fa {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.964776] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d540f722-5acb-4a70-9c1e-772e3ab2a9e8 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.007408] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56da4041-57bb-4237-ba25-1e5eca105482 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.015145] env[63515]: DEBUG oslo_vmware.api [None req-561efd36-c7e8-4824-9224-7669ba1ee132 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Task: {'id': task-1111455, 'name': Rename_Task} progress is 99%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.018475] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5f5db98-8254-448e-9ec9-344643807d14 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.034384] env[63515]: DEBUG nova.compute.provider_tree [None req-9a80c171-6c9b-44ae-ab0a-573a8347fcf4 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 880.109646] env[63515]: DEBUG nova.network.neutron [None req-593b35b4-5d74-4c58-a93e-9cad00fc0b2d tempest-ServerGroupTestJSON-385159797 tempest-ServerGroupTestJSON-385159797-project-member] [instance: a2e014a2-9e0f-4250-aa38-3a7154149903] Updating instance_info_cache with network_info: [{"id": "a453f43a-c524-4eb7-8dcc-9edc8aabdd2c", "address": "fa:16:3e:ff:f7:bc", "network": {"id": "0be4a890-7ace-450f-8ab2-86e89a9531e9", "bridge": "br-int", "label": "tempest-ServerGroupTestJSON-298628005-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "07921f3cdb3140b9b5703c418cb6a90e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6eb7e3e9-5cc2-40f1-a6eb-f70f06531667", "external-id": "nsx-vlan-transportzone-938", "segmentation_id": 938, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa453f43a-c5", "ovs_interfaceid": "a453f43a-c524-4eb7-8dcc-9edc8aabdd2c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 880.118953] env[63515]: DEBUG nova.compute.manager [None req-d7e3c195-4890-454e-9f6b-af1a7cc4ee6a tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: ae5d6f34-bac6-4ba0-b0c3-8205982ddc88] Start building block device mappings for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 880.216573] env[63515]: DEBUG nova.compute.manager [req-21b90731-a451-4a70-8d93-20a33856ac8e req-b68f5f7f-89c3-4411-9c48-e4cf801c1716 service nova] [instance: 2ea99c7a-8b61-4718-bee0-f4ce4fc021af] Received event network-changed-1ddac51a-f040-4ab9-b3bd-3f75ed5e1f5b {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 880.216836] env[63515]: DEBUG nova.compute.manager [req-21b90731-a451-4a70-8d93-20a33856ac8e req-b68f5f7f-89c3-4411-9c48-e4cf801c1716 service nova] [instance: 2ea99c7a-8b61-4718-bee0-f4ce4fc021af] Refreshing instance network info cache due to event network-changed-1ddac51a-f040-4ab9-b3bd-3f75ed5e1f5b. {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 880.217325] env[63515]: DEBUG oslo_concurrency.lockutils [req-21b90731-a451-4a70-8d93-20a33856ac8e req-b68f5f7f-89c3-4411-9c48-e4cf801c1716 service nova] Acquiring lock "refresh_cache-2ea99c7a-8b61-4718-bee0-f4ce4fc021af" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 880.217562] env[63515]: DEBUG oslo_concurrency.lockutils [req-21b90731-a451-4a70-8d93-20a33856ac8e req-b68f5f7f-89c3-4411-9c48-e4cf801c1716 service nova] Acquired lock "refresh_cache-2ea99c7a-8b61-4718-bee0-f4ce4fc021af" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 880.217764] env[63515]: DEBUG nova.network.neutron [req-21b90731-a451-4a70-8d93-20a33856ac8e req-b68f5f7f-89c3-4411-9c48-e4cf801c1716 service nova] [instance: 2ea99c7a-8b61-4718-bee0-f4ce4fc021af] Refreshing network info cache for port 1ddac51a-f040-4ab9-b3bd-3f75ed5e1f5b {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 880.261673] env[63515]: DEBUG oslo_vmware.api [None req-3fd65342-5e9e-46b6-830a-47df91928006 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Task: {'id': task-1111456, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.294634] env[63515]: DEBUG oslo_vmware.api [None req-82894ab5-858a-446e-a869-69b8a9c84386 tempest-InstanceActionsTestJSON-247065106 tempest-InstanceActionsTestJSON-247065106-project-member] Task: {'id': task-1111457, 'name': ResetVM_Task, 'duration_secs': 0.12035} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.294920] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-82894ab5-858a-446e-a869-69b8a9c84386 tempest-InstanceActionsTestJSON-247065106 tempest-InstanceActionsTestJSON-247065106-project-member] [instance: 0ee9cce2-94b9-4038-ab38-1b3f96a864bc] Did hard reboot of VM {{(pid=63515) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1067}} [ 880.295162] env[63515]: DEBUG nova.compute.manager [None req-82894ab5-858a-446e-a869-69b8a9c84386 tempest-InstanceActionsTestJSON-247065106 tempest-InstanceActionsTestJSON-247065106-project-member] [instance: 0ee9cce2-94b9-4038-ab38-1b3f96a864bc] Checking state {{(pid=63515) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 880.296074] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71d17c53-6ff1-4e0b-a69f-44c53a2399b4 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.307758] env[63515]: DEBUG oslo_concurrency.lockutils [None req-cf46911e-9c80-4bd3-a167-7e0168ef6827 tempest-ServerDiagnosticsV248Test-1736022867 tempest-ServerDiagnosticsV248Test-1736022867-project-member] Releasing lock "refresh_cache-705b7e9b-3ab6-40ec-8cf6-d86bf0a1faf0" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 880.308233] env[63515]: DEBUG nova.compute.manager [None req-cf46911e-9c80-4bd3-a167-7e0168ef6827 tempest-ServerDiagnosticsV248Test-1736022867 tempest-ServerDiagnosticsV248Test-1736022867-project-member] [instance: 705b7e9b-3ab6-40ec-8cf6-d86bf0a1faf0] Start destroying the instance on the hypervisor. {{(pid=63515) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 880.308426] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-cf46911e-9c80-4bd3-a167-7e0168ef6827 tempest-ServerDiagnosticsV248Test-1736022867 tempest-ServerDiagnosticsV248Test-1736022867-project-member] [instance: 705b7e9b-3ab6-40ec-8cf6-d86bf0a1faf0] Destroying instance {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 880.310431] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-139cd229-d1f8-40c5-b21a-950a2bfe5d56 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.317379] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-cf46911e-9c80-4bd3-a167-7e0168ef6827 tempest-ServerDiagnosticsV248Test-1736022867 tempest-ServerDiagnosticsV248Test-1736022867-project-member] [instance: 705b7e9b-3ab6-40ec-8cf6-d86bf0a1faf0] Powering off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 880.317803] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-60d833dc-7e38-4e15-a064-5b23d8c4d57f {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.324402] env[63515]: DEBUG oslo_vmware.api [None req-cf46911e-9c80-4bd3-a167-7e0168ef6827 tempest-ServerDiagnosticsV248Test-1736022867 tempest-ServerDiagnosticsV248Test-1736022867-project-member] Waiting for the task: (returnval){ [ 880.324402] env[63515]: value = "task-1111458" [ 880.324402] env[63515]: _type = "Task" [ 880.324402] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.332930] env[63515]: DEBUG oslo_vmware.api [None req-cf46911e-9c80-4bd3-a167-7e0168ef6827 tempest-ServerDiagnosticsV248Test-1736022867 tempest-ServerDiagnosticsV248Test-1736022867-project-member] Task: {'id': task-1111458, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.495195] env[63515]: DEBUG nova.network.neutron [None req-d7e3c195-4890-454e-9f6b-af1a7cc4ee6a tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: ae5d6f34-bac6-4ba0-b0c3-8205982ddc88] Successfully created port: dbc4bf29-34a5-424b-bf71-8c80ba74be85 {{(pid=63515) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 880.515977] env[63515]: DEBUG oslo_vmware.api [None req-561efd36-c7e8-4824-9224-7669ba1ee132 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Task: {'id': task-1111455, 'name': Rename_Task} progress is 99%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.537344] env[63515]: DEBUG nova.scheduler.client.report [None req-9a80c171-6c9b-44ae-ab0a-573a8347fcf4 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 880.610678] env[63515]: DEBUG oslo_concurrency.lockutils [None req-593b35b4-5d74-4c58-a93e-9cad00fc0b2d tempest-ServerGroupTestJSON-385159797 tempest-ServerGroupTestJSON-385159797-project-member] Releasing lock "refresh_cache-a2e014a2-9e0f-4250-aa38-3a7154149903" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 880.611044] env[63515]: DEBUG nova.compute.manager [None req-593b35b4-5d74-4c58-a93e-9cad00fc0b2d tempest-ServerGroupTestJSON-385159797 tempest-ServerGroupTestJSON-385159797-project-member] [instance: a2e014a2-9e0f-4250-aa38-3a7154149903] Instance network_info: |[{"id": "a453f43a-c524-4eb7-8dcc-9edc8aabdd2c", "address": "fa:16:3e:ff:f7:bc", "network": {"id": "0be4a890-7ace-450f-8ab2-86e89a9531e9", "bridge": "br-int", "label": "tempest-ServerGroupTestJSON-298628005-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "07921f3cdb3140b9b5703c418cb6a90e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6eb7e3e9-5cc2-40f1-a6eb-f70f06531667", "external-id": "nsx-vlan-transportzone-938", "segmentation_id": 938, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa453f43a-c5", "ovs_interfaceid": "a453f43a-c524-4eb7-8dcc-9edc8aabdd2c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 880.611631] env[63515]: DEBUG oslo_concurrency.lockutils [req-85bff543-95b6-4e87-8673-1c77db611942 req-bda2bce6-201c-4892-a6c3-f55478cee521 service nova] Acquired lock "refresh_cache-a2e014a2-9e0f-4250-aa38-3a7154149903" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 880.611631] env[63515]: DEBUG nova.network.neutron [req-85bff543-95b6-4e87-8673-1c77db611942 req-bda2bce6-201c-4892-a6c3-f55478cee521 service nova] [instance: a2e014a2-9e0f-4250-aa38-3a7154149903] Refreshing network info cache for port a453f43a-c524-4eb7-8dcc-9edc8aabdd2c {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 880.612808] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-593b35b4-5d74-4c58-a93e-9cad00fc0b2d tempest-ServerGroupTestJSON-385159797 tempest-ServerGroupTestJSON-385159797-project-member] [instance: a2e014a2-9e0f-4250-aa38-3a7154149903] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ff:f7:bc', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '6eb7e3e9-5cc2-40f1-a6eb-f70f06531667', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a453f43a-c524-4eb7-8dcc-9edc8aabdd2c', 'vif_model': 'vmxnet3'}] {{(pid=63515) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 880.622294] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-593b35b4-5d74-4c58-a93e-9cad00fc0b2d tempest-ServerGroupTestJSON-385159797 tempest-ServerGroupTestJSON-385159797-project-member] Creating folder: Project (07921f3cdb3140b9b5703c418cb6a90e). Parent ref: group-v243370. {{(pid=63515) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 880.622878] env[63515]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b822c5c8-c1a8-46ad-b88a-96576bb5a65f {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.640249] env[63515]: INFO nova.virt.vmwareapi.vm_util [None req-593b35b4-5d74-4c58-a93e-9cad00fc0b2d tempest-ServerGroupTestJSON-385159797 tempest-ServerGroupTestJSON-385159797-project-member] Created folder: Project (07921f3cdb3140b9b5703c418cb6a90e) in parent group-v243370. [ 880.640596] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-593b35b4-5d74-4c58-a93e-9cad00fc0b2d tempest-ServerGroupTestJSON-385159797 tempest-ServerGroupTestJSON-385159797-project-member] Creating folder: Instances. Parent ref: group-v243509. {{(pid=63515) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 880.640991] env[63515]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-11cac109-dc86-406a-bccb-46ddbc4c91a8 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.653879] env[63515]: INFO nova.virt.vmwareapi.vm_util [None req-593b35b4-5d74-4c58-a93e-9cad00fc0b2d tempest-ServerGroupTestJSON-385159797 tempest-ServerGroupTestJSON-385159797-project-member] Created folder: Instances in parent group-v243509. [ 880.653879] env[63515]: DEBUG oslo.service.loopingcall [None req-593b35b4-5d74-4c58-a93e-9cad00fc0b2d tempest-ServerGroupTestJSON-385159797 tempest-ServerGroupTestJSON-385159797-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 880.653879] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a2e014a2-9e0f-4250-aa38-3a7154149903] Creating VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 880.653879] env[63515]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ed3a33f8-9193-4827-9c16-1a1948919e90 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.678401] env[63515]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 880.678401] env[63515]: value = "task-1111461" [ 880.678401] env[63515]: _type = "Task" [ 880.678401] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.690930] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1111461, 'name': CreateVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.757113] env[63515]: DEBUG oslo_vmware.api [None req-3fd65342-5e9e-46b6-830a-47df91928006 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Task: {'id': task-1111456, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.873461} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.757462] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-3fd65342-5e9e-46b6-830a-47df91928006 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk to [datastore2] 0967283f-8dda-4692-b6b7-601545cabf2f/0967283f-8dda-4692-b6b7-601545cabf2f.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 880.757690] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-3fd65342-5e9e-46b6-830a-47df91928006 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: 0967283f-8dda-4692-b6b7-601545cabf2f] Extending root virtual disk to 1048576 {{(pid=63515) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 880.757980] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-7fdb4c3c-628e-4c9a-ab9b-c1ec7d91294f {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.767912] env[63515]: DEBUG oslo_vmware.api [None req-3fd65342-5e9e-46b6-830a-47df91928006 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Waiting for the task: (returnval){ [ 880.767912] env[63515]: value = "task-1111462" [ 880.767912] env[63515]: _type = "Task" [ 880.767912] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.776494] env[63515]: DEBUG oslo_vmware.api [None req-3fd65342-5e9e-46b6-830a-47df91928006 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Task: {'id': task-1111462, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.813068] env[63515]: DEBUG oslo_concurrency.lockutils [None req-82894ab5-858a-446e-a869-69b8a9c84386 tempest-InstanceActionsTestJSON-247065106 tempest-InstanceActionsTestJSON-247065106-project-member] Lock "0ee9cce2-94b9-4038-ab38-1b3f96a864bc" "released" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: held 4.276s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 880.834870] env[63515]: DEBUG oslo_vmware.api [None req-cf46911e-9c80-4bd3-a167-7e0168ef6827 tempest-ServerDiagnosticsV248Test-1736022867 tempest-ServerDiagnosticsV248Test-1736022867-project-member] Task: {'id': task-1111458, 'name': PowerOffVM_Task, 'duration_secs': 0.301165} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.835174] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-cf46911e-9c80-4bd3-a167-7e0168ef6827 tempest-ServerDiagnosticsV248Test-1736022867 tempest-ServerDiagnosticsV248Test-1736022867-project-member] [instance: 705b7e9b-3ab6-40ec-8cf6-d86bf0a1faf0] Powered off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 880.835351] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-cf46911e-9c80-4bd3-a167-7e0168ef6827 tempest-ServerDiagnosticsV248Test-1736022867 tempest-ServerDiagnosticsV248Test-1736022867-project-member] [instance: 705b7e9b-3ab6-40ec-8cf6-d86bf0a1faf0] Unregistering the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 880.835616] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3f7552e0-e6ea-4f3f-8df8-afcecd9f2271 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.863110] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-cf46911e-9c80-4bd3-a167-7e0168ef6827 tempest-ServerDiagnosticsV248Test-1736022867 tempest-ServerDiagnosticsV248Test-1736022867-project-member] [instance: 705b7e9b-3ab6-40ec-8cf6-d86bf0a1faf0] Unregistered the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 880.863391] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-cf46911e-9c80-4bd3-a167-7e0168ef6827 tempest-ServerDiagnosticsV248Test-1736022867 tempest-ServerDiagnosticsV248Test-1736022867-project-member] [instance: 705b7e9b-3ab6-40ec-8cf6-d86bf0a1faf0] Deleting contents of the VM from datastore datastore1 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 880.863604] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-cf46911e-9c80-4bd3-a167-7e0168ef6827 tempest-ServerDiagnosticsV248Test-1736022867 tempest-ServerDiagnosticsV248Test-1736022867-project-member] Deleting the datastore file [datastore1] 705b7e9b-3ab6-40ec-8cf6-d86bf0a1faf0 {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 880.863892] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-753a0324-cae0-4541-a541-f1a8fbde40d9 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.870347] env[63515]: DEBUG oslo_vmware.api [None req-cf46911e-9c80-4bd3-a167-7e0168ef6827 tempest-ServerDiagnosticsV248Test-1736022867 tempest-ServerDiagnosticsV248Test-1736022867-project-member] Waiting for the task: (returnval){ [ 880.870347] env[63515]: value = "task-1111464" [ 880.870347] env[63515]: _type = "Task" [ 880.870347] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.878602] env[63515]: DEBUG oslo_vmware.api [None req-cf46911e-9c80-4bd3-a167-7e0168ef6827 tempest-ServerDiagnosticsV248Test-1736022867 tempest-ServerDiagnosticsV248Test-1736022867-project-member] Task: {'id': task-1111464, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.013853] env[63515]: DEBUG oslo_vmware.api [None req-561efd36-c7e8-4824-9224-7669ba1ee132 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Task: {'id': task-1111455, 'name': Rename_Task, 'duration_secs': 1.181871} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 881.016583] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-561efd36-c7e8-4824-9224-7669ba1ee132 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 68766cd4-84be-475b-8494-d7ab43a9e969] Powering on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 881.016583] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d39e6d41-05ed-4cb3-8e62-05dca9f998cf {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.021270] env[63515]: DEBUG oslo_vmware.api [None req-561efd36-c7e8-4824-9224-7669ba1ee132 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Waiting for the task: (returnval){ [ 881.021270] env[63515]: value = "task-1111465" [ 881.021270] env[63515]: _type = "Task" [ 881.021270] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 881.031095] env[63515]: DEBUG oslo_vmware.api [None req-561efd36-c7e8-4824-9224-7669ba1ee132 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Task: {'id': task-1111465, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.043166] env[63515]: DEBUG oslo_concurrency.lockutils [None req-9a80c171-6c9b-44ae-ab0a-573a8347fcf4 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 1.944s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 881.043649] env[63515]: DEBUG nova.compute.manager [None req-9a80c171-6c9b-44ae-ab0a-573a8347fcf4 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: b1bf5a0a-42e7-41a6-90fb-a0c311623fdd] Start building networks asynchronously for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 881.049048] env[63515]: DEBUG oslo_concurrency.lockutils [None req-bfac7ffd-4aa9-41c4-9bda-3f289bbda0a2 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 16.101s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 881.136088] env[63515]: DEBUG nova.compute.manager [None req-d7e3c195-4890-454e-9f6b-af1a7cc4ee6a tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: ae5d6f34-bac6-4ba0-b0c3-8205982ddc88] Start spawning the instance on the hypervisor. {{(pid=63515) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 881.173111] env[63515]: DEBUG nova.virt.hardware [None req-d7e3c195-4890-454e-9f6b-af1a7cc4ee6a tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T02:52:14Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T02:51:56Z,direct_url=,disk_format='vmdk',id=8a120570-cb06-4099-b262-554ca0ad15c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b569255cc43e42e7ae3f2b4ad37c6ef4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T02:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 881.173111] env[63515]: DEBUG nova.virt.hardware [None req-d7e3c195-4890-454e-9f6b-af1a7cc4ee6a tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Flavor limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 881.173111] env[63515]: DEBUG nova.virt.hardware [None req-d7e3c195-4890-454e-9f6b-af1a7cc4ee6a tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Image limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 881.173111] env[63515]: DEBUG nova.virt.hardware [None req-d7e3c195-4890-454e-9f6b-af1a7cc4ee6a tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Flavor pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 881.173111] env[63515]: DEBUG nova.virt.hardware [None req-d7e3c195-4890-454e-9f6b-af1a7cc4ee6a tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Image pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 881.173111] env[63515]: DEBUG nova.virt.hardware [None req-d7e3c195-4890-454e-9f6b-af1a7cc4ee6a tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 881.173362] env[63515]: DEBUG nova.virt.hardware [None req-d7e3c195-4890-454e-9f6b-af1a7cc4ee6a tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 881.173749] env[63515]: DEBUG nova.virt.hardware [None req-d7e3c195-4890-454e-9f6b-af1a7cc4ee6a tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 881.173749] env[63515]: DEBUG nova.virt.hardware [None req-d7e3c195-4890-454e-9f6b-af1a7cc4ee6a tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Got 1 possible topologies {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 881.173749] env[63515]: DEBUG nova.virt.hardware [None req-d7e3c195-4890-454e-9f6b-af1a7cc4ee6a tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 881.173898] env[63515]: DEBUG nova.virt.hardware [None req-d7e3c195-4890-454e-9f6b-af1a7cc4ee6a tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 881.174781] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a1a5ded-b585-443e-a127-b60bf7801d05 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.192346] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4003d532-5bd2-4c8b-ac9f-7887bf395bbd {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.212806] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1111461, 'name': CreateVM_Task} progress is 99%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.213776] env[63515]: DEBUG nova.network.neutron [req-21b90731-a451-4a70-8d93-20a33856ac8e req-b68f5f7f-89c3-4411-9c48-e4cf801c1716 service nova] [instance: 2ea99c7a-8b61-4718-bee0-f4ce4fc021af] Updated VIF entry in instance network info cache for port 1ddac51a-f040-4ab9-b3bd-3f75ed5e1f5b. {{(pid=63515) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 881.214124] env[63515]: DEBUG nova.network.neutron [req-21b90731-a451-4a70-8d93-20a33856ac8e req-b68f5f7f-89c3-4411-9c48-e4cf801c1716 service nova] [instance: 2ea99c7a-8b61-4718-bee0-f4ce4fc021af] Updating instance_info_cache with network_info: [{"id": "1ddac51a-f040-4ab9-b3bd-3f75ed5e1f5b", "address": "fa:16:3e:f0:73:5e", "network": {"id": "4803e9b2-07bf-4e7d-b71a-5ea8fd8cde6c", "bridge": "br-int", "label": "tempest-ServerActionsV293TestJSON-439688245-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.195", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ae6fb15713db493793797f015fe6fcf9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "48512b02-ad5c-4105-ba7d-fd4775acf8e1", "external-id": "nsx-vlan-transportzone-516", "segmentation_id": 516, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1ddac51a-f0", "ovs_interfaceid": "1ddac51a-f040-4ab9-b3bd-3f75ed5e1f5b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 881.276536] env[63515]: DEBUG oslo_vmware.api [None req-3fd65342-5e9e-46b6-830a-47df91928006 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Task: {'id': task-1111462, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.151883} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 881.276824] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-3fd65342-5e9e-46b6-830a-47df91928006 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: 0967283f-8dda-4692-b6b7-601545cabf2f] Extended root virtual disk {{(pid=63515) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 881.277626] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6cdeb316-e9f2-44ff-9bcc-88ee7213d2b9 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.302423] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-3fd65342-5e9e-46b6-830a-47df91928006 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: 0967283f-8dda-4692-b6b7-601545cabf2f] Reconfiguring VM instance instance-0000004a to attach disk [datastore2] 0967283f-8dda-4692-b6b7-601545cabf2f/0967283f-8dda-4692-b6b7-601545cabf2f.vmdk or device None with type sparse {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 881.305424] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f155d968-429a-46e9-93ce-049765dd3b32 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.324942] env[63515]: DEBUG oslo_vmware.api [None req-3fd65342-5e9e-46b6-830a-47df91928006 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Waiting for the task: (returnval){ [ 881.324942] env[63515]: value = "task-1111466" [ 881.324942] env[63515]: _type = "Task" [ 881.324942] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 881.336943] env[63515]: DEBUG oslo_vmware.api [None req-3fd65342-5e9e-46b6-830a-47df91928006 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Task: {'id': task-1111466, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.380101] env[63515]: DEBUG oslo_vmware.api [None req-cf46911e-9c80-4bd3-a167-7e0168ef6827 tempest-ServerDiagnosticsV248Test-1736022867 tempest-ServerDiagnosticsV248Test-1736022867-project-member] Task: {'id': task-1111464, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.264964} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 881.380391] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-cf46911e-9c80-4bd3-a167-7e0168ef6827 tempest-ServerDiagnosticsV248Test-1736022867 tempest-ServerDiagnosticsV248Test-1736022867-project-member] Deleted the datastore file {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 881.380630] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-cf46911e-9c80-4bd3-a167-7e0168ef6827 tempest-ServerDiagnosticsV248Test-1736022867 tempest-ServerDiagnosticsV248Test-1736022867-project-member] [instance: 705b7e9b-3ab6-40ec-8cf6-d86bf0a1faf0] Deleted contents of the VM from datastore datastore1 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 881.380862] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-cf46911e-9c80-4bd3-a167-7e0168ef6827 tempest-ServerDiagnosticsV248Test-1736022867 tempest-ServerDiagnosticsV248Test-1736022867-project-member] [instance: 705b7e9b-3ab6-40ec-8cf6-d86bf0a1faf0] Instance destroyed {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 881.381088] env[63515]: INFO nova.compute.manager [None req-cf46911e-9c80-4bd3-a167-7e0168ef6827 tempest-ServerDiagnosticsV248Test-1736022867 tempest-ServerDiagnosticsV248Test-1736022867-project-member] [instance: 705b7e9b-3ab6-40ec-8cf6-d86bf0a1faf0] Took 1.07 seconds to destroy the instance on the hypervisor. [ 881.381376] env[63515]: DEBUG oslo.service.loopingcall [None req-cf46911e-9c80-4bd3-a167-7e0168ef6827 tempest-ServerDiagnosticsV248Test-1736022867 tempest-ServerDiagnosticsV248Test-1736022867-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 881.381626] env[63515]: DEBUG nova.compute.manager [-] [instance: 705b7e9b-3ab6-40ec-8cf6-d86bf0a1faf0] Deallocating network for instance {{(pid=63515) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 881.381762] env[63515]: DEBUG nova.network.neutron [-] [instance: 705b7e9b-3ab6-40ec-8cf6-d86bf0a1faf0] deallocate_for_instance() {{(pid=63515) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 881.403361] env[63515]: DEBUG nova.network.neutron [-] [instance: 705b7e9b-3ab6-40ec-8cf6-d86bf0a1faf0] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 881.474625] env[63515]: DEBUG nova.network.neutron [req-85bff543-95b6-4e87-8673-1c77db611942 req-bda2bce6-201c-4892-a6c3-f55478cee521 service nova] [instance: a2e014a2-9e0f-4250-aa38-3a7154149903] Updated VIF entry in instance network info cache for port a453f43a-c524-4eb7-8dcc-9edc8aabdd2c. {{(pid=63515) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 881.475012] env[63515]: DEBUG nova.network.neutron [req-85bff543-95b6-4e87-8673-1c77db611942 req-bda2bce6-201c-4892-a6c3-f55478cee521 service nova] [instance: a2e014a2-9e0f-4250-aa38-3a7154149903] Updating instance_info_cache with network_info: [{"id": "a453f43a-c524-4eb7-8dcc-9edc8aabdd2c", "address": "fa:16:3e:ff:f7:bc", "network": {"id": "0be4a890-7ace-450f-8ab2-86e89a9531e9", "bridge": "br-int", "label": "tempest-ServerGroupTestJSON-298628005-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "07921f3cdb3140b9b5703c418cb6a90e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6eb7e3e9-5cc2-40f1-a6eb-f70f06531667", "external-id": "nsx-vlan-transportzone-938", "segmentation_id": 938, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa453f43a-c5", "ovs_interfaceid": "a453f43a-c524-4eb7-8dcc-9edc8aabdd2c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 881.537564] env[63515]: DEBUG oslo_vmware.api [None req-561efd36-c7e8-4824-9224-7669ba1ee132 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Task: {'id': task-1111465, 'name': PowerOnVM_Task} progress is 88%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.552596] env[63515]: DEBUG nova.compute.utils [None req-9a80c171-6c9b-44ae-ab0a-573a8347fcf4 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Using /dev/sd instead of None {{(pid=63515) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 881.558018] env[63515]: DEBUG nova.compute.manager [None req-9a80c171-6c9b-44ae-ab0a-573a8347fcf4 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: b1bf5a0a-42e7-41a6-90fb-a0c311623fdd] Allocating IP information in the background. {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 881.558142] env[63515]: DEBUG nova.network.neutron [None req-9a80c171-6c9b-44ae-ab0a-573a8347fcf4 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: b1bf5a0a-42e7-41a6-90fb-a0c311623fdd] allocate_for_instance() {{(pid=63515) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 881.577391] env[63515]: DEBUG nova.compute.manager [req-c2145e60-549b-4620-885a-924a7e3c7401 req-d6472db9-ad4a-4a64-b649-95543f24afb4 service nova] [instance: eab16df9-7bb5-4576-bca0-769a561c5fe9] Received event network-changed-4168f5a5-c282-4f13-9c97-0b0236bd659b {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 881.577694] env[63515]: DEBUG nova.compute.manager [req-c2145e60-549b-4620-885a-924a7e3c7401 req-d6472db9-ad4a-4a64-b649-95543f24afb4 service nova] [instance: eab16df9-7bb5-4576-bca0-769a561c5fe9] Refreshing instance network info cache due to event network-changed-4168f5a5-c282-4f13-9c97-0b0236bd659b. {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 881.577977] env[63515]: DEBUG oslo_concurrency.lockutils [req-c2145e60-549b-4620-885a-924a7e3c7401 req-d6472db9-ad4a-4a64-b649-95543f24afb4 service nova] Acquiring lock "refresh_cache-eab16df9-7bb5-4576-bca0-769a561c5fe9" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 881.578174] env[63515]: DEBUG oslo_concurrency.lockutils [req-c2145e60-549b-4620-885a-924a7e3c7401 req-d6472db9-ad4a-4a64-b649-95543f24afb4 service nova] Acquired lock "refresh_cache-eab16df9-7bb5-4576-bca0-769a561c5fe9" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 881.578354] env[63515]: DEBUG nova.network.neutron [req-c2145e60-549b-4620-885a-924a7e3c7401 req-d6472db9-ad4a-4a64-b649-95543f24afb4 service nova] [instance: eab16df9-7bb5-4576-bca0-769a561c5fe9] Refreshing network info cache for port 4168f5a5-c282-4f13-9c97-0b0236bd659b {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 881.623342] env[63515]: DEBUG nova.policy [None req-9a80c171-6c9b-44ae-ab0a-573a8347fcf4 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '002c1a9364184b82b649fc18d017d97f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2259667ddc0c42639ea1af3b6e6fa0e8', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63515) authorize /opt/stack/nova/nova/policy.py:201}} [ 881.700559] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1111461, 'name': CreateVM_Task, 'duration_secs': 0.557889} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 881.705273] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a2e014a2-9e0f-4250-aa38-3a7154149903] Created VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 881.706145] env[63515]: DEBUG oslo_concurrency.lockutils [None req-593b35b4-5d74-4c58-a93e-9cad00fc0b2d tempest-ServerGroupTestJSON-385159797 tempest-ServerGroupTestJSON-385159797-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 881.706495] env[63515]: DEBUG oslo_concurrency.lockutils [None req-593b35b4-5d74-4c58-a93e-9cad00fc0b2d tempest-ServerGroupTestJSON-385159797 tempest-ServerGroupTestJSON-385159797-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 881.706637] env[63515]: DEBUG oslo_concurrency.lockutils [None req-593b35b4-5d74-4c58-a93e-9cad00fc0b2d tempest-ServerGroupTestJSON-385159797 tempest-ServerGroupTestJSON-385159797-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 881.707235] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4e7c0af1-b330-40a7-a821-e7a19960e412 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.713030] env[63515]: DEBUG oslo_vmware.api [None req-593b35b4-5d74-4c58-a93e-9cad00fc0b2d tempest-ServerGroupTestJSON-385159797 tempest-ServerGroupTestJSON-385159797-project-member] Waiting for the task: (returnval){ [ 881.713030] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]52924c6a-61f8-a383-3df4-697c0740e104" [ 881.713030] env[63515]: _type = "Task" [ 881.713030] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 881.720987] env[63515]: DEBUG oslo_concurrency.lockutils [req-21b90731-a451-4a70-8d93-20a33856ac8e req-b68f5f7f-89c3-4411-9c48-e4cf801c1716 service nova] Releasing lock "refresh_cache-2ea99c7a-8b61-4718-bee0-f4ce4fc021af" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 881.726036] env[63515]: DEBUG oslo_vmware.api [None req-593b35b4-5d74-4c58-a93e-9cad00fc0b2d tempest-ServerGroupTestJSON-385159797 tempest-ServerGroupTestJSON-385159797-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52924c6a-61f8-a383-3df4-697c0740e104, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.778282] env[63515]: DEBUG oslo_concurrency.lockutils [None req-ceee2529-e58a-42aa-a26e-32b1ce124ccb tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] Acquiring lock "324f7e89-3c17-474d-af1b-6d7a6d042510" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 881.778830] env[63515]: DEBUG oslo_concurrency.lockutils [None req-ceee2529-e58a-42aa-a26e-32b1ce124ccb tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] Lock "324f7e89-3c17-474d-af1b-6d7a6d042510" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 881.842531] env[63515]: DEBUG oslo_vmware.api [None req-3fd65342-5e9e-46b6-830a-47df91928006 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Task: {'id': task-1111466, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.906419] env[63515]: DEBUG nova.network.neutron [-] [instance: 705b7e9b-3ab6-40ec-8cf6-d86bf0a1faf0] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 881.910750] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-003750f2-42b3-468a-b3d4-aa6b37460dba {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.919367] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b59deefa-f6bf-47d8-904a-8e74406ea8aa {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.950550] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a2f1e35-752c-4018-aab7-38affdc944c8 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.958472] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e59eb4a8-93f4-48ce-9c1d-5efdbadd97b0 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.973920] env[63515]: DEBUG nova.compute.provider_tree [None req-bfac7ffd-4aa9-41c4-9bda-3f289bbda0a2 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 881.981926] env[63515]: DEBUG oslo_concurrency.lockutils [req-85bff543-95b6-4e87-8673-1c77db611942 req-bda2bce6-201c-4892-a6c3-f55478cee521 service nova] Releasing lock "refresh_cache-a2e014a2-9e0f-4250-aa38-3a7154149903" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 882.018659] env[63515]: DEBUG nova.network.neutron [None req-9a80c171-6c9b-44ae-ab0a-573a8347fcf4 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: b1bf5a0a-42e7-41a6-90fb-a0c311623fdd] Successfully created port: 8ab9e788-50a9-4669-8448-77cd0983386e {{(pid=63515) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 882.032276] env[63515]: DEBUG oslo_vmware.api [None req-561efd36-c7e8-4824-9224-7669ba1ee132 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Task: {'id': task-1111465, 'name': PowerOnVM_Task, 'duration_secs': 0.628284} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 882.032276] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-561efd36-c7e8-4824-9224-7669ba1ee132 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 68766cd4-84be-475b-8494-d7ab43a9e969] Powered on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 882.032435] env[63515]: INFO nova.compute.manager [None req-561efd36-c7e8-4824-9224-7669ba1ee132 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 68766cd4-84be-475b-8494-d7ab43a9e969] Took 10.30 seconds to spawn the instance on the hypervisor. [ 882.032648] env[63515]: DEBUG nova.compute.manager [None req-561efd36-c7e8-4824-9224-7669ba1ee132 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 68766cd4-84be-475b-8494-d7ab43a9e969] Checking state {{(pid=63515) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 882.033449] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a220771d-fc50-4098-9192-a8c8aa6e5dd2 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.058856] env[63515]: DEBUG nova.compute.manager [None req-9a80c171-6c9b-44ae-ab0a-573a8347fcf4 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: b1bf5a0a-42e7-41a6-90fb-a0c311623fdd] Start building block device mappings for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 882.128752] env[63515]: DEBUG oslo_concurrency.lockutils [None req-1d02c1b8-b71d-4e69-b38f-4c4377a39898 tempest-InstanceActionsTestJSON-247065106 tempest-InstanceActionsTestJSON-247065106-project-member] Acquiring lock "0ee9cce2-94b9-4038-ab38-1b3f96a864bc" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 882.129063] env[63515]: DEBUG oslo_concurrency.lockutils [None req-1d02c1b8-b71d-4e69-b38f-4c4377a39898 tempest-InstanceActionsTestJSON-247065106 tempest-InstanceActionsTestJSON-247065106-project-member] Lock "0ee9cce2-94b9-4038-ab38-1b3f96a864bc" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 882.129288] env[63515]: DEBUG oslo_concurrency.lockutils [None req-1d02c1b8-b71d-4e69-b38f-4c4377a39898 tempest-InstanceActionsTestJSON-247065106 tempest-InstanceActionsTestJSON-247065106-project-member] Acquiring lock "0ee9cce2-94b9-4038-ab38-1b3f96a864bc-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 882.129595] env[63515]: DEBUG oslo_concurrency.lockutils [None req-1d02c1b8-b71d-4e69-b38f-4c4377a39898 tempest-InstanceActionsTestJSON-247065106 tempest-InstanceActionsTestJSON-247065106-project-member] Lock "0ee9cce2-94b9-4038-ab38-1b3f96a864bc-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 882.129663] env[63515]: DEBUG oslo_concurrency.lockutils [None req-1d02c1b8-b71d-4e69-b38f-4c4377a39898 tempest-InstanceActionsTestJSON-247065106 tempest-InstanceActionsTestJSON-247065106-project-member] Lock "0ee9cce2-94b9-4038-ab38-1b3f96a864bc-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 882.134050] env[63515]: INFO nova.compute.manager [None req-1d02c1b8-b71d-4e69-b38f-4c4377a39898 tempest-InstanceActionsTestJSON-247065106 tempest-InstanceActionsTestJSON-247065106-project-member] [instance: 0ee9cce2-94b9-4038-ab38-1b3f96a864bc] Terminating instance [ 882.136135] env[63515]: DEBUG nova.compute.manager [None req-1d02c1b8-b71d-4e69-b38f-4c4377a39898 tempest-InstanceActionsTestJSON-247065106 tempest-InstanceActionsTestJSON-247065106-project-member] [instance: 0ee9cce2-94b9-4038-ab38-1b3f96a864bc] Start destroying the instance on the hypervisor. {{(pid=63515) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 882.136338] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-1d02c1b8-b71d-4e69-b38f-4c4377a39898 tempest-InstanceActionsTestJSON-247065106 tempest-InstanceActionsTestJSON-247065106-project-member] [instance: 0ee9cce2-94b9-4038-ab38-1b3f96a864bc] Destroying instance {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 882.137217] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67d32bcd-6c6c-42b3-a036-8c5a5b475361 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.146238] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-1d02c1b8-b71d-4e69-b38f-4c4377a39898 tempest-InstanceActionsTestJSON-247065106 tempest-InstanceActionsTestJSON-247065106-project-member] [instance: 0ee9cce2-94b9-4038-ab38-1b3f96a864bc] Powering off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 882.146496] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a3150284-08fd-4f67-821c-c08cfaef84ee {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.153782] env[63515]: DEBUG oslo_vmware.api [None req-1d02c1b8-b71d-4e69-b38f-4c4377a39898 tempest-InstanceActionsTestJSON-247065106 tempest-InstanceActionsTestJSON-247065106-project-member] Waiting for the task: (returnval){ [ 882.153782] env[63515]: value = "task-1111467" [ 882.153782] env[63515]: _type = "Task" [ 882.153782] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.161876] env[63515]: DEBUG oslo_vmware.api [None req-1d02c1b8-b71d-4e69-b38f-4c4377a39898 tempest-InstanceActionsTestJSON-247065106 tempest-InstanceActionsTestJSON-247065106-project-member] Task: {'id': task-1111467, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.225582] env[63515]: DEBUG oslo_vmware.api [None req-593b35b4-5d74-4c58-a93e-9cad00fc0b2d tempest-ServerGroupTestJSON-385159797 tempest-ServerGroupTestJSON-385159797-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52924c6a-61f8-a383-3df4-697c0740e104, 'name': SearchDatastore_Task, 'duration_secs': 0.024824} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 882.225991] env[63515]: DEBUG oslo_concurrency.lockutils [None req-593b35b4-5d74-4c58-a93e-9cad00fc0b2d tempest-ServerGroupTestJSON-385159797 tempest-ServerGroupTestJSON-385159797-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 882.226261] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-593b35b4-5d74-4c58-a93e-9cad00fc0b2d tempest-ServerGroupTestJSON-385159797 tempest-ServerGroupTestJSON-385159797-project-member] [instance: a2e014a2-9e0f-4250-aa38-3a7154149903] Processing image 8a120570-cb06-4099-b262-554ca0ad15c5 {{(pid=63515) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 882.226500] env[63515]: DEBUG oslo_concurrency.lockutils [None req-593b35b4-5d74-4c58-a93e-9cad00fc0b2d tempest-ServerGroupTestJSON-385159797 tempest-ServerGroupTestJSON-385159797-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 882.226646] env[63515]: DEBUG oslo_concurrency.lockutils [None req-593b35b4-5d74-4c58-a93e-9cad00fc0b2d tempest-ServerGroupTestJSON-385159797 tempest-ServerGroupTestJSON-385159797-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 882.226824] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-593b35b4-5d74-4c58-a93e-9cad00fc0b2d tempest-ServerGroupTestJSON-385159797 tempest-ServerGroupTestJSON-385159797-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 882.227126] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6cdb705a-e275-4d7e-9937-96abfdbb53d0 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.240061] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-593b35b4-5d74-4c58-a93e-9cad00fc0b2d tempest-ServerGroupTestJSON-385159797 tempest-ServerGroupTestJSON-385159797-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 882.240385] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-593b35b4-5d74-4c58-a93e-9cad00fc0b2d tempest-ServerGroupTestJSON-385159797 tempest-ServerGroupTestJSON-385159797-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63515) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 882.241202] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c1375936-8017-4df8-a6bf-d9e1c28f0f84 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.247018] env[63515]: DEBUG oslo_vmware.api [None req-593b35b4-5d74-4c58-a93e-9cad00fc0b2d tempest-ServerGroupTestJSON-385159797 tempest-ServerGroupTestJSON-385159797-project-member] Waiting for the task: (returnval){ [ 882.247018] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]52c2b93d-79db-06e7-ad79-2f4f26af18a6" [ 882.247018] env[63515]: _type = "Task" [ 882.247018] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.256280] env[63515]: DEBUG oslo_vmware.api [None req-593b35b4-5d74-4c58-a93e-9cad00fc0b2d tempest-ServerGroupTestJSON-385159797 tempest-ServerGroupTestJSON-385159797-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52c2b93d-79db-06e7-ad79-2f4f26af18a6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.284509] env[63515]: DEBUG nova.compute.manager [None req-ceee2529-e58a-42aa-a26e-32b1ce124ccb tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] [instance: 324f7e89-3c17-474d-af1b-6d7a6d042510] Starting instance... {{(pid=63515) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 882.339767] env[63515]: DEBUG oslo_vmware.api [None req-3fd65342-5e9e-46b6-830a-47df91928006 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Task: {'id': task-1111466, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.409817] env[63515]: INFO nova.compute.manager [-] [instance: 705b7e9b-3ab6-40ec-8cf6-d86bf0a1faf0] Took 1.03 seconds to deallocate network for instance. [ 882.476737] env[63515]: DEBUG nova.scheduler.client.report [None req-bfac7ffd-4aa9-41c4-9bda-3f289bbda0a2 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 882.552536] env[63515]: INFO nova.compute.manager [None req-561efd36-c7e8-4824-9224-7669ba1ee132 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 68766cd4-84be-475b-8494-d7ab43a9e969] Took 32.78 seconds to build instance. [ 882.607524] env[63515]: DEBUG nova.network.neutron [req-c2145e60-549b-4620-885a-924a7e3c7401 req-d6472db9-ad4a-4a64-b649-95543f24afb4 service nova] [instance: eab16df9-7bb5-4576-bca0-769a561c5fe9] Updated VIF entry in instance network info cache for port 4168f5a5-c282-4f13-9c97-0b0236bd659b. {{(pid=63515) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 882.607945] env[63515]: DEBUG nova.network.neutron [req-c2145e60-549b-4620-885a-924a7e3c7401 req-d6472db9-ad4a-4a64-b649-95543f24afb4 service nova] [instance: eab16df9-7bb5-4576-bca0-769a561c5fe9] Updating instance_info_cache with network_info: [{"id": "4168f5a5-c282-4f13-9c97-0b0236bd659b", "address": "fa:16:3e:a0:6a:50", "network": {"id": "88dfb29d-41e0-4df7-8f70-e52cda05620c", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1355578400-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.203", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7d4ada9ec35f42b19c6480a9101d21a4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "357d2811-e990-4985-9f9e-b158d10d3699", "external-id": "nsx-vlan-transportzone-641", "segmentation_id": 641, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4168f5a5-c2", "ovs_interfaceid": "4168f5a5-c282-4f13-9c97-0b0236bd659b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 882.665425] env[63515]: DEBUG oslo_vmware.api [None req-1d02c1b8-b71d-4e69-b38f-4c4377a39898 tempest-InstanceActionsTestJSON-247065106 tempest-InstanceActionsTestJSON-247065106-project-member] Task: {'id': task-1111467, 'name': PowerOffVM_Task, 'duration_secs': 0.310699} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 882.665646] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-1d02c1b8-b71d-4e69-b38f-4c4377a39898 tempest-InstanceActionsTestJSON-247065106 tempest-InstanceActionsTestJSON-247065106-project-member] [instance: 0ee9cce2-94b9-4038-ab38-1b3f96a864bc] Powered off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 882.665849] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-1d02c1b8-b71d-4e69-b38f-4c4377a39898 tempest-InstanceActionsTestJSON-247065106 tempest-InstanceActionsTestJSON-247065106-project-member] [instance: 0ee9cce2-94b9-4038-ab38-1b3f96a864bc] Unregistering the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 882.666133] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-2b733ebf-0949-448d-b855-7a02011b9385 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.763791] env[63515]: DEBUG nova.compute.manager [req-00e2b9a6-2f1f-4896-8977-aca92900a11d req-6fd7024d-5ec6-4ef5-b1ff-dc2e0646f124 service nova] [instance: ae5d6f34-bac6-4ba0-b0c3-8205982ddc88] Received event network-vif-plugged-dbc4bf29-34a5-424b-bf71-8c80ba74be85 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 882.763791] env[63515]: DEBUG oslo_concurrency.lockutils [req-00e2b9a6-2f1f-4896-8977-aca92900a11d req-6fd7024d-5ec6-4ef5-b1ff-dc2e0646f124 service nova] Acquiring lock "ae5d6f34-bac6-4ba0-b0c3-8205982ddc88-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 882.763791] env[63515]: DEBUG oslo_concurrency.lockutils [req-00e2b9a6-2f1f-4896-8977-aca92900a11d req-6fd7024d-5ec6-4ef5-b1ff-dc2e0646f124 service nova] Lock "ae5d6f34-bac6-4ba0-b0c3-8205982ddc88-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 882.763791] env[63515]: DEBUG oslo_concurrency.lockutils [req-00e2b9a6-2f1f-4896-8977-aca92900a11d req-6fd7024d-5ec6-4ef5-b1ff-dc2e0646f124 service nova] Lock "ae5d6f34-bac6-4ba0-b0c3-8205982ddc88-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 882.763791] env[63515]: DEBUG nova.compute.manager [req-00e2b9a6-2f1f-4896-8977-aca92900a11d req-6fd7024d-5ec6-4ef5-b1ff-dc2e0646f124 service nova] [instance: ae5d6f34-bac6-4ba0-b0c3-8205982ddc88] No waiting events found dispatching network-vif-plugged-dbc4bf29-34a5-424b-bf71-8c80ba74be85 {{(pid=63515) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 882.764044] env[63515]: WARNING nova.compute.manager [req-00e2b9a6-2f1f-4896-8977-aca92900a11d req-6fd7024d-5ec6-4ef5-b1ff-dc2e0646f124 service nova] [instance: ae5d6f34-bac6-4ba0-b0c3-8205982ddc88] Received unexpected event network-vif-plugged-dbc4bf29-34a5-424b-bf71-8c80ba74be85 for instance with vm_state building and task_state spawning. [ 882.764577] env[63515]: DEBUG oslo_vmware.api [None req-593b35b4-5d74-4c58-a93e-9cad00fc0b2d tempest-ServerGroupTestJSON-385159797 tempest-ServerGroupTestJSON-385159797-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52c2b93d-79db-06e7-ad79-2f4f26af18a6, 'name': SearchDatastore_Task, 'duration_secs': 0.014044} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 882.765375] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2d158bd7-98e6-41a2-a749-ed739a31f871 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.771425] env[63515]: DEBUG oslo_vmware.api [None req-593b35b4-5d74-4c58-a93e-9cad00fc0b2d tempest-ServerGroupTestJSON-385159797 tempest-ServerGroupTestJSON-385159797-project-member] Waiting for the task: (returnval){ [ 882.771425] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]52e21db6-d08e-62ff-59bc-4ce28d022a19" [ 882.771425] env[63515]: _type = "Task" [ 882.771425] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.779419] env[63515]: DEBUG oslo_vmware.api [None req-593b35b4-5d74-4c58-a93e-9cad00fc0b2d tempest-ServerGroupTestJSON-385159797 tempest-ServerGroupTestJSON-385159797-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52e21db6-d08e-62ff-59bc-4ce28d022a19, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.805652] env[63515]: DEBUG oslo_concurrency.lockutils [None req-ceee2529-e58a-42aa-a26e-32b1ce124ccb tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 882.837346] env[63515]: DEBUG oslo_vmware.api [None req-3fd65342-5e9e-46b6-830a-47df91928006 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Task: {'id': task-1111466, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.880483] env[63515]: DEBUG nova.network.neutron [None req-d7e3c195-4890-454e-9f6b-af1a7cc4ee6a tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: ae5d6f34-bac6-4ba0-b0c3-8205982ddc88] Successfully updated port: dbc4bf29-34a5-424b-bf71-8c80ba74be85 {{(pid=63515) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 882.916301] env[63515]: DEBUG oslo_concurrency.lockutils [None req-cf46911e-9c80-4bd3-a167-7e0168ef6827 tempest-ServerDiagnosticsV248Test-1736022867 tempest-ServerDiagnosticsV248Test-1736022867-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 883.056065] env[63515]: DEBUG oslo_concurrency.lockutils [None req-561efd36-c7e8-4824-9224-7669ba1ee132 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Lock "68766cd4-84be-475b-8494-d7ab43a9e969" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 34.294s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 883.067568] env[63515]: DEBUG nova.compute.manager [None req-9a80c171-6c9b-44ae-ab0a-573a8347fcf4 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: b1bf5a0a-42e7-41a6-90fb-a0c311623fdd] Start spawning the instance on the hypervisor. {{(pid=63515) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 883.095809] env[63515]: DEBUG nova.virt.hardware [None req-9a80c171-6c9b-44ae-ab0a-573a8347fcf4 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T02:52:14Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='9cf9c44f3eabd3c47f9b9deafd58751a',container_format='bare',created_at=2024-10-03T02:59:09Z,direct_url=,disk_format='vmdk',id=d1918637-9f8a-4d89-9bc2-49e7d0fdaeee,min_disk=1,min_ram=0,name='tempest-test-snap-1115221592',owner='2259667ddc0c42639ea1af3b6e6fa0e8',properties=ImageMetaProps,protected=,size=21334016,status='active',tags=,updated_at=2024-10-03T02:59:26Z,virtual_size=,visibility=), allow threads: False {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 883.096152] env[63515]: DEBUG nova.virt.hardware [None req-9a80c171-6c9b-44ae-ab0a-573a8347fcf4 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Flavor limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 883.096319] env[63515]: DEBUG nova.virt.hardware [None req-9a80c171-6c9b-44ae-ab0a-573a8347fcf4 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Image limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 883.096505] env[63515]: DEBUG nova.virt.hardware [None req-9a80c171-6c9b-44ae-ab0a-573a8347fcf4 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Flavor pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 883.096656] env[63515]: DEBUG nova.virt.hardware [None req-9a80c171-6c9b-44ae-ab0a-573a8347fcf4 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Image pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 883.096847] env[63515]: DEBUG nova.virt.hardware [None req-9a80c171-6c9b-44ae-ab0a-573a8347fcf4 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 883.097119] env[63515]: DEBUG nova.virt.hardware [None req-9a80c171-6c9b-44ae-ab0a-573a8347fcf4 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 883.097307] env[63515]: DEBUG nova.virt.hardware [None req-9a80c171-6c9b-44ae-ab0a-573a8347fcf4 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 883.097486] env[63515]: DEBUG nova.virt.hardware [None req-9a80c171-6c9b-44ae-ab0a-573a8347fcf4 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Got 1 possible topologies {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 883.097657] env[63515]: DEBUG nova.virt.hardware [None req-9a80c171-6c9b-44ae-ab0a-573a8347fcf4 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 883.097847] env[63515]: DEBUG nova.virt.hardware [None req-9a80c171-6c9b-44ae-ab0a-573a8347fcf4 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 883.098744] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2fd4503-c84f-4bdc-b95f-aaa0982f5249 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.107254] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7aa3a7aa-3c4d-43f7-aa8c-f6de90655c45 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.111808] env[63515]: DEBUG oslo_concurrency.lockutils [req-c2145e60-549b-4620-885a-924a7e3c7401 req-d6472db9-ad4a-4a64-b649-95543f24afb4 service nova] Releasing lock "refresh_cache-eab16df9-7bb5-4576-bca0-769a561c5fe9" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 883.281895] env[63515]: DEBUG oslo_vmware.api [None req-593b35b4-5d74-4c58-a93e-9cad00fc0b2d tempest-ServerGroupTestJSON-385159797 tempest-ServerGroupTestJSON-385159797-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52e21db6-d08e-62ff-59bc-4ce28d022a19, 'name': SearchDatastore_Task, 'duration_secs': 0.009993} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 883.282198] env[63515]: DEBUG oslo_concurrency.lockutils [None req-593b35b4-5d74-4c58-a93e-9cad00fc0b2d tempest-ServerGroupTestJSON-385159797 tempest-ServerGroupTestJSON-385159797-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 883.282469] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-593b35b4-5d74-4c58-a93e-9cad00fc0b2d tempest-ServerGroupTestJSON-385159797 tempest-ServerGroupTestJSON-385159797-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk to [datastore2] a2e014a2-9e0f-4250-aa38-3a7154149903/a2e014a2-9e0f-4250-aa38-3a7154149903.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 883.282725] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-1b642c51-01e1-4d47-b5ec-fd8e699d7209 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.289346] env[63515]: DEBUG oslo_vmware.api [None req-593b35b4-5d74-4c58-a93e-9cad00fc0b2d tempest-ServerGroupTestJSON-385159797 tempest-ServerGroupTestJSON-385159797-project-member] Waiting for the task: (returnval){ [ 883.289346] env[63515]: value = "task-1111469" [ 883.289346] env[63515]: _type = "Task" [ 883.289346] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 883.298561] env[63515]: DEBUG oslo_vmware.api [None req-593b35b4-5d74-4c58-a93e-9cad00fc0b2d tempest-ServerGroupTestJSON-385159797 tempest-ServerGroupTestJSON-385159797-project-member] Task: {'id': task-1111469, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.339897] env[63515]: DEBUG oslo_vmware.api [None req-3fd65342-5e9e-46b6-830a-47df91928006 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Task: {'id': task-1111466, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.384848] env[63515]: DEBUG oslo_concurrency.lockutils [None req-d7e3c195-4890-454e-9f6b-af1a7cc4ee6a tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Acquiring lock "refresh_cache-ae5d6f34-bac6-4ba0-b0c3-8205982ddc88" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 883.385389] env[63515]: DEBUG oslo_concurrency.lockutils [None req-d7e3c195-4890-454e-9f6b-af1a7cc4ee6a tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Acquired lock "refresh_cache-ae5d6f34-bac6-4ba0-b0c3-8205982ddc88" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 883.385684] env[63515]: DEBUG nova.network.neutron [None req-d7e3c195-4890-454e-9f6b-af1a7cc4ee6a tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: ae5d6f34-bac6-4ba0-b0c3-8205982ddc88] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 883.490667] env[63515]: DEBUG oslo_concurrency.lockutils [None req-bfac7ffd-4aa9-41c4-9bda-3f289bbda0a2 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.442s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 883.494012] env[63515]: DEBUG oslo_concurrency.lockutils [None req-dbe24c17-88c2-40e6-987c-d70326fd4bd3 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 6.749s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 883.496275] env[63515]: INFO nova.compute.claims [None req-dbe24c17-88c2-40e6-987c-d70326fd4bd3 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: e13da90c-28e6-43d2-99b7-19c5095954ca] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 883.575149] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-1d02c1b8-b71d-4e69-b38f-4c4377a39898 tempest-InstanceActionsTestJSON-247065106 tempest-InstanceActionsTestJSON-247065106-project-member] [instance: 0ee9cce2-94b9-4038-ab38-1b3f96a864bc] Unregistered the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 883.575876] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-1d02c1b8-b71d-4e69-b38f-4c4377a39898 tempest-InstanceActionsTestJSON-247065106 tempest-InstanceActionsTestJSON-247065106-project-member] [instance: 0ee9cce2-94b9-4038-ab38-1b3f96a864bc] Deleting contents of the VM from datastore datastore1 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 883.575876] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-1d02c1b8-b71d-4e69-b38f-4c4377a39898 tempest-InstanceActionsTestJSON-247065106 tempest-InstanceActionsTestJSON-247065106-project-member] Deleting the datastore file [datastore1] 0ee9cce2-94b9-4038-ab38-1b3f96a864bc {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 883.576241] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-663f7b04-3e97-4580-9ff5-1186d91367cf {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.587339] env[63515]: DEBUG oslo_vmware.api [None req-1d02c1b8-b71d-4e69-b38f-4c4377a39898 tempest-InstanceActionsTestJSON-247065106 tempest-InstanceActionsTestJSON-247065106-project-member] Waiting for the task: (returnval){ [ 883.587339] env[63515]: value = "task-1111470" [ 883.587339] env[63515]: _type = "Task" [ 883.587339] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 883.595975] env[63515]: DEBUG oslo_vmware.api [None req-1d02c1b8-b71d-4e69-b38f-4c4377a39898 tempest-InstanceActionsTestJSON-247065106 tempest-InstanceActionsTestJSON-247065106-project-member] Task: {'id': task-1111470, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.706811] env[63515]: DEBUG nova.compute.manager [req-e58309d0-882c-43fb-8a76-8beec3010ae8 req-338397ef-4cf0-4ca7-b469-0e2050a99320 service nova] [instance: 68766cd4-84be-475b-8494-d7ab43a9e969] Received event network-changed-28b59076-140f-46b5-b4a9-0bf530a1d43d {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 883.707117] env[63515]: DEBUG nova.compute.manager [req-e58309d0-882c-43fb-8a76-8beec3010ae8 req-338397ef-4cf0-4ca7-b469-0e2050a99320 service nova] [instance: 68766cd4-84be-475b-8494-d7ab43a9e969] Refreshing instance network info cache due to event network-changed-28b59076-140f-46b5-b4a9-0bf530a1d43d. {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 883.707371] env[63515]: DEBUG oslo_concurrency.lockutils [req-e58309d0-882c-43fb-8a76-8beec3010ae8 req-338397ef-4cf0-4ca7-b469-0e2050a99320 service nova] Acquiring lock "refresh_cache-68766cd4-84be-475b-8494-d7ab43a9e969" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 883.707521] env[63515]: DEBUG oslo_concurrency.lockutils [req-e58309d0-882c-43fb-8a76-8beec3010ae8 req-338397ef-4cf0-4ca7-b469-0e2050a99320 service nova] Acquired lock "refresh_cache-68766cd4-84be-475b-8494-d7ab43a9e969" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 883.707689] env[63515]: DEBUG nova.network.neutron [req-e58309d0-882c-43fb-8a76-8beec3010ae8 req-338397ef-4cf0-4ca7-b469-0e2050a99320 service nova] [instance: 68766cd4-84be-475b-8494-d7ab43a9e969] Refreshing network info cache for port 28b59076-140f-46b5-b4a9-0bf530a1d43d {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 883.807136] env[63515]: DEBUG oslo_vmware.api [None req-593b35b4-5d74-4c58-a93e-9cad00fc0b2d tempest-ServerGroupTestJSON-385159797 tempest-ServerGroupTestJSON-385159797-project-member] Task: {'id': task-1111469, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.506608} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 883.807136] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-593b35b4-5d74-4c58-a93e-9cad00fc0b2d tempest-ServerGroupTestJSON-385159797 tempest-ServerGroupTestJSON-385159797-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk to [datastore2] a2e014a2-9e0f-4250-aa38-3a7154149903/a2e014a2-9e0f-4250-aa38-3a7154149903.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 883.807136] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-593b35b4-5d74-4c58-a93e-9cad00fc0b2d tempest-ServerGroupTestJSON-385159797 tempest-ServerGroupTestJSON-385159797-project-member] [instance: a2e014a2-9e0f-4250-aa38-3a7154149903] Extending root virtual disk to 1048576 {{(pid=63515) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 883.807136] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-573a18ea-99e7-4576-a700-d553623fa655 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.813080] env[63515]: DEBUG oslo_vmware.api [None req-593b35b4-5d74-4c58-a93e-9cad00fc0b2d tempest-ServerGroupTestJSON-385159797 tempest-ServerGroupTestJSON-385159797-project-member] Waiting for the task: (returnval){ [ 883.813080] env[63515]: value = "task-1111471" [ 883.813080] env[63515]: _type = "Task" [ 883.813080] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 883.822845] env[63515]: DEBUG oslo_vmware.api [None req-593b35b4-5d74-4c58-a93e-9cad00fc0b2d tempest-ServerGroupTestJSON-385159797 tempest-ServerGroupTestJSON-385159797-project-member] Task: {'id': task-1111471, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.841327] env[63515]: DEBUG oslo_vmware.api [None req-3fd65342-5e9e-46b6-830a-47df91928006 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Task: {'id': task-1111466, 'name': ReconfigVM_Task, 'duration_secs': 2.116909} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 883.842120] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-3fd65342-5e9e-46b6-830a-47df91928006 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: 0967283f-8dda-4692-b6b7-601545cabf2f] Reconfigured VM instance instance-0000004a to attach disk [datastore2] 0967283f-8dda-4692-b6b7-601545cabf2f/0967283f-8dda-4692-b6b7-601545cabf2f.vmdk or device None with type sparse {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 883.843549] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-40e0b50b-5ec5-4c14-930d-775ac928c451 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.850936] env[63515]: DEBUG oslo_vmware.api [None req-3fd65342-5e9e-46b6-830a-47df91928006 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Waiting for the task: (returnval){ [ 883.850936] env[63515]: value = "task-1111472" [ 883.850936] env[63515]: _type = "Task" [ 883.850936] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 883.860082] env[63515]: DEBUG oslo_vmware.api [None req-3fd65342-5e9e-46b6-830a-47df91928006 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Task: {'id': task-1111472, 'name': Rename_Task} progress is 5%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.934413] env[63515]: DEBUG nova.network.neutron [None req-d7e3c195-4890-454e-9f6b-af1a7cc4ee6a tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: ae5d6f34-bac6-4ba0-b0c3-8205982ddc88] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 884.049705] env[63515]: DEBUG nova.network.neutron [None req-9a80c171-6c9b-44ae-ab0a-573a8347fcf4 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: b1bf5a0a-42e7-41a6-90fb-a0c311623fdd] Successfully updated port: 8ab9e788-50a9-4669-8448-77cd0983386e {{(pid=63515) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 884.076804] env[63515]: INFO nova.scheduler.client.report [None req-bfac7ffd-4aa9-41c4-9bda-3f289bbda0a2 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Deleted allocation for migration ae31ad09-2d25-42de-82b4-75e2937a1149 [ 884.099730] env[63515]: DEBUG oslo_vmware.api [None req-1d02c1b8-b71d-4e69-b38f-4c4377a39898 tempest-InstanceActionsTestJSON-247065106 tempest-InstanceActionsTestJSON-247065106-project-member] Task: {'id': task-1111470, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.275928} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 884.100014] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-1d02c1b8-b71d-4e69-b38f-4c4377a39898 tempest-InstanceActionsTestJSON-247065106 tempest-InstanceActionsTestJSON-247065106-project-member] Deleted the datastore file {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 884.101809] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-1d02c1b8-b71d-4e69-b38f-4c4377a39898 tempest-InstanceActionsTestJSON-247065106 tempest-InstanceActionsTestJSON-247065106-project-member] [instance: 0ee9cce2-94b9-4038-ab38-1b3f96a864bc] Deleted contents of the VM from datastore datastore1 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 884.102030] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-1d02c1b8-b71d-4e69-b38f-4c4377a39898 tempest-InstanceActionsTestJSON-247065106 tempest-InstanceActionsTestJSON-247065106-project-member] [instance: 0ee9cce2-94b9-4038-ab38-1b3f96a864bc] Instance destroyed {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 884.102233] env[63515]: INFO nova.compute.manager [None req-1d02c1b8-b71d-4e69-b38f-4c4377a39898 tempest-InstanceActionsTestJSON-247065106 tempest-InstanceActionsTestJSON-247065106-project-member] [instance: 0ee9cce2-94b9-4038-ab38-1b3f96a864bc] Took 1.97 seconds to destroy the instance on the hypervisor. [ 884.102488] env[63515]: DEBUG oslo.service.loopingcall [None req-1d02c1b8-b71d-4e69-b38f-4c4377a39898 tempest-InstanceActionsTestJSON-247065106 tempest-InstanceActionsTestJSON-247065106-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 884.104639] env[63515]: DEBUG nova.compute.manager [-] [instance: 0ee9cce2-94b9-4038-ab38-1b3f96a864bc] Deallocating network for instance {{(pid=63515) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 884.104738] env[63515]: DEBUG nova.network.neutron [-] [instance: 0ee9cce2-94b9-4038-ab38-1b3f96a864bc] deallocate_for_instance() {{(pid=63515) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 884.258841] env[63515]: DEBUG nova.network.neutron [None req-d7e3c195-4890-454e-9f6b-af1a7cc4ee6a tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: ae5d6f34-bac6-4ba0-b0c3-8205982ddc88] Updating instance_info_cache with network_info: [{"id": "dbc4bf29-34a5-424b-bf71-8c80ba74be85", "address": "fa:16:3e:af:57:b6", "network": {"id": "3e5914c8-0cd4-49ee-81b4-1f3e280440bb", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-641632277-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ea099ce07bfb4a8da014d9303cf552eb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "225b6979-9329-403b-91fa-138bd41f6e83", "external-id": "nsx-vlan-transportzone-38", "segmentation_id": 38, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdbc4bf29-34", "ovs_interfaceid": "dbc4bf29-34a5-424b-bf71-8c80ba74be85", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 884.324073] env[63515]: DEBUG oslo_vmware.api [None req-593b35b4-5d74-4c58-a93e-9cad00fc0b2d tempest-ServerGroupTestJSON-385159797 tempest-ServerGroupTestJSON-385159797-project-member] Task: {'id': task-1111471, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.3103} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 884.324381] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-593b35b4-5d74-4c58-a93e-9cad00fc0b2d tempest-ServerGroupTestJSON-385159797 tempest-ServerGroupTestJSON-385159797-project-member] [instance: a2e014a2-9e0f-4250-aa38-3a7154149903] Extended root virtual disk {{(pid=63515) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 884.325243] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73dfc51c-ff13-47ee-a36a-2e3542d2c409 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.351233] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-593b35b4-5d74-4c58-a93e-9cad00fc0b2d tempest-ServerGroupTestJSON-385159797 tempest-ServerGroupTestJSON-385159797-project-member] [instance: a2e014a2-9e0f-4250-aa38-3a7154149903] Reconfiguring VM instance instance-0000004b to attach disk [datastore2] a2e014a2-9e0f-4250-aa38-3a7154149903/a2e014a2-9e0f-4250-aa38-3a7154149903.vmdk or device None with type sparse {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 884.351233] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cf796df4-4f45-46ce-96e7-67df8c94cadb {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.381183] env[63515]: DEBUG oslo_vmware.api [None req-3fd65342-5e9e-46b6-830a-47df91928006 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Task: {'id': task-1111472, 'name': Rename_Task, 'duration_secs': 0.156985} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 884.382240] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-3fd65342-5e9e-46b6-830a-47df91928006 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: 0967283f-8dda-4692-b6b7-601545cabf2f] Powering on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 884.382571] env[63515]: DEBUG oslo_vmware.api [None req-593b35b4-5d74-4c58-a93e-9cad00fc0b2d tempest-ServerGroupTestJSON-385159797 tempest-ServerGroupTestJSON-385159797-project-member] Waiting for the task: (returnval){ [ 884.382571] env[63515]: value = "task-1111473" [ 884.382571] env[63515]: _type = "Task" [ 884.382571] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 884.383204] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a93162cf-1f0a-4425-93ff-b68093001d77 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.397293] env[63515]: DEBUG oslo_vmware.api [None req-593b35b4-5d74-4c58-a93e-9cad00fc0b2d tempest-ServerGroupTestJSON-385159797 tempest-ServerGroupTestJSON-385159797-project-member] Task: {'id': task-1111473, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.398775] env[63515]: DEBUG oslo_vmware.api [None req-3fd65342-5e9e-46b6-830a-47df91928006 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Waiting for the task: (returnval){ [ 884.398775] env[63515]: value = "task-1111474" [ 884.398775] env[63515]: _type = "Task" [ 884.398775] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 884.413160] env[63515]: DEBUG oslo_vmware.api [None req-3fd65342-5e9e-46b6-830a-47df91928006 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Task: {'id': task-1111474, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.552264] env[63515]: DEBUG oslo_concurrency.lockutils [None req-9a80c171-6c9b-44ae-ab0a-573a8347fcf4 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Acquiring lock "refresh_cache-b1bf5a0a-42e7-41a6-90fb-a0c311623fdd" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 884.552264] env[63515]: DEBUG oslo_concurrency.lockutils [None req-9a80c171-6c9b-44ae-ab0a-573a8347fcf4 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Acquired lock "refresh_cache-b1bf5a0a-42e7-41a6-90fb-a0c311623fdd" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 884.552390] env[63515]: DEBUG nova.network.neutron [None req-9a80c171-6c9b-44ae-ab0a-573a8347fcf4 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: b1bf5a0a-42e7-41a6-90fb-a0c311623fdd] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 884.587348] env[63515]: DEBUG oslo_concurrency.lockutils [None req-bfac7ffd-4aa9-41c4-9bda-3f289bbda0a2 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Lock "d926c699-a64a-4942-9ef4-f0166414661d" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 23.040s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 884.764041] env[63515]: DEBUG oslo_concurrency.lockutils [None req-d7e3c195-4890-454e-9f6b-af1a7cc4ee6a tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Releasing lock "refresh_cache-ae5d6f34-bac6-4ba0-b0c3-8205982ddc88" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 884.764041] env[63515]: DEBUG nova.compute.manager [None req-d7e3c195-4890-454e-9f6b-af1a7cc4ee6a tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: ae5d6f34-bac6-4ba0-b0c3-8205982ddc88] Instance network_info: |[{"id": "dbc4bf29-34a5-424b-bf71-8c80ba74be85", "address": "fa:16:3e:af:57:b6", "network": {"id": "3e5914c8-0cd4-49ee-81b4-1f3e280440bb", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-641632277-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ea099ce07bfb4a8da014d9303cf552eb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "225b6979-9329-403b-91fa-138bd41f6e83", "external-id": "nsx-vlan-transportzone-38", "segmentation_id": 38, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdbc4bf29-34", "ovs_interfaceid": "dbc4bf29-34a5-424b-bf71-8c80ba74be85", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 884.764041] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-d7e3c195-4890-454e-9f6b-af1a7cc4ee6a tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: ae5d6f34-bac6-4ba0-b0c3-8205982ddc88] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:af:57:b6', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '225b6979-9329-403b-91fa-138bd41f6e83', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'dbc4bf29-34a5-424b-bf71-8c80ba74be85', 'vif_model': 'vmxnet3'}] {{(pid=63515) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 884.772815] env[63515]: DEBUG oslo.service.loopingcall [None req-d7e3c195-4890-454e-9f6b-af1a7cc4ee6a tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 884.778161] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ae5d6f34-bac6-4ba0-b0c3-8205982ddc88] Creating VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 884.780023] env[63515]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c049e3f9-0f43-4eda-a7d2-544c7db97922 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.807391] env[63515]: DEBUG nova.compute.manager [req-8a28f08d-d58a-43f1-900f-6c95ee3315f6 req-1d05abb6-cda8-4201-82c5-b077f346f67d service nova] [instance: ae5d6f34-bac6-4ba0-b0c3-8205982ddc88] Received event network-changed-dbc4bf29-34a5-424b-bf71-8c80ba74be85 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 884.807727] env[63515]: DEBUG nova.compute.manager [req-8a28f08d-d58a-43f1-900f-6c95ee3315f6 req-1d05abb6-cda8-4201-82c5-b077f346f67d service nova] [instance: ae5d6f34-bac6-4ba0-b0c3-8205982ddc88] Refreshing instance network info cache due to event network-changed-dbc4bf29-34a5-424b-bf71-8c80ba74be85. {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 884.808063] env[63515]: DEBUG oslo_concurrency.lockutils [req-8a28f08d-d58a-43f1-900f-6c95ee3315f6 req-1d05abb6-cda8-4201-82c5-b077f346f67d service nova] Acquiring lock "refresh_cache-ae5d6f34-bac6-4ba0-b0c3-8205982ddc88" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 884.808504] env[63515]: DEBUG oslo_concurrency.lockutils [req-8a28f08d-d58a-43f1-900f-6c95ee3315f6 req-1d05abb6-cda8-4201-82c5-b077f346f67d service nova] Acquired lock "refresh_cache-ae5d6f34-bac6-4ba0-b0c3-8205982ddc88" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 884.808504] env[63515]: DEBUG nova.network.neutron [req-8a28f08d-d58a-43f1-900f-6c95ee3315f6 req-1d05abb6-cda8-4201-82c5-b077f346f67d service nova] [instance: ae5d6f34-bac6-4ba0-b0c3-8205982ddc88] Refreshing network info cache for port dbc4bf29-34a5-424b-bf71-8c80ba74be85 {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 884.810706] env[63515]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 884.810706] env[63515]: value = "task-1111475" [ 884.810706] env[63515]: _type = "Task" [ 884.810706] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 884.836442] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1111475, 'name': CreateVM_Task} progress is 10%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.866878] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2365209-0ea6-4c69-84ce-9e2330bafdf6 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.878827] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8c773e7-10d7-4419-a5e5-a984d352f9cc {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.926405] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e13098f5-f842-41cf-865c-2dd49606c5fc {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.929422] env[63515]: DEBUG oslo_vmware.api [None req-593b35b4-5d74-4c58-a93e-9cad00fc0b2d tempest-ServerGroupTestJSON-385159797 tempest-ServerGroupTestJSON-385159797-project-member] Task: {'id': task-1111473, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.934928] env[63515]: DEBUG oslo_vmware.api [None req-3fd65342-5e9e-46b6-830a-47df91928006 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Task: {'id': task-1111474, 'name': PowerOnVM_Task} progress is 87%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.938136] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5eb5edda-f1e3-4566-a8c8-4377ddea6b0c {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.951771] env[63515]: DEBUG nova.compute.provider_tree [None req-dbe24c17-88c2-40e6-987c-d70326fd4bd3 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Updating inventory in ProviderTree for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 884.962020] env[63515]: DEBUG nova.network.neutron [req-e58309d0-882c-43fb-8a76-8beec3010ae8 req-338397ef-4cf0-4ca7-b469-0e2050a99320 service nova] [instance: 68766cd4-84be-475b-8494-d7ab43a9e969] Updated VIF entry in instance network info cache for port 28b59076-140f-46b5-b4a9-0bf530a1d43d. {{(pid=63515) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 884.962020] env[63515]: DEBUG nova.network.neutron [req-e58309d0-882c-43fb-8a76-8beec3010ae8 req-338397ef-4cf0-4ca7-b469-0e2050a99320 service nova] [instance: 68766cd4-84be-475b-8494-d7ab43a9e969] Updating instance_info_cache with network_info: [{"id": "28b59076-140f-46b5-b4a9-0bf530a1d43d", "address": "fa:16:3e:a3:55:0c", "network": {"id": "5758c885-e1c9-45c8-86aa-27a802101f52", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-734458138-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.204", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b18d4daacbc84e758f9f9ae4f3f09d28", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69351262-8d39-441a-85ba-3a78df436d17", "external-id": "nsx-vlan-transportzone-205", "segmentation_id": 205, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap28b59076-14", "ovs_interfaceid": "28b59076-140f-46b5-b4a9-0bf530a1d43d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 885.084653] env[63515]: DEBUG nova.network.neutron [None req-9a80c171-6c9b-44ae-ab0a-573a8347fcf4 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: b1bf5a0a-42e7-41a6-90fb-a0c311623fdd] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 885.266675] env[63515]: DEBUG nova.network.neutron [None req-9a80c171-6c9b-44ae-ab0a-573a8347fcf4 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: b1bf5a0a-42e7-41a6-90fb-a0c311623fdd] Updating instance_info_cache with network_info: [{"id": "8ab9e788-50a9-4669-8448-77cd0983386e", "address": "fa:16:3e:0f:e7:cc", "network": {"id": "320921bd-a6c9-4726-a801-19009d70d16e", "bridge": "br-int", "label": "tempest-ImagesTestJSON-304158429-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2259667ddc0c42639ea1af3b6e6fa0e8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2d94740a-bce8-4103-8ecf-230d02ec0a44", "external-id": "nsx-vlan-transportzone-149", "segmentation_id": 149, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8ab9e788-50", "ovs_interfaceid": "8ab9e788-50a9-4669-8448-77cd0983386e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 885.278772] env[63515]: DEBUG nova.network.neutron [-] [instance: 0ee9cce2-94b9-4038-ab38-1b3f96a864bc] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 885.320590] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1111475, 'name': CreateVM_Task, 'duration_secs': 0.386995} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 885.320751] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ae5d6f34-bac6-4ba0-b0c3-8205982ddc88] Created VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 885.321443] env[63515]: DEBUG oslo_concurrency.lockutils [None req-d7e3c195-4890-454e-9f6b-af1a7cc4ee6a tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 885.321598] env[63515]: DEBUG oslo_concurrency.lockutils [None req-d7e3c195-4890-454e-9f6b-af1a7cc4ee6a tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 885.321916] env[63515]: DEBUG oslo_concurrency.lockutils [None req-d7e3c195-4890-454e-9f6b-af1a7cc4ee6a tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 885.328725] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8e161ce7-5271-49f9-9762-1b64f3926175 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.334998] env[63515]: DEBUG oslo_vmware.api [None req-d7e3c195-4890-454e-9f6b-af1a7cc4ee6a tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Waiting for the task: (returnval){ [ 885.334998] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]52c80fb5-c9df-0cc8-5325-1b9b5e843f90" [ 885.334998] env[63515]: _type = "Task" [ 885.334998] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 885.345033] env[63515]: DEBUG oslo_vmware.api [None req-d7e3c195-4890-454e-9f6b-af1a7cc4ee6a tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52c80fb5-c9df-0cc8-5325-1b9b5e843f90, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.395911] env[63515]: DEBUG oslo_vmware.api [None req-593b35b4-5d74-4c58-a93e-9cad00fc0b2d tempest-ServerGroupTestJSON-385159797 tempest-ServerGroupTestJSON-385159797-project-member] Task: {'id': task-1111473, 'name': ReconfigVM_Task} progress is 99%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.429415] env[63515]: DEBUG oslo_vmware.api [None req-3fd65342-5e9e-46b6-830a-47df91928006 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Task: {'id': task-1111474, 'name': PowerOnVM_Task, 'duration_secs': 0.879474} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 885.430098] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-3fd65342-5e9e-46b6-830a-47df91928006 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: 0967283f-8dda-4692-b6b7-601545cabf2f] Powered on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 885.430366] env[63515]: INFO nova.compute.manager [None req-3fd65342-5e9e-46b6-830a-47df91928006 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: 0967283f-8dda-4692-b6b7-601545cabf2f] Took 11.19 seconds to spawn the instance on the hypervisor. [ 885.430560] env[63515]: DEBUG nova.compute.manager [None req-3fd65342-5e9e-46b6-830a-47df91928006 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: 0967283f-8dda-4692-b6b7-601545cabf2f] Checking state {{(pid=63515) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 885.431535] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7b318cc-1712-4f87-ba3e-6f4636a64a06 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.465349] env[63515]: DEBUG oslo_concurrency.lockutils [req-e58309d0-882c-43fb-8a76-8beec3010ae8 req-338397ef-4cf0-4ca7-b469-0e2050a99320 service nova] Releasing lock "refresh_cache-68766cd4-84be-475b-8494-d7ab43a9e969" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 885.486370] env[63515]: ERROR nova.scheduler.client.report [None req-dbe24c17-88c2-40e6-987c-d70326fd4bd3 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [req-338cc99a-cf91-4917-82a1-7f7e2432644b] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 2bd86232-2b6f-44d5-9057-1a3a6b27185a. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-338cc99a-cf91-4917-82a1-7f7e2432644b"}]} [ 885.512777] env[63515]: DEBUG nova.scheduler.client.report [None req-dbe24c17-88c2-40e6-987c-d70326fd4bd3 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Refreshing inventories for resource provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 885.529810] env[63515]: DEBUG nova.scheduler.client.report [None req-dbe24c17-88c2-40e6-987c-d70326fd4bd3 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Updating ProviderTree inventory for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 885.530019] env[63515]: DEBUG nova.compute.provider_tree [None req-dbe24c17-88c2-40e6-987c-d70326fd4bd3 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Updating inventory in ProviderTree for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 885.545914] env[63515]: DEBUG nova.scheduler.client.report [None req-dbe24c17-88c2-40e6-987c-d70326fd4bd3 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Refreshing aggregate associations for resource provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a, aggregates: None {{(pid=63515) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 885.567014] env[63515]: DEBUG nova.scheduler.client.report [None req-dbe24c17-88c2-40e6-987c-d70326fd4bd3 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Refreshing trait associations for resource provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a, traits: COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=63515) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 885.595702] env[63515]: DEBUG nova.network.neutron [req-8a28f08d-d58a-43f1-900f-6c95ee3315f6 req-1d05abb6-cda8-4201-82c5-b077f346f67d service nova] [instance: ae5d6f34-bac6-4ba0-b0c3-8205982ddc88] Updated VIF entry in instance network info cache for port dbc4bf29-34a5-424b-bf71-8c80ba74be85. {{(pid=63515) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 885.596213] env[63515]: DEBUG nova.network.neutron [req-8a28f08d-d58a-43f1-900f-6c95ee3315f6 req-1d05abb6-cda8-4201-82c5-b077f346f67d service nova] [instance: ae5d6f34-bac6-4ba0-b0c3-8205982ddc88] Updating instance_info_cache with network_info: [{"id": "dbc4bf29-34a5-424b-bf71-8c80ba74be85", "address": "fa:16:3e:af:57:b6", "network": {"id": "3e5914c8-0cd4-49ee-81b4-1f3e280440bb", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-641632277-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ea099ce07bfb4a8da014d9303cf552eb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "225b6979-9329-403b-91fa-138bd41f6e83", "external-id": "nsx-vlan-transportzone-38", "segmentation_id": 38, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdbc4bf29-34", "ovs_interfaceid": "dbc4bf29-34a5-424b-bf71-8c80ba74be85", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 885.770925] env[63515]: DEBUG oslo_concurrency.lockutils [None req-9a80c171-6c9b-44ae-ab0a-573a8347fcf4 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Releasing lock "refresh_cache-b1bf5a0a-42e7-41a6-90fb-a0c311623fdd" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 885.770925] env[63515]: DEBUG nova.compute.manager [None req-9a80c171-6c9b-44ae-ab0a-573a8347fcf4 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: b1bf5a0a-42e7-41a6-90fb-a0c311623fdd] Instance network_info: |[{"id": "8ab9e788-50a9-4669-8448-77cd0983386e", "address": "fa:16:3e:0f:e7:cc", "network": {"id": "320921bd-a6c9-4726-a801-19009d70d16e", "bridge": "br-int", "label": "tempest-ImagesTestJSON-304158429-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2259667ddc0c42639ea1af3b6e6fa0e8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2d94740a-bce8-4103-8ecf-230d02ec0a44", "external-id": "nsx-vlan-transportzone-149", "segmentation_id": 149, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8ab9e788-50", "ovs_interfaceid": "8ab9e788-50a9-4669-8448-77cd0983386e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 885.772071] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-9a80c171-6c9b-44ae-ab0a-573a8347fcf4 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: b1bf5a0a-42e7-41a6-90fb-a0c311623fdd] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:0f:e7:cc', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '2d94740a-bce8-4103-8ecf-230d02ec0a44', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '8ab9e788-50a9-4669-8448-77cd0983386e', 'vif_model': 'vmxnet3'}] {{(pid=63515) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 885.779172] env[63515]: DEBUG oslo.service.loopingcall [None req-9a80c171-6c9b-44ae-ab0a-573a8347fcf4 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 885.783314] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b1bf5a0a-42e7-41a6-90fb-a0c311623fdd] Creating VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 885.783466] env[63515]: INFO nova.compute.manager [-] [instance: 0ee9cce2-94b9-4038-ab38-1b3f96a864bc] Took 1.68 seconds to deallocate network for instance. [ 885.784163] env[63515]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-770bdf98-d70b-4cca-b389-c0041f9af076 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.809627] env[63515]: DEBUG nova.compute.manager [req-4298f941-73bb-4a1d-b8b6-9633bdfece12 req-40bc454f-bf4a-408e-9b42-32152322fc80 service nova] [instance: b1bf5a0a-42e7-41a6-90fb-a0c311623fdd] Received event network-vif-plugged-8ab9e788-50a9-4669-8448-77cd0983386e {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 885.809627] env[63515]: DEBUG oslo_concurrency.lockutils [req-4298f941-73bb-4a1d-b8b6-9633bdfece12 req-40bc454f-bf4a-408e-9b42-32152322fc80 service nova] Acquiring lock "b1bf5a0a-42e7-41a6-90fb-a0c311623fdd-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 885.809627] env[63515]: DEBUG oslo_concurrency.lockutils [req-4298f941-73bb-4a1d-b8b6-9633bdfece12 req-40bc454f-bf4a-408e-9b42-32152322fc80 service nova] Lock "b1bf5a0a-42e7-41a6-90fb-a0c311623fdd-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 885.809820] env[63515]: DEBUG oslo_concurrency.lockutils [req-4298f941-73bb-4a1d-b8b6-9633bdfece12 req-40bc454f-bf4a-408e-9b42-32152322fc80 service nova] Lock "b1bf5a0a-42e7-41a6-90fb-a0c311623fdd-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 885.809893] env[63515]: DEBUG nova.compute.manager [req-4298f941-73bb-4a1d-b8b6-9633bdfece12 req-40bc454f-bf4a-408e-9b42-32152322fc80 service nova] [instance: b1bf5a0a-42e7-41a6-90fb-a0c311623fdd] No waiting events found dispatching network-vif-plugged-8ab9e788-50a9-4669-8448-77cd0983386e {{(pid=63515) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 885.810078] env[63515]: WARNING nova.compute.manager [req-4298f941-73bb-4a1d-b8b6-9633bdfece12 req-40bc454f-bf4a-408e-9b42-32152322fc80 service nova] [instance: b1bf5a0a-42e7-41a6-90fb-a0c311623fdd] Received unexpected event network-vif-plugged-8ab9e788-50a9-4669-8448-77cd0983386e for instance with vm_state building and task_state spawning. [ 885.810258] env[63515]: DEBUG nova.compute.manager [req-4298f941-73bb-4a1d-b8b6-9633bdfece12 req-40bc454f-bf4a-408e-9b42-32152322fc80 service nova] [instance: b1bf5a0a-42e7-41a6-90fb-a0c311623fdd] Received event network-changed-8ab9e788-50a9-4669-8448-77cd0983386e {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 885.810418] env[63515]: DEBUG nova.compute.manager [req-4298f941-73bb-4a1d-b8b6-9633bdfece12 req-40bc454f-bf4a-408e-9b42-32152322fc80 service nova] [instance: b1bf5a0a-42e7-41a6-90fb-a0c311623fdd] Refreshing instance network info cache due to event network-changed-8ab9e788-50a9-4669-8448-77cd0983386e. {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 885.810744] env[63515]: DEBUG oslo_concurrency.lockutils [req-4298f941-73bb-4a1d-b8b6-9633bdfece12 req-40bc454f-bf4a-408e-9b42-32152322fc80 service nova] Acquiring lock "refresh_cache-b1bf5a0a-42e7-41a6-90fb-a0c311623fdd" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 885.810744] env[63515]: DEBUG oslo_concurrency.lockutils [req-4298f941-73bb-4a1d-b8b6-9633bdfece12 req-40bc454f-bf4a-408e-9b42-32152322fc80 service nova] Acquired lock "refresh_cache-b1bf5a0a-42e7-41a6-90fb-a0c311623fdd" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 885.810898] env[63515]: DEBUG nova.network.neutron [req-4298f941-73bb-4a1d-b8b6-9633bdfece12 req-40bc454f-bf4a-408e-9b42-32152322fc80 service nova] [instance: b1bf5a0a-42e7-41a6-90fb-a0c311623fdd] Refreshing network info cache for port 8ab9e788-50a9-4669-8448-77cd0983386e {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 885.819706] env[63515]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 885.819706] env[63515]: value = "task-1111476" [ 885.819706] env[63515]: _type = "Task" [ 885.819706] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 885.833754] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1111476, 'name': CreateVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.847055] env[63515]: DEBUG oslo_vmware.api [None req-d7e3c195-4890-454e-9f6b-af1a7cc4ee6a tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52c80fb5-c9df-0cc8-5325-1b9b5e843f90, 'name': SearchDatastore_Task, 'duration_secs': 0.00964} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 885.847629] env[63515]: DEBUG oslo_concurrency.lockutils [None req-d7e3c195-4890-454e-9f6b-af1a7cc4ee6a tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 885.847879] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-d7e3c195-4890-454e-9f6b-af1a7cc4ee6a tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: ae5d6f34-bac6-4ba0-b0c3-8205982ddc88] Processing image 8a120570-cb06-4099-b262-554ca0ad15c5 {{(pid=63515) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 885.848134] env[63515]: DEBUG oslo_concurrency.lockutils [None req-d7e3c195-4890-454e-9f6b-af1a7cc4ee6a tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 885.848290] env[63515]: DEBUG oslo_concurrency.lockutils [None req-d7e3c195-4890-454e-9f6b-af1a7cc4ee6a tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 885.848600] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-d7e3c195-4890-454e-9f6b-af1a7cc4ee6a tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 885.848950] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-18c76e05-848d-40c9-992c-a455454986fd {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.856977] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-d7e3c195-4890-454e-9f6b-af1a7cc4ee6a tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 885.857410] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-d7e3c195-4890-454e-9f6b-af1a7cc4ee6a tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63515) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 885.859043] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-07ea6bd9-1eb9-4b5b-8f30-bc7400f88cb8 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.861822] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df88465b-8dda-4e6f-bbb2-4f3ef6a1893f {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.869786] env[63515]: DEBUG oslo_vmware.api [None req-d7e3c195-4890-454e-9f6b-af1a7cc4ee6a tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Waiting for the task: (returnval){ [ 885.869786] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]52a90fda-1dcc-9fd3-547a-1e63d3da8003" [ 885.869786] env[63515]: _type = "Task" [ 885.869786] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 885.871103] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2828671-f4b3-415d-a9b4-4a0442f12899 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.909616] env[63515]: DEBUG oslo_vmware.api [None req-d7e3c195-4890-454e-9f6b-af1a7cc4ee6a tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52a90fda-1dcc-9fd3-547a-1e63d3da8003, 'name': SearchDatastore_Task, 'duration_secs': 0.0132} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 885.914474] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc0038ac-c09c-4d48-bab4-5fcd46ee8a1f {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.917275] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4b04e55b-d865-43c0-9eaf-7e8857b6dc54 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.928060] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b1ef5b9-48ef-43ca-8e6b-751c968e9448 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.931954] env[63515]: DEBUG oslo_vmware.api [None req-d7e3c195-4890-454e-9f6b-af1a7cc4ee6a tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Waiting for the task: (returnval){ [ 885.931954] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]52d3e2cc-1639-3d16-b07f-c3a42b27c358" [ 885.931954] env[63515]: _type = "Task" [ 885.931954] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 885.932274] env[63515]: DEBUG oslo_vmware.api [None req-593b35b4-5d74-4c58-a93e-9cad00fc0b2d tempest-ServerGroupTestJSON-385159797 tempest-ServerGroupTestJSON-385159797-project-member] Task: {'id': task-1111473, 'name': ReconfigVM_Task} progress is 99%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.942755] env[63515]: DEBUG nova.compute.provider_tree [None req-dbe24c17-88c2-40e6-987c-d70326fd4bd3 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Updating inventory in ProviderTree for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 885.955197] env[63515]: DEBUG oslo_vmware.api [None req-d7e3c195-4890-454e-9f6b-af1a7cc4ee6a tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52d3e2cc-1639-3d16-b07f-c3a42b27c358, 'name': SearchDatastore_Task, 'duration_secs': 0.010156} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 885.955622] env[63515]: INFO nova.compute.manager [None req-3fd65342-5e9e-46b6-830a-47df91928006 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: 0967283f-8dda-4692-b6b7-601545cabf2f] Took 35.69 seconds to build instance. [ 885.957453] env[63515]: DEBUG oslo_concurrency.lockutils [None req-d7e3c195-4890-454e-9f6b-af1a7cc4ee6a tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 885.957716] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-d7e3c195-4890-454e-9f6b-af1a7cc4ee6a tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk to [datastore2] ae5d6f34-bac6-4ba0-b0c3-8205982ddc88/ae5d6f34-bac6-4ba0-b0c3-8205982ddc88.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 885.957985] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3a41511e-793c-426f-a05d-112a6e748cd7 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.967203] env[63515]: DEBUG oslo_vmware.api [None req-d7e3c195-4890-454e-9f6b-af1a7cc4ee6a tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Waiting for the task: (returnval){ [ 885.967203] env[63515]: value = "task-1111477" [ 885.967203] env[63515]: _type = "Task" [ 885.967203] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 885.976457] env[63515]: DEBUG oslo_vmware.api [None req-d7e3c195-4890-454e-9f6b-af1a7cc4ee6a tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Task: {'id': task-1111477, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.098592] env[63515]: DEBUG oslo_concurrency.lockutils [req-8a28f08d-d58a-43f1-900f-6c95ee3315f6 req-1d05abb6-cda8-4201-82c5-b077f346f67d service nova] Releasing lock "refresh_cache-ae5d6f34-bac6-4ba0-b0c3-8205982ddc88" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 886.253732] env[63515]: DEBUG oslo_concurrency.lockutils [None req-b4e6fa31-2603-4190-94fe-f49ca8c1e386 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Acquiring lock "d926c699-a64a-4942-9ef4-f0166414661d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 886.254197] env[63515]: DEBUG oslo_concurrency.lockutils [None req-b4e6fa31-2603-4190-94fe-f49ca8c1e386 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Lock "d926c699-a64a-4942-9ef4-f0166414661d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 886.254446] env[63515]: DEBUG oslo_concurrency.lockutils [None req-b4e6fa31-2603-4190-94fe-f49ca8c1e386 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Acquiring lock "d926c699-a64a-4942-9ef4-f0166414661d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 886.254687] env[63515]: DEBUG oslo_concurrency.lockutils [None req-b4e6fa31-2603-4190-94fe-f49ca8c1e386 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Lock "d926c699-a64a-4942-9ef4-f0166414661d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 886.255193] env[63515]: DEBUG oslo_concurrency.lockutils [None req-b4e6fa31-2603-4190-94fe-f49ca8c1e386 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Lock "d926c699-a64a-4942-9ef4-f0166414661d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 886.260607] env[63515]: INFO nova.compute.manager [None req-b4e6fa31-2603-4190-94fe-f49ca8c1e386 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: d926c699-a64a-4942-9ef4-f0166414661d] Terminating instance [ 886.262997] env[63515]: DEBUG nova.compute.manager [None req-b4e6fa31-2603-4190-94fe-f49ca8c1e386 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: d926c699-a64a-4942-9ef4-f0166414661d] Start destroying the instance on the hypervisor. {{(pid=63515) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 886.263337] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-b4e6fa31-2603-4190-94fe-f49ca8c1e386 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: d926c699-a64a-4942-9ef4-f0166414661d] Destroying instance {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 886.264136] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3a2172f-494f-4fea-a47d-3671f9599455 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.275511] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-b4e6fa31-2603-4190-94fe-f49ca8c1e386 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: d926c699-a64a-4942-9ef4-f0166414661d] Powering off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 886.275752] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c47d330c-7403-4826-8895-146d8b64adb7 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.283762] env[63515]: DEBUG oslo_vmware.api [None req-b4e6fa31-2603-4190-94fe-f49ca8c1e386 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Waiting for the task: (returnval){ [ 886.283762] env[63515]: value = "task-1111478" [ 886.283762] env[63515]: _type = "Task" [ 886.283762] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 886.292691] env[63515]: DEBUG oslo_vmware.api [None req-b4e6fa31-2603-4190-94fe-f49ca8c1e386 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Task: {'id': task-1111478, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.305170] env[63515]: DEBUG oslo_concurrency.lockutils [None req-1d02c1b8-b71d-4e69-b38f-4c4377a39898 tempest-InstanceActionsTestJSON-247065106 tempest-InstanceActionsTestJSON-247065106-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 886.337737] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1111476, 'name': CreateVM_Task, 'duration_secs': 0.375079} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 886.337916] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b1bf5a0a-42e7-41a6-90fb-a0c311623fdd] Created VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 886.339958] env[63515]: DEBUG oslo_concurrency.lockutils [None req-9a80c171-6c9b-44ae-ab0a-573a8347fcf4 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d1918637-9f8a-4d89-9bc2-49e7d0fdaeee" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 886.339958] env[63515]: DEBUG oslo_concurrency.lockutils [None req-9a80c171-6c9b-44ae-ab0a-573a8347fcf4 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d1918637-9f8a-4d89-9bc2-49e7d0fdaeee" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 886.339958] env[63515]: DEBUG oslo_concurrency.lockutils [None req-9a80c171-6c9b-44ae-ab0a-573a8347fcf4 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d1918637-9f8a-4d89-9bc2-49e7d0fdaeee" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 886.339958] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6f32b415-1b09-4d52-a0eb-32da1c25bbc8 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.349029] env[63515]: DEBUG oslo_vmware.api [None req-9a80c171-6c9b-44ae-ab0a-573a8347fcf4 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Waiting for the task: (returnval){ [ 886.349029] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]52dd24d3-3221-cf99-f637-84d67597df9d" [ 886.349029] env[63515]: _type = "Task" [ 886.349029] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 886.360150] env[63515]: DEBUG oslo_vmware.api [None req-9a80c171-6c9b-44ae-ab0a-573a8347fcf4 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52dd24d3-3221-cf99-f637-84d67597df9d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.426652] env[63515]: DEBUG oslo_vmware.api [None req-593b35b4-5d74-4c58-a93e-9cad00fc0b2d tempest-ServerGroupTestJSON-385159797 tempest-ServerGroupTestJSON-385159797-project-member] Task: {'id': task-1111473, 'name': ReconfigVM_Task, 'duration_secs': 1.697267} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 886.426997] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-593b35b4-5d74-4c58-a93e-9cad00fc0b2d tempest-ServerGroupTestJSON-385159797 tempest-ServerGroupTestJSON-385159797-project-member] [instance: a2e014a2-9e0f-4250-aa38-3a7154149903] Reconfigured VM instance instance-0000004b to attach disk [datastore2] a2e014a2-9e0f-4250-aa38-3a7154149903/a2e014a2-9e0f-4250-aa38-3a7154149903.vmdk or device None with type sparse {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 886.427696] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-2c483aec-4c5c-4d39-a9f6-898949c2e7d8 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.435454] env[63515]: DEBUG oslo_vmware.api [None req-593b35b4-5d74-4c58-a93e-9cad00fc0b2d tempest-ServerGroupTestJSON-385159797 tempest-ServerGroupTestJSON-385159797-project-member] Waiting for the task: (returnval){ [ 886.435454] env[63515]: value = "task-1111479" [ 886.435454] env[63515]: _type = "Task" [ 886.435454] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 886.449574] env[63515]: DEBUG oslo_vmware.api [None req-593b35b4-5d74-4c58-a93e-9cad00fc0b2d tempest-ServerGroupTestJSON-385159797 tempest-ServerGroupTestJSON-385159797-project-member] Task: {'id': task-1111479, 'name': Rename_Task} progress is 5%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.459697] env[63515]: DEBUG oslo_concurrency.lockutils [None req-3fd65342-5e9e-46b6-830a-47df91928006 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Lock "0967283f-8dda-4692-b6b7-601545cabf2f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 37.203s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 886.468478] env[63515]: ERROR nova.scheduler.client.report [None req-dbe24c17-88c2-40e6-987c-d70326fd4bd3 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [req-f05907fb-64ed-4eaa-8a2c-209af024abf8] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 2bd86232-2b6f-44d5-9057-1a3a6b27185a. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-f05907fb-64ed-4eaa-8a2c-209af024abf8"}]} [ 886.487680] env[63515]: DEBUG oslo_vmware.api [None req-d7e3c195-4890-454e-9f6b-af1a7cc4ee6a tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Task: {'id': task-1111477, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.489180] env[63515]: DEBUG nova.scheduler.client.report [None req-dbe24c17-88c2-40e6-987c-d70326fd4bd3 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Refreshing inventories for resource provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 886.507119] env[63515]: DEBUG nova.scheduler.client.report [None req-dbe24c17-88c2-40e6-987c-d70326fd4bd3 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Updating ProviderTree inventory for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 886.507568] env[63515]: DEBUG nova.compute.provider_tree [None req-dbe24c17-88c2-40e6-987c-d70326fd4bd3 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Updating inventory in ProviderTree for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 886.520657] env[63515]: DEBUG nova.scheduler.client.report [None req-dbe24c17-88c2-40e6-987c-d70326fd4bd3 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Refreshing aggregate associations for resource provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a, aggregates: None {{(pid=63515) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 886.545700] env[63515]: DEBUG nova.scheduler.client.report [None req-dbe24c17-88c2-40e6-987c-d70326fd4bd3 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Refreshing trait associations for resource provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a, traits: COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=63515) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 886.616875] env[63515]: DEBUG nova.network.neutron [req-4298f941-73bb-4a1d-b8b6-9633bdfece12 req-40bc454f-bf4a-408e-9b42-32152322fc80 service nova] [instance: b1bf5a0a-42e7-41a6-90fb-a0c311623fdd] Updated VIF entry in instance network info cache for port 8ab9e788-50a9-4669-8448-77cd0983386e. {{(pid=63515) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 886.617328] env[63515]: DEBUG nova.network.neutron [req-4298f941-73bb-4a1d-b8b6-9633bdfece12 req-40bc454f-bf4a-408e-9b42-32152322fc80 service nova] [instance: b1bf5a0a-42e7-41a6-90fb-a0c311623fdd] Updating instance_info_cache with network_info: [{"id": "8ab9e788-50a9-4669-8448-77cd0983386e", "address": "fa:16:3e:0f:e7:cc", "network": {"id": "320921bd-a6c9-4726-a801-19009d70d16e", "bridge": "br-int", "label": "tempest-ImagesTestJSON-304158429-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2259667ddc0c42639ea1af3b6e6fa0e8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2d94740a-bce8-4103-8ecf-230d02ec0a44", "external-id": "nsx-vlan-transportzone-149", "segmentation_id": 149, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8ab9e788-50", "ovs_interfaceid": "8ab9e788-50a9-4669-8448-77cd0983386e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 886.796770] env[63515]: DEBUG oslo_vmware.api [None req-b4e6fa31-2603-4190-94fe-f49ca8c1e386 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Task: {'id': task-1111478, 'name': PowerOffVM_Task, 'duration_secs': 0.319381} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 886.797080] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-b4e6fa31-2603-4190-94fe-f49ca8c1e386 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: d926c699-a64a-4942-9ef4-f0166414661d] Powered off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 886.797260] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-b4e6fa31-2603-4190-94fe-f49ca8c1e386 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: d926c699-a64a-4942-9ef4-f0166414661d] Unregistering the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 886.797512] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-800a7e1e-535d-4c1a-86b9-03652d192182 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.804711] env[63515]: DEBUG oslo_concurrency.lockutils [None req-3173d497-cfe4-4001-b365-7bedc6c45742 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Acquiring lock "829e72d7-a56a-451c-b98a-94253e5900d5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 886.804947] env[63515]: DEBUG oslo_concurrency.lockutils [None req-3173d497-cfe4-4001-b365-7bedc6c45742 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Lock "829e72d7-a56a-451c-b98a-94253e5900d5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 886.806556] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c8de0e5-ca2a-48b7-a79b-46de0d1563dc {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.816111] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d50c4c9-5149-47b9-99c3-c313a0e5c652 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.849661] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-384df374-b05b-4e50-bd10-94e28d3c6388 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.865429] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb8e7d8a-0a78-4d9e-9a65-15c108f661e9 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.869739] env[63515]: DEBUG oslo_concurrency.lockutils [None req-9a80c171-6c9b-44ae-ab0a-573a8347fcf4 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d1918637-9f8a-4d89-9bc2-49e7d0fdaeee" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 886.870028] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-9a80c171-6c9b-44ae-ab0a-573a8347fcf4 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: b1bf5a0a-42e7-41a6-90fb-a0c311623fdd] Processing image d1918637-9f8a-4d89-9bc2-49e7d0fdaeee {{(pid=63515) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 886.870292] env[63515]: DEBUG oslo_concurrency.lockutils [None req-9a80c171-6c9b-44ae-ab0a-573a8347fcf4 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d1918637-9f8a-4d89-9bc2-49e7d0fdaeee/d1918637-9f8a-4d89-9bc2-49e7d0fdaeee.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 886.870444] env[63515]: DEBUG oslo_concurrency.lockutils [None req-9a80c171-6c9b-44ae-ab0a-573a8347fcf4 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d1918637-9f8a-4d89-9bc2-49e7d0fdaeee/d1918637-9f8a-4d89-9bc2-49e7d0fdaeee.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 886.870622] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-9a80c171-6c9b-44ae-ab0a-573a8347fcf4 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 886.871252] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-319b112b-f1ed-4b06-89fb-fc10d37bfe1e {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.881632] env[63515]: DEBUG nova.compute.provider_tree [None req-dbe24c17-88c2-40e6-987c-d70326fd4bd3 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Updating inventory in ProviderTree for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 886.884411] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-b4e6fa31-2603-4190-94fe-f49ca8c1e386 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: d926c699-a64a-4942-9ef4-f0166414661d] Unregistered the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 886.884525] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-b4e6fa31-2603-4190-94fe-f49ca8c1e386 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: d926c699-a64a-4942-9ef4-f0166414661d] Deleting contents of the VM from datastore datastore2 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 886.885164] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-b4e6fa31-2603-4190-94fe-f49ca8c1e386 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Deleting the datastore file [datastore2] d926c699-a64a-4942-9ef4-f0166414661d {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 886.885164] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b2544af0-2687-4002-beff-f7df04fbecb8 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.892446] env[63515]: DEBUG oslo_vmware.api [None req-b4e6fa31-2603-4190-94fe-f49ca8c1e386 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Waiting for the task: (returnval){ [ 886.892446] env[63515]: value = "task-1111481" [ 886.892446] env[63515]: _type = "Task" [ 886.892446] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 886.893220] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-9a80c171-6c9b-44ae-ab0a-573a8347fcf4 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 886.893452] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-9a80c171-6c9b-44ae-ab0a-573a8347fcf4 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63515) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 886.894784] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-af72efa2-11b7-4653-abd2-d4efad138237 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.903336] env[63515]: DEBUG oslo_vmware.api [None req-9a80c171-6c9b-44ae-ab0a-573a8347fcf4 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Waiting for the task: (returnval){ [ 886.903336] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]52a84b8d-aff8-09b0-4deb-1453c6403571" [ 886.903336] env[63515]: _type = "Task" [ 886.903336] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 886.906631] env[63515]: DEBUG oslo_vmware.api [None req-b4e6fa31-2603-4190-94fe-f49ca8c1e386 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Task: {'id': task-1111481, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.918372] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-9a80c171-6c9b-44ae-ab0a-573a8347fcf4 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: b1bf5a0a-42e7-41a6-90fb-a0c311623fdd] Preparing fetch location {{(pid=63515) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 886.918634] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-9a80c171-6c9b-44ae-ab0a-573a8347fcf4 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: b1bf5a0a-42e7-41a6-90fb-a0c311623fdd] Fetch image to [datastore2] OSTACK_IMG_570d79c9-5dbb-48eb-b6b9-46e09f2d30ca/OSTACK_IMG_570d79c9-5dbb-48eb-b6b9-46e09f2d30ca.vmdk {{(pid=63515) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 886.918818] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-9a80c171-6c9b-44ae-ab0a-573a8347fcf4 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: b1bf5a0a-42e7-41a6-90fb-a0c311623fdd] Downloading stream optimized image d1918637-9f8a-4d89-9bc2-49e7d0fdaeee to [datastore2] OSTACK_IMG_570d79c9-5dbb-48eb-b6b9-46e09f2d30ca/OSTACK_IMG_570d79c9-5dbb-48eb-b6b9-46e09f2d30ca.vmdk on the data store datastore2 as vApp {{(pid=63515) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 886.920914] env[63515]: DEBUG nova.virt.vmwareapi.images [None req-9a80c171-6c9b-44ae-ab0a-573a8347fcf4 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: b1bf5a0a-42e7-41a6-90fb-a0c311623fdd] Downloading image file data d1918637-9f8a-4d89-9bc2-49e7d0fdaeee to the ESX as VM named 'OSTACK_IMG_570d79c9-5dbb-48eb-b6b9-46e09f2d30ca' {{(pid=63515) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 886.947450] env[63515]: DEBUG oslo_vmware.api [None req-593b35b4-5d74-4c58-a93e-9cad00fc0b2d tempest-ServerGroupTestJSON-385159797 tempest-ServerGroupTestJSON-385159797-project-member] Task: {'id': task-1111479, 'name': Rename_Task, 'duration_secs': 0.267206} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 886.963523] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-593b35b4-5d74-4c58-a93e-9cad00fc0b2d tempest-ServerGroupTestJSON-385159797 tempest-ServerGroupTestJSON-385159797-project-member] [instance: a2e014a2-9e0f-4250-aa38-3a7154149903] Powering on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 886.964019] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-78a3a170-f362-4efd-aa2d-d6210e231f8d {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.974659] env[63515]: DEBUG oslo_vmware.api [None req-593b35b4-5d74-4c58-a93e-9cad00fc0b2d tempest-ServerGroupTestJSON-385159797 tempest-ServerGroupTestJSON-385159797-project-member] Waiting for the task: (returnval){ [ 886.974659] env[63515]: value = "task-1111482" [ 886.974659] env[63515]: _type = "Task" [ 886.974659] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 886.983898] env[63515]: DEBUG oslo_vmware.api [None req-d7e3c195-4890-454e-9f6b-af1a7cc4ee6a tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Task: {'id': task-1111477, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.58089} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 886.986879] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-d7e3c195-4890-454e-9f6b-af1a7cc4ee6a tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk to [datastore2] ae5d6f34-bac6-4ba0-b0c3-8205982ddc88/ae5d6f34-bac6-4ba0-b0c3-8205982ddc88.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 886.987125] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-d7e3c195-4890-454e-9f6b-af1a7cc4ee6a tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: ae5d6f34-bac6-4ba0-b0c3-8205982ddc88] Extending root virtual disk to 1048576 {{(pid=63515) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 886.989772] env[63515]: DEBUG oslo_vmware.api [None req-593b35b4-5d74-4c58-a93e-9cad00fc0b2d tempest-ServerGroupTestJSON-385159797 tempest-ServerGroupTestJSON-385159797-project-member] Task: {'id': task-1111482, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.989772] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-45ad3f48-2a23-4ae0-bb89-668ed1d64d33 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.994433] env[63515]: DEBUG oslo_vmware.api [None req-d7e3c195-4890-454e-9f6b-af1a7cc4ee6a tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Waiting for the task: (returnval){ [ 886.994433] env[63515]: value = "task-1111483" [ 886.994433] env[63515]: _type = "Task" [ 886.994433] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 887.001415] env[63515]: DEBUG oslo_vmware.api [None req-d7e3c195-4890-454e-9f6b-af1a7cc4ee6a tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Task: {'id': task-1111483, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.014853] env[63515]: DEBUG oslo_vmware.rw_handles [None req-9a80c171-6c9b-44ae-ab0a-573a8347fcf4 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 887.014853] env[63515]: value = "resgroup-9" [ 887.014853] env[63515]: _type = "ResourcePool" [ 887.014853] env[63515]: }. {{(pid=63515) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 887.015192] env[63515]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-f063e35a-dfd7-4727-87f4-78d09c2274b3 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.035600] env[63515]: DEBUG oslo_vmware.rw_handles [None req-9a80c171-6c9b-44ae-ab0a-573a8347fcf4 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Lease: (returnval){ [ 887.035600] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]5255a7c2-6798-a69c-e24b-7072185c2281" [ 887.035600] env[63515]: _type = "HttpNfcLease" [ 887.035600] env[63515]: } obtained for vApp import into resource pool (val){ [ 887.035600] env[63515]: value = "resgroup-9" [ 887.035600] env[63515]: _type = "ResourcePool" [ 887.035600] env[63515]: }. {{(pid=63515) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 887.036231] env[63515]: DEBUG oslo_vmware.api [None req-9a80c171-6c9b-44ae-ab0a-573a8347fcf4 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Waiting for the lease: (returnval){ [ 887.036231] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]5255a7c2-6798-a69c-e24b-7072185c2281" [ 887.036231] env[63515]: _type = "HttpNfcLease" [ 887.036231] env[63515]: } to be ready. {{(pid=63515) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 887.042437] env[63515]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 887.042437] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]5255a7c2-6798-a69c-e24b-7072185c2281" [ 887.042437] env[63515]: _type = "HttpNfcLease" [ 887.042437] env[63515]: } is initializing. {{(pid=63515) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 887.122191] env[63515]: DEBUG oslo_concurrency.lockutils [req-4298f941-73bb-4a1d-b8b6-9633bdfece12 req-40bc454f-bf4a-408e-9b42-32152322fc80 service nova] Releasing lock "refresh_cache-b1bf5a0a-42e7-41a6-90fb-a0c311623fdd" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 887.122531] env[63515]: DEBUG nova.compute.manager [req-4298f941-73bb-4a1d-b8b6-9633bdfece12 req-40bc454f-bf4a-408e-9b42-32152322fc80 service nova] [instance: 0ee9cce2-94b9-4038-ab38-1b3f96a864bc] Received event network-vif-deleted-2fc6eecd-e0cd-41f2-8707-6eb57c2fef5a {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 887.310232] env[63515]: DEBUG nova.compute.manager [None req-3173d497-cfe4-4001-b365-7bedc6c45742 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: 829e72d7-a56a-451c-b98a-94253e5900d5] Starting instance... {{(pid=63515) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 887.348161] env[63515]: DEBUG oslo_concurrency.lockutils [None req-07515409-459b-4c3e-992f-c4e070f0ae85 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Acquiring lock "0967283f-8dda-4692-b6b7-601545cabf2f" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 887.348446] env[63515]: DEBUG oslo_concurrency.lockutils [None req-07515409-459b-4c3e-992f-c4e070f0ae85 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Lock "0967283f-8dda-4692-b6b7-601545cabf2f" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 887.348661] env[63515]: DEBUG oslo_concurrency.lockutils [None req-07515409-459b-4c3e-992f-c4e070f0ae85 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Acquiring lock "0967283f-8dda-4692-b6b7-601545cabf2f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 887.348853] env[63515]: DEBUG oslo_concurrency.lockutils [None req-07515409-459b-4c3e-992f-c4e070f0ae85 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Lock "0967283f-8dda-4692-b6b7-601545cabf2f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 887.349040] env[63515]: DEBUG oslo_concurrency.lockutils [None req-07515409-459b-4c3e-992f-c4e070f0ae85 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Lock "0967283f-8dda-4692-b6b7-601545cabf2f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 887.351177] env[63515]: INFO nova.compute.manager [None req-07515409-459b-4c3e-992f-c4e070f0ae85 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: 0967283f-8dda-4692-b6b7-601545cabf2f] Terminating instance [ 887.353120] env[63515]: DEBUG nova.compute.manager [None req-07515409-459b-4c3e-992f-c4e070f0ae85 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: 0967283f-8dda-4692-b6b7-601545cabf2f] Start destroying the instance on the hypervisor. {{(pid=63515) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 887.353334] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-07515409-459b-4c3e-992f-c4e070f0ae85 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: 0967283f-8dda-4692-b6b7-601545cabf2f] Destroying instance {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 887.354448] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b913c21-2a55-412e-aeb7-9ad6e135cf0a {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.362275] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-07515409-459b-4c3e-992f-c4e070f0ae85 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: 0967283f-8dda-4692-b6b7-601545cabf2f] Powering off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 887.362524] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-40838994-1cea-4c6b-a5c0-f763e390ed38 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.369465] env[63515]: DEBUG oslo_vmware.api [None req-07515409-459b-4c3e-992f-c4e070f0ae85 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Waiting for the task: (returnval){ [ 887.369465] env[63515]: value = "task-1111485" [ 887.369465] env[63515]: _type = "Task" [ 887.369465] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 887.379197] env[63515]: DEBUG oslo_vmware.api [None req-07515409-459b-4c3e-992f-c4e070f0ae85 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Task: {'id': task-1111485, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.405448] env[63515]: DEBUG oslo_vmware.api [None req-b4e6fa31-2603-4190-94fe-f49ca8c1e386 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Task: {'id': task-1111481, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.147183} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 887.405765] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-b4e6fa31-2603-4190-94fe-f49ca8c1e386 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Deleted the datastore file {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 887.406044] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-b4e6fa31-2603-4190-94fe-f49ca8c1e386 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: d926c699-a64a-4942-9ef4-f0166414661d] Deleted contents of the VM from datastore datastore2 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 887.406184] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-b4e6fa31-2603-4190-94fe-f49ca8c1e386 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: d926c699-a64a-4942-9ef4-f0166414661d] Instance destroyed {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 887.406436] env[63515]: INFO nova.compute.manager [None req-b4e6fa31-2603-4190-94fe-f49ca8c1e386 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: d926c699-a64a-4942-9ef4-f0166414661d] Took 1.14 seconds to destroy the instance on the hypervisor. [ 887.406643] env[63515]: DEBUG oslo.service.loopingcall [None req-b4e6fa31-2603-4190-94fe-f49ca8c1e386 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 887.406898] env[63515]: DEBUG nova.compute.manager [-] [instance: d926c699-a64a-4942-9ef4-f0166414661d] Deallocating network for instance {{(pid=63515) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 887.407010] env[63515]: DEBUG nova.network.neutron [-] [instance: d926c699-a64a-4942-9ef4-f0166414661d] deallocate_for_instance() {{(pid=63515) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 887.415333] env[63515]: DEBUG nova.scheduler.client.report [None req-dbe24c17-88c2-40e6-987c-d70326fd4bd3 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Updated inventory for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a with generation 104 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 887.415601] env[63515]: DEBUG nova.compute.provider_tree [None req-dbe24c17-88c2-40e6-987c-d70326fd4bd3 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Updating resource provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a generation from 104 to 105 during operation: update_inventory {{(pid=63515) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 887.415783] env[63515]: DEBUG nova.compute.provider_tree [None req-dbe24c17-88c2-40e6-987c-d70326fd4bd3 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Updating inventory in ProviderTree for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 887.486285] env[63515]: DEBUG oslo_vmware.api [None req-593b35b4-5d74-4c58-a93e-9cad00fc0b2d tempest-ServerGroupTestJSON-385159797 tempest-ServerGroupTestJSON-385159797-project-member] Task: {'id': task-1111482, 'name': PowerOnVM_Task} progress is 94%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.503673] env[63515]: DEBUG oslo_vmware.api [None req-d7e3c195-4890-454e-9f6b-af1a7cc4ee6a tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Task: {'id': task-1111483, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.073622} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 887.503977] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-d7e3c195-4890-454e-9f6b-af1a7cc4ee6a tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: ae5d6f34-bac6-4ba0-b0c3-8205982ddc88] Extended root virtual disk {{(pid=63515) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 887.504812] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28951d19-21ef-4c61-9e95-a08b1de8d824 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.528975] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-d7e3c195-4890-454e-9f6b-af1a7cc4ee6a tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: ae5d6f34-bac6-4ba0-b0c3-8205982ddc88] Reconfiguring VM instance instance-0000004c to attach disk [datastore2] ae5d6f34-bac6-4ba0-b0c3-8205982ddc88/ae5d6f34-bac6-4ba0-b0c3-8205982ddc88.vmdk or device None with type sparse {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 887.529317] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f78081e1-67fe-40aa-8d54-d19e579fd475 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.549900] env[63515]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 887.549900] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]5255a7c2-6798-a69c-e24b-7072185c2281" [ 887.549900] env[63515]: _type = "HttpNfcLease" [ 887.549900] env[63515]: } is initializing. {{(pid=63515) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 887.551233] env[63515]: DEBUG oslo_vmware.api [None req-d7e3c195-4890-454e-9f6b-af1a7cc4ee6a tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Waiting for the task: (returnval){ [ 887.551233] env[63515]: value = "task-1111486" [ 887.551233] env[63515]: _type = "Task" [ 887.551233] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 887.558448] env[63515]: DEBUG oslo_vmware.api [None req-d7e3c195-4890-454e-9f6b-af1a7cc4ee6a tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Task: {'id': task-1111486, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.707310] env[63515]: DEBUG nova.compute.manager [req-324b3486-67d3-4767-8d99-264311423bf7 req-4b0c7a05-2b1e-4257-898b-e0bd70da07a2 service nova] [instance: d926c699-a64a-4942-9ef4-f0166414661d] Received event network-vif-deleted-fdec91d2-43e8-4d40-9aa9-885a82fd9b42 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 887.707548] env[63515]: INFO nova.compute.manager [req-324b3486-67d3-4767-8d99-264311423bf7 req-4b0c7a05-2b1e-4257-898b-e0bd70da07a2 service nova] [instance: d926c699-a64a-4942-9ef4-f0166414661d] Neutron deleted interface fdec91d2-43e8-4d40-9aa9-885a82fd9b42; detaching it from the instance and deleting it from the info cache [ 887.707674] env[63515]: DEBUG nova.network.neutron [req-324b3486-67d3-4767-8d99-264311423bf7 req-4b0c7a05-2b1e-4257-898b-e0bd70da07a2 service nova] [instance: d926c699-a64a-4942-9ef4-f0166414661d] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 887.831895] env[63515]: DEBUG oslo_concurrency.lockutils [None req-3173d497-cfe4-4001-b365-7bedc6c45742 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 887.879035] env[63515]: DEBUG oslo_vmware.api [None req-07515409-459b-4c3e-992f-c4e070f0ae85 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Task: {'id': task-1111485, 'name': PowerOffVM_Task, 'duration_secs': 0.29298} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 887.879323] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-07515409-459b-4c3e-992f-c4e070f0ae85 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: 0967283f-8dda-4692-b6b7-601545cabf2f] Powered off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 887.879498] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-07515409-459b-4c3e-992f-c4e070f0ae85 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: 0967283f-8dda-4692-b6b7-601545cabf2f] Unregistering the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 887.879752] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-cc604377-aa78-43c9-aa44-295ae3fe45f4 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.920548] env[63515]: DEBUG oslo_concurrency.lockutils [None req-dbe24c17-88c2-40e6-987c-d70326fd4bd3 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 4.427s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 887.921069] env[63515]: DEBUG nova.compute.manager [None req-dbe24c17-88c2-40e6-987c-d70326fd4bd3 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: e13da90c-28e6-43d2-99b7-19c5095954ca] Start building networks asynchronously for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 887.923950] env[63515]: DEBUG oslo_concurrency.lockutils [None req-87ca4373-5e35-4af3-b945-64d29205298c tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 10.370s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 887.924208] env[63515]: DEBUG nova.objects.instance [None req-87ca4373-5e35-4af3-b945-64d29205298c tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Lazy-loading 'resources' on Instance uuid d7e8dcdb-41cb-46fb-8b61-d251e7c2d372 {{(pid=63515) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 887.959964] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-07515409-459b-4c3e-992f-c4e070f0ae85 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: 0967283f-8dda-4692-b6b7-601545cabf2f] Unregistered the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 887.959964] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-07515409-459b-4c3e-992f-c4e070f0ae85 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: 0967283f-8dda-4692-b6b7-601545cabf2f] Deleting contents of the VM from datastore datastore2 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 887.960162] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-07515409-459b-4c3e-992f-c4e070f0ae85 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Deleting the datastore file [datastore2] 0967283f-8dda-4692-b6b7-601545cabf2f {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 887.960453] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-fe6ceda6-30fd-4b4d-b5e5-2bbe09da7704 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.967392] env[63515]: DEBUG oslo_vmware.api [None req-07515409-459b-4c3e-992f-c4e070f0ae85 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Waiting for the task: (returnval){ [ 887.967392] env[63515]: value = "task-1111488" [ 887.967392] env[63515]: _type = "Task" [ 887.967392] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 887.976668] env[63515]: DEBUG oslo_vmware.api [None req-07515409-459b-4c3e-992f-c4e070f0ae85 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Task: {'id': task-1111488, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.984862] env[63515]: DEBUG oslo_vmware.api [None req-593b35b4-5d74-4c58-a93e-9cad00fc0b2d tempest-ServerGroupTestJSON-385159797 tempest-ServerGroupTestJSON-385159797-project-member] Task: {'id': task-1111482, 'name': PowerOnVM_Task, 'duration_secs': 0.583812} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 887.985195] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-593b35b4-5d74-4c58-a93e-9cad00fc0b2d tempest-ServerGroupTestJSON-385159797 tempest-ServerGroupTestJSON-385159797-project-member] [instance: a2e014a2-9e0f-4250-aa38-3a7154149903] Powered on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 887.985452] env[63515]: INFO nova.compute.manager [None req-593b35b4-5d74-4c58-a93e-9cad00fc0b2d tempest-ServerGroupTestJSON-385159797 tempest-ServerGroupTestJSON-385159797-project-member] [instance: a2e014a2-9e0f-4250-aa38-3a7154149903] Took 10.20 seconds to spawn the instance on the hypervisor. [ 887.985681] env[63515]: DEBUG nova.compute.manager [None req-593b35b4-5d74-4c58-a93e-9cad00fc0b2d tempest-ServerGroupTestJSON-385159797 tempest-ServerGroupTestJSON-385159797-project-member] [instance: a2e014a2-9e0f-4250-aa38-3a7154149903] Checking state {{(pid=63515) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 887.986780] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c29ce8b-130c-4643-9222-fc3babf117a0 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.052534] env[63515]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 888.052534] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]5255a7c2-6798-a69c-e24b-7072185c2281" [ 888.052534] env[63515]: _type = "HttpNfcLease" [ 888.052534] env[63515]: } is initializing. {{(pid=63515) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 888.060123] env[63515]: DEBUG oslo_vmware.api [None req-d7e3c195-4890-454e-9f6b-af1a7cc4ee6a tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Task: {'id': task-1111486, 'name': ReconfigVM_Task, 'duration_secs': 0.271756} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 888.060417] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-d7e3c195-4890-454e-9f6b-af1a7cc4ee6a tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: ae5d6f34-bac6-4ba0-b0c3-8205982ddc88] Reconfigured VM instance instance-0000004c to attach disk [datastore2] ae5d6f34-bac6-4ba0-b0c3-8205982ddc88/ae5d6f34-bac6-4ba0-b0c3-8205982ddc88.vmdk or device None with type sparse {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 888.061040] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-1385505c-bd52-408b-b147-0a9c6cbe4050 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.066626] env[63515]: DEBUG oslo_vmware.api [None req-d7e3c195-4890-454e-9f6b-af1a7cc4ee6a tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Waiting for the task: (returnval){ [ 888.066626] env[63515]: value = "task-1111489" [ 888.066626] env[63515]: _type = "Task" [ 888.066626] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.073975] env[63515]: DEBUG oslo_vmware.api [None req-d7e3c195-4890-454e-9f6b-af1a7cc4ee6a tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Task: {'id': task-1111489, 'name': Rename_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.181801] env[63515]: DEBUG nova.network.neutron [-] [instance: d926c699-a64a-4942-9ef4-f0166414661d] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 888.210706] env[63515]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9b7872fb-e0f3-447b-98b1-aa267b94a74a {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.220598] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93b2e9ee-3766-4559-a137-b321210c9172 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.250743] env[63515]: DEBUG nova.compute.manager [req-324b3486-67d3-4767-8d99-264311423bf7 req-4b0c7a05-2b1e-4257-898b-e0bd70da07a2 service nova] [instance: d926c699-a64a-4942-9ef4-f0166414661d] Detach interface failed, port_id=fdec91d2-43e8-4d40-9aa9-885a82fd9b42, reason: Instance d926c699-a64a-4942-9ef4-f0166414661d could not be found. {{(pid=63515) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 888.426858] env[63515]: DEBUG nova.compute.utils [None req-dbe24c17-88c2-40e6-987c-d70326fd4bd3 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Using /dev/sd instead of None {{(pid=63515) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 888.431273] env[63515]: DEBUG nova.compute.manager [None req-dbe24c17-88c2-40e6-987c-d70326fd4bd3 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: e13da90c-28e6-43d2-99b7-19c5095954ca] Allocating IP information in the background. {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 888.431578] env[63515]: DEBUG nova.network.neutron [None req-dbe24c17-88c2-40e6-987c-d70326fd4bd3 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: e13da90c-28e6-43d2-99b7-19c5095954ca] allocate_for_instance() {{(pid=63515) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 888.474959] env[63515]: DEBUG nova.policy [None req-dbe24c17-88c2-40e6-987c-d70326fd4bd3 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ed295ef8f0134bcaa668149ac9f4dd74', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'aef947bfaf7d4180b46b4be662e3d903', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63515) authorize /opt/stack/nova/nova/policy.py:201}} [ 888.482936] env[63515]: DEBUG oslo_vmware.api [None req-07515409-459b-4c3e-992f-c4e070f0ae85 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Task: {'id': task-1111488, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.138844} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 888.483467] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-07515409-459b-4c3e-992f-c4e070f0ae85 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Deleted the datastore file {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 888.483696] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-07515409-459b-4c3e-992f-c4e070f0ae85 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: 0967283f-8dda-4692-b6b7-601545cabf2f] Deleted contents of the VM from datastore datastore2 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 888.483896] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-07515409-459b-4c3e-992f-c4e070f0ae85 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: 0967283f-8dda-4692-b6b7-601545cabf2f] Instance destroyed {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 888.484123] env[63515]: INFO nova.compute.manager [None req-07515409-459b-4c3e-992f-c4e070f0ae85 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: 0967283f-8dda-4692-b6b7-601545cabf2f] Took 1.13 seconds to destroy the instance on the hypervisor. [ 888.484447] env[63515]: DEBUG oslo.service.loopingcall [None req-07515409-459b-4c3e-992f-c4e070f0ae85 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 888.484667] env[63515]: DEBUG nova.compute.manager [-] [instance: 0967283f-8dda-4692-b6b7-601545cabf2f] Deallocating network for instance {{(pid=63515) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 888.484774] env[63515]: DEBUG nova.network.neutron [-] [instance: 0967283f-8dda-4692-b6b7-601545cabf2f] deallocate_for_instance() {{(pid=63515) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 888.505728] env[63515]: INFO nova.compute.manager [None req-593b35b4-5d74-4c58-a93e-9cad00fc0b2d tempest-ServerGroupTestJSON-385159797 tempest-ServerGroupTestJSON-385159797-project-member] [instance: a2e014a2-9e0f-4250-aa38-3a7154149903] Took 33.63 seconds to build instance. [ 888.552815] env[63515]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 888.552815] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]5255a7c2-6798-a69c-e24b-7072185c2281" [ 888.552815] env[63515]: _type = "HttpNfcLease" [ 888.552815] env[63515]: } is ready. {{(pid=63515) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 888.554635] env[63515]: DEBUG oslo_vmware.rw_handles [None req-9a80c171-6c9b-44ae-ab0a-573a8347fcf4 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 888.554635] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]5255a7c2-6798-a69c-e24b-7072185c2281" [ 888.554635] env[63515]: _type = "HttpNfcLease" [ 888.554635] env[63515]: }. {{(pid=63515) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 888.556933] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18a5b2c9-1850-49bc-9bce-728b2f37c1eb {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.562581] env[63515]: DEBUG oslo_vmware.rw_handles [None req-9a80c171-6c9b-44ae-ab0a-573a8347fcf4 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52b89a6d-8418-c376-a113-c8e03bf9c4de/disk-0.vmdk from lease info. {{(pid=63515) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 888.562758] env[63515]: DEBUG oslo_vmware.rw_handles [None req-9a80c171-6c9b-44ae-ab0a-573a8347fcf4 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Creating HTTP connection to write to file with size = 21334016 and URL = https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52b89a6d-8418-c376-a113-c8e03bf9c4de/disk-0.vmdk. {{(pid=63515) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 888.637382] env[63515]: DEBUG oslo_vmware.api [None req-d7e3c195-4890-454e-9f6b-af1a7cc4ee6a tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Task: {'id': task-1111489, 'name': Rename_Task, 'duration_secs': 0.154984} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 888.638155] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-d7e3c195-4890-454e-9f6b-af1a7cc4ee6a tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: ae5d6f34-bac6-4ba0-b0c3-8205982ddc88] Powering on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 888.641024] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1e1703bd-c93b-45e6-bce3-0bd4196cb2e4 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.644481] env[63515]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-d6d1d8fc-ec8b-4cf2-92df-9b6b4bf2e7e1 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.650998] env[63515]: DEBUG oslo_vmware.api [None req-d7e3c195-4890-454e-9f6b-af1a7cc4ee6a tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Waiting for the task: (returnval){ [ 888.650998] env[63515]: value = "task-1111490" [ 888.650998] env[63515]: _type = "Task" [ 888.650998] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.661726] env[63515]: DEBUG oslo_vmware.api [None req-d7e3c195-4890-454e-9f6b-af1a7cc4ee6a tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Task: {'id': task-1111490, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.686109] env[63515]: INFO nova.compute.manager [-] [instance: d926c699-a64a-4942-9ef4-f0166414661d] Took 1.28 seconds to deallocate network for instance. [ 888.798718] env[63515]: DEBUG nova.compute.manager [req-87e062cb-9333-4df7-affb-1e0c2f9700ec req-e0de0861-b943-4dd5-97d5-8453be51b3f9 service nova] [instance: 0967283f-8dda-4692-b6b7-601545cabf2f] Received event network-vif-deleted-66571248-83b1-46c5-b892-f84ab3882407 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 888.800088] env[63515]: INFO nova.compute.manager [req-87e062cb-9333-4df7-affb-1e0c2f9700ec req-e0de0861-b943-4dd5-97d5-8453be51b3f9 service nova] [instance: 0967283f-8dda-4692-b6b7-601545cabf2f] Neutron deleted interface 66571248-83b1-46c5-b892-f84ab3882407; detaching it from the instance and deleting it from the info cache [ 888.800088] env[63515]: DEBUG nova.network.neutron [req-87e062cb-9333-4df7-affb-1e0c2f9700ec req-e0de0861-b943-4dd5-97d5-8453be51b3f9 service nova] [instance: 0967283f-8dda-4692-b6b7-601545cabf2f] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 888.810208] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14dbb097-604c-4a96-8e60-de1ed17a728e {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.819055] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e9eb2f4-d59f-4cad-9fde-9bbbe1d5fe45 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.850663] env[63515]: DEBUG nova.network.neutron [None req-dbe24c17-88c2-40e6-987c-d70326fd4bd3 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: e13da90c-28e6-43d2-99b7-19c5095954ca] Successfully created port: d9363b80-630e-4262-901f-c13e5c8b8c2a {{(pid=63515) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 888.853257] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc76a46c-eab3-4e7f-9f23-313ec5e7cb51 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.861398] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd73645b-5ba0-4d0b-92ad-c1147afdc7b9 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.879252] env[63515]: DEBUG nova.compute.provider_tree [None req-87ca4373-5e35-4af3-b945-64d29205298c tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 888.932285] env[63515]: DEBUG nova.compute.manager [None req-dbe24c17-88c2-40e6-987c-d70326fd4bd3 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: e13da90c-28e6-43d2-99b7-19c5095954ca] Start building block device mappings for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 889.011268] env[63515]: DEBUG oslo_concurrency.lockutils [None req-593b35b4-5d74-4c58-a93e-9cad00fc0b2d tempest-ServerGroupTestJSON-385159797 tempest-ServerGroupTestJSON-385159797-project-member] Lock "a2e014a2-9e0f-4250-aa38-3a7154149903" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 35.138s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 889.139355] env[63515]: DEBUG oslo_concurrency.lockutils [None req-060c62db-6550-44b8-abf2-82f4d769cd48 tempest-ServerGroupTestJSON-385159797 tempest-ServerGroupTestJSON-385159797-project-member] Acquiring lock "a2e014a2-9e0f-4250-aa38-3a7154149903" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 889.139355] env[63515]: DEBUG oslo_concurrency.lockutils [None req-060c62db-6550-44b8-abf2-82f4d769cd48 tempest-ServerGroupTestJSON-385159797 tempest-ServerGroupTestJSON-385159797-project-member] Lock "a2e014a2-9e0f-4250-aa38-3a7154149903" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 889.139355] env[63515]: DEBUG oslo_concurrency.lockutils [None req-060c62db-6550-44b8-abf2-82f4d769cd48 tempest-ServerGroupTestJSON-385159797 tempest-ServerGroupTestJSON-385159797-project-member] Acquiring lock "a2e014a2-9e0f-4250-aa38-3a7154149903-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 889.139494] env[63515]: DEBUG oslo_concurrency.lockutils [None req-060c62db-6550-44b8-abf2-82f4d769cd48 tempest-ServerGroupTestJSON-385159797 tempest-ServerGroupTestJSON-385159797-project-member] Lock "a2e014a2-9e0f-4250-aa38-3a7154149903-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 889.139613] env[63515]: DEBUG oslo_concurrency.lockutils [None req-060c62db-6550-44b8-abf2-82f4d769cd48 tempest-ServerGroupTestJSON-385159797 tempest-ServerGroupTestJSON-385159797-project-member] Lock "a2e014a2-9e0f-4250-aa38-3a7154149903-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 889.143260] env[63515]: INFO nova.compute.manager [None req-060c62db-6550-44b8-abf2-82f4d769cd48 tempest-ServerGroupTestJSON-385159797 tempest-ServerGroupTestJSON-385159797-project-member] [instance: a2e014a2-9e0f-4250-aa38-3a7154149903] Terminating instance [ 889.145354] env[63515]: DEBUG nova.compute.manager [None req-060c62db-6550-44b8-abf2-82f4d769cd48 tempest-ServerGroupTestJSON-385159797 tempest-ServerGroupTestJSON-385159797-project-member] [instance: a2e014a2-9e0f-4250-aa38-3a7154149903] Start destroying the instance on the hypervisor. {{(pid=63515) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 889.145562] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-060c62db-6550-44b8-abf2-82f4d769cd48 tempest-ServerGroupTestJSON-385159797 tempest-ServerGroupTestJSON-385159797-project-member] [instance: a2e014a2-9e0f-4250-aa38-3a7154149903] Destroying instance {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 889.146523] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e92f4ce-1d9c-4387-87aa-a185d11c58bc {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.156705] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-060c62db-6550-44b8-abf2-82f4d769cd48 tempest-ServerGroupTestJSON-385159797 tempest-ServerGroupTestJSON-385159797-project-member] [instance: a2e014a2-9e0f-4250-aa38-3a7154149903] Powering off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 889.158564] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9abee5ad-cfca-4901-a0f9-d880ee5cdc37 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.163566] env[63515]: DEBUG oslo_vmware.api [None req-d7e3c195-4890-454e-9f6b-af1a7cc4ee6a tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Task: {'id': task-1111490, 'name': PowerOnVM_Task, 'duration_secs': 0.464391} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 889.166477] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-d7e3c195-4890-454e-9f6b-af1a7cc4ee6a tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: ae5d6f34-bac6-4ba0-b0c3-8205982ddc88] Powered on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 889.166637] env[63515]: INFO nova.compute.manager [None req-d7e3c195-4890-454e-9f6b-af1a7cc4ee6a tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: ae5d6f34-bac6-4ba0-b0c3-8205982ddc88] Took 8.03 seconds to spawn the instance on the hypervisor. [ 889.166820] env[63515]: DEBUG nova.compute.manager [None req-d7e3c195-4890-454e-9f6b-af1a7cc4ee6a tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: ae5d6f34-bac6-4ba0-b0c3-8205982ddc88] Checking state {{(pid=63515) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 889.167218] env[63515]: DEBUG oslo_vmware.api [None req-060c62db-6550-44b8-abf2-82f4d769cd48 tempest-ServerGroupTestJSON-385159797 tempest-ServerGroupTestJSON-385159797-project-member] Waiting for the task: (returnval){ [ 889.167218] env[63515]: value = "task-1111491" [ 889.167218] env[63515]: _type = "Task" [ 889.167218] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 889.167956] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27fd6ad8-893c-424e-b3fd-eafbc46e4e1c {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.187920] env[63515]: DEBUG oslo_vmware.api [None req-060c62db-6550-44b8-abf2-82f4d769cd48 tempest-ServerGroupTestJSON-385159797 tempest-ServerGroupTestJSON-385159797-project-member] Task: {'id': task-1111491, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.191740] env[63515]: DEBUG oslo_concurrency.lockutils [None req-b4e6fa31-2603-4190-94fe-f49ca8c1e386 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 889.275966] env[63515]: DEBUG nova.network.neutron [-] [instance: 0967283f-8dda-4692-b6b7-601545cabf2f] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 889.303486] env[63515]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e2993634-3fd0-4992-941a-d4176c2cf5be {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.313128] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0fcab0c4-1f35-416a-bdcd-0becd5e4badc {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.345638] env[63515]: DEBUG nova.compute.manager [req-87e062cb-9333-4df7-affb-1e0c2f9700ec req-e0de0861-b943-4dd5-97d5-8453be51b3f9 service nova] [instance: 0967283f-8dda-4692-b6b7-601545cabf2f] Detach interface failed, port_id=66571248-83b1-46c5-b892-f84ab3882407, reason: Instance 0967283f-8dda-4692-b6b7-601545cabf2f could not be found. {{(pid=63515) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 889.382116] env[63515]: DEBUG nova.scheduler.client.report [None req-87ca4373-5e35-4af3-b945-64d29205298c tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 889.541960] env[63515]: DEBUG oslo_vmware.rw_handles [None req-9a80c171-6c9b-44ae-ab0a-573a8347fcf4 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Completed reading data from the image iterator. {{(pid=63515) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 889.542227] env[63515]: DEBUG oslo_vmware.rw_handles [None req-9a80c171-6c9b-44ae-ab0a-573a8347fcf4 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52b89a6d-8418-c376-a113-c8e03bf9c4de/disk-0.vmdk. {{(pid=63515) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 889.543151] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f312736c-ff76-46f0-8cbb-c20867ae7197 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.550295] env[63515]: DEBUG oslo_vmware.rw_handles [None req-9a80c171-6c9b-44ae-ab0a-573a8347fcf4 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52b89a6d-8418-c376-a113-c8e03bf9c4de/disk-0.vmdk is in state: ready. {{(pid=63515) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 889.550468] env[63515]: DEBUG oslo_vmware.rw_handles [None req-9a80c171-6c9b-44ae-ab0a-573a8347fcf4 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Releasing lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52b89a6d-8418-c376-a113-c8e03bf9c4de/disk-0.vmdk. {{(pid=63515) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 889.550744] env[63515]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-75182246-1ebc-4878-b1cd-27c83c0eadf6 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.684499] env[63515]: DEBUG oslo_vmware.api [None req-060c62db-6550-44b8-abf2-82f4d769cd48 tempest-ServerGroupTestJSON-385159797 tempest-ServerGroupTestJSON-385159797-project-member] Task: {'id': task-1111491, 'name': PowerOffVM_Task, 'duration_secs': 0.227721} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 889.685089] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-060c62db-6550-44b8-abf2-82f4d769cd48 tempest-ServerGroupTestJSON-385159797 tempest-ServerGroupTestJSON-385159797-project-member] [instance: a2e014a2-9e0f-4250-aa38-3a7154149903] Powered off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 889.685832] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-060c62db-6550-44b8-abf2-82f4d769cd48 tempest-ServerGroupTestJSON-385159797 tempest-ServerGroupTestJSON-385159797-project-member] [instance: a2e014a2-9e0f-4250-aa38-3a7154149903] Unregistering the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 889.685832] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a2c7ec89-9599-4097-8f12-7b7046bc3b40 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.697427] env[63515]: INFO nova.compute.manager [None req-d7e3c195-4890-454e-9f6b-af1a7cc4ee6a tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: ae5d6f34-bac6-4ba0-b0c3-8205982ddc88] Took 31.10 seconds to build instance. [ 889.753745] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-060c62db-6550-44b8-abf2-82f4d769cd48 tempest-ServerGroupTestJSON-385159797 tempest-ServerGroupTestJSON-385159797-project-member] [instance: a2e014a2-9e0f-4250-aa38-3a7154149903] Unregistered the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 889.754023] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-060c62db-6550-44b8-abf2-82f4d769cd48 tempest-ServerGroupTestJSON-385159797 tempest-ServerGroupTestJSON-385159797-project-member] [instance: a2e014a2-9e0f-4250-aa38-3a7154149903] Deleting contents of the VM from datastore datastore2 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 889.754223] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-060c62db-6550-44b8-abf2-82f4d769cd48 tempest-ServerGroupTestJSON-385159797 tempest-ServerGroupTestJSON-385159797-project-member] Deleting the datastore file [datastore2] a2e014a2-9e0f-4250-aa38-3a7154149903 {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 889.754572] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-99aa4baa-831c-4d31-9adc-6a1875d57292 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.761747] env[63515]: DEBUG oslo_vmware.api [None req-060c62db-6550-44b8-abf2-82f4d769cd48 tempest-ServerGroupTestJSON-385159797 tempest-ServerGroupTestJSON-385159797-project-member] Waiting for the task: (returnval){ [ 889.761747] env[63515]: value = "task-1111493" [ 889.761747] env[63515]: _type = "Task" [ 889.761747] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 889.772484] env[63515]: DEBUG oslo_vmware.api [None req-060c62db-6550-44b8-abf2-82f4d769cd48 tempest-ServerGroupTestJSON-385159797 tempest-ServerGroupTestJSON-385159797-project-member] Task: {'id': task-1111493, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.779214] env[63515]: INFO nova.compute.manager [-] [instance: 0967283f-8dda-4692-b6b7-601545cabf2f] Took 1.29 seconds to deallocate network for instance. [ 889.889036] env[63515]: DEBUG oslo_concurrency.lockutils [None req-87ca4373-5e35-4af3-b945-64d29205298c tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.965s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 889.891982] env[63515]: DEBUG oslo_vmware.rw_handles [None req-9a80c171-6c9b-44ae-ab0a-573a8347fcf4 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Closed VMDK write handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52b89a6d-8418-c376-a113-c8e03bf9c4de/disk-0.vmdk. {{(pid=63515) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 889.892309] env[63515]: INFO nova.virt.vmwareapi.images [None req-9a80c171-6c9b-44ae-ab0a-573a8347fcf4 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: b1bf5a0a-42e7-41a6-90fb-a0c311623fdd] Downloaded image file data d1918637-9f8a-4d89-9bc2-49e7d0fdaeee [ 889.893029] env[63515]: DEBUG oslo_concurrency.lockutils [None req-ceee2529-e58a-42aa-a26e-32b1ce124ccb tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 7.087s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 889.894836] env[63515]: INFO nova.compute.claims [None req-ceee2529-e58a-42aa-a26e-32b1ce124ccb tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] [instance: 324f7e89-3c17-474d-af1b-6d7a6d042510] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 889.899321] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e883597-3bdd-4d6d-8dd5-17f79dfd8bca {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.918694] env[63515]: INFO nova.scheduler.client.report [None req-87ca4373-5e35-4af3-b945-64d29205298c tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Deleted allocations for instance d7e8dcdb-41cb-46fb-8b61-d251e7c2d372 [ 889.921740] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-0a28190f-0457-4d25-8967-affba2e0379c {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.945056] env[63515]: DEBUG nova.compute.manager [None req-dbe24c17-88c2-40e6-987c-d70326fd4bd3 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: e13da90c-28e6-43d2-99b7-19c5095954ca] Start spawning the instance on the hypervisor. {{(pid=63515) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 889.947313] env[63515]: INFO nova.virt.vmwareapi.images [None req-9a80c171-6c9b-44ae-ab0a-573a8347fcf4 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: b1bf5a0a-42e7-41a6-90fb-a0c311623fdd] The imported VM was unregistered [ 889.949570] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-9a80c171-6c9b-44ae-ab0a-573a8347fcf4 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: b1bf5a0a-42e7-41a6-90fb-a0c311623fdd] Caching image {{(pid=63515) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 889.949781] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-9a80c171-6c9b-44ae-ab0a-573a8347fcf4 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Creating directory with path [datastore2] devstack-image-cache_base/d1918637-9f8a-4d89-9bc2-49e7d0fdaeee {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 889.950516] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-3428efcd-b3be-4c51-af65-6265b4c50ee1 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.968311] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-9a80c171-6c9b-44ae-ab0a-573a8347fcf4 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Created directory with path [datastore2] devstack-image-cache_base/d1918637-9f8a-4d89-9bc2-49e7d0fdaeee {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 889.968721] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-9a80c171-6c9b-44ae-ab0a-573a8347fcf4 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Moving virtual disk from [datastore2] OSTACK_IMG_570d79c9-5dbb-48eb-b6b9-46e09f2d30ca/OSTACK_IMG_570d79c9-5dbb-48eb-b6b9-46e09f2d30ca.vmdk to [datastore2] devstack-image-cache_base/d1918637-9f8a-4d89-9bc2-49e7d0fdaeee/d1918637-9f8a-4d89-9bc2-49e7d0fdaeee.vmdk. {{(pid=63515) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 889.968891] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-85070e72-615b-4715-90b2-ad88fe6589a4 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.973514] env[63515]: DEBUG nova.virt.hardware [None req-dbe24c17-88c2-40e6-987c-d70326fd4bd3 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T02:52:14Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T02:51:56Z,direct_url=,disk_format='vmdk',id=8a120570-cb06-4099-b262-554ca0ad15c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b569255cc43e42e7ae3f2b4ad37c6ef4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T02:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 889.973754] env[63515]: DEBUG nova.virt.hardware [None req-dbe24c17-88c2-40e6-987c-d70326fd4bd3 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Flavor limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 889.974040] env[63515]: DEBUG nova.virt.hardware [None req-dbe24c17-88c2-40e6-987c-d70326fd4bd3 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Image limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 889.974265] env[63515]: DEBUG nova.virt.hardware [None req-dbe24c17-88c2-40e6-987c-d70326fd4bd3 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Flavor pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 889.974420] env[63515]: DEBUG nova.virt.hardware [None req-dbe24c17-88c2-40e6-987c-d70326fd4bd3 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Image pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 889.974688] env[63515]: DEBUG nova.virt.hardware [None req-dbe24c17-88c2-40e6-987c-d70326fd4bd3 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 889.974816] env[63515]: DEBUG nova.virt.hardware [None req-dbe24c17-88c2-40e6-987c-d70326fd4bd3 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 889.974984] env[63515]: DEBUG nova.virt.hardware [None req-dbe24c17-88c2-40e6-987c-d70326fd4bd3 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 889.975171] env[63515]: DEBUG nova.virt.hardware [None req-dbe24c17-88c2-40e6-987c-d70326fd4bd3 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Got 1 possible topologies {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 889.975350] env[63515]: DEBUG nova.virt.hardware [None req-dbe24c17-88c2-40e6-987c-d70326fd4bd3 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 889.975548] env[63515]: DEBUG nova.virt.hardware [None req-dbe24c17-88c2-40e6-987c-d70326fd4bd3 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 889.976573] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d440608c-0dc0-4302-a11f-81827b98a345 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.985984] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8acdab6-d091-4ad5-8712-f305b1096f47 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.990305] env[63515]: DEBUG oslo_vmware.api [None req-9a80c171-6c9b-44ae-ab0a-573a8347fcf4 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Waiting for the task: (returnval){ [ 889.990305] env[63515]: value = "task-1111495" [ 889.990305] env[63515]: _type = "Task" [ 889.990305] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 890.006505] env[63515]: DEBUG oslo_vmware.api [None req-9a80c171-6c9b-44ae-ab0a-573a8347fcf4 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Task: {'id': task-1111495, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.201501] env[63515]: DEBUG oslo_concurrency.lockutils [None req-d7e3c195-4890-454e-9f6b-af1a7cc4ee6a tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Lock "ae5d6f34-bac6-4ba0-b0c3-8205982ddc88" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 32.625s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 890.271958] env[63515]: DEBUG oslo_vmware.api [None req-060c62db-6550-44b8-abf2-82f4d769cd48 tempest-ServerGroupTestJSON-385159797 tempest-ServerGroupTestJSON-385159797-project-member] Task: {'id': task-1111493, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.3193} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 890.272247] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-060c62db-6550-44b8-abf2-82f4d769cd48 tempest-ServerGroupTestJSON-385159797 tempest-ServerGroupTestJSON-385159797-project-member] Deleted the datastore file {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 890.272439] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-060c62db-6550-44b8-abf2-82f4d769cd48 tempest-ServerGroupTestJSON-385159797 tempest-ServerGroupTestJSON-385159797-project-member] [instance: a2e014a2-9e0f-4250-aa38-3a7154149903] Deleted contents of the VM from datastore datastore2 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 890.273029] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-060c62db-6550-44b8-abf2-82f4d769cd48 tempest-ServerGroupTestJSON-385159797 tempest-ServerGroupTestJSON-385159797-project-member] [instance: a2e014a2-9e0f-4250-aa38-3a7154149903] Instance destroyed {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 890.273029] env[63515]: INFO nova.compute.manager [None req-060c62db-6550-44b8-abf2-82f4d769cd48 tempest-ServerGroupTestJSON-385159797 tempest-ServerGroupTestJSON-385159797-project-member] [instance: a2e014a2-9e0f-4250-aa38-3a7154149903] Took 1.13 seconds to destroy the instance on the hypervisor. [ 890.273149] env[63515]: DEBUG oslo.service.loopingcall [None req-060c62db-6550-44b8-abf2-82f4d769cd48 tempest-ServerGroupTestJSON-385159797 tempest-ServerGroupTestJSON-385159797-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 890.273318] env[63515]: DEBUG nova.compute.manager [-] [instance: a2e014a2-9e0f-4250-aa38-3a7154149903] Deallocating network for instance {{(pid=63515) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 890.273414] env[63515]: DEBUG nova.network.neutron [-] [instance: a2e014a2-9e0f-4250-aa38-3a7154149903] deallocate_for_instance() {{(pid=63515) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 890.286361] env[63515]: DEBUG oslo_concurrency.lockutils [None req-07515409-459b-4c3e-992f-c4e070f0ae85 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 890.368406] env[63515]: DEBUG oslo_concurrency.lockutils [None req-ef0f5ab2-1246-4de1-af39-f1efd470cd76 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Acquiring lock "ae5d6f34-bac6-4ba0-b0c3-8205982ddc88" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 890.368703] env[63515]: DEBUG oslo_concurrency.lockutils [None req-ef0f5ab2-1246-4de1-af39-f1efd470cd76 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Lock "ae5d6f34-bac6-4ba0-b0c3-8205982ddc88" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.001s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 890.368894] env[63515]: INFO nova.compute.manager [None req-ef0f5ab2-1246-4de1-af39-f1efd470cd76 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: ae5d6f34-bac6-4ba0-b0c3-8205982ddc88] Shelving [ 890.409751] env[63515]: DEBUG nova.compute.manager [req-8e16cba1-da68-4b8f-9f88-e24f532a6449 req-5cfebfd5-7dc2-49e0-866d-be6cbe69275a service nova] [instance: e13da90c-28e6-43d2-99b7-19c5095954ca] Received event network-vif-plugged-d9363b80-630e-4262-901f-c13e5c8b8c2a {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 890.410500] env[63515]: DEBUG oslo_concurrency.lockutils [req-8e16cba1-da68-4b8f-9f88-e24f532a6449 req-5cfebfd5-7dc2-49e0-866d-be6cbe69275a service nova] Acquiring lock "e13da90c-28e6-43d2-99b7-19c5095954ca-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 890.410735] env[63515]: DEBUG oslo_concurrency.lockutils [req-8e16cba1-da68-4b8f-9f88-e24f532a6449 req-5cfebfd5-7dc2-49e0-866d-be6cbe69275a service nova] Lock "e13da90c-28e6-43d2-99b7-19c5095954ca-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 890.410915] env[63515]: DEBUG oslo_concurrency.lockutils [req-8e16cba1-da68-4b8f-9f88-e24f532a6449 req-5cfebfd5-7dc2-49e0-866d-be6cbe69275a service nova] Lock "e13da90c-28e6-43d2-99b7-19c5095954ca-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 890.411172] env[63515]: DEBUG nova.compute.manager [req-8e16cba1-da68-4b8f-9f88-e24f532a6449 req-5cfebfd5-7dc2-49e0-866d-be6cbe69275a service nova] [instance: e13da90c-28e6-43d2-99b7-19c5095954ca] No waiting events found dispatching network-vif-plugged-d9363b80-630e-4262-901f-c13e5c8b8c2a {{(pid=63515) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 890.411373] env[63515]: WARNING nova.compute.manager [req-8e16cba1-da68-4b8f-9f88-e24f532a6449 req-5cfebfd5-7dc2-49e0-866d-be6cbe69275a service nova] [instance: e13da90c-28e6-43d2-99b7-19c5095954ca] Received unexpected event network-vif-plugged-d9363b80-630e-4262-901f-c13e5c8b8c2a for instance with vm_state building and task_state spawning. [ 890.429092] env[63515]: DEBUG oslo_concurrency.lockutils [None req-87ca4373-5e35-4af3-b945-64d29205298c tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Lock "d7e8dcdb-41cb-46fb-8b61-d251e7c2d372" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 16.170s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 890.501199] env[63515]: DEBUG oslo_vmware.api [None req-9a80c171-6c9b-44ae-ab0a-573a8347fcf4 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Task: {'id': task-1111495, 'name': MoveVirtualDisk_Task} progress is 9%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.564661] env[63515]: DEBUG nova.network.neutron [None req-dbe24c17-88c2-40e6-987c-d70326fd4bd3 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: e13da90c-28e6-43d2-99b7-19c5095954ca] Successfully updated port: d9363b80-630e-4262-901f-c13e5c8b8c2a {{(pid=63515) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 890.837216] env[63515]: DEBUG nova.compute.manager [req-b7934f09-de5f-45be-a057-dc946d170d68 req-11e88bac-c685-409c-84f6-2081535e67d1 service nova] [instance: a2e014a2-9e0f-4250-aa38-3a7154149903] Received event network-vif-deleted-a453f43a-c524-4eb7-8dcc-9edc8aabdd2c {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 890.837216] env[63515]: INFO nova.compute.manager [req-b7934f09-de5f-45be-a057-dc946d170d68 req-11e88bac-c685-409c-84f6-2081535e67d1 service nova] [instance: a2e014a2-9e0f-4250-aa38-3a7154149903] Neutron deleted interface a453f43a-c524-4eb7-8dcc-9edc8aabdd2c; detaching it from the instance and deleting it from the info cache [ 890.837391] env[63515]: DEBUG nova.network.neutron [req-b7934f09-de5f-45be-a057-dc946d170d68 req-11e88bac-c685-409c-84f6-2081535e67d1 service nova] [instance: a2e014a2-9e0f-4250-aa38-3a7154149903] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 890.879853] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-ef0f5ab2-1246-4de1-af39-f1efd470cd76 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: ae5d6f34-bac6-4ba0-b0c3-8205982ddc88] Powering off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 890.880201] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-40c72af7-d8b9-42e8-9a4e-f21570fc3c5a {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.888275] env[63515]: DEBUG oslo_vmware.api [None req-ef0f5ab2-1246-4de1-af39-f1efd470cd76 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Waiting for the task: (returnval){ [ 890.888275] env[63515]: value = "task-1111496" [ 890.888275] env[63515]: _type = "Task" [ 890.888275] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 890.897526] env[63515]: DEBUG oslo_vmware.api [None req-ef0f5ab2-1246-4de1-af39-f1efd470cd76 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Task: {'id': task-1111496, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.004710] env[63515]: DEBUG oslo_vmware.api [None req-9a80c171-6c9b-44ae-ab0a-573a8347fcf4 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Task: {'id': task-1111495, 'name': MoveVirtualDisk_Task} progress is 21%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.047828] env[63515]: DEBUG nova.network.neutron [-] [instance: a2e014a2-9e0f-4250-aa38-3a7154149903] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 891.067487] env[63515]: DEBUG oslo_concurrency.lockutils [None req-dbe24c17-88c2-40e6-987c-d70326fd4bd3 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Acquiring lock "refresh_cache-e13da90c-28e6-43d2-99b7-19c5095954ca" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 891.067487] env[63515]: DEBUG oslo_concurrency.lockutils [None req-dbe24c17-88c2-40e6-987c-d70326fd4bd3 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Acquired lock "refresh_cache-e13da90c-28e6-43d2-99b7-19c5095954ca" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 891.067487] env[63515]: DEBUG nova.network.neutron [None req-dbe24c17-88c2-40e6-987c-d70326fd4bd3 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: e13da90c-28e6-43d2-99b7-19c5095954ca] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 891.254840] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84bad3f2-b2f1-4f6d-8460-54f0151cc2b6 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.264090] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42848eb8-cadc-445a-b020-a00362958740 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.299049] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69f4877c-79fb-4a3b-acd5-f5a3bd9cf58f {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.307349] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d640e8d3-5e9a-4134-9899-fc173c3099aa {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.322101] env[63515]: DEBUG nova.compute.provider_tree [None req-ceee2529-e58a-42aa-a26e-32b1ce124ccb tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 891.341256] env[63515]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ca950cbd-16f9-47c1-bad4-d4f34cc84d7b {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.351477] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-601836e1-527e-4a23-8962-f862f18b91c3 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.385374] env[63515]: DEBUG nova.compute.manager [req-b7934f09-de5f-45be-a057-dc946d170d68 req-11e88bac-c685-409c-84f6-2081535e67d1 service nova] [instance: a2e014a2-9e0f-4250-aa38-3a7154149903] Detach interface failed, port_id=a453f43a-c524-4eb7-8dcc-9edc8aabdd2c, reason: Instance a2e014a2-9e0f-4250-aa38-3a7154149903 could not be found. {{(pid=63515) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 891.400325] env[63515]: DEBUG oslo_vmware.api [None req-ef0f5ab2-1246-4de1-af39-f1efd470cd76 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Task: {'id': task-1111496, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.503362] env[63515]: DEBUG oslo_vmware.api [None req-9a80c171-6c9b-44ae-ab0a-573a8347fcf4 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Task: {'id': task-1111495, 'name': MoveVirtualDisk_Task} progress is 43%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.552209] env[63515]: INFO nova.compute.manager [-] [instance: a2e014a2-9e0f-4250-aa38-3a7154149903] Took 1.28 seconds to deallocate network for instance. [ 891.603863] env[63515]: DEBUG nova.network.neutron [None req-dbe24c17-88c2-40e6-987c-d70326fd4bd3 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: e13da90c-28e6-43d2-99b7-19c5095954ca] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 891.774074] env[63515]: DEBUG nova.network.neutron [None req-dbe24c17-88c2-40e6-987c-d70326fd4bd3 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: e13da90c-28e6-43d2-99b7-19c5095954ca] Updating instance_info_cache with network_info: [{"id": "d9363b80-630e-4262-901f-c13e5c8b8c2a", "address": "fa:16:3e:dd:fe:76", "network": {"id": "cec614cb-7a5b-4b16-a83d-061ff62d6677", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1501468300-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "aef947bfaf7d4180b46b4be662e3d903", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3a0ddd7d-c321-4187-bdd8-b19044ea2c4a", "external-id": "nsx-vlan-transportzone-747", "segmentation_id": 747, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd9363b80-63", "ovs_interfaceid": "d9363b80-630e-4262-901f-c13e5c8b8c2a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 891.826893] env[63515]: DEBUG nova.scheduler.client.report [None req-ceee2529-e58a-42aa-a26e-32b1ce124ccb tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 891.898988] env[63515]: DEBUG oslo_vmware.api [None req-ef0f5ab2-1246-4de1-af39-f1efd470cd76 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Task: {'id': task-1111496, 'name': PowerOffVM_Task, 'duration_secs': 0.933105} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.899364] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-ef0f5ab2-1246-4de1-af39-f1efd470cd76 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: ae5d6f34-bac6-4ba0-b0c3-8205982ddc88] Powered off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 891.900358] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1672af1a-e5e9-4be2-99d5-3b9c99acbc86 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.920792] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b31a5d19-55ee-44de-aefe-a99190dd0108 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.002909] env[63515]: DEBUG oslo_vmware.api [None req-9a80c171-6c9b-44ae-ab0a-573a8347fcf4 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Task: {'id': task-1111495, 'name': MoveVirtualDisk_Task} progress is 63%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.058850] env[63515]: DEBUG oslo_concurrency.lockutils [None req-060c62db-6550-44b8-abf2-82f4d769cd48 tempest-ServerGroupTestJSON-385159797 tempest-ServerGroupTestJSON-385159797-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 892.277374] env[63515]: DEBUG oslo_concurrency.lockutils [None req-dbe24c17-88c2-40e6-987c-d70326fd4bd3 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Releasing lock "refresh_cache-e13da90c-28e6-43d2-99b7-19c5095954ca" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 892.277742] env[63515]: DEBUG nova.compute.manager [None req-dbe24c17-88c2-40e6-987c-d70326fd4bd3 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: e13da90c-28e6-43d2-99b7-19c5095954ca] Instance network_info: |[{"id": "d9363b80-630e-4262-901f-c13e5c8b8c2a", "address": "fa:16:3e:dd:fe:76", "network": {"id": "cec614cb-7a5b-4b16-a83d-061ff62d6677", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1501468300-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "aef947bfaf7d4180b46b4be662e3d903", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3a0ddd7d-c321-4187-bdd8-b19044ea2c4a", "external-id": "nsx-vlan-transportzone-747", "segmentation_id": 747, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd9363b80-63", "ovs_interfaceid": "d9363b80-630e-4262-901f-c13e5c8b8c2a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 892.278258] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-dbe24c17-88c2-40e6-987c-d70326fd4bd3 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: e13da90c-28e6-43d2-99b7-19c5095954ca] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:dd:fe:76', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '3a0ddd7d-c321-4187-bdd8-b19044ea2c4a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'd9363b80-630e-4262-901f-c13e5c8b8c2a', 'vif_model': 'vmxnet3'}] {{(pid=63515) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 892.286107] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-dbe24c17-88c2-40e6-987c-d70326fd4bd3 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Creating folder: Project (aef947bfaf7d4180b46b4be662e3d903). Parent ref: group-v243370. {{(pid=63515) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 892.286461] env[63515]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-803e2119-41bf-4636-b93e-9e5b9ee0965b {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.297765] env[63515]: INFO nova.virt.vmwareapi.vm_util [None req-dbe24c17-88c2-40e6-987c-d70326fd4bd3 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Created folder: Project (aef947bfaf7d4180b46b4be662e3d903) in parent group-v243370. [ 892.297997] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-dbe24c17-88c2-40e6-987c-d70326fd4bd3 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Creating folder: Instances. Parent ref: group-v243515. {{(pid=63515) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 892.298291] env[63515]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-dc3c1e28-3e54-4e90-ac14-1c4180cacd44 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.308659] env[63515]: INFO nova.virt.vmwareapi.vm_util [None req-dbe24c17-88c2-40e6-987c-d70326fd4bd3 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Created folder: Instances in parent group-v243515. [ 892.308927] env[63515]: DEBUG oslo.service.loopingcall [None req-dbe24c17-88c2-40e6-987c-d70326fd4bd3 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 892.309164] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e13da90c-28e6-43d2-99b7-19c5095954ca] Creating VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 892.309388] env[63515]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-8891d3d0-fc67-46c7-94e7-a0befb304078 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.329511] env[63515]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 892.329511] env[63515]: value = "task-1111499" [ 892.329511] env[63515]: _type = "Task" [ 892.329511] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 892.333464] env[63515]: DEBUG oslo_concurrency.lockutils [None req-ceee2529-e58a-42aa-a26e-32b1ce124ccb tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.440s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 892.334147] env[63515]: DEBUG nova.compute.manager [None req-ceee2529-e58a-42aa-a26e-32b1ce124ccb tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] [instance: 324f7e89-3c17-474d-af1b-6d7a6d042510] Start building networks asynchronously for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 892.336724] env[63515]: DEBUG oslo_concurrency.lockutils [None req-cf46911e-9c80-4bd3-a167-7e0168ef6827 tempest-ServerDiagnosticsV248Test-1736022867 tempest-ServerDiagnosticsV248Test-1736022867-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 9.421s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 892.337013] env[63515]: DEBUG nova.objects.instance [None req-cf46911e-9c80-4bd3-a167-7e0168ef6827 tempest-ServerDiagnosticsV248Test-1736022867 tempest-ServerDiagnosticsV248Test-1736022867-project-member] Lazy-loading 'resources' on Instance uuid 705b7e9b-3ab6-40ec-8cf6-d86bf0a1faf0 {{(pid=63515) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 892.346963] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1111499, 'name': CreateVM_Task} progress is 6%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.431933] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-ef0f5ab2-1246-4de1-af39-f1efd470cd76 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: ae5d6f34-bac6-4ba0-b0c3-8205982ddc88] Creating Snapshot of the VM instance {{(pid=63515) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 892.432540] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-db337b83-25c9-4527-997c-c054bb1cfa30 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.439559] env[63515]: DEBUG oslo_vmware.api [None req-ef0f5ab2-1246-4de1-af39-f1efd470cd76 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Waiting for the task: (returnval){ [ 892.439559] env[63515]: value = "task-1111500" [ 892.439559] env[63515]: _type = "Task" [ 892.439559] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 892.446511] env[63515]: DEBUG nova.compute.manager [req-459b67e3-e477-47d2-84c1-5684e7fccd68 req-853cd7b2-d045-47ac-b936-a61a5b20f2f8 service nova] [instance: e13da90c-28e6-43d2-99b7-19c5095954ca] Received event network-changed-d9363b80-630e-4262-901f-c13e5c8b8c2a {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 892.446738] env[63515]: DEBUG nova.compute.manager [req-459b67e3-e477-47d2-84c1-5684e7fccd68 req-853cd7b2-d045-47ac-b936-a61a5b20f2f8 service nova] [instance: e13da90c-28e6-43d2-99b7-19c5095954ca] Refreshing instance network info cache due to event network-changed-d9363b80-630e-4262-901f-c13e5c8b8c2a. {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 892.446964] env[63515]: DEBUG oslo_concurrency.lockutils [req-459b67e3-e477-47d2-84c1-5684e7fccd68 req-853cd7b2-d045-47ac-b936-a61a5b20f2f8 service nova] Acquiring lock "refresh_cache-e13da90c-28e6-43d2-99b7-19c5095954ca" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 892.447219] env[63515]: DEBUG oslo_concurrency.lockutils [req-459b67e3-e477-47d2-84c1-5684e7fccd68 req-853cd7b2-d045-47ac-b936-a61a5b20f2f8 service nova] Acquired lock "refresh_cache-e13da90c-28e6-43d2-99b7-19c5095954ca" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 892.447785] env[63515]: DEBUG nova.network.neutron [req-459b67e3-e477-47d2-84c1-5684e7fccd68 req-853cd7b2-d045-47ac-b936-a61a5b20f2f8 service nova] [instance: e13da90c-28e6-43d2-99b7-19c5095954ca] Refreshing network info cache for port d9363b80-630e-4262-901f-c13e5c8b8c2a {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 892.455771] env[63515]: DEBUG oslo_vmware.api [None req-ef0f5ab2-1246-4de1-af39-f1efd470cd76 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Task: {'id': task-1111500, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.504342] env[63515]: DEBUG oslo_vmware.api [None req-9a80c171-6c9b-44ae-ab0a-573a8347fcf4 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Task: {'id': task-1111495, 'name': MoveVirtualDisk_Task} progress is 85%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.776286] env[63515]: DEBUG oslo_concurrency.lockutils [None req-29ee5601-80a8-4cbb-90e9-98d09454abfa tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Acquiring lock "86b1ddcf-585c-41d9-84fe-17b794f1abbc" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 892.776526] env[63515]: DEBUG oslo_concurrency.lockutils [None req-29ee5601-80a8-4cbb-90e9-98d09454abfa tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Lock "86b1ddcf-585c-41d9-84fe-17b794f1abbc" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 892.843406] env[63515]: DEBUG nova.compute.utils [None req-ceee2529-e58a-42aa-a26e-32b1ce124ccb tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] Using /dev/sd instead of None {{(pid=63515) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 892.845027] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1111499, 'name': CreateVM_Task, 'duration_secs': 0.393994} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 892.845829] env[63515]: DEBUG nova.compute.manager [None req-ceee2529-e58a-42aa-a26e-32b1ce124ccb tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] [instance: 324f7e89-3c17-474d-af1b-6d7a6d042510] Allocating IP information in the background. {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 892.846024] env[63515]: DEBUG nova.network.neutron [None req-ceee2529-e58a-42aa-a26e-32b1ce124ccb tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] [instance: 324f7e89-3c17-474d-af1b-6d7a6d042510] allocate_for_instance() {{(pid=63515) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 892.847629] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e13da90c-28e6-43d2-99b7-19c5095954ca] Created VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 892.848225] env[63515]: DEBUG oslo_concurrency.lockutils [None req-dbe24c17-88c2-40e6-987c-d70326fd4bd3 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 892.848389] env[63515]: DEBUG oslo_concurrency.lockutils [None req-dbe24c17-88c2-40e6-987c-d70326fd4bd3 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 892.848692] env[63515]: DEBUG oslo_concurrency.lockutils [None req-dbe24c17-88c2-40e6-987c-d70326fd4bd3 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 892.849235] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8b23f732-421e-4e94-8973-b98cec91dd16 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.855922] env[63515]: DEBUG oslo_vmware.api [None req-dbe24c17-88c2-40e6-987c-d70326fd4bd3 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Waiting for the task: (returnval){ [ 892.855922] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]52fb1c41-2b3d-320c-dffd-099e488589c2" [ 892.855922] env[63515]: _type = "Task" [ 892.855922] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 892.866702] env[63515]: DEBUG oslo_vmware.api [None req-dbe24c17-88c2-40e6-987c-d70326fd4bd3 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52fb1c41-2b3d-320c-dffd-099e488589c2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.896651] env[63515]: DEBUG nova.policy [None req-ceee2529-e58a-42aa-a26e-32b1ce124ccb tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b0aec3663b22457f8b87f797ccb41af6', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '26a52fb1944a4f949f6779829e163eaa', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63515) authorize /opt/stack/nova/nova/policy.py:201}} [ 892.948601] env[63515]: DEBUG oslo_vmware.api [None req-ef0f5ab2-1246-4de1-af39-f1efd470cd76 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Task: {'id': task-1111500, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.006834] env[63515]: DEBUG oslo_vmware.api [None req-9a80c171-6c9b-44ae-ab0a-573a8347fcf4 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Task: {'id': task-1111495, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.88984} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.007525] env[63515]: INFO nova.virt.vmwareapi.ds_util [None req-9a80c171-6c9b-44ae-ab0a-573a8347fcf4 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Moved virtual disk from [datastore2] OSTACK_IMG_570d79c9-5dbb-48eb-b6b9-46e09f2d30ca/OSTACK_IMG_570d79c9-5dbb-48eb-b6b9-46e09f2d30ca.vmdk to [datastore2] devstack-image-cache_base/d1918637-9f8a-4d89-9bc2-49e7d0fdaeee/d1918637-9f8a-4d89-9bc2-49e7d0fdaeee.vmdk. [ 893.007876] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-9a80c171-6c9b-44ae-ab0a-573a8347fcf4 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: b1bf5a0a-42e7-41a6-90fb-a0c311623fdd] Cleaning up location [datastore2] OSTACK_IMG_570d79c9-5dbb-48eb-b6b9-46e09f2d30ca {{(pid=63515) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 893.008192] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-9a80c171-6c9b-44ae-ab0a-573a8347fcf4 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Deleting the datastore file [datastore2] OSTACK_IMG_570d79c9-5dbb-48eb-b6b9-46e09f2d30ca {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 893.008697] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-6f34571d-ce96-4400-b96f-c20a8f543a11 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.018551] env[63515]: DEBUG oslo_vmware.api [None req-9a80c171-6c9b-44ae-ab0a-573a8347fcf4 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Waiting for the task: (returnval){ [ 893.018551] env[63515]: value = "task-1111501" [ 893.018551] env[63515]: _type = "Task" [ 893.018551] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 893.026561] env[63515]: DEBUG oslo_vmware.api [None req-9a80c171-6c9b-44ae-ab0a-573a8347fcf4 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Task: {'id': task-1111501, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.137110] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d2602fb-4834-4ead-90bd-3b7cf7c75b19 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.144264] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7fabe485-e82b-49ef-b1a7-39ab9adbbf2f {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.174509] env[63515]: DEBUG nova.network.neutron [None req-ceee2529-e58a-42aa-a26e-32b1ce124ccb tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] [instance: 324f7e89-3c17-474d-af1b-6d7a6d042510] Successfully created port: 0ea22959-737c-478d-84d8-17fbb175b186 {{(pid=63515) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 893.178350] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08d754b5-902a-434b-9591-5edbb437eeb2 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.185762] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89f4fcaf-3009-40e8-aab4-521f29005d7b {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.199982] env[63515]: DEBUG nova.compute.provider_tree [None req-cf46911e-9c80-4bd3-a167-7e0168ef6827 tempest-ServerDiagnosticsV248Test-1736022867 tempest-ServerDiagnosticsV248Test-1736022867-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 893.213779] env[63515]: DEBUG nova.network.neutron [req-459b67e3-e477-47d2-84c1-5684e7fccd68 req-853cd7b2-d045-47ac-b936-a61a5b20f2f8 service nova] [instance: e13da90c-28e6-43d2-99b7-19c5095954ca] Updated VIF entry in instance network info cache for port d9363b80-630e-4262-901f-c13e5c8b8c2a. {{(pid=63515) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 893.214209] env[63515]: DEBUG nova.network.neutron [req-459b67e3-e477-47d2-84c1-5684e7fccd68 req-853cd7b2-d045-47ac-b936-a61a5b20f2f8 service nova] [instance: e13da90c-28e6-43d2-99b7-19c5095954ca] Updating instance_info_cache with network_info: [{"id": "d9363b80-630e-4262-901f-c13e5c8b8c2a", "address": "fa:16:3e:dd:fe:76", "network": {"id": "cec614cb-7a5b-4b16-a83d-061ff62d6677", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1501468300-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "aef947bfaf7d4180b46b4be662e3d903", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3a0ddd7d-c321-4187-bdd8-b19044ea2c4a", "external-id": "nsx-vlan-transportzone-747", "segmentation_id": 747, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd9363b80-63", "ovs_interfaceid": "d9363b80-630e-4262-901f-c13e5c8b8c2a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 893.279051] env[63515]: DEBUG nova.compute.manager [None req-29ee5601-80a8-4cbb-90e9-98d09454abfa tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] [instance: 86b1ddcf-585c-41d9-84fe-17b794f1abbc] Starting instance... {{(pid=63515) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 893.345695] env[63515]: DEBUG nova.compute.manager [None req-ceee2529-e58a-42aa-a26e-32b1ce124ccb tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] [instance: 324f7e89-3c17-474d-af1b-6d7a6d042510] Start building block device mappings for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 893.367851] env[63515]: DEBUG oslo_vmware.api [None req-dbe24c17-88c2-40e6-987c-d70326fd4bd3 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52fb1c41-2b3d-320c-dffd-099e488589c2, 'name': SearchDatastore_Task, 'duration_secs': 0.01906} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.368178] env[63515]: DEBUG oslo_concurrency.lockutils [None req-dbe24c17-88c2-40e6-987c-d70326fd4bd3 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 893.368451] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-dbe24c17-88c2-40e6-987c-d70326fd4bd3 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: e13da90c-28e6-43d2-99b7-19c5095954ca] Processing image 8a120570-cb06-4099-b262-554ca0ad15c5 {{(pid=63515) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 893.368735] env[63515]: DEBUG oslo_concurrency.lockutils [None req-dbe24c17-88c2-40e6-987c-d70326fd4bd3 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 893.368847] env[63515]: DEBUG oslo_concurrency.lockutils [None req-dbe24c17-88c2-40e6-987c-d70326fd4bd3 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 893.369039] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-dbe24c17-88c2-40e6-987c-d70326fd4bd3 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 893.369319] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-57c62ecf-b67f-486e-8723-a3e5644c4dc7 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.382059] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-dbe24c17-88c2-40e6-987c-d70326fd4bd3 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 893.382337] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-dbe24c17-88c2-40e6-987c-d70326fd4bd3 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63515) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 893.383603] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e893a404-c30e-41c1-9c5d-edb1dfb4340c {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.389206] env[63515]: DEBUG oslo_vmware.api [None req-dbe24c17-88c2-40e6-987c-d70326fd4bd3 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Waiting for the task: (returnval){ [ 893.389206] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]52b8ec3c-c48d-497c-3800-7654cf285b75" [ 893.389206] env[63515]: _type = "Task" [ 893.389206] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 893.398523] env[63515]: DEBUG oslo_vmware.api [None req-dbe24c17-88c2-40e6-987c-d70326fd4bd3 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52b8ec3c-c48d-497c-3800-7654cf285b75, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.441884] env[63515]: DEBUG nova.network.neutron [None req-ceee2529-e58a-42aa-a26e-32b1ce124ccb tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] [instance: 324f7e89-3c17-474d-af1b-6d7a6d042510] Successfully created port: c18e8fdc-6f1f-4fd6-8401-cab93bb67126 {{(pid=63515) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 893.454835] env[63515]: DEBUG oslo_vmware.api [None req-ef0f5ab2-1246-4de1-af39-f1efd470cd76 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Task: {'id': task-1111500, 'name': CreateSnapshot_Task, 'duration_secs': 0.80789} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.454835] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-ef0f5ab2-1246-4de1-af39-f1efd470cd76 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: ae5d6f34-bac6-4ba0-b0c3-8205982ddc88] Created Snapshot of the VM instance {{(pid=63515) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 893.454835] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1777aace-68cc-4b8c-ab8a-aa932b9d2eb7 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.529065] env[63515]: DEBUG oslo_vmware.api [None req-9a80c171-6c9b-44ae-ab0a-573a8347fcf4 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Task: {'id': task-1111501, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.043989} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.529364] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-9a80c171-6c9b-44ae-ab0a-573a8347fcf4 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Deleted the datastore file {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 893.529535] env[63515]: DEBUG oslo_concurrency.lockutils [None req-9a80c171-6c9b-44ae-ab0a-573a8347fcf4 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d1918637-9f8a-4d89-9bc2-49e7d0fdaeee/d1918637-9f8a-4d89-9bc2-49e7d0fdaeee.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 893.529787] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-9a80c171-6c9b-44ae-ab0a-573a8347fcf4 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d1918637-9f8a-4d89-9bc2-49e7d0fdaeee/d1918637-9f8a-4d89-9bc2-49e7d0fdaeee.vmdk to [datastore2] b1bf5a0a-42e7-41a6-90fb-a0c311623fdd/b1bf5a0a-42e7-41a6-90fb-a0c311623fdd.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 893.530062] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b07d80bd-8e6c-474d-90d2-9317dcda0b0d {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.537282] env[63515]: DEBUG oslo_vmware.api [None req-9a80c171-6c9b-44ae-ab0a-573a8347fcf4 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Waiting for the task: (returnval){ [ 893.537282] env[63515]: value = "task-1111502" [ 893.537282] env[63515]: _type = "Task" [ 893.537282] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 893.544991] env[63515]: DEBUG oslo_vmware.api [None req-9a80c171-6c9b-44ae-ab0a-573a8347fcf4 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Task: {'id': task-1111502, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.703396] env[63515]: DEBUG nova.scheduler.client.report [None req-cf46911e-9c80-4bd3-a167-7e0168ef6827 tempest-ServerDiagnosticsV248Test-1736022867 tempest-ServerDiagnosticsV248Test-1736022867-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 893.717736] env[63515]: DEBUG oslo_concurrency.lockutils [req-459b67e3-e477-47d2-84c1-5684e7fccd68 req-853cd7b2-d045-47ac-b936-a61a5b20f2f8 service nova] Releasing lock "refresh_cache-e13da90c-28e6-43d2-99b7-19c5095954ca" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 893.809185] env[63515]: DEBUG oslo_concurrency.lockutils [None req-29ee5601-80a8-4cbb-90e9-98d09454abfa tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 893.900220] env[63515]: DEBUG oslo_vmware.api [None req-dbe24c17-88c2-40e6-987c-d70326fd4bd3 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52b8ec3c-c48d-497c-3800-7654cf285b75, 'name': SearchDatastore_Task, 'duration_secs': 0.008752} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.901117] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-778e2c5d-0eb1-4ae6-a380-1601f13a996f {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.907110] env[63515]: DEBUG oslo_vmware.api [None req-dbe24c17-88c2-40e6-987c-d70326fd4bd3 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Waiting for the task: (returnval){ [ 893.907110] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]525bd207-3d69-e145-7353-5504bc7705e7" [ 893.907110] env[63515]: _type = "Task" [ 893.907110] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 893.915162] env[63515]: DEBUG oslo_vmware.api [None req-dbe24c17-88c2-40e6-987c-d70326fd4bd3 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]525bd207-3d69-e145-7353-5504bc7705e7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.972604] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-ef0f5ab2-1246-4de1-af39-f1efd470cd76 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: ae5d6f34-bac6-4ba0-b0c3-8205982ddc88] Creating linked-clone VM from snapshot {{(pid=63515) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 893.973012] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-fe45edf2-9949-43dd-9835-fc464f475ee1 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.981143] env[63515]: DEBUG oslo_vmware.api [None req-ef0f5ab2-1246-4de1-af39-f1efd470cd76 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Waiting for the task: (returnval){ [ 893.981143] env[63515]: value = "task-1111503" [ 893.981143] env[63515]: _type = "Task" [ 893.981143] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 893.989924] env[63515]: DEBUG oslo_vmware.api [None req-ef0f5ab2-1246-4de1-af39-f1efd470cd76 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Task: {'id': task-1111503, 'name': CloneVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.048440] env[63515]: DEBUG oslo_vmware.api [None req-9a80c171-6c9b-44ae-ab0a-573a8347fcf4 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Task: {'id': task-1111502, 'name': CopyVirtualDisk_Task} progress is 24%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.208982] env[63515]: DEBUG oslo_concurrency.lockutils [None req-cf46911e-9c80-4bd3-a167-7e0168ef6827 tempest-ServerDiagnosticsV248Test-1736022867 tempest-ServerDiagnosticsV248Test-1736022867-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.872s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 894.211572] env[63515]: DEBUG oslo_concurrency.lockutils [None req-1d02c1b8-b71d-4e69-b38f-4c4377a39898 tempest-InstanceActionsTestJSON-247065106 tempest-InstanceActionsTestJSON-247065106-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 7.907s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 894.211845] env[63515]: DEBUG nova.objects.instance [None req-1d02c1b8-b71d-4e69-b38f-4c4377a39898 tempest-InstanceActionsTestJSON-247065106 tempest-InstanceActionsTestJSON-247065106-project-member] Lazy-loading 'resources' on Instance uuid 0ee9cce2-94b9-4038-ab38-1b3f96a864bc {{(pid=63515) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 894.238457] env[63515]: INFO nova.scheduler.client.report [None req-cf46911e-9c80-4bd3-a167-7e0168ef6827 tempest-ServerDiagnosticsV248Test-1736022867 tempest-ServerDiagnosticsV248Test-1736022867-project-member] Deleted allocations for instance 705b7e9b-3ab6-40ec-8cf6-d86bf0a1faf0 [ 894.354988] env[63515]: DEBUG nova.compute.manager [None req-ceee2529-e58a-42aa-a26e-32b1ce124ccb tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] [instance: 324f7e89-3c17-474d-af1b-6d7a6d042510] Start spawning the instance on the hypervisor. {{(pid=63515) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 894.381973] env[63515]: DEBUG nova.virt.hardware [None req-ceee2529-e58a-42aa-a26e-32b1ce124ccb tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T02:52:14Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T02:51:56Z,direct_url=,disk_format='vmdk',id=8a120570-cb06-4099-b262-554ca0ad15c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b569255cc43e42e7ae3f2b4ad37c6ef4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T02:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 894.382174] env[63515]: DEBUG nova.virt.hardware [None req-ceee2529-e58a-42aa-a26e-32b1ce124ccb tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] Flavor limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 894.382416] env[63515]: DEBUG nova.virt.hardware [None req-ceee2529-e58a-42aa-a26e-32b1ce124ccb tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] Image limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 894.382616] env[63515]: DEBUG nova.virt.hardware [None req-ceee2529-e58a-42aa-a26e-32b1ce124ccb tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] Flavor pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 894.382770] env[63515]: DEBUG nova.virt.hardware [None req-ceee2529-e58a-42aa-a26e-32b1ce124ccb tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] Image pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 894.382924] env[63515]: DEBUG nova.virt.hardware [None req-ceee2529-e58a-42aa-a26e-32b1ce124ccb tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 894.383212] env[63515]: DEBUG nova.virt.hardware [None req-ceee2529-e58a-42aa-a26e-32b1ce124ccb tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 894.383321] env[63515]: DEBUG nova.virt.hardware [None req-ceee2529-e58a-42aa-a26e-32b1ce124ccb tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 894.383497] env[63515]: DEBUG nova.virt.hardware [None req-ceee2529-e58a-42aa-a26e-32b1ce124ccb tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] Got 1 possible topologies {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 894.383664] env[63515]: DEBUG nova.virt.hardware [None req-ceee2529-e58a-42aa-a26e-32b1ce124ccb tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 894.384195] env[63515]: DEBUG nova.virt.hardware [None req-ceee2529-e58a-42aa-a26e-32b1ce124ccb tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 894.384751] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-458cf839-321b-4a62-953a-fda2540992bf {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.393874] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8888cbf-2e32-4ab5-b4df-286fd2877f1a {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.416512] env[63515]: DEBUG oslo_vmware.api [None req-dbe24c17-88c2-40e6-987c-d70326fd4bd3 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]525bd207-3d69-e145-7353-5504bc7705e7, 'name': SearchDatastore_Task, 'duration_secs': 0.016383} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 894.416850] env[63515]: DEBUG oslo_concurrency.lockutils [None req-dbe24c17-88c2-40e6-987c-d70326fd4bd3 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 894.417147] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-dbe24c17-88c2-40e6-987c-d70326fd4bd3 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk to [datastore1] e13da90c-28e6-43d2-99b7-19c5095954ca/e13da90c-28e6-43d2-99b7-19c5095954ca.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 894.417429] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-9aaee480-c5d2-457b-91af-9bd946a542c7 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.426396] env[63515]: DEBUG oslo_vmware.api [None req-dbe24c17-88c2-40e6-987c-d70326fd4bd3 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Waiting for the task: (returnval){ [ 894.426396] env[63515]: value = "task-1111504" [ 894.426396] env[63515]: _type = "Task" [ 894.426396] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 894.433234] env[63515]: DEBUG oslo_vmware.api [None req-dbe24c17-88c2-40e6-987c-d70326fd4bd3 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Task: {'id': task-1111504, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.491397] env[63515]: DEBUG oslo_vmware.api [None req-ef0f5ab2-1246-4de1-af39-f1efd470cd76 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Task: {'id': task-1111503, 'name': CloneVM_Task} progress is 93%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.547977] env[63515]: DEBUG oslo_vmware.api [None req-9a80c171-6c9b-44ae-ab0a-573a8347fcf4 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Task: {'id': task-1111502, 'name': CopyVirtualDisk_Task} progress is 43%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.749602] env[63515]: DEBUG oslo_concurrency.lockutils [None req-cf46911e-9c80-4bd3-a167-7e0168ef6827 tempest-ServerDiagnosticsV248Test-1736022867 tempest-ServerDiagnosticsV248Test-1736022867-project-member] Lock "705b7e9b-3ab6-40ec-8cf6-d86bf0a1faf0" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 15.693s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 894.939603] env[63515]: DEBUG oslo_vmware.api [None req-dbe24c17-88c2-40e6-987c-d70326fd4bd3 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Task: {'id': task-1111504, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.997996] env[63515]: DEBUG oslo_vmware.api [None req-ef0f5ab2-1246-4de1-af39-f1efd470cd76 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Task: {'id': task-1111503, 'name': CloneVM_Task} progress is 94%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.019289] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dec9d7b5-5cc6-45f2-8d78-56478f03fa15 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.028537] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16fc842c-06dd-4c05-b736-8b73d5e52a56 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.071569] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2b96646-5992-4830-a3cc-50502a9d7835 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.084729] env[63515]: DEBUG oslo_vmware.api [None req-9a80c171-6c9b-44ae-ab0a-573a8347fcf4 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Task: {'id': task-1111502, 'name': CopyVirtualDisk_Task} progress is 66%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.088401] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c617129b-ae83-48c9-a24c-40e736833967 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.105154] env[63515]: DEBUG nova.compute.provider_tree [None req-1d02c1b8-b71d-4e69-b38f-4c4377a39898 tempest-InstanceActionsTestJSON-247065106 tempest-InstanceActionsTestJSON-247065106-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 895.127831] env[63515]: DEBUG nova.compute.manager [req-b2d3c5ce-027e-4946-b2f2-2e6127ca33a4 req-62e95ea2-d708-45cd-b2e3-374a1778c210 service nova] [instance: 324f7e89-3c17-474d-af1b-6d7a6d042510] Received event network-vif-plugged-0ea22959-737c-478d-84d8-17fbb175b186 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 895.128614] env[63515]: DEBUG oslo_concurrency.lockutils [req-b2d3c5ce-027e-4946-b2f2-2e6127ca33a4 req-62e95ea2-d708-45cd-b2e3-374a1778c210 service nova] Acquiring lock "324f7e89-3c17-474d-af1b-6d7a6d042510-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 895.128976] env[63515]: DEBUG oslo_concurrency.lockutils [req-b2d3c5ce-027e-4946-b2f2-2e6127ca33a4 req-62e95ea2-d708-45cd-b2e3-374a1778c210 service nova] Lock "324f7e89-3c17-474d-af1b-6d7a6d042510-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 895.129286] env[63515]: DEBUG oslo_concurrency.lockutils [req-b2d3c5ce-027e-4946-b2f2-2e6127ca33a4 req-62e95ea2-d708-45cd-b2e3-374a1778c210 service nova] Lock "324f7e89-3c17-474d-af1b-6d7a6d042510-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 895.129494] env[63515]: DEBUG nova.compute.manager [req-b2d3c5ce-027e-4946-b2f2-2e6127ca33a4 req-62e95ea2-d708-45cd-b2e3-374a1778c210 service nova] [instance: 324f7e89-3c17-474d-af1b-6d7a6d042510] No waiting events found dispatching network-vif-plugged-0ea22959-737c-478d-84d8-17fbb175b186 {{(pid=63515) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 895.129678] env[63515]: WARNING nova.compute.manager [req-b2d3c5ce-027e-4946-b2f2-2e6127ca33a4 req-62e95ea2-d708-45cd-b2e3-374a1778c210 service nova] [instance: 324f7e89-3c17-474d-af1b-6d7a6d042510] Received unexpected event network-vif-plugged-0ea22959-737c-478d-84d8-17fbb175b186 for instance with vm_state building and task_state spawning. [ 895.257839] env[63515]: DEBUG nova.network.neutron [None req-ceee2529-e58a-42aa-a26e-32b1ce124ccb tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] [instance: 324f7e89-3c17-474d-af1b-6d7a6d042510] Successfully updated port: 0ea22959-737c-478d-84d8-17fbb175b186 {{(pid=63515) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 895.440109] env[63515]: DEBUG oslo_vmware.api [None req-dbe24c17-88c2-40e6-987c-d70326fd4bd3 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Task: {'id': task-1111504, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.738} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 895.440381] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-dbe24c17-88c2-40e6-987c-d70326fd4bd3 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk to [datastore1] e13da90c-28e6-43d2-99b7-19c5095954ca/e13da90c-28e6-43d2-99b7-19c5095954ca.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 895.440548] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-dbe24c17-88c2-40e6-987c-d70326fd4bd3 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: e13da90c-28e6-43d2-99b7-19c5095954ca] Extending root virtual disk to 1048576 {{(pid=63515) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 895.440823] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ec9ec1ef-bc9c-4c72-abef-41ad421e929e {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.449477] env[63515]: DEBUG oslo_vmware.api [None req-dbe24c17-88c2-40e6-987c-d70326fd4bd3 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Waiting for the task: (returnval){ [ 895.449477] env[63515]: value = "task-1111505" [ 895.449477] env[63515]: _type = "Task" [ 895.449477] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 895.464463] env[63515]: DEBUG oslo_vmware.api [None req-dbe24c17-88c2-40e6-987c-d70326fd4bd3 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Task: {'id': task-1111505, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.496849] env[63515]: DEBUG oslo_vmware.api [None req-ef0f5ab2-1246-4de1-af39-f1efd470cd76 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Task: {'id': task-1111503, 'name': CloneVM_Task} progress is 94%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.579428] env[63515]: DEBUG oslo_vmware.api [None req-9a80c171-6c9b-44ae-ab0a-573a8347fcf4 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Task: {'id': task-1111502, 'name': CopyVirtualDisk_Task} progress is 85%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.610449] env[63515]: DEBUG nova.scheduler.client.report [None req-1d02c1b8-b71d-4e69-b38f-4c4377a39898 tempest-InstanceActionsTestJSON-247065106 tempest-InstanceActionsTestJSON-247065106-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 895.962284] env[63515]: DEBUG oslo_vmware.api [None req-dbe24c17-88c2-40e6-987c-d70326fd4bd3 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Task: {'id': task-1111505, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.311547} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 895.963046] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-dbe24c17-88c2-40e6-987c-d70326fd4bd3 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: e13da90c-28e6-43d2-99b7-19c5095954ca] Extended root virtual disk {{(pid=63515) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 895.963340] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-470190d4-169a-4d6a-b0ff-ae1b55365372 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.986072] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-dbe24c17-88c2-40e6-987c-d70326fd4bd3 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: e13da90c-28e6-43d2-99b7-19c5095954ca] Reconfiguring VM instance instance-0000004e to attach disk [datastore1] e13da90c-28e6-43d2-99b7-19c5095954ca/e13da90c-28e6-43d2-99b7-19c5095954ca.vmdk or device None with type sparse {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 895.986467] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-767094a9-9ab8-4745-9b83-13013981a780 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.009706] env[63515]: DEBUG oslo_vmware.api [None req-ef0f5ab2-1246-4de1-af39-f1efd470cd76 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Task: {'id': task-1111503, 'name': CloneVM_Task} progress is 94%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.011199] env[63515]: DEBUG oslo_vmware.api [None req-dbe24c17-88c2-40e6-987c-d70326fd4bd3 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Waiting for the task: (returnval){ [ 896.011199] env[63515]: value = "task-1111506" [ 896.011199] env[63515]: _type = "Task" [ 896.011199] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 896.020429] env[63515]: DEBUG oslo_vmware.api [None req-dbe24c17-88c2-40e6-987c-d70326fd4bd3 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Task: {'id': task-1111506, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.074682] env[63515]: DEBUG oslo_vmware.api [None req-9a80c171-6c9b-44ae-ab0a-573a8347fcf4 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Task: {'id': task-1111502, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.448124} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 896.074966] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-9a80c171-6c9b-44ae-ab0a-573a8347fcf4 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d1918637-9f8a-4d89-9bc2-49e7d0fdaeee/d1918637-9f8a-4d89-9bc2-49e7d0fdaeee.vmdk to [datastore2] b1bf5a0a-42e7-41a6-90fb-a0c311623fdd/b1bf5a0a-42e7-41a6-90fb-a0c311623fdd.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 896.075863] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9efdfe9e-4c28-43ec-b884-7e97ee10ee55 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.099238] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-9a80c171-6c9b-44ae-ab0a-573a8347fcf4 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: b1bf5a0a-42e7-41a6-90fb-a0c311623fdd] Reconfiguring VM instance instance-0000004d to attach disk [datastore2] b1bf5a0a-42e7-41a6-90fb-a0c311623fdd/b1bf5a0a-42e7-41a6-90fb-a0c311623fdd.vmdk or device None with type streamOptimized {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 896.099238] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-df45645f-acb7-41c0-b7f2-92fe052aa318 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.116862] env[63515]: DEBUG oslo_concurrency.lockutils [None req-1d02c1b8-b71d-4e69-b38f-4c4377a39898 tempest-InstanceActionsTestJSON-247065106 tempest-InstanceActionsTestJSON-247065106-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.905s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 896.120326] env[63515]: DEBUG oslo_concurrency.lockutils [None req-3173d497-cfe4-4001-b365-7bedc6c45742 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 8.289s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 896.122351] env[63515]: INFO nova.compute.claims [None req-3173d497-cfe4-4001-b365-7bedc6c45742 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: 829e72d7-a56a-451c-b98a-94253e5900d5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 896.124671] env[63515]: DEBUG oslo_vmware.api [None req-9a80c171-6c9b-44ae-ab0a-573a8347fcf4 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Waiting for the task: (returnval){ [ 896.124671] env[63515]: value = "task-1111507" [ 896.124671] env[63515]: _type = "Task" [ 896.124671] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 896.134342] env[63515]: DEBUG oslo_vmware.api [None req-9a80c171-6c9b-44ae-ab0a-573a8347fcf4 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Task: {'id': task-1111507, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.152350] env[63515]: INFO nova.scheduler.client.report [None req-1d02c1b8-b71d-4e69-b38f-4c4377a39898 tempest-InstanceActionsTestJSON-247065106 tempest-InstanceActionsTestJSON-247065106-project-member] Deleted allocations for instance 0ee9cce2-94b9-4038-ab38-1b3f96a864bc [ 896.498174] env[63515]: DEBUG oslo_vmware.api [None req-ef0f5ab2-1246-4de1-af39-f1efd470cd76 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Task: {'id': task-1111503, 'name': CloneVM_Task} progress is 94%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.520773] env[63515]: DEBUG oslo_vmware.api [None req-dbe24c17-88c2-40e6-987c-d70326fd4bd3 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Task: {'id': task-1111506, 'name': ReconfigVM_Task, 'duration_secs': 0.276884} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 896.521106] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-dbe24c17-88c2-40e6-987c-d70326fd4bd3 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: e13da90c-28e6-43d2-99b7-19c5095954ca] Reconfigured VM instance instance-0000004e to attach disk [datastore1] e13da90c-28e6-43d2-99b7-19c5095954ca/e13da90c-28e6-43d2-99b7-19c5095954ca.vmdk or device None with type sparse {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 896.521748] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-51412186-0641-492a-8c1d-ecf833aa4ced {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.528207] env[63515]: DEBUG oslo_vmware.api [None req-dbe24c17-88c2-40e6-987c-d70326fd4bd3 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Waiting for the task: (returnval){ [ 896.528207] env[63515]: value = "task-1111508" [ 896.528207] env[63515]: _type = "Task" [ 896.528207] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 896.535543] env[63515]: DEBUG oslo_vmware.api [None req-dbe24c17-88c2-40e6-987c-d70326fd4bd3 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Task: {'id': task-1111508, 'name': Rename_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.638410] env[63515]: DEBUG oslo_vmware.api [None req-9a80c171-6c9b-44ae-ab0a-573a8347fcf4 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Task: {'id': task-1111507, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.667486] env[63515]: DEBUG oslo_concurrency.lockutils [None req-1d02c1b8-b71d-4e69-b38f-4c4377a39898 tempest-InstanceActionsTestJSON-247065106 tempest-InstanceActionsTestJSON-247065106-project-member] Lock "0ee9cce2-94b9-4038-ab38-1b3f96a864bc" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 14.538s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 896.996683] env[63515]: DEBUG oslo_vmware.api [None req-ef0f5ab2-1246-4de1-af39-f1efd470cd76 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Task: {'id': task-1111503, 'name': CloneVM_Task} progress is 95%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.039928] env[63515]: DEBUG oslo_vmware.api [None req-dbe24c17-88c2-40e6-987c-d70326fd4bd3 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Task: {'id': task-1111508, 'name': Rename_Task, 'duration_secs': 0.139647} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 897.040236] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-dbe24c17-88c2-40e6-987c-d70326fd4bd3 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: e13da90c-28e6-43d2-99b7-19c5095954ca] Powering on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 897.040477] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-763666fe-2daa-429a-966d-21dea1bfcc31 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.049221] env[63515]: DEBUG oslo_vmware.api [None req-dbe24c17-88c2-40e6-987c-d70326fd4bd3 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Waiting for the task: (returnval){ [ 897.049221] env[63515]: value = "task-1111509" [ 897.049221] env[63515]: _type = "Task" [ 897.049221] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.055059] env[63515]: DEBUG oslo_vmware.api [None req-dbe24c17-88c2-40e6-987c-d70326fd4bd3 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Task: {'id': task-1111509, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.137186] env[63515]: DEBUG oslo_vmware.api [None req-9a80c171-6c9b-44ae-ab0a-573a8347fcf4 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Task: {'id': task-1111507, 'name': ReconfigVM_Task, 'duration_secs': 0.684121} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 897.137681] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-9a80c171-6c9b-44ae-ab0a-573a8347fcf4 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: b1bf5a0a-42e7-41a6-90fb-a0c311623fdd] Reconfigured VM instance instance-0000004d to attach disk [datastore2] b1bf5a0a-42e7-41a6-90fb-a0c311623fdd/b1bf5a0a-42e7-41a6-90fb-a0c311623fdd.vmdk or device None with type streamOptimized {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 897.138492] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a147a9a6-2027-4541-8670-b85f6e361992 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.145014] env[63515]: DEBUG oslo_vmware.api [None req-9a80c171-6c9b-44ae-ab0a-573a8347fcf4 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Waiting for the task: (returnval){ [ 897.145014] env[63515]: value = "task-1111510" [ 897.145014] env[63515]: _type = "Task" [ 897.145014] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.153230] env[63515]: DEBUG oslo_vmware.api [None req-9a80c171-6c9b-44ae-ab0a-573a8347fcf4 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Task: {'id': task-1111510, 'name': Rename_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.173823] env[63515]: DEBUG nova.compute.manager [req-5e930bab-6422-497e-8ecd-74a516dd09f6 req-08e23b53-5915-4713-b818-13b30d2d4477 service nova] [instance: 324f7e89-3c17-474d-af1b-6d7a6d042510] Received event network-changed-0ea22959-737c-478d-84d8-17fbb175b186 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 897.174064] env[63515]: DEBUG nova.compute.manager [req-5e930bab-6422-497e-8ecd-74a516dd09f6 req-08e23b53-5915-4713-b818-13b30d2d4477 service nova] [instance: 324f7e89-3c17-474d-af1b-6d7a6d042510] Refreshing instance network info cache due to event network-changed-0ea22959-737c-478d-84d8-17fbb175b186. {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 897.174332] env[63515]: DEBUG oslo_concurrency.lockutils [req-5e930bab-6422-497e-8ecd-74a516dd09f6 req-08e23b53-5915-4713-b818-13b30d2d4477 service nova] Acquiring lock "refresh_cache-324f7e89-3c17-474d-af1b-6d7a6d042510" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 897.174520] env[63515]: DEBUG oslo_concurrency.lockutils [req-5e930bab-6422-497e-8ecd-74a516dd09f6 req-08e23b53-5915-4713-b818-13b30d2d4477 service nova] Acquired lock "refresh_cache-324f7e89-3c17-474d-af1b-6d7a6d042510" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 897.174726] env[63515]: DEBUG nova.network.neutron [req-5e930bab-6422-497e-8ecd-74a516dd09f6 req-08e23b53-5915-4713-b818-13b30d2d4477 service nova] [instance: 324f7e89-3c17-474d-af1b-6d7a6d042510] Refreshing network info cache for port 0ea22959-737c-478d-84d8-17fbb175b186 {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 897.416028] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37ebe6a1-c0be-4183-9202-e20e721c7a96 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.424718] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-622093c2-2d2f-43b7-8f48-bc0bba89bf2f {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.461213] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c236b6c-627f-4692-bd12-64afb84543ef {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.469437] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f80edb08-eb57-42ec-a436-d2617f8547d4 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.485943] env[63515]: DEBUG nova.compute.provider_tree [None req-3173d497-cfe4-4001-b365-7bedc6c45742 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Updating inventory in ProviderTree for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 897.499882] env[63515]: DEBUG oslo_vmware.api [None req-ef0f5ab2-1246-4de1-af39-f1efd470cd76 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Task: {'id': task-1111503, 'name': CloneVM_Task, 'duration_secs': 3.127401} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 897.500890] env[63515]: INFO nova.virt.vmwareapi.vmops [None req-ef0f5ab2-1246-4de1-af39-f1efd470cd76 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: ae5d6f34-bac6-4ba0-b0c3-8205982ddc88] Created linked-clone VM from snapshot [ 897.501682] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e64e6272-4bc8-4417-9ae1-ab2eda3da18f {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.511962] env[63515]: DEBUG nova.virt.vmwareapi.images [None req-ef0f5ab2-1246-4de1-af39-f1efd470cd76 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: ae5d6f34-bac6-4ba0-b0c3-8205982ddc88] Uploading image 9a616ff7-f282-45bd-9090-bf9b0dfb5f2d {{(pid=63515) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 897.543137] env[63515]: DEBUG oslo_vmware.rw_handles [None req-ef0f5ab2-1246-4de1-af39-f1efd470cd76 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 897.543137] env[63515]: value = "vm-243519" [ 897.543137] env[63515]: _type = "VirtualMachine" [ 897.543137] env[63515]: }. {{(pid=63515) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 897.543730] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-db792e85-9b9b-4396-902c-042f9e66252b {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.552345] env[63515]: DEBUG oslo_vmware.rw_handles [None req-ef0f5ab2-1246-4de1-af39-f1efd470cd76 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Lease: (returnval){ [ 897.552345] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]52933364-682d-5ade-8631-8b4848aff371" [ 897.552345] env[63515]: _type = "HttpNfcLease" [ 897.552345] env[63515]: } obtained for exporting VM: (result){ [ 897.552345] env[63515]: value = "vm-243519" [ 897.552345] env[63515]: _type = "VirtualMachine" [ 897.552345] env[63515]: }. {{(pid=63515) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 897.552627] env[63515]: DEBUG oslo_vmware.api [None req-ef0f5ab2-1246-4de1-af39-f1efd470cd76 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Waiting for the lease: (returnval){ [ 897.552627] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]52933364-682d-5ade-8631-8b4848aff371" [ 897.552627] env[63515]: _type = "HttpNfcLease" [ 897.552627] env[63515]: } to be ready. {{(pid=63515) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 897.559032] env[63515]: DEBUG oslo_vmware.api [None req-dbe24c17-88c2-40e6-987c-d70326fd4bd3 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Task: {'id': task-1111509, 'name': PowerOnVM_Task, 'duration_secs': 0.471323} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 897.559635] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-dbe24c17-88c2-40e6-987c-d70326fd4bd3 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: e13da90c-28e6-43d2-99b7-19c5095954ca] Powered on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 897.559854] env[63515]: INFO nova.compute.manager [None req-dbe24c17-88c2-40e6-987c-d70326fd4bd3 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: e13da90c-28e6-43d2-99b7-19c5095954ca] Took 7.61 seconds to spawn the instance on the hypervisor. [ 897.560076] env[63515]: DEBUG nova.compute.manager [None req-dbe24c17-88c2-40e6-987c-d70326fd4bd3 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: e13da90c-28e6-43d2-99b7-19c5095954ca] Checking state {{(pid=63515) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 897.560828] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7133121-ac97-43ad-9879-e030774d093c {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.564311] env[63515]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 897.564311] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]52933364-682d-5ade-8631-8b4848aff371" [ 897.564311] env[63515]: _type = "HttpNfcLease" [ 897.564311] env[63515]: } is ready. {{(pid=63515) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 897.564862] env[63515]: DEBUG oslo_vmware.rw_handles [None req-ef0f5ab2-1246-4de1-af39-f1efd470cd76 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 897.564862] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]52933364-682d-5ade-8631-8b4848aff371" [ 897.564862] env[63515]: _type = "HttpNfcLease" [ 897.564862] env[63515]: }. {{(pid=63515) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 897.565592] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-316d4aba-6579-41c4-b3ce-a8e686f7e9a7 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.576547] env[63515]: DEBUG oslo_vmware.rw_handles [None req-ef0f5ab2-1246-4de1-af39-f1efd470cd76 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5223dc08-c895-6d10-cbda-75409b05588d/disk-0.vmdk from lease info. {{(pid=63515) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 897.576728] env[63515]: DEBUG oslo_vmware.rw_handles [None req-ef0f5ab2-1246-4de1-af39-f1efd470cd76 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5223dc08-c895-6d10-cbda-75409b05588d/disk-0.vmdk for reading. {{(pid=63515) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 897.654437] env[63515]: DEBUG oslo_vmware.api [None req-9a80c171-6c9b-44ae-ab0a-573a8347fcf4 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Task: {'id': task-1111510, 'name': Rename_Task, 'duration_secs': 0.144233} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 897.654732] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-9a80c171-6c9b-44ae-ab0a-573a8347fcf4 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: b1bf5a0a-42e7-41a6-90fb-a0c311623fdd] Powering on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 897.654998] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-af5b6932-d9ec-4fa1-8fb2-8980a6f3880f {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.660953] env[63515]: DEBUG oslo_vmware.api [None req-9a80c171-6c9b-44ae-ab0a-573a8347fcf4 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Waiting for the task: (returnval){ [ 897.660953] env[63515]: value = "task-1111512" [ 897.660953] env[63515]: _type = "Task" [ 897.660953] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.668926] env[63515]: DEBUG oslo_vmware.api [None req-9a80c171-6c9b-44ae-ab0a-573a8347fcf4 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Task: {'id': task-1111512, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.672106] env[63515]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-915dabcc-6aa9-4946-8927-e11b60965a33 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.713552] env[63515]: DEBUG nova.network.neutron [req-5e930bab-6422-497e-8ecd-74a516dd09f6 req-08e23b53-5915-4713-b818-13b30d2d4477 service nova] [instance: 324f7e89-3c17-474d-af1b-6d7a6d042510] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 897.804421] env[63515]: DEBUG nova.network.neutron [req-5e930bab-6422-497e-8ecd-74a516dd09f6 req-08e23b53-5915-4713-b818-13b30d2d4477 service nova] [instance: 324f7e89-3c17-474d-af1b-6d7a6d042510] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 898.021418] env[63515]: ERROR nova.scheduler.client.report [None req-3173d497-cfe4-4001-b365-7bedc6c45742 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [req-143df268-7b47-4e6e-b81f-1b48093fccb2] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 2bd86232-2b6f-44d5-9057-1a3a6b27185a. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-143df268-7b47-4e6e-b81f-1b48093fccb2"}]} [ 898.039346] env[63515]: DEBUG nova.scheduler.client.report [None req-3173d497-cfe4-4001-b365-7bedc6c45742 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Refreshing inventories for resource provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 898.062354] env[63515]: DEBUG nova.scheduler.client.report [None req-3173d497-cfe4-4001-b365-7bedc6c45742 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Updating ProviderTree inventory for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 898.062924] env[63515]: DEBUG nova.compute.provider_tree [None req-3173d497-cfe4-4001-b365-7bedc6c45742 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Updating inventory in ProviderTree for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 898.083014] env[63515]: DEBUG nova.scheduler.client.report [None req-3173d497-cfe4-4001-b365-7bedc6c45742 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Refreshing aggregate associations for resource provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a, aggregates: None {{(pid=63515) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 898.089985] env[63515]: INFO nova.compute.manager [None req-dbe24c17-88c2-40e6-987c-d70326fd4bd3 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: e13da90c-28e6-43d2-99b7-19c5095954ca] Took 21.38 seconds to build instance. [ 898.112961] env[63515]: DEBUG nova.scheduler.client.report [None req-3173d497-cfe4-4001-b365-7bedc6c45742 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Refreshing trait associations for resource provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a, traits: COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=63515) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 898.139830] env[63515]: DEBUG nova.network.neutron [None req-ceee2529-e58a-42aa-a26e-32b1ce124ccb tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] [instance: 324f7e89-3c17-474d-af1b-6d7a6d042510] Successfully updated port: c18e8fdc-6f1f-4fd6-8401-cab93bb67126 {{(pid=63515) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 898.175379] env[63515]: DEBUG oslo_vmware.api [None req-9a80c171-6c9b-44ae-ab0a-573a8347fcf4 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Task: {'id': task-1111512, 'name': PowerOnVM_Task} progress is 88%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.307462] env[63515]: DEBUG oslo_concurrency.lockutils [req-5e930bab-6422-497e-8ecd-74a516dd09f6 req-08e23b53-5915-4713-b818-13b30d2d4477 service nova] Releasing lock "refresh_cache-324f7e89-3c17-474d-af1b-6d7a6d042510" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 898.441405] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c8c791a-cf05-40be-8f93-2cace7b7f6bc {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.452208] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-055bef7b-5f79-4392-8ad4-0100fd4c6873 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.493049] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bdc860d2-8a09-4ea6-b15b-c22b719691e6 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.503155] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d24f226-7ce2-47ec-b005-31b227c5a35e {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.520031] env[63515]: DEBUG nova.compute.provider_tree [None req-3173d497-cfe4-4001-b365-7bedc6c45742 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Updating inventory in ProviderTree for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 898.593076] env[63515]: DEBUG oslo_concurrency.lockutils [None req-dbe24c17-88c2-40e6-987c-d70326fd4bd3 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Lock "e13da90c-28e6-43d2-99b7-19c5095954ca" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 22.906s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 898.647097] env[63515]: DEBUG oslo_concurrency.lockutils [None req-ceee2529-e58a-42aa-a26e-32b1ce124ccb tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] Acquiring lock "refresh_cache-324f7e89-3c17-474d-af1b-6d7a6d042510" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 898.647343] env[63515]: DEBUG oslo_concurrency.lockutils [None req-ceee2529-e58a-42aa-a26e-32b1ce124ccb tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] Acquired lock "refresh_cache-324f7e89-3c17-474d-af1b-6d7a6d042510" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 898.647482] env[63515]: DEBUG nova.network.neutron [None req-ceee2529-e58a-42aa-a26e-32b1ce124ccb tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] [instance: 324f7e89-3c17-474d-af1b-6d7a6d042510] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 898.676994] env[63515]: DEBUG oslo_vmware.api [None req-9a80c171-6c9b-44ae-ab0a-573a8347fcf4 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Task: {'id': task-1111512, 'name': PowerOnVM_Task, 'duration_secs': 0.670369} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 898.678091] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-9a80c171-6c9b-44ae-ab0a-573a8347fcf4 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: b1bf5a0a-42e7-41a6-90fb-a0c311623fdd] Powered on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 898.678317] env[63515]: INFO nova.compute.manager [None req-9a80c171-6c9b-44ae-ab0a-573a8347fcf4 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: b1bf5a0a-42e7-41a6-90fb-a0c311623fdd] Took 15.61 seconds to spawn the instance on the hypervisor. [ 898.678527] env[63515]: DEBUG nova.compute.manager [None req-9a80c171-6c9b-44ae-ab0a-573a8347fcf4 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: b1bf5a0a-42e7-41a6-90fb-a0c311623fdd] Checking state {{(pid=63515) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 898.680544] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-338dbfa0-4755-4947-aea0-d98d6ce52a1a {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.079812] env[63515]: DEBUG nova.scheduler.client.report [None req-3173d497-cfe4-4001-b365-7bedc6c45742 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Updated inventory for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a with generation 106 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 899.082362] env[63515]: DEBUG nova.compute.provider_tree [None req-3173d497-cfe4-4001-b365-7bedc6c45742 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Updating resource provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a generation from 106 to 107 during operation: update_inventory {{(pid=63515) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 899.082362] env[63515]: DEBUG nova.compute.provider_tree [None req-3173d497-cfe4-4001-b365-7bedc6c45742 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Updating inventory in ProviderTree for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 899.205741] env[63515]: DEBUG nova.compute.manager [req-fd3d0e68-bde8-447c-8043-c21c7cf7008d req-3794daf3-f198-4b4f-bed0-b54a9900c503 service nova] [instance: 324f7e89-3c17-474d-af1b-6d7a6d042510] Received event network-vif-plugged-c18e8fdc-6f1f-4fd6-8401-cab93bb67126 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 899.205741] env[63515]: DEBUG oslo_concurrency.lockutils [req-fd3d0e68-bde8-447c-8043-c21c7cf7008d req-3794daf3-f198-4b4f-bed0-b54a9900c503 service nova] Acquiring lock "324f7e89-3c17-474d-af1b-6d7a6d042510-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 899.205741] env[63515]: DEBUG oslo_concurrency.lockutils [req-fd3d0e68-bde8-447c-8043-c21c7cf7008d req-3794daf3-f198-4b4f-bed0-b54a9900c503 service nova] Lock "324f7e89-3c17-474d-af1b-6d7a6d042510-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 899.205741] env[63515]: DEBUG oslo_concurrency.lockutils [req-fd3d0e68-bde8-447c-8043-c21c7cf7008d req-3794daf3-f198-4b4f-bed0-b54a9900c503 service nova] Lock "324f7e89-3c17-474d-af1b-6d7a6d042510-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 899.205741] env[63515]: DEBUG nova.compute.manager [req-fd3d0e68-bde8-447c-8043-c21c7cf7008d req-3794daf3-f198-4b4f-bed0-b54a9900c503 service nova] [instance: 324f7e89-3c17-474d-af1b-6d7a6d042510] No waiting events found dispatching network-vif-plugged-c18e8fdc-6f1f-4fd6-8401-cab93bb67126 {{(pid=63515) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 899.205741] env[63515]: WARNING nova.compute.manager [req-fd3d0e68-bde8-447c-8043-c21c7cf7008d req-3794daf3-f198-4b4f-bed0-b54a9900c503 service nova] [instance: 324f7e89-3c17-474d-af1b-6d7a6d042510] Received unexpected event network-vif-plugged-c18e8fdc-6f1f-4fd6-8401-cab93bb67126 for instance with vm_state building and task_state spawning. [ 899.205741] env[63515]: DEBUG nova.compute.manager [req-fd3d0e68-bde8-447c-8043-c21c7cf7008d req-3794daf3-f198-4b4f-bed0-b54a9900c503 service nova] [instance: 324f7e89-3c17-474d-af1b-6d7a6d042510] Received event network-changed-c18e8fdc-6f1f-4fd6-8401-cab93bb67126 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 899.205741] env[63515]: DEBUG nova.compute.manager [req-fd3d0e68-bde8-447c-8043-c21c7cf7008d req-3794daf3-f198-4b4f-bed0-b54a9900c503 service nova] [instance: 324f7e89-3c17-474d-af1b-6d7a6d042510] Refreshing instance network info cache due to event network-changed-c18e8fdc-6f1f-4fd6-8401-cab93bb67126. {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 899.205741] env[63515]: DEBUG oslo_concurrency.lockutils [req-fd3d0e68-bde8-447c-8043-c21c7cf7008d req-3794daf3-f198-4b4f-bed0-b54a9900c503 service nova] Acquiring lock "refresh_cache-324f7e89-3c17-474d-af1b-6d7a6d042510" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 899.206853] env[63515]: INFO nova.compute.manager [None req-9a80c171-6c9b-44ae-ab0a-573a8347fcf4 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: b1bf5a0a-42e7-41a6-90fb-a0c311623fdd] Took 39.93 seconds to build instance. [ 899.213525] env[63515]: DEBUG nova.network.neutron [None req-ceee2529-e58a-42aa-a26e-32b1ce124ccb tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] [instance: 324f7e89-3c17-474d-af1b-6d7a6d042510] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 899.589818] env[63515]: DEBUG oslo_concurrency.lockutils [None req-3173d497-cfe4-4001-b365-7bedc6c45742 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.469s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 899.590451] env[63515]: DEBUG nova.compute.manager [None req-3173d497-cfe4-4001-b365-7bedc6c45742 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: 829e72d7-a56a-451c-b98a-94253e5900d5] Start building networks asynchronously for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 899.596022] env[63515]: DEBUG oslo_concurrency.lockutils [None req-b4e6fa31-2603-4190-94fe-f49ca8c1e386 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 10.403s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 899.596022] env[63515]: DEBUG oslo_concurrency.lockutils [None req-b4e6fa31-2603-4190-94fe-f49ca8c1e386 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 899.601170] env[63515]: DEBUG oslo_concurrency.lockutils [None req-07515409-459b-4c3e-992f-c4e070f0ae85 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 9.314s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 899.601170] env[63515]: DEBUG nova.objects.instance [None req-07515409-459b-4c3e-992f-c4e070f0ae85 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Lazy-loading 'resources' on Instance uuid 0967283f-8dda-4692-b6b7-601545cabf2f {{(pid=63515) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 899.644017] env[63515]: INFO nova.scheduler.client.report [None req-b4e6fa31-2603-4190-94fe-f49ca8c1e386 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Deleted allocations for instance d926c699-a64a-4942-9ef4-f0166414661d [ 899.662443] env[63515]: DEBUG nova.network.neutron [None req-ceee2529-e58a-42aa-a26e-32b1ce124ccb tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] [instance: 324f7e89-3c17-474d-af1b-6d7a6d042510] Updating instance_info_cache with network_info: [{"id": "0ea22959-737c-478d-84d8-17fbb175b186", "address": "fa:16:3e:62:f2:c0", "network": {"id": "a3055a87-44b4-4b02-ae7d-87eb877088ba", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-343503251", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.37", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "26a52fb1944a4f949f6779829e163eaa", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "86a35d07-53d3-46b3-92cb-ae34236c0f41", "external-id": "nsx-vlan-transportzone-811", "segmentation_id": 811, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0ea22959-73", "ovs_interfaceid": "0ea22959-737c-478d-84d8-17fbb175b186", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "c18e8fdc-6f1f-4fd6-8401-cab93bb67126", "address": "fa:16:3e:2c:ac:b5", "network": {"id": "a54ddb5c-3376-40aa-aa48-055659aba84f", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1564152417", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.23", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "26a52fb1944a4f949f6779829e163eaa", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5116f690-f825-4fee-8a47-42b073e716c5", "external-id": "nsx-vlan-transportzone-692", "segmentation_id": 692, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc18e8fdc-6f", "ovs_interfaceid": "c18e8fdc-6f1f-4fd6-8401-cab93bb67126", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 899.712165] env[63515]: DEBUG oslo_concurrency.lockutils [None req-9a80c171-6c9b-44ae-ab0a-573a8347fcf4 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Lock "b1bf5a0a-42e7-41a6-90fb-a0c311623fdd" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 41.450s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 900.104411] env[63515]: DEBUG nova.compute.utils [None req-3173d497-cfe4-4001-b365-7bedc6c45742 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Using /dev/sd instead of None {{(pid=63515) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 900.110070] env[63515]: DEBUG nova.compute.manager [None req-3173d497-cfe4-4001-b365-7bedc6c45742 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: 829e72d7-a56a-451c-b98a-94253e5900d5] Allocating IP information in the background. {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 900.110220] env[63515]: DEBUG nova.network.neutron [None req-3173d497-cfe4-4001-b365-7bedc6c45742 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: 829e72d7-a56a-451c-b98a-94253e5900d5] allocate_for_instance() {{(pid=63515) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 900.155821] env[63515]: DEBUG oslo_concurrency.lockutils [None req-b4e6fa31-2603-4190-94fe-f49ca8c1e386 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Lock "d926c699-a64a-4942-9ef4-f0166414661d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 13.902s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 900.166288] env[63515]: DEBUG oslo_concurrency.lockutils [None req-ceee2529-e58a-42aa-a26e-32b1ce124ccb tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] Releasing lock "refresh_cache-324f7e89-3c17-474d-af1b-6d7a6d042510" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 900.166490] env[63515]: DEBUG nova.compute.manager [None req-ceee2529-e58a-42aa-a26e-32b1ce124ccb tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] [instance: 324f7e89-3c17-474d-af1b-6d7a6d042510] Instance network_info: |[{"id": "0ea22959-737c-478d-84d8-17fbb175b186", "address": "fa:16:3e:62:f2:c0", "network": {"id": "a3055a87-44b4-4b02-ae7d-87eb877088ba", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-343503251", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.37", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "26a52fb1944a4f949f6779829e163eaa", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "86a35d07-53d3-46b3-92cb-ae34236c0f41", "external-id": "nsx-vlan-transportzone-811", "segmentation_id": 811, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0ea22959-73", "ovs_interfaceid": "0ea22959-737c-478d-84d8-17fbb175b186", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "c18e8fdc-6f1f-4fd6-8401-cab93bb67126", "address": "fa:16:3e:2c:ac:b5", "network": {"id": "a54ddb5c-3376-40aa-aa48-055659aba84f", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1564152417", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.23", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "26a52fb1944a4f949f6779829e163eaa", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5116f690-f825-4fee-8a47-42b073e716c5", "external-id": "nsx-vlan-transportzone-692", "segmentation_id": 692, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc18e8fdc-6f", "ovs_interfaceid": "c18e8fdc-6f1f-4fd6-8401-cab93bb67126", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 900.166982] env[63515]: DEBUG oslo_concurrency.lockutils [req-fd3d0e68-bde8-447c-8043-c21c7cf7008d req-3794daf3-f198-4b4f-bed0-b54a9900c503 service nova] Acquired lock "refresh_cache-324f7e89-3c17-474d-af1b-6d7a6d042510" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 900.166982] env[63515]: DEBUG nova.network.neutron [req-fd3d0e68-bde8-447c-8043-c21c7cf7008d req-3794daf3-f198-4b4f-bed0-b54a9900c503 service nova] [instance: 324f7e89-3c17-474d-af1b-6d7a6d042510] Refreshing network info cache for port c18e8fdc-6f1f-4fd6-8401-cab93bb67126 {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 900.168316] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-ceee2529-e58a-42aa-a26e-32b1ce124ccb tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] [instance: 324f7e89-3c17-474d-af1b-6d7a6d042510] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:62:f2:c0', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '86a35d07-53d3-46b3-92cb-ae34236c0f41', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '0ea22959-737c-478d-84d8-17fbb175b186', 'vif_model': 'vmxnet3'}, {'network_name': 'br-int', 'mac_address': 'fa:16:3e:2c:ac:b5', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '5116f690-f825-4fee-8a47-42b073e716c5', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c18e8fdc-6f1f-4fd6-8401-cab93bb67126', 'vif_model': 'vmxnet3'}] {{(pid=63515) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 900.188527] env[63515]: DEBUG oslo.service.loopingcall [None req-ceee2529-e58a-42aa-a26e-32b1ce124ccb tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 900.190145] env[63515]: DEBUG nova.policy [None req-3173d497-cfe4-4001-b365-7bedc6c45742 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b5de3e36ae88482eb795894592e76c7c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f9c98e9f6020475490aaa7e76d907ab7', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63515) authorize /opt/stack/nova/nova/policy.py:201}} [ 900.197463] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 324f7e89-3c17-474d-af1b-6d7a6d042510] Creating VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 900.198275] env[63515]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-7dc73e1f-1681-4ee8-a4a3-561fa1746d25 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.225532] env[63515]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 900.225532] env[63515]: value = "task-1111513" [ 900.225532] env[63515]: _type = "Task" [ 900.225532] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 900.243988] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1111513, 'name': CreateVM_Task} progress is 6%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.373322] env[63515]: DEBUG nova.compute.manager [req-99d7b96c-2344-44b7-bd74-68cd6622fb83 req-4ab4dce3-762a-4ef5-b515-4ed28d8837ef service nova] [instance: e13da90c-28e6-43d2-99b7-19c5095954ca] Received event network-changed-d9363b80-630e-4262-901f-c13e5c8b8c2a {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 900.373702] env[63515]: DEBUG nova.compute.manager [req-99d7b96c-2344-44b7-bd74-68cd6622fb83 req-4ab4dce3-762a-4ef5-b515-4ed28d8837ef service nova] [instance: e13da90c-28e6-43d2-99b7-19c5095954ca] Refreshing instance network info cache due to event network-changed-d9363b80-630e-4262-901f-c13e5c8b8c2a. {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 900.375332] env[63515]: DEBUG oslo_concurrency.lockutils [req-99d7b96c-2344-44b7-bd74-68cd6622fb83 req-4ab4dce3-762a-4ef5-b515-4ed28d8837ef service nova] Acquiring lock "refresh_cache-e13da90c-28e6-43d2-99b7-19c5095954ca" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 900.375332] env[63515]: DEBUG oslo_concurrency.lockutils [req-99d7b96c-2344-44b7-bd74-68cd6622fb83 req-4ab4dce3-762a-4ef5-b515-4ed28d8837ef service nova] Acquired lock "refresh_cache-e13da90c-28e6-43d2-99b7-19c5095954ca" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 900.375332] env[63515]: DEBUG nova.network.neutron [req-99d7b96c-2344-44b7-bd74-68cd6622fb83 req-4ab4dce3-762a-4ef5-b515-4ed28d8837ef service nova] [instance: e13da90c-28e6-43d2-99b7-19c5095954ca] Refreshing network info cache for port d9363b80-630e-4262-901f-c13e5c8b8c2a {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 900.526433] env[63515]: DEBUG oslo_concurrency.lockutils [None req-8c6d15be-1f5e-44a9-a3e4-dab9aa82e26a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Acquiring lock "9236607d-47af-4272-93db-79826e918aa1" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 900.526672] env[63515]: DEBUG oslo_concurrency.lockutils [None req-8c6d15be-1f5e-44a9-a3e4-dab9aa82e26a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Lock "9236607d-47af-4272-93db-79826e918aa1" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 900.533153] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-533493e9-5909-4253-bc76-49b7f3b12580 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.542048] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b49199b3-13c0-469b-9a18-875bf1a01fc3 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.596561] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97fe10c6-9552-4ac1-81f0-16e75cf29e2b {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.600974] env[63515]: DEBUG oslo_concurrency.lockutils [None req-8c6d15be-1f5e-44a9-a3e4-dab9aa82e26a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Acquiring lock "ca6a8ac3-d880-4333-9816-2f076e423844" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 900.601656] env[63515]: DEBUG oslo_concurrency.lockutils [None req-8c6d15be-1f5e-44a9-a3e4-dab9aa82e26a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Lock "ca6a8ac3-d880-4333-9816-2f076e423844" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 900.609552] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-174da82d-7e7f-405a-aff7-8fa9fc403045 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.615164] env[63515]: DEBUG nova.compute.manager [None req-3173d497-cfe4-4001-b365-7bedc6c45742 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: 829e72d7-a56a-451c-b98a-94253e5900d5] Start building block device mappings for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 900.633853] env[63515]: DEBUG nova.compute.provider_tree [None req-07515409-459b-4c3e-992f-c4e070f0ae85 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 900.660023] env[63515]: DEBUG nova.network.neutron [req-fd3d0e68-bde8-447c-8043-c21c7cf7008d req-3794daf3-f198-4b4f-bed0-b54a9900c503 service nova] [instance: 324f7e89-3c17-474d-af1b-6d7a6d042510] Updated VIF entry in instance network info cache for port c18e8fdc-6f1f-4fd6-8401-cab93bb67126. {{(pid=63515) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 900.660023] env[63515]: DEBUG nova.network.neutron [req-fd3d0e68-bde8-447c-8043-c21c7cf7008d req-3794daf3-f198-4b4f-bed0-b54a9900c503 service nova] [instance: 324f7e89-3c17-474d-af1b-6d7a6d042510] Updating instance_info_cache with network_info: [{"id": "0ea22959-737c-478d-84d8-17fbb175b186", "address": "fa:16:3e:62:f2:c0", "network": {"id": "a3055a87-44b4-4b02-ae7d-87eb877088ba", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-343503251", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.37", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "26a52fb1944a4f949f6779829e163eaa", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "86a35d07-53d3-46b3-92cb-ae34236c0f41", "external-id": "nsx-vlan-transportzone-811", "segmentation_id": 811, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0ea22959-73", "ovs_interfaceid": "0ea22959-737c-478d-84d8-17fbb175b186", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "c18e8fdc-6f1f-4fd6-8401-cab93bb67126", "address": "fa:16:3e:2c:ac:b5", "network": {"id": "a54ddb5c-3376-40aa-aa48-055659aba84f", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1564152417", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.23", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "26a52fb1944a4f949f6779829e163eaa", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5116f690-f825-4fee-8a47-42b073e716c5", "external-id": "nsx-vlan-transportzone-692", "segmentation_id": 692, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc18e8fdc-6f", "ovs_interfaceid": "c18e8fdc-6f1f-4fd6-8401-cab93bb67126", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 900.684625] env[63515]: DEBUG nova.network.neutron [None req-3173d497-cfe4-4001-b365-7bedc6c45742 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: 829e72d7-a56a-451c-b98a-94253e5900d5] Successfully created port: 856de677-b1c8-479b-91d9-34d31a703a63 {{(pid=63515) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 900.737397] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1111513, 'name': CreateVM_Task} progress is 99%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.831387] env[63515]: DEBUG oslo_concurrency.lockutils [None req-1e59a109-1efb-4ebd-8738-fea0e8fa8e8b tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Acquiring lock "b1bf5a0a-42e7-41a6-90fb-a0c311623fdd" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 900.831695] env[63515]: DEBUG oslo_concurrency.lockutils [None req-1e59a109-1efb-4ebd-8738-fea0e8fa8e8b tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Lock "b1bf5a0a-42e7-41a6-90fb-a0c311623fdd" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 900.831917] env[63515]: DEBUG oslo_concurrency.lockutils [None req-1e59a109-1efb-4ebd-8738-fea0e8fa8e8b tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Acquiring lock "b1bf5a0a-42e7-41a6-90fb-a0c311623fdd-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 900.832133] env[63515]: DEBUG oslo_concurrency.lockutils [None req-1e59a109-1efb-4ebd-8738-fea0e8fa8e8b tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Lock "b1bf5a0a-42e7-41a6-90fb-a0c311623fdd-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 900.832465] env[63515]: DEBUG oslo_concurrency.lockutils [None req-1e59a109-1efb-4ebd-8738-fea0e8fa8e8b tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Lock "b1bf5a0a-42e7-41a6-90fb-a0c311623fdd-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 900.834799] env[63515]: INFO nova.compute.manager [None req-1e59a109-1efb-4ebd-8738-fea0e8fa8e8b tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: b1bf5a0a-42e7-41a6-90fb-a0c311623fdd] Terminating instance [ 900.836968] env[63515]: DEBUG nova.compute.manager [None req-1e59a109-1efb-4ebd-8738-fea0e8fa8e8b tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: b1bf5a0a-42e7-41a6-90fb-a0c311623fdd] Start destroying the instance on the hypervisor. {{(pid=63515) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 900.837224] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-1e59a109-1efb-4ebd-8738-fea0e8fa8e8b tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: b1bf5a0a-42e7-41a6-90fb-a0c311623fdd] Destroying instance {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 900.838084] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-116f2bad-e6fe-4a25-9b42-223432afa6f5 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.846753] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-1e59a109-1efb-4ebd-8738-fea0e8fa8e8b tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: b1bf5a0a-42e7-41a6-90fb-a0c311623fdd] Powering off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 900.847053] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9b727dc5-ad9e-4942-a59c-c7267f2ae4b8 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.854070] env[63515]: DEBUG oslo_vmware.api [None req-1e59a109-1efb-4ebd-8738-fea0e8fa8e8b tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Waiting for the task: (returnval){ [ 900.854070] env[63515]: value = "task-1111514" [ 900.854070] env[63515]: _type = "Task" [ 900.854070] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 900.863613] env[63515]: DEBUG oslo_vmware.api [None req-1e59a109-1efb-4ebd-8738-fea0e8fa8e8b tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Task: {'id': task-1111514, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.030745] env[63515]: DEBUG nova.compute.manager [None req-8c6d15be-1f5e-44a9-a3e4-dab9aa82e26a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] [instance: 9236607d-47af-4272-93db-79826e918aa1] Starting instance... {{(pid=63515) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 901.104864] env[63515]: DEBUG nova.compute.manager [None req-8c6d15be-1f5e-44a9-a3e4-dab9aa82e26a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] [instance: ca6a8ac3-d880-4333-9816-2f076e423844] Starting instance... {{(pid=63515) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 901.140670] env[63515]: DEBUG nova.scheduler.client.report [None req-07515409-459b-4c3e-992f-c4e070f0ae85 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 901.163329] env[63515]: DEBUG oslo_concurrency.lockutils [req-fd3d0e68-bde8-447c-8043-c21c7cf7008d req-3794daf3-f198-4b4f-bed0-b54a9900c503 service nova] Releasing lock "refresh_cache-324f7e89-3c17-474d-af1b-6d7a6d042510" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 901.187021] env[63515]: DEBUG nova.network.neutron [req-99d7b96c-2344-44b7-bd74-68cd6622fb83 req-4ab4dce3-762a-4ef5-b515-4ed28d8837ef service nova] [instance: e13da90c-28e6-43d2-99b7-19c5095954ca] Updated VIF entry in instance network info cache for port d9363b80-630e-4262-901f-c13e5c8b8c2a. {{(pid=63515) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 901.187530] env[63515]: DEBUG nova.network.neutron [req-99d7b96c-2344-44b7-bd74-68cd6622fb83 req-4ab4dce3-762a-4ef5-b515-4ed28d8837ef service nova] [instance: e13da90c-28e6-43d2-99b7-19c5095954ca] Updating instance_info_cache with network_info: [{"id": "d9363b80-630e-4262-901f-c13e5c8b8c2a", "address": "fa:16:3e:dd:fe:76", "network": {"id": "cec614cb-7a5b-4b16-a83d-061ff62d6677", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1501468300-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.155", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "aef947bfaf7d4180b46b4be662e3d903", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3a0ddd7d-c321-4187-bdd8-b19044ea2c4a", "external-id": "nsx-vlan-transportzone-747", "segmentation_id": 747, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd9363b80-63", "ovs_interfaceid": "d9363b80-630e-4262-901f-c13e5c8b8c2a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 901.239051] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1111513, 'name': CreateVM_Task, 'duration_secs': 0.511118} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 901.239448] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 324f7e89-3c17-474d-af1b-6d7a6d042510] Created VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 901.240011] env[63515]: DEBUG oslo_concurrency.lockutils [None req-ceee2529-e58a-42aa-a26e-32b1ce124ccb tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 901.240201] env[63515]: DEBUG oslo_concurrency.lockutils [None req-ceee2529-e58a-42aa-a26e-32b1ce124ccb tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 901.240526] env[63515]: DEBUG oslo_concurrency.lockutils [None req-ceee2529-e58a-42aa-a26e-32b1ce124ccb tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 901.240783] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-846bd27c-181a-4fa2-85c2-8b0fceea9dd8 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.245769] env[63515]: DEBUG oslo_vmware.api [None req-ceee2529-e58a-42aa-a26e-32b1ce124ccb tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] Waiting for the task: (returnval){ [ 901.245769] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]52608bd1-f31d-a7d9-fd23-753baf3af2d0" [ 901.245769] env[63515]: _type = "Task" [ 901.245769] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.254202] env[63515]: DEBUG oslo_vmware.api [None req-ceee2529-e58a-42aa-a26e-32b1ce124ccb tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52608bd1-f31d-a7d9-fd23-753baf3af2d0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.364992] env[63515]: DEBUG oslo_vmware.api [None req-1e59a109-1efb-4ebd-8738-fea0e8fa8e8b tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Task: {'id': task-1111514, 'name': PowerOffVM_Task, 'duration_secs': 0.224844} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 901.365343] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-1e59a109-1efb-4ebd-8738-fea0e8fa8e8b tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: b1bf5a0a-42e7-41a6-90fb-a0c311623fdd] Powered off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 901.365549] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-1e59a109-1efb-4ebd-8738-fea0e8fa8e8b tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: b1bf5a0a-42e7-41a6-90fb-a0c311623fdd] Unregistering the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 901.365839] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-caeb9969-902e-4fc0-b175-e6c7f07e52c4 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.433607] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-1e59a109-1efb-4ebd-8738-fea0e8fa8e8b tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: b1bf5a0a-42e7-41a6-90fb-a0c311623fdd] Unregistered the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 901.433892] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-1e59a109-1efb-4ebd-8738-fea0e8fa8e8b tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: b1bf5a0a-42e7-41a6-90fb-a0c311623fdd] Deleting contents of the VM from datastore datastore2 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 901.434056] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-1e59a109-1efb-4ebd-8738-fea0e8fa8e8b tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Deleting the datastore file [datastore2] b1bf5a0a-42e7-41a6-90fb-a0c311623fdd {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 901.434345] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-32ccd650-c2fb-400e-bf6d-cfb9ba165d68 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.441087] env[63515]: DEBUG oslo_vmware.api [None req-1e59a109-1efb-4ebd-8738-fea0e8fa8e8b tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Waiting for the task: (returnval){ [ 901.441087] env[63515]: value = "task-1111516" [ 901.441087] env[63515]: _type = "Task" [ 901.441087] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.449481] env[63515]: DEBUG oslo_vmware.api [None req-1e59a109-1efb-4ebd-8738-fea0e8fa8e8b tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Task: {'id': task-1111516, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.553887] env[63515]: DEBUG oslo_concurrency.lockutils [None req-8c6d15be-1f5e-44a9-a3e4-dab9aa82e26a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 901.625819] env[63515]: DEBUG nova.compute.manager [None req-3173d497-cfe4-4001-b365-7bedc6c45742 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: 829e72d7-a56a-451c-b98a-94253e5900d5] Start spawning the instance on the hypervisor. {{(pid=63515) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 901.635664] env[63515]: DEBUG oslo_concurrency.lockutils [None req-8c6d15be-1f5e-44a9-a3e4-dab9aa82e26a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 901.650570] env[63515]: DEBUG oslo_concurrency.lockutils [None req-07515409-459b-4c3e-992f-c4e070f0ae85 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.051s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 901.655320] env[63515]: DEBUG nova.virt.hardware [None req-3173d497-cfe4-4001-b365-7bedc6c45742 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T02:52:14Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T02:51:56Z,direct_url=,disk_format='vmdk',id=8a120570-cb06-4099-b262-554ca0ad15c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b569255cc43e42e7ae3f2b4ad37c6ef4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T02:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 901.655582] env[63515]: DEBUG nova.virt.hardware [None req-3173d497-cfe4-4001-b365-7bedc6c45742 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Flavor limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 901.658124] env[63515]: DEBUG nova.virt.hardware [None req-3173d497-cfe4-4001-b365-7bedc6c45742 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Image limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 901.658124] env[63515]: DEBUG nova.virt.hardware [None req-3173d497-cfe4-4001-b365-7bedc6c45742 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Flavor pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 901.658124] env[63515]: DEBUG nova.virt.hardware [None req-3173d497-cfe4-4001-b365-7bedc6c45742 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Image pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 901.658124] env[63515]: DEBUG nova.virt.hardware [None req-3173d497-cfe4-4001-b365-7bedc6c45742 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 901.658124] env[63515]: DEBUG nova.virt.hardware [None req-3173d497-cfe4-4001-b365-7bedc6c45742 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 901.658124] env[63515]: DEBUG nova.virt.hardware [None req-3173d497-cfe4-4001-b365-7bedc6c45742 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 901.658124] env[63515]: DEBUG nova.virt.hardware [None req-3173d497-cfe4-4001-b365-7bedc6c45742 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Got 1 possible topologies {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 901.658124] env[63515]: DEBUG nova.virt.hardware [None req-3173d497-cfe4-4001-b365-7bedc6c45742 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 901.658124] env[63515]: DEBUG nova.virt.hardware [None req-3173d497-cfe4-4001-b365-7bedc6c45742 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 901.658124] env[63515]: DEBUG oslo_concurrency.lockutils [None req-060c62db-6550-44b8-abf2-82f4d769cd48 tempest-ServerGroupTestJSON-385159797 tempest-ServerGroupTestJSON-385159797-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 9.599s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 901.658124] env[63515]: DEBUG nova.objects.instance [None req-060c62db-6550-44b8-abf2-82f4d769cd48 tempest-ServerGroupTestJSON-385159797 tempest-ServerGroupTestJSON-385159797-project-member] Lazy-loading 'resources' on Instance uuid a2e014a2-9e0f-4250-aa38-3a7154149903 {{(pid=63515) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 901.659835] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7765ab8e-fee5-4c54-b421-5fb9ff917085 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.669998] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ae10c96-0f48-49e1-a034-74c83d3ccc04 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.675448] env[63515]: INFO nova.scheduler.client.report [None req-07515409-459b-4c3e-992f-c4e070f0ae85 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Deleted allocations for instance 0967283f-8dda-4692-b6b7-601545cabf2f [ 901.690745] env[63515]: DEBUG oslo_concurrency.lockutils [req-99d7b96c-2344-44b7-bd74-68cd6622fb83 req-4ab4dce3-762a-4ef5-b515-4ed28d8837ef service nova] Releasing lock "refresh_cache-e13da90c-28e6-43d2-99b7-19c5095954ca" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 901.757914] env[63515]: DEBUG oslo_vmware.api [None req-ceee2529-e58a-42aa-a26e-32b1ce124ccb tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52608bd1-f31d-a7d9-fd23-753baf3af2d0, 'name': SearchDatastore_Task, 'duration_secs': 0.011798} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 901.757914] env[63515]: DEBUG oslo_concurrency.lockutils [None req-ceee2529-e58a-42aa-a26e-32b1ce124ccb tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 901.757914] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-ceee2529-e58a-42aa-a26e-32b1ce124ccb tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] [instance: 324f7e89-3c17-474d-af1b-6d7a6d042510] Processing image 8a120570-cb06-4099-b262-554ca0ad15c5 {{(pid=63515) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 901.758056] env[63515]: DEBUG oslo_concurrency.lockutils [None req-ceee2529-e58a-42aa-a26e-32b1ce124ccb tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 901.759045] env[63515]: DEBUG oslo_concurrency.lockutils [None req-ceee2529-e58a-42aa-a26e-32b1ce124ccb tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 901.759045] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-ceee2529-e58a-42aa-a26e-32b1ce124ccb tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 901.759045] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-16ab2e5c-3db2-43f0-a3f4-c33edd03240b {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.767441] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-ceee2529-e58a-42aa-a26e-32b1ce124ccb tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 901.767638] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-ceee2529-e58a-42aa-a26e-32b1ce124ccb tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63515) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 901.768451] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-81db482c-ba27-445f-bc34-cce55effc444 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.774600] env[63515]: DEBUG oslo_vmware.api [None req-ceee2529-e58a-42aa-a26e-32b1ce124ccb tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] Waiting for the task: (returnval){ [ 901.774600] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]5231ff1d-8fa4-c4ba-f103-c7392822d2ae" [ 901.774600] env[63515]: _type = "Task" [ 901.774600] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.782349] env[63515]: DEBUG oslo_vmware.api [None req-ceee2529-e58a-42aa-a26e-32b1ce124ccb tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]5231ff1d-8fa4-c4ba-f103-c7392822d2ae, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.952062] env[63515]: DEBUG oslo_vmware.api [None req-1e59a109-1efb-4ebd-8738-fea0e8fa8e8b tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Task: {'id': task-1111516, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.203394} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 901.952401] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-1e59a109-1efb-4ebd-8738-fea0e8fa8e8b tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Deleted the datastore file {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 901.952630] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-1e59a109-1efb-4ebd-8738-fea0e8fa8e8b tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: b1bf5a0a-42e7-41a6-90fb-a0c311623fdd] Deleted contents of the VM from datastore datastore2 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 901.952846] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-1e59a109-1efb-4ebd-8738-fea0e8fa8e8b tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: b1bf5a0a-42e7-41a6-90fb-a0c311623fdd] Instance destroyed {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 901.953083] env[63515]: INFO nova.compute.manager [None req-1e59a109-1efb-4ebd-8738-fea0e8fa8e8b tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: b1bf5a0a-42e7-41a6-90fb-a0c311623fdd] Took 1.12 seconds to destroy the instance on the hypervisor. [ 901.953388] env[63515]: DEBUG oslo.service.loopingcall [None req-1e59a109-1efb-4ebd-8738-fea0e8fa8e8b tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 901.953615] env[63515]: DEBUG nova.compute.manager [-] [instance: b1bf5a0a-42e7-41a6-90fb-a0c311623fdd] Deallocating network for instance {{(pid=63515) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 901.953742] env[63515]: DEBUG nova.network.neutron [-] [instance: b1bf5a0a-42e7-41a6-90fb-a0c311623fdd] deallocate_for_instance() {{(pid=63515) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 902.188106] env[63515]: DEBUG oslo_concurrency.lockutils [None req-07515409-459b-4c3e-992f-c4e070f0ae85 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Lock "0967283f-8dda-4692-b6b7-601545cabf2f" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 14.839s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 902.296750] env[63515]: DEBUG oslo_vmware.api [None req-ceee2529-e58a-42aa-a26e-32b1ce124ccb tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]5231ff1d-8fa4-c4ba-f103-c7392822d2ae, 'name': SearchDatastore_Task, 'duration_secs': 0.011456} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.301880] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-612ee61c-7080-4ac4-b0e3-d789e3d26a61 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.306591] env[63515]: DEBUG oslo_vmware.api [None req-ceee2529-e58a-42aa-a26e-32b1ce124ccb tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] Waiting for the task: (returnval){ [ 902.306591] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]520e8a7c-dcde-e312-9b25-8f30be3b4682" [ 902.306591] env[63515]: _type = "Task" [ 902.306591] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 902.316091] env[63515]: DEBUG nova.compute.manager [req-ab3b35db-a96c-4938-a293-dd4d08767b33 req-a000844a-668a-49c7-ad90-388da7d70091 service nova] [instance: 829e72d7-a56a-451c-b98a-94253e5900d5] Received event network-vif-plugged-856de677-b1c8-479b-91d9-34d31a703a63 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 902.316320] env[63515]: DEBUG oslo_concurrency.lockutils [req-ab3b35db-a96c-4938-a293-dd4d08767b33 req-a000844a-668a-49c7-ad90-388da7d70091 service nova] Acquiring lock "829e72d7-a56a-451c-b98a-94253e5900d5-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 902.316536] env[63515]: DEBUG oslo_concurrency.lockutils [req-ab3b35db-a96c-4938-a293-dd4d08767b33 req-a000844a-668a-49c7-ad90-388da7d70091 service nova] Lock "829e72d7-a56a-451c-b98a-94253e5900d5-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 902.316705] env[63515]: DEBUG oslo_concurrency.lockutils [req-ab3b35db-a96c-4938-a293-dd4d08767b33 req-a000844a-668a-49c7-ad90-388da7d70091 service nova] Lock "829e72d7-a56a-451c-b98a-94253e5900d5-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 902.316880] env[63515]: DEBUG nova.compute.manager [req-ab3b35db-a96c-4938-a293-dd4d08767b33 req-a000844a-668a-49c7-ad90-388da7d70091 service nova] [instance: 829e72d7-a56a-451c-b98a-94253e5900d5] No waiting events found dispatching network-vif-plugged-856de677-b1c8-479b-91d9-34d31a703a63 {{(pid=63515) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 902.317065] env[63515]: WARNING nova.compute.manager [req-ab3b35db-a96c-4938-a293-dd4d08767b33 req-a000844a-668a-49c7-ad90-388da7d70091 service nova] [instance: 829e72d7-a56a-451c-b98a-94253e5900d5] Received unexpected event network-vif-plugged-856de677-b1c8-479b-91d9-34d31a703a63 for instance with vm_state building and task_state spawning. [ 902.324568] env[63515]: DEBUG oslo_vmware.api [None req-ceee2529-e58a-42aa-a26e-32b1ce124ccb tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]520e8a7c-dcde-e312-9b25-8f30be3b4682, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.369442] env[63515]: DEBUG nova.network.neutron [None req-3173d497-cfe4-4001-b365-7bedc6c45742 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: 829e72d7-a56a-451c-b98a-94253e5900d5] Successfully updated port: 856de677-b1c8-479b-91d9-34d31a703a63 {{(pid=63515) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 902.452307] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12712071-427a-4b88-8963-522feae4c0d7 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.462836] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e884226-4907-4a33-a81e-5315a9b81386 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.494945] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71a1739a-4b6b-4919-a36b-015b93b2ab81 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.499482] env[63515]: DEBUG nova.compute.manager [req-c895101b-933c-49e8-8767-05512f2c3d85 req-6a351366-2204-499c-8e38-c312f4455793 service nova] [instance: b1bf5a0a-42e7-41a6-90fb-a0c311623fdd] Received event network-vif-deleted-8ab9e788-50a9-4669-8448-77cd0983386e {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 902.499482] env[63515]: INFO nova.compute.manager [req-c895101b-933c-49e8-8767-05512f2c3d85 req-6a351366-2204-499c-8e38-c312f4455793 service nova] [instance: b1bf5a0a-42e7-41a6-90fb-a0c311623fdd] Neutron deleted interface 8ab9e788-50a9-4669-8448-77cd0983386e; detaching it from the instance and deleting it from the info cache [ 902.499482] env[63515]: DEBUG nova.network.neutron [req-c895101b-933c-49e8-8767-05512f2c3d85 req-6a351366-2204-499c-8e38-c312f4455793 service nova] [instance: b1bf5a0a-42e7-41a6-90fb-a0c311623fdd] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 902.507386] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d780aa20-a874-46e9-9eea-38dbfbf451a3 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.522287] env[63515]: DEBUG nova.compute.provider_tree [None req-060c62db-6550-44b8-abf2-82f4d769cd48 tempest-ServerGroupTestJSON-385159797 tempest-ServerGroupTestJSON-385159797-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 902.686029] env[63515]: DEBUG nova.network.neutron [-] [instance: b1bf5a0a-42e7-41a6-90fb-a0c311623fdd] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 902.819066] env[63515]: DEBUG oslo_vmware.api [None req-ceee2529-e58a-42aa-a26e-32b1ce124ccb tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]520e8a7c-dcde-e312-9b25-8f30be3b4682, 'name': SearchDatastore_Task, 'duration_secs': 0.017027} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.819374] env[63515]: DEBUG oslo_concurrency.lockutils [None req-ceee2529-e58a-42aa-a26e-32b1ce124ccb tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 902.819643] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-ceee2529-e58a-42aa-a26e-32b1ce124ccb tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk to [datastore2] 324f7e89-3c17-474d-af1b-6d7a6d042510/324f7e89-3c17-474d-af1b-6d7a6d042510.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 902.819923] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-215c6f43-7228-4a55-8ec4-ddaf307cba96 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.828438] env[63515]: DEBUG oslo_vmware.api [None req-ceee2529-e58a-42aa-a26e-32b1ce124ccb tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] Waiting for the task: (returnval){ [ 902.828438] env[63515]: value = "task-1111517" [ 902.828438] env[63515]: _type = "Task" [ 902.828438] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 902.836373] env[63515]: DEBUG oslo_vmware.api [None req-ceee2529-e58a-42aa-a26e-32b1ce124ccb tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] Task: {'id': task-1111517, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.873388] env[63515]: DEBUG oslo_concurrency.lockutils [None req-3173d497-cfe4-4001-b365-7bedc6c45742 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Acquiring lock "refresh_cache-829e72d7-a56a-451c-b98a-94253e5900d5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 902.873388] env[63515]: DEBUG oslo_concurrency.lockutils [None req-3173d497-cfe4-4001-b365-7bedc6c45742 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Acquired lock "refresh_cache-829e72d7-a56a-451c-b98a-94253e5900d5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 902.873388] env[63515]: DEBUG nova.network.neutron [None req-3173d497-cfe4-4001-b365-7bedc6c45742 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: 829e72d7-a56a-451c-b98a-94253e5900d5] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 903.002559] env[63515]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-5ada64f2-1365-48ca-96f9-8b1d9433241b {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.012462] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7dcb1bf6-0f51-426b-86fb-aa1f2154a122 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.024990] env[63515]: DEBUG nova.scheduler.client.report [None req-060c62db-6550-44b8-abf2-82f4d769cd48 tempest-ServerGroupTestJSON-385159797 tempest-ServerGroupTestJSON-385159797-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 903.049065] env[63515]: DEBUG nova.compute.manager [req-c895101b-933c-49e8-8767-05512f2c3d85 req-6a351366-2204-499c-8e38-c312f4455793 service nova] [instance: b1bf5a0a-42e7-41a6-90fb-a0c311623fdd] Detach interface failed, port_id=8ab9e788-50a9-4669-8448-77cd0983386e, reason: Instance b1bf5a0a-42e7-41a6-90fb-a0c311623fdd could not be found. {{(pid=63515) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 903.081483] env[63515]: DEBUG oslo_concurrency.lockutils [None req-0bb813f4-ce4d-4144-b80d-b179b1c5fd20 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Acquiring lock "a2203c5e-66b9-4c7e-8d76-e12a08b57b1d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 903.081811] env[63515]: DEBUG oslo_concurrency.lockutils [None req-0bb813f4-ce4d-4144-b80d-b179b1c5fd20 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Lock "a2203c5e-66b9-4c7e-8d76-e12a08b57b1d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 903.188493] env[63515]: INFO nova.compute.manager [-] [instance: b1bf5a0a-42e7-41a6-90fb-a0c311623fdd] Took 1.23 seconds to deallocate network for instance. [ 903.341186] env[63515]: DEBUG oslo_vmware.api [None req-ceee2529-e58a-42aa-a26e-32b1ce124ccb tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] Task: {'id': task-1111517, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.406711] env[63515]: DEBUG nova.network.neutron [None req-3173d497-cfe4-4001-b365-7bedc6c45742 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: 829e72d7-a56a-451c-b98a-94253e5900d5] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 903.534415] env[63515]: DEBUG oslo_concurrency.lockutils [None req-060c62db-6550-44b8-abf2-82f4d769cd48 tempest-ServerGroupTestJSON-385159797 tempest-ServerGroupTestJSON-385159797-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.876s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 903.537481] env[63515]: DEBUG oslo_concurrency.lockutils [None req-29ee5601-80a8-4cbb-90e9-98d09454abfa tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 9.728s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 903.540039] env[63515]: INFO nova.compute.claims [None req-29ee5601-80a8-4cbb-90e9-98d09454abfa tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] [instance: 86b1ddcf-585c-41d9-84fe-17b794f1abbc] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 903.554155] env[63515]: DEBUG nova.network.neutron [None req-3173d497-cfe4-4001-b365-7bedc6c45742 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: 829e72d7-a56a-451c-b98a-94253e5900d5] Updating instance_info_cache with network_info: [{"id": "856de677-b1c8-479b-91d9-34d31a703a63", "address": "fa:16:3e:13:27:b4", "network": {"id": "15fdaabe-6793-400f-9737-2cdff8f07238", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1175204866-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f9c98e9f6020475490aaa7e76d907ab7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7d377d75-3add-4a15-8691-74b2eb010924", "external-id": "nsx-vlan-transportzone-71", "segmentation_id": 71, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap856de677-b1", "ovs_interfaceid": "856de677-b1c8-479b-91d9-34d31a703a63", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 903.570334] env[63515]: INFO nova.scheduler.client.report [None req-060c62db-6550-44b8-abf2-82f4d769cd48 tempest-ServerGroupTestJSON-385159797 tempest-ServerGroupTestJSON-385159797-project-member] Deleted allocations for instance a2e014a2-9e0f-4250-aa38-3a7154149903 [ 903.587299] env[63515]: DEBUG nova.compute.manager [None req-0bb813f4-ce4d-4144-b80d-b179b1c5fd20 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: a2203c5e-66b9-4c7e-8d76-e12a08b57b1d] Starting instance... {{(pid=63515) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 903.656982] env[63515]: DEBUG oslo_service.periodic_task [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=63515) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 903.656982] env[63515]: DEBUG oslo_service.periodic_task [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=63515) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 903.695801] env[63515]: DEBUG oslo_concurrency.lockutils [None req-1e59a109-1efb-4ebd-8738-fea0e8fa8e8b tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 903.839142] env[63515]: DEBUG oslo_vmware.api [None req-ceee2529-e58a-42aa-a26e-32b1ce124ccb tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] Task: {'id': task-1111517, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.567247} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 903.839490] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-ceee2529-e58a-42aa-a26e-32b1ce124ccb tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk to [datastore2] 324f7e89-3c17-474d-af1b-6d7a6d042510/324f7e89-3c17-474d-af1b-6d7a6d042510.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 903.839755] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-ceee2529-e58a-42aa-a26e-32b1ce124ccb tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] [instance: 324f7e89-3c17-474d-af1b-6d7a6d042510] Extending root virtual disk to 1048576 {{(pid=63515) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 903.840051] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-f40cc88e-72b6-462c-80ec-f26c08502ac5 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.846384] env[63515]: DEBUG oslo_vmware.api [None req-ceee2529-e58a-42aa-a26e-32b1ce124ccb tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] Waiting for the task: (returnval){ [ 903.846384] env[63515]: value = "task-1111518" [ 903.846384] env[63515]: _type = "Task" [ 903.846384] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 903.855493] env[63515]: DEBUG oslo_vmware.api [None req-ceee2529-e58a-42aa-a26e-32b1ce124ccb tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] Task: {'id': task-1111518, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.058129] env[63515]: DEBUG oslo_concurrency.lockutils [None req-3173d497-cfe4-4001-b365-7bedc6c45742 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Releasing lock "refresh_cache-829e72d7-a56a-451c-b98a-94253e5900d5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 904.058441] env[63515]: DEBUG nova.compute.manager [None req-3173d497-cfe4-4001-b365-7bedc6c45742 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: 829e72d7-a56a-451c-b98a-94253e5900d5] Instance network_info: |[{"id": "856de677-b1c8-479b-91d9-34d31a703a63", "address": "fa:16:3e:13:27:b4", "network": {"id": "15fdaabe-6793-400f-9737-2cdff8f07238", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1175204866-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f9c98e9f6020475490aaa7e76d907ab7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7d377d75-3add-4a15-8691-74b2eb010924", "external-id": "nsx-vlan-transportzone-71", "segmentation_id": 71, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap856de677-b1", "ovs_interfaceid": "856de677-b1c8-479b-91d9-34d31a703a63", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 904.058852] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-3173d497-cfe4-4001-b365-7bedc6c45742 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: 829e72d7-a56a-451c-b98a-94253e5900d5] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:13:27:b4', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '7d377d75-3add-4a15-8691-74b2eb010924', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '856de677-b1c8-479b-91d9-34d31a703a63', 'vif_model': 'vmxnet3'}] {{(pid=63515) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 904.066476] env[63515]: DEBUG oslo.service.loopingcall [None req-3173d497-cfe4-4001-b365-7bedc6c45742 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 904.067044] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 829e72d7-a56a-451c-b98a-94253e5900d5] Creating VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 904.067297] env[63515]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-7231e3dd-e1da-49cd-b8f1-73644d61f130 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.085140] env[63515]: DEBUG oslo_concurrency.lockutils [None req-060c62db-6550-44b8-abf2-82f4d769cd48 tempest-ServerGroupTestJSON-385159797 tempest-ServerGroupTestJSON-385159797-project-member] Lock "a2e014a2-9e0f-4250-aa38-3a7154149903" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 14.946s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 904.091325] env[63515]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 904.091325] env[63515]: value = "task-1111519" [ 904.091325] env[63515]: _type = "Task" [ 904.091325] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 904.106700] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1111519, 'name': CreateVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.117532] env[63515]: DEBUG oslo_concurrency.lockutils [None req-0bb813f4-ce4d-4144-b80d-b179b1c5fd20 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 904.168251] env[63515]: DEBUG oslo_service.periodic_task [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=63515) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 904.168514] env[63515]: DEBUG nova.compute.manager [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Starting heal instance info cache {{(pid=63515) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9926}} [ 904.339053] env[63515]: DEBUG nova.compute.manager [req-de1b57c2-ec2d-41f1-bd9e-011805e02d15 req-9c84eef0-a262-433b-a6f0-bf9031dafef3 service nova] [instance: 829e72d7-a56a-451c-b98a-94253e5900d5] Received event network-changed-856de677-b1c8-479b-91d9-34d31a703a63 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 904.339269] env[63515]: DEBUG nova.compute.manager [req-de1b57c2-ec2d-41f1-bd9e-011805e02d15 req-9c84eef0-a262-433b-a6f0-bf9031dafef3 service nova] [instance: 829e72d7-a56a-451c-b98a-94253e5900d5] Refreshing instance network info cache due to event network-changed-856de677-b1c8-479b-91d9-34d31a703a63. {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 904.339497] env[63515]: DEBUG oslo_concurrency.lockutils [req-de1b57c2-ec2d-41f1-bd9e-011805e02d15 req-9c84eef0-a262-433b-a6f0-bf9031dafef3 service nova] Acquiring lock "refresh_cache-829e72d7-a56a-451c-b98a-94253e5900d5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 904.339649] env[63515]: DEBUG oslo_concurrency.lockutils [req-de1b57c2-ec2d-41f1-bd9e-011805e02d15 req-9c84eef0-a262-433b-a6f0-bf9031dafef3 service nova] Acquired lock "refresh_cache-829e72d7-a56a-451c-b98a-94253e5900d5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 904.339818] env[63515]: DEBUG nova.network.neutron [req-de1b57c2-ec2d-41f1-bd9e-011805e02d15 req-9c84eef0-a262-433b-a6f0-bf9031dafef3 service nova] [instance: 829e72d7-a56a-451c-b98a-94253e5900d5] Refreshing network info cache for port 856de677-b1c8-479b-91d9-34d31a703a63 {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 904.357637] env[63515]: DEBUG oslo_vmware.api [None req-ceee2529-e58a-42aa-a26e-32b1ce124ccb tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] Task: {'id': task-1111518, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.086888} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 904.357780] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-ceee2529-e58a-42aa-a26e-32b1ce124ccb tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] [instance: 324f7e89-3c17-474d-af1b-6d7a6d042510] Extended root virtual disk {{(pid=63515) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 904.358592] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d68f990-4ec3-4a68-8197-d1caa8b850ee {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.383796] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-ceee2529-e58a-42aa-a26e-32b1ce124ccb tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] [instance: 324f7e89-3c17-474d-af1b-6d7a6d042510] Reconfiguring VM instance instance-0000004f to attach disk [datastore2] 324f7e89-3c17-474d-af1b-6d7a6d042510/324f7e89-3c17-474d-af1b-6d7a6d042510.vmdk or device None with type sparse {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 904.384594] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-64d2c44a-74f7-4589-92be-b50e190e4aca {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.405497] env[63515]: DEBUG oslo_vmware.api [None req-ceee2529-e58a-42aa-a26e-32b1ce124ccb tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] Waiting for the task: (returnval){ [ 904.405497] env[63515]: value = "task-1111520" [ 904.405497] env[63515]: _type = "Task" [ 904.405497] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 904.413529] env[63515]: DEBUG oslo_vmware.api [None req-ceee2529-e58a-42aa-a26e-32b1ce124ccb tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] Task: {'id': task-1111520, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.603641] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1111519, 'name': CreateVM_Task, 'duration_secs': 0.457815} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 904.606185] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 829e72d7-a56a-451c-b98a-94253e5900d5] Created VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 904.607068] env[63515]: DEBUG oslo_concurrency.lockutils [None req-3173d497-cfe4-4001-b365-7bedc6c45742 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 904.607273] env[63515]: DEBUG oslo_concurrency.lockutils [None req-3173d497-cfe4-4001-b365-7bedc6c45742 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 904.607589] env[63515]: DEBUG oslo_concurrency.lockutils [None req-3173d497-cfe4-4001-b365-7bedc6c45742 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 904.607843] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9b669cd1-8845-4254-981f-67d1ff3051a1 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.612563] env[63515]: DEBUG oslo_vmware.api [None req-3173d497-cfe4-4001-b365-7bedc6c45742 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Waiting for the task: (returnval){ [ 904.612563] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]52bccba3-b825-4dee-e4bf-9b0dd537d1c8" [ 904.612563] env[63515]: _type = "Task" [ 904.612563] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 904.621554] env[63515]: DEBUG oslo_vmware.api [None req-3173d497-cfe4-4001-b365-7bedc6c45742 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52bccba3-b825-4dee-e4bf-9b0dd537d1c8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.813832] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7321a947-cd09-4ada-bc19-3382d7b0280f {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.829336] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c24e55b-e38c-4bcc-9cd0-da2ad661ca28 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.868232] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5040f022-1f21-4df1-9285-e4b80b63e381 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.877479] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7143fa0-b597-4602-b3b2-b19a29b4a89e {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.892879] env[63515]: DEBUG nova.compute.provider_tree [None req-29ee5601-80a8-4cbb-90e9-98d09454abfa tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 904.915747] env[63515]: DEBUG oslo_vmware.api [None req-ceee2529-e58a-42aa-a26e-32b1ce124ccb tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] Task: {'id': task-1111520, 'name': ReconfigVM_Task, 'duration_secs': 0.465588} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 904.916045] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-ceee2529-e58a-42aa-a26e-32b1ce124ccb tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] [instance: 324f7e89-3c17-474d-af1b-6d7a6d042510] Reconfigured VM instance instance-0000004f to attach disk [datastore2] 324f7e89-3c17-474d-af1b-6d7a6d042510/324f7e89-3c17-474d-af1b-6d7a6d042510.vmdk or device None with type sparse {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 904.916711] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-51eba56e-3337-4660-982e-cb3dcc905ae2 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.922878] env[63515]: DEBUG oslo_vmware.api [None req-ceee2529-e58a-42aa-a26e-32b1ce124ccb tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] Waiting for the task: (returnval){ [ 904.922878] env[63515]: value = "task-1111521" [ 904.922878] env[63515]: _type = "Task" [ 904.922878] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 904.931598] env[63515]: DEBUG oslo_vmware.api [None req-ceee2529-e58a-42aa-a26e-32b1ce124ccb tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] Task: {'id': task-1111521, 'name': Rename_Task} progress is 5%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.094923] env[63515]: DEBUG nova.network.neutron [req-de1b57c2-ec2d-41f1-bd9e-011805e02d15 req-9c84eef0-a262-433b-a6f0-bf9031dafef3 service nova] [instance: 829e72d7-a56a-451c-b98a-94253e5900d5] Updated VIF entry in instance network info cache for port 856de677-b1c8-479b-91d9-34d31a703a63. {{(pid=63515) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 905.095443] env[63515]: DEBUG nova.network.neutron [req-de1b57c2-ec2d-41f1-bd9e-011805e02d15 req-9c84eef0-a262-433b-a6f0-bf9031dafef3 service nova] [instance: 829e72d7-a56a-451c-b98a-94253e5900d5] Updating instance_info_cache with network_info: [{"id": "856de677-b1c8-479b-91d9-34d31a703a63", "address": "fa:16:3e:13:27:b4", "network": {"id": "15fdaabe-6793-400f-9737-2cdff8f07238", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1175204866-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f9c98e9f6020475490aaa7e76d907ab7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7d377d75-3add-4a15-8691-74b2eb010924", "external-id": "nsx-vlan-transportzone-71", "segmentation_id": 71, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap856de677-b1", "ovs_interfaceid": "856de677-b1c8-479b-91d9-34d31a703a63", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 905.126405] env[63515]: DEBUG oslo_vmware.api [None req-3173d497-cfe4-4001-b365-7bedc6c45742 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52bccba3-b825-4dee-e4bf-9b0dd537d1c8, 'name': SearchDatastore_Task, 'duration_secs': 0.011118} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 905.126764] env[63515]: DEBUG oslo_concurrency.lockutils [None req-3173d497-cfe4-4001-b365-7bedc6c45742 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 905.127058] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-3173d497-cfe4-4001-b365-7bedc6c45742 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: 829e72d7-a56a-451c-b98a-94253e5900d5] Processing image 8a120570-cb06-4099-b262-554ca0ad15c5 {{(pid=63515) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 905.127381] env[63515]: DEBUG oslo_concurrency.lockutils [None req-3173d497-cfe4-4001-b365-7bedc6c45742 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 905.127683] env[63515]: DEBUG oslo_concurrency.lockutils [None req-3173d497-cfe4-4001-b365-7bedc6c45742 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 905.127783] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-3173d497-cfe4-4001-b365-7bedc6c45742 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 905.128421] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-81262e50-2339-48eb-a4ba-2cc085d3df04 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.137685] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-3173d497-cfe4-4001-b365-7bedc6c45742 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 905.137896] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-3173d497-cfe4-4001-b365-7bedc6c45742 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63515) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 905.138671] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f028920a-1e0c-4b70-b38b-9bd3f7883cec {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.143859] env[63515]: DEBUG oslo_vmware.api [None req-3173d497-cfe4-4001-b365-7bedc6c45742 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Waiting for the task: (returnval){ [ 905.143859] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]528350d5-6d2a-716c-3064-76609c279b86" [ 905.143859] env[63515]: _type = "Task" [ 905.143859] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.152337] env[63515]: DEBUG oslo_vmware.api [None req-3173d497-cfe4-4001-b365-7bedc6c45742 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]528350d5-6d2a-716c-3064-76609c279b86, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.399035] env[63515]: DEBUG nova.scheduler.client.report [None req-29ee5601-80a8-4cbb-90e9-98d09454abfa tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 905.433179] env[63515]: DEBUG oslo_vmware.api [None req-ceee2529-e58a-42aa-a26e-32b1ce124ccb tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] Task: {'id': task-1111521, 'name': Rename_Task, 'duration_secs': 0.211593} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 905.433499] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-ceee2529-e58a-42aa-a26e-32b1ce124ccb tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] [instance: 324f7e89-3c17-474d-af1b-6d7a6d042510] Powering on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 905.433787] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a1ec8334-9499-4bcb-85c0-31089eda0ff2 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.441024] env[63515]: DEBUG oslo_vmware.api [None req-ceee2529-e58a-42aa-a26e-32b1ce124ccb tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] Waiting for the task: (returnval){ [ 905.441024] env[63515]: value = "task-1111522" [ 905.441024] env[63515]: _type = "Task" [ 905.441024] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.449763] env[63515]: DEBUG oslo_vmware.api [None req-ceee2529-e58a-42aa-a26e-32b1ce124ccb tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] Task: {'id': task-1111522, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.597909] env[63515]: DEBUG oslo_concurrency.lockutils [req-de1b57c2-ec2d-41f1-bd9e-011805e02d15 req-9c84eef0-a262-433b-a6f0-bf9031dafef3 service nova] Releasing lock "refresh_cache-829e72d7-a56a-451c-b98a-94253e5900d5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 905.655225] env[63515]: DEBUG oslo_vmware.api [None req-3173d497-cfe4-4001-b365-7bedc6c45742 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]528350d5-6d2a-716c-3064-76609c279b86, 'name': SearchDatastore_Task, 'duration_secs': 0.011314} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 905.656096] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d296c21c-71ac-4d8a-bd09-6a607f765a61 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.661372] env[63515]: DEBUG oslo_vmware.api [None req-3173d497-cfe4-4001-b365-7bedc6c45742 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Waiting for the task: (returnval){ [ 905.661372] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]52706b8c-3299-272c-848c-f853ca49ec1b" [ 905.661372] env[63515]: _type = "Task" [ 905.661372] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.670572] env[63515]: DEBUG oslo_vmware.api [None req-3173d497-cfe4-4001-b365-7bedc6c45742 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52706b8c-3299-272c-848c-f853ca49ec1b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.695612] env[63515]: DEBUG oslo_vmware.rw_handles [None req-ef0f5ab2-1246-4de1-af39-f1efd470cd76 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5223dc08-c895-6d10-cbda-75409b05588d/disk-0.vmdk. {{(pid=63515) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 905.696693] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8b35859-8e7a-4d43-becf-d4050596fd6d {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.703117] env[63515]: DEBUG oslo_vmware.rw_handles [None req-ef0f5ab2-1246-4de1-af39-f1efd470cd76 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5223dc08-c895-6d10-cbda-75409b05588d/disk-0.vmdk is in state: ready. {{(pid=63515) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 905.703349] env[63515]: ERROR oslo_vmware.rw_handles [None req-ef0f5ab2-1246-4de1-af39-f1efd470cd76 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5223dc08-c895-6d10-cbda-75409b05588d/disk-0.vmdk due to incomplete transfer. [ 905.703597] env[63515]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-b68c5c01-50c2-490e-83c8-29d21e798fa0 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.710258] env[63515]: DEBUG oslo_vmware.rw_handles [None req-ef0f5ab2-1246-4de1-af39-f1efd470cd76 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5223dc08-c895-6d10-cbda-75409b05588d/disk-0.vmdk. {{(pid=63515) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 905.710465] env[63515]: DEBUG nova.virt.vmwareapi.images [None req-ef0f5ab2-1246-4de1-af39-f1efd470cd76 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: ae5d6f34-bac6-4ba0-b0c3-8205982ddc88] Uploaded image 9a616ff7-f282-45bd-9090-bf9b0dfb5f2d to the Glance image server {{(pid=63515) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 905.712829] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-ef0f5ab2-1246-4de1-af39-f1efd470cd76 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: ae5d6f34-bac6-4ba0-b0c3-8205982ddc88] Destroying the VM {{(pid=63515) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 905.713357] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-c1891e31-1621-4318-8b12-912c3b23319a {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.720220] env[63515]: DEBUG oslo_vmware.api [None req-ef0f5ab2-1246-4de1-af39-f1efd470cd76 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Waiting for the task: (returnval){ [ 905.720220] env[63515]: value = "task-1111523" [ 905.720220] env[63515]: _type = "Task" [ 905.720220] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.728652] env[63515]: DEBUG oslo_vmware.api [None req-ef0f5ab2-1246-4de1-af39-f1efd470cd76 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Task: {'id': task-1111523, 'name': Destroy_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.901608] env[63515]: DEBUG oslo_concurrency.lockutils [None req-29ee5601-80a8-4cbb-90e9-98d09454abfa tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.364s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 905.902228] env[63515]: DEBUG nova.compute.manager [None req-29ee5601-80a8-4cbb-90e9-98d09454abfa tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] [instance: 86b1ddcf-585c-41d9-84fe-17b794f1abbc] Start building networks asynchronously for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 905.905682] env[63515]: DEBUG oslo_concurrency.lockutils [None req-8c6d15be-1f5e-44a9-a3e4-dab9aa82e26a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.351s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 905.907451] env[63515]: INFO nova.compute.claims [None req-8c6d15be-1f5e-44a9-a3e4-dab9aa82e26a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] [instance: 9236607d-47af-4272-93db-79826e918aa1] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 905.957140] env[63515]: DEBUG oslo_vmware.api [None req-ceee2529-e58a-42aa-a26e-32b1ce124ccb tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] Task: {'id': task-1111522, 'name': PowerOnVM_Task} progress is 89%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.171900] env[63515]: DEBUG oslo_vmware.api [None req-3173d497-cfe4-4001-b365-7bedc6c45742 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52706b8c-3299-272c-848c-f853ca49ec1b, 'name': SearchDatastore_Task, 'duration_secs': 0.010828} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 906.172225] env[63515]: DEBUG oslo_concurrency.lockutils [None req-3173d497-cfe4-4001-b365-7bedc6c45742 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 906.172490] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-3173d497-cfe4-4001-b365-7bedc6c45742 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk to [datastore1] 829e72d7-a56a-451c-b98a-94253e5900d5/829e72d7-a56a-451c-b98a-94253e5900d5.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 906.172751] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-9c384254-3701-4cf5-8dac-e2c7dbb55b4e {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.179158] env[63515]: DEBUG oslo_vmware.api [None req-3173d497-cfe4-4001-b365-7bedc6c45742 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Waiting for the task: (returnval){ [ 906.179158] env[63515]: value = "task-1111524" [ 906.179158] env[63515]: _type = "Task" [ 906.179158] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 906.186316] env[63515]: DEBUG oslo_vmware.api [None req-3173d497-cfe4-4001-b365-7bedc6c45742 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Task: {'id': task-1111524, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.217330] env[63515]: DEBUG oslo_concurrency.lockutils [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Acquiring lock "refresh_cache-87c468d9-9594-4804-b461-527f01f6118f" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 906.217523] env[63515]: DEBUG oslo_concurrency.lockutils [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Acquired lock "refresh_cache-87c468d9-9594-4804-b461-527f01f6118f" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 906.217694] env[63515]: DEBUG nova.network.neutron [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] [instance: 87c468d9-9594-4804-b461-527f01f6118f] Forcefully refreshing network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2004}} [ 906.230405] env[63515]: DEBUG oslo_vmware.api [None req-ef0f5ab2-1246-4de1-af39-f1efd470cd76 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Task: {'id': task-1111523, 'name': Destroy_Task, 'duration_secs': 0.34824} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 906.230695] env[63515]: INFO nova.virt.vmwareapi.vm_util [None req-ef0f5ab2-1246-4de1-af39-f1efd470cd76 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: ae5d6f34-bac6-4ba0-b0c3-8205982ddc88] Destroyed the VM [ 906.231015] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-ef0f5ab2-1246-4de1-af39-f1efd470cd76 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: ae5d6f34-bac6-4ba0-b0c3-8205982ddc88] Deleting Snapshot of the VM instance {{(pid=63515) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 906.231288] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-94cfbcbd-25b5-4efe-b512-c577b6d76ea5 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.237613] env[63515]: DEBUG oslo_vmware.api [None req-ef0f5ab2-1246-4de1-af39-f1efd470cd76 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Waiting for the task: (returnval){ [ 906.237613] env[63515]: value = "task-1111525" [ 906.237613] env[63515]: _type = "Task" [ 906.237613] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 906.247842] env[63515]: DEBUG oslo_vmware.api [None req-ef0f5ab2-1246-4de1-af39-f1efd470cd76 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Task: {'id': task-1111525, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.412749] env[63515]: DEBUG nova.compute.utils [None req-29ee5601-80a8-4cbb-90e9-98d09454abfa tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Using /dev/sd instead of None {{(pid=63515) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 906.416400] env[63515]: DEBUG nova.compute.manager [None req-29ee5601-80a8-4cbb-90e9-98d09454abfa tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] [instance: 86b1ddcf-585c-41d9-84fe-17b794f1abbc] Allocating IP information in the background. {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 906.416584] env[63515]: DEBUG nova.network.neutron [None req-29ee5601-80a8-4cbb-90e9-98d09454abfa tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] [instance: 86b1ddcf-585c-41d9-84fe-17b794f1abbc] allocate_for_instance() {{(pid=63515) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 906.456067] env[63515]: DEBUG oslo_vmware.api [None req-ceee2529-e58a-42aa-a26e-32b1ce124ccb tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] Task: {'id': task-1111522, 'name': PowerOnVM_Task, 'duration_secs': 0.623476} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 906.456489] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-ceee2529-e58a-42aa-a26e-32b1ce124ccb tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] [instance: 324f7e89-3c17-474d-af1b-6d7a6d042510] Powered on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 906.456746] env[63515]: INFO nova.compute.manager [None req-ceee2529-e58a-42aa-a26e-32b1ce124ccb tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] [instance: 324f7e89-3c17-474d-af1b-6d7a6d042510] Took 12.10 seconds to spawn the instance on the hypervisor. [ 906.457043] env[63515]: DEBUG nova.compute.manager [None req-ceee2529-e58a-42aa-a26e-32b1ce124ccb tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] [instance: 324f7e89-3c17-474d-af1b-6d7a6d042510] Checking state {{(pid=63515) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 906.458290] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9e33223-093c-41b5-bd9f-c6f3481280a0 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.462866] env[63515]: DEBUG nova.policy [None req-29ee5601-80a8-4cbb-90e9-98d09454abfa tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c7dc08223aa943e1968a9c62da3a117d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '12a90929c98e40d9a1263929e2b97532', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63515) authorize /opt/stack/nova/nova/policy.py:201}} [ 906.691522] env[63515]: DEBUG oslo_vmware.api [None req-3173d497-cfe4-4001-b365-7bedc6c45742 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Task: {'id': task-1111524, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.473784} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 906.692156] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-3173d497-cfe4-4001-b365-7bedc6c45742 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk to [datastore1] 829e72d7-a56a-451c-b98a-94253e5900d5/829e72d7-a56a-451c-b98a-94253e5900d5.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 906.693132] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-3173d497-cfe4-4001-b365-7bedc6c45742 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: 829e72d7-a56a-451c-b98a-94253e5900d5] Extending root virtual disk to 1048576 {{(pid=63515) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 906.693132] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-907b963a-85e3-4c03-a883-6087fe154fc7 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.702025] env[63515]: DEBUG oslo_vmware.api [None req-3173d497-cfe4-4001-b365-7bedc6c45742 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Waiting for the task: (returnval){ [ 906.702025] env[63515]: value = "task-1111526" [ 906.702025] env[63515]: _type = "Task" [ 906.702025] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 906.710406] env[63515]: DEBUG oslo_vmware.api [None req-3173d497-cfe4-4001-b365-7bedc6c45742 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Task: {'id': task-1111526, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.748908] env[63515]: DEBUG oslo_vmware.api [None req-ef0f5ab2-1246-4de1-af39-f1efd470cd76 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Task: {'id': task-1111525, 'name': RemoveSnapshot_Task} progress is 100%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.833931] env[63515]: DEBUG nova.network.neutron [None req-29ee5601-80a8-4cbb-90e9-98d09454abfa tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] [instance: 86b1ddcf-585c-41d9-84fe-17b794f1abbc] Successfully created port: 43c4b108-fec8-46ea-942f-d3e734fe2be6 {{(pid=63515) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 906.920923] env[63515]: DEBUG nova.compute.manager [None req-29ee5601-80a8-4cbb-90e9-98d09454abfa tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] [instance: 86b1ddcf-585c-41d9-84fe-17b794f1abbc] Start building block device mappings for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 906.989625] env[63515]: INFO nova.compute.manager [None req-ceee2529-e58a-42aa-a26e-32b1ce124ccb tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] [instance: 324f7e89-3c17-474d-af1b-6d7a6d042510] Took 24.20 seconds to build instance. [ 907.199039] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-424bb6b7-69f6-48fb-9dbf-3405886eab63 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.215196] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ba1785e-3789-4ae3-ba42-945976ea28f0 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.219789] env[63515]: DEBUG oslo_vmware.api [None req-3173d497-cfe4-4001-b365-7bedc6c45742 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Task: {'id': task-1111526, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.073775} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 907.220159] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-3173d497-cfe4-4001-b365-7bedc6c45742 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: 829e72d7-a56a-451c-b98a-94253e5900d5] Extended root virtual disk {{(pid=63515) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 907.221665] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24d739f8-406f-46e2-9d43-0c845f0441f3 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.271089] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-129016a6-fa1a-4dc9-ad7d-0aa8b5836592 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.297708] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-3173d497-cfe4-4001-b365-7bedc6c45742 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: 829e72d7-a56a-451c-b98a-94253e5900d5] Reconfiguring VM instance instance-00000050 to attach disk [datastore1] 829e72d7-a56a-451c-b98a-94253e5900d5/829e72d7-a56a-451c-b98a-94253e5900d5.vmdk or device None with type sparse {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 907.298597] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-27ac4462-a7df-4c81-a302-9f97b67301a5 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.317034] env[63515]: DEBUG oslo_vmware.api [None req-ef0f5ab2-1246-4de1-af39-f1efd470cd76 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Task: {'id': task-1111525, 'name': RemoveSnapshot_Task, 'duration_secs': 0.526542} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 907.318510] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6cd70a21-9eff-46ea-9126-e65aec04d8ce {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.322089] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-ef0f5ab2-1246-4de1-af39-f1efd470cd76 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: ae5d6f34-bac6-4ba0-b0c3-8205982ddc88] Deleted Snapshot of the VM instance {{(pid=63515) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 907.322390] env[63515]: DEBUG nova.compute.manager [None req-ef0f5ab2-1246-4de1-af39-f1efd470cd76 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: ae5d6f34-bac6-4ba0-b0c3-8205982ddc88] Checking state {{(pid=63515) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 907.324125] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9862bf38-00b3-450c-bdf2-be6e6d38b7cc {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.326192] env[63515]: DEBUG oslo_vmware.api [None req-3173d497-cfe4-4001-b365-7bedc6c45742 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Waiting for the task: (returnval){ [ 907.326192] env[63515]: value = "task-1111527" [ 907.326192] env[63515]: _type = "Task" [ 907.326192] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 907.338031] env[63515]: DEBUG nova.compute.provider_tree [None req-8c6d15be-1f5e-44a9-a3e4-dab9aa82e26a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Updating inventory in ProviderTree for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 907.346191] env[63515]: DEBUG oslo_vmware.api [None req-3173d497-cfe4-4001-b365-7bedc6c45742 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Task: {'id': task-1111527, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.492246] env[63515]: DEBUG oslo_concurrency.lockutils [None req-ceee2529-e58a-42aa-a26e-32b1ce124ccb tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] Lock "324f7e89-3c17-474d-af1b-6d7a6d042510" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 25.713s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 907.511160] env[63515]: DEBUG oslo_concurrency.lockutils [None req-358377cb-a02d-42b1-aec7-e1ec3bad941d tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] Acquiring lock "324f7e89-3c17-474d-af1b-6d7a6d042510" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 907.511477] env[63515]: DEBUG oslo_concurrency.lockutils [None req-358377cb-a02d-42b1-aec7-e1ec3bad941d tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] Lock "324f7e89-3c17-474d-af1b-6d7a6d042510" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 907.511681] env[63515]: DEBUG oslo_concurrency.lockutils [None req-358377cb-a02d-42b1-aec7-e1ec3bad941d tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] Acquiring lock "324f7e89-3c17-474d-af1b-6d7a6d042510-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 907.511880] env[63515]: DEBUG oslo_concurrency.lockutils [None req-358377cb-a02d-42b1-aec7-e1ec3bad941d tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] Lock "324f7e89-3c17-474d-af1b-6d7a6d042510-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 907.512066] env[63515]: DEBUG oslo_concurrency.lockutils [None req-358377cb-a02d-42b1-aec7-e1ec3bad941d tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] Lock "324f7e89-3c17-474d-af1b-6d7a6d042510-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 907.514304] env[63515]: INFO nova.compute.manager [None req-358377cb-a02d-42b1-aec7-e1ec3bad941d tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] [instance: 324f7e89-3c17-474d-af1b-6d7a6d042510] Terminating instance [ 907.518489] env[63515]: DEBUG nova.compute.manager [None req-358377cb-a02d-42b1-aec7-e1ec3bad941d tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] [instance: 324f7e89-3c17-474d-af1b-6d7a6d042510] Start destroying the instance on the hypervisor. {{(pid=63515) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 907.518489] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-358377cb-a02d-42b1-aec7-e1ec3bad941d tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] [instance: 324f7e89-3c17-474d-af1b-6d7a6d042510] Destroying instance {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 907.518489] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1b798f2-67ad-4107-8bb8-1be3849c30c1 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.525167] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-358377cb-a02d-42b1-aec7-e1ec3bad941d tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] [instance: 324f7e89-3c17-474d-af1b-6d7a6d042510] Powering off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 907.525485] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9057303a-4b15-4327-af5a-3735b6184731 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.533971] env[63515]: DEBUG oslo_vmware.api [None req-358377cb-a02d-42b1-aec7-e1ec3bad941d tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] Waiting for the task: (returnval){ [ 907.533971] env[63515]: value = "task-1111528" [ 907.533971] env[63515]: _type = "Task" [ 907.533971] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 907.543022] env[63515]: DEBUG oslo_vmware.api [None req-358377cb-a02d-42b1-aec7-e1ec3bad941d tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] Task: {'id': task-1111528, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.607536] env[63515]: DEBUG nova.network.neutron [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] [instance: 87c468d9-9594-4804-b461-527f01f6118f] Updating instance_info_cache with network_info: [{"id": "d2098868-0475-4b06-bae3-785c608a861e", "address": "fa:16:3e:af:f9:f0", "network": {"id": "09fb20ca-2b17-4d46-b915-6aff85b862d8", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1815529427-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.212", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "19ce3b05eec64475b685faad1fd801f6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ded18042-834c-4792-b3e8-b1c377446432", "external-id": "nsx-vlan-transportzone-293", "segmentation_id": 293, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd2098868-04", "ovs_interfaceid": "d2098868-0475-4b06-bae3-785c608a861e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 907.837028] env[63515]: DEBUG oslo_vmware.api [None req-3173d497-cfe4-4001-b365-7bedc6c45742 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Task: {'id': task-1111527, 'name': ReconfigVM_Task, 'duration_secs': 0.482096} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 907.837028] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-3173d497-cfe4-4001-b365-7bedc6c45742 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: 829e72d7-a56a-451c-b98a-94253e5900d5] Reconfigured VM instance instance-00000050 to attach disk [datastore1] 829e72d7-a56a-451c-b98a-94253e5900d5/829e72d7-a56a-451c-b98a-94253e5900d5.vmdk or device None with type sparse {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 907.837239] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-667085ec-1f67-46a1-8a27-fb171e01d73d {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.845950] env[63515]: DEBUG oslo_vmware.api [None req-3173d497-cfe4-4001-b365-7bedc6c45742 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Waiting for the task: (returnval){ [ 907.845950] env[63515]: value = "task-1111529" [ 907.845950] env[63515]: _type = "Task" [ 907.845950] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 907.851757] env[63515]: INFO nova.compute.manager [None req-ef0f5ab2-1246-4de1-af39-f1efd470cd76 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: ae5d6f34-bac6-4ba0-b0c3-8205982ddc88] Shelve offloading [ 907.856604] env[63515]: DEBUG oslo_vmware.api [None req-3173d497-cfe4-4001-b365-7bedc6c45742 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Task: {'id': task-1111529, 'name': Rename_Task} progress is 6%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.857163] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-ef0f5ab2-1246-4de1-af39-f1efd470cd76 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: ae5d6f34-bac6-4ba0-b0c3-8205982ddc88] Powering off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 907.857433] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d30ae916-3f01-4e9d-998d-cb8f5b5cbb6e {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.863105] env[63515]: DEBUG oslo_vmware.api [None req-ef0f5ab2-1246-4de1-af39-f1efd470cd76 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Waiting for the task: (returnval){ [ 907.863105] env[63515]: value = "task-1111530" [ 907.863105] env[63515]: _type = "Task" [ 907.863105] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 907.866623] env[63515]: ERROR nova.scheduler.client.report [None req-8c6d15be-1f5e-44a9-a3e4-dab9aa82e26a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] [req-b9a0a921-3815-4504-84cf-59f74b38d3ff] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 2bd86232-2b6f-44d5-9057-1a3a6b27185a. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-b9a0a921-3815-4504-84cf-59f74b38d3ff"}]} [ 907.874187] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-ef0f5ab2-1246-4de1-af39-f1efd470cd76 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: ae5d6f34-bac6-4ba0-b0c3-8205982ddc88] VM already powered off {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 907.874397] env[63515]: DEBUG nova.compute.manager [None req-ef0f5ab2-1246-4de1-af39-f1efd470cd76 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: ae5d6f34-bac6-4ba0-b0c3-8205982ddc88] Checking state {{(pid=63515) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 907.875597] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f62cd622-69b9-4b80-a854-4b5271117f46 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.881517] env[63515]: DEBUG oslo_concurrency.lockutils [None req-ef0f5ab2-1246-4de1-af39-f1efd470cd76 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Acquiring lock "refresh_cache-ae5d6f34-bac6-4ba0-b0c3-8205982ddc88" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 907.881698] env[63515]: DEBUG oslo_concurrency.lockutils [None req-ef0f5ab2-1246-4de1-af39-f1efd470cd76 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Acquired lock "refresh_cache-ae5d6f34-bac6-4ba0-b0c3-8205982ddc88" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 907.881874] env[63515]: DEBUG nova.network.neutron [None req-ef0f5ab2-1246-4de1-af39-f1efd470cd76 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: ae5d6f34-bac6-4ba0-b0c3-8205982ddc88] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 907.883785] env[63515]: DEBUG nova.scheduler.client.report [None req-8c6d15be-1f5e-44a9-a3e4-dab9aa82e26a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Refreshing inventories for resource provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 907.897955] env[63515]: DEBUG nova.scheduler.client.report [None req-8c6d15be-1f5e-44a9-a3e4-dab9aa82e26a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Updating ProviderTree inventory for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 907.898205] env[63515]: DEBUG nova.compute.provider_tree [None req-8c6d15be-1f5e-44a9-a3e4-dab9aa82e26a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Updating inventory in ProviderTree for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 907.910406] env[63515]: DEBUG nova.scheduler.client.report [None req-8c6d15be-1f5e-44a9-a3e4-dab9aa82e26a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Refreshing aggregate associations for resource provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a, aggregates: None {{(pid=63515) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 907.929268] env[63515]: DEBUG nova.scheduler.client.report [None req-8c6d15be-1f5e-44a9-a3e4-dab9aa82e26a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Refreshing trait associations for resource provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a, traits: COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=63515) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 907.933328] env[63515]: DEBUG nova.compute.manager [None req-29ee5601-80a8-4cbb-90e9-98d09454abfa tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] [instance: 86b1ddcf-585c-41d9-84fe-17b794f1abbc] Start spawning the instance on the hypervisor. {{(pid=63515) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 907.957885] env[63515]: DEBUG nova.virt.hardware [None req-29ee5601-80a8-4cbb-90e9-98d09454abfa tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T02:52:14Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T02:51:56Z,direct_url=,disk_format='vmdk',id=8a120570-cb06-4099-b262-554ca0ad15c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b569255cc43e42e7ae3f2b4ad37c6ef4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T02:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 907.958154] env[63515]: DEBUG nova.virt.hardware [None req-29ee5601-80a8-4cbb-90e9-98d09454abfa tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Flavor limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 907.958341] env[63515]: DEBUG nova.virt.hardware [None req-29ee5601-80a8-4cbb-90e9-98d09454abfa tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Image limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 907.958553] env[63515]: DEBUG nova.virt.hardware [None req-29ee5601-80a8-4cbb-90e9-98d09454abfa tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Flavor pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 907.958707] env[63515]: DEBUG nova.virt.hardware [None req-29ee5601-80a8-4cbb-90e9-98d09454abfa tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Image pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 907.958859] env[63515]: DEBUG nova.virt.hardware [None req-29ee5601-80a8-4cbb-90e9-98d09454abfa tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 907.959089] env[63515]: DEBUG nova.virt.hardware [None req-29ee5601-80a8-4cbb-90e9-98d09454abfa tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 907.959258] env[63515]: DEBUG nova.virt.hardware [None req-29ee5601-80a8-4cbb-90e9-98d09454abfa tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 907.959426] env[63515]: DEBUG nova.virt.hardware [None req-29ee5601-80a8-4cbb-90e9-98d09454abfa tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Got 1 possible topologies {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 907.959590] env[63515]: DEBUG nova.virt.hardware [None req-29ee5601-80a8-4cbb-90e9-98d09454abfa tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 907.959762] env[63515]: DEBUG nova.virt.hardware [None req-29ee5601-80a8-4cbb-90e9-98d09454abfa tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 907.961269] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f896dee-ee83-4645-8c6c-8980e9c7d71d {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.971846] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7fce84ac-f6ef-4fe7-a5a2-72df19d6f16a {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.042365] env[63515]: DEBUG oslo_vmware.api [None req-358377cb-a02d-42b1-aec7-e1ec3bad941d tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] Task: {'id': task-1111528, 'name': PowerOffVM_Task, 'duration_secs': 0.494954} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 908.042644] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-358377cb-a02d-42b1-aec7-e1ec3bad941d tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] [instance: 324f7e89-3c17-474d-af1b-6d7a6d042510] Powered off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 908.042820] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-358377cb-a02d-42b1-aec7-e1ec3bad941d tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] [instance: 324f7e89-3c17-474d-af1b-6d7a6d042510] Unregistering the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 908.043084] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c4efa20a-432a-4fdf-8f84-dd7e2d32439f {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.110578] env[63515]: DEBUG oslo_concurrency.lockutils [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Releasing lock "refresh_cache-87c468d9-9594-4804-b461-527f01f6118f" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 908.110793] env[63515]: DEBUG nova.compute.manager [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] [instance: 87c468d9-9594-4804-b461-527f01f6118f] Updated the network info_cache for instance {{(pid=63515) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9997}} [ 908.110997] env[63515]: DEBUG oslo_service.periodic_task [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=63515) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 908.111220] env[63515]: DEBUG oslo_service.periodic_task [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=63515) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 908.111402] env[63515]: DEBUG oslo_service.periodic_task [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=63515) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 908.111560] env[63515]: DEBUG oslo_service.periodic_task [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=63515) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 908.111797] env[63515]: DEBUG oslo_service.periodic_task [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=63515) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 908.112026] env[63515]: DEBUG oslo_service.periodic_task [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=63515) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 908.112144] env[63515]: DEBUG nova.compute.manager [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=63515) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10545}} [ 908.112276] env[63515]: DEBUG oslo_service.periodic_task [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Running periodic task ComputeManager.update_available_resource {{(pid=63515) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 908.162064] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-870acf39-c64a-4cbb-bb76-5fecf85dafd5 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.170937] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04b2f9de-4b52-4ee1-b6a2-db084cf4d4b2 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.205138] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66e5d270-d0dd-4938-958a-871617f015d1 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.207859] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-358377cb-a02d-42b1-aec7-e1ec3bad941d tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] [instance: 324f7e89-3c17-474d-af1b-6d7a6d042510] Unregistered the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 908.208076] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-358377cb-a02d-42b1-aec7-e1ec3bad941d tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] [instance: 324f7e89-3c17-474d-af1b-6d7a6d042510] Deleting contents of the VM from datastore datastore2 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 908.208344] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-358377cb-a02d-42b1-aec7-e1ec3bad941d tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] Deleting the datastore file [datastore2] 324f7e89-3c17-474d-af1b-6d7a6d042510 {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 908.208547] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4045fa0b-ec7b-4658-ba87-244ea8bc0db1 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.215845] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-050c115c-311b-42ce-a5ff-00bb3af13a53 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.220453] env[63515]: DEBUG oslo_vmware.api [None req-358377cb-a02d-42b1-aec7-e1ec3bad941d tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] Waiting for the task: (returnval){ [ 908.220453] env[63515]: value = "task-1111532" [ 908.220453] env[63515]: _type = "Task" [ 908.220453] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 908.232311] env[63515]: DEBUG nova.compute.provider_tree [None req-8c6d15be-1f5e-44a9-a3e4-dab9aa82e26a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Updating inventory in ProviderTree for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 908.238509] env[63515]: DEBUG oslo_vmware.api [None req-358377cb-a02d-42b1-aec7-e1ec3bad941d tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] Task: {'id': task-1111532, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.356540] env[63515]: DEBUG oslo_vmware.api [None req-3173d497-cfe4-4001-b365-7bedc6c45742 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Task: {'id': task-1111529, 'name': Rename_Task, 'duration_secs': 0.239155} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 908.356916] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-3173d497-cfe4-4001-b365-7bedc6c45742 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: 829e72d7-a56a-451c-b98a-94253e5900d5] Powering on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 908.357231] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-957b6d88-9312-4825-bbfd-2a702b2cdac3 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.363660] env[63515]: DEBUG oslo_vmware.api [None req-3173d497-cfe4-4001-b365-7bedc6c45742 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Waiting for the task: (returnval){ [ 908.363660] env[63515]: value = "task-1111533" [ 908.363660] env[63515]: _type = "Task" [ 908.363660] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 908.372309] env[63515]: DEBUG oslo_vmware.api [None req-3173d497-cfe4-4001-b365-7bedc6c45742 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Task: {'id': task-1111533, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.549008] env[63515]: DEBUG nova.compute.manager [req-3ff9a2dd-577d-46ba-8af6-54ace8843444 req-d8697fb3-b33a-40e8-bfa6-0b92c253ba93 service nova] [instance: 86b1ddcf-585c-41d9-84fe-17b794f1abbc] Received event network-vif-plugged-43c4b108-fec8-46ea-942f-d3e734fe2be6 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 908.549254] env[63515]: DEBUG oslo_concurrency.lockutils [req-3ff9a2dd-577d-46ba-8af6-54ace8843444 req-d8697fb3-b33a-40e8-bfa6-0b92c253ba93 service nova] Acquiring lock "86b1ddcf-585c-41d9-84fe-17b794f1abbc-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 908.549495] env[63515]: DEBUG oslo_concurrency.lockutils [req-3ff9a2dd-577d-46ba-8af6-54ace8843444 req-d8697fb3-b33a-40e8-bfa6-0b92c253ba93 service nova] Lock "86b1ddcf-585c-41d9-84fe-17b794f1abbc-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 908.549672] env[63515]: DEBUG oslo_concurrency.lockutils [req-3ff9a2dd-577d-46ba-8af6-54ace8843444 req-d8697fb3-b33a-40e8-bfa6-0b92c253ba93 service nova] Lock "86b1ddcf-585c-41d9-84fe-17b794f1abbc-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 908.549785] env[63515]: DEBUG nova.compute.manager [req-3ff9a2dd-577d-46ba-8af6-54ace8843444 req-d8697fb3-b33a-40e8-bfa6-0b92c253ba93 service nova] [instance: 86b1ddcf-585c-41d9-84fe-17b794f1abbc] No waiting events found dispatching network-vif-plugged-43c4b108-fec8-46ea-942f-d3e734fe2be6 {{(pid=63515) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 908.549950] env[63515]: WARNING nova.compute.manager [req-3ff9a2dd-577d-46ba-8af6-54ace8843444 req-d8697fb3-b33a-40e8-bfa6-0b92c253ba93 service nova] [instance: 86b1ddcf-585c-41d9-84fe-17b794f1abbc] Received unexpected event network-vif-plugged-43c4b108-fec8-46ea-942f-d3e734fe2be6 for instance with vm_state building and task_state spawning. [ 908.617671] env[63515]: DEBUG oslo_concurrency.lockutils [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 908.632292] env[63515]: DEBUG nova.network.neutron [None req-29ee5601-80a8-4cbb-90e9-98d09454abfa tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] [instance: 86b1ddcf-585c-41d9-84fe-17b794f1abbc] Successfully updated port: 43c4b108-fec8-46ea-942f-d3e734fe2be6 {{(pid=63515) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 908.653881] env[63515]: DEBUG nova.network.neutron [None req-ef0f5ab2-1246-4de1-af39-f1efd470cd76 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: ae5d6f34-bac6-4ba0-b0c3-8205982ddc88] Updating instance_info_cache with network_info: [{"id": "dbc4bf29-34a5-424b-bf71-8c80ba74be85", "address": "fa:16:3e:af:57:b6", "network": {"id": "3e5914c8-0cd4-49ee-81b4-1f3e280440bb", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-641632277-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ea099ce07bfb4a8da014d9303cf552eb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "225b6979-9329-403b-91fa-138bd41f6e83", "external-id": "nsx-vlan-transportzone-38", "segmentation_id": 38, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdbc4bf29-34", "ovs_interfaceid": "dbc4bf29-34a5-424b-bf71-8c80ba74be85", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 908.730637] env[63515]: DEBUG oslo_vmware.api [None req-358377cb-a02d-42b1-aec7-e1ec3bad941d tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] Task: {'id': task-1111532, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.161969} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 908.730932] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-358377cb-a02d-42b1-aec7-e1ec3bad941d tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] Deleted the datastore file {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 908.731151] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-358377cb-a02d-42b1-aec7-e1ec3bad941d tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] [instance: 324f7e89-3c17-474d-af1b-6d7a6d042510] Deleted contents of the VM from datastore datastore2 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 908.731362] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-358377cb-a02d-42b1-aec7-e1ec3bad941d tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] [instance: 324f7e89-3c17-474d-af1b-6d7a6d042510] Instance destroyed {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 908.731549] env[63515]: INFO nova.compute.manager [None req-358377cb-a02d-42b1-aec7-e1ec3bad941d tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] [instance: 324f7e89-3c17-474d-af1b-6d7a6d042510] Took 1.22 seconds to destroy the instance on the hypervisor. [ 908.731810] env[63515]: DEBUG oslo.service.loopingcall [None req-358377cb-a02d-42b1-aec7-e1ec3bad941d tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 908.732027] env[63515]: DEBUG nova.compute.manager [-] [instance: 324f7e89-3c17-474d-af1b-6d7a6d042510] Deallocating network for instance {{(pid=63515) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 908.732574] env[63515]: DEBUG nova.network.neutron [-] [instance: 324f7e89-3c17-474d-af1b-6d7a6d042510] deallocate_for_instance() {{(pid=63515) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 908.767490] env[63515]: DEBUG nova.scheduler.client.report [None req-8c6d15be-1f5e-44a9-a3e4-dab9aa82e26a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Updated inventory for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a with generation 110 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 908.767706] env[63515]: DEBUG nova.compute.provider_tree [None req-8c6d15be-1f5e-44a9-a3e4-dab9aa82e26a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Updating resource provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a generation from 110 to 111 during operation: update_inventory {{(pid=63515) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 908.767793] env[63515]: DEBUG nova.compute.provider_tree [None req-8c6d15be-1f5e-44a9-a3e4-dab9aa82e26a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Updating inventory in ProviderTree for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 908.874808] env[63515]: DEBUG oslo_vmware.api [None req-3173d497-cfe4-4001-b365-7bedc6c45742 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Task: {'id': task-1111533, 'name': PowerOnVM_Task} progress is 76%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.218102] env[63515]: DEBUG oslo_concurrency.lockutils [None req-29ee5601-80a8-4cbb-90e9-98d09454abfa tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Acquiring lock "refresh_cache-86b1ddcf-585c-41d9-84fe-17b794f1abbc" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 909.218102] env[63515]: DEBUG oslo_concurrency.lockutils [None req-29ee5601-80a8-4cbb-90e9-98d09454abfa tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Acquired lock "refresh_cache-86b1ddcf-585c-41d9-84fe-17b794f1abbc" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 909.218102] env[63515]: DEBUG nova.network.neutron [None req-29ee5601-80a8-4cbb-90e9-98d09454abfa tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] [instance: 86b1ddcf-585c-41d9-84fe-17b794f1abbc] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 909.218102] env[63515]: DEBUG oslo_concurrency.lockutils [None req-ef0f5ab2-1246-4de1-af39-f1efd470cd76 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Releasing lock "refresh_cache-ae5d6f34-bac6-4ba0-b0c3-8205982ddc88" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 909.275170] env[63515]: DEBUG oslo_concurrency.lockutils [None req-8c6d15be-1f5e-44a9-a3e4-dab9aa82e26a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.369s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 909.275636] env[63515]: DEBUG nova.compute.manager [None req-8c6d15be-1f5e-44a9-a3e4-dab9aa82e26a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] [instance: 9236607d-47af-4272-93db-79826e918aa1] Start building networks asynchronously for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 909.280954] env[63515]: DEBUG oslo_concurrency.lockutils [None req-8c6d15be-1f5e-44a9-a3e4-dab9aa82e26a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 7.645s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 909.283260] env[63515]: INFO nova.compute.claims [None req-8c6d15be-1f5e-44a9-a3e4-dab9aa82e26a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] [instance: ca6a8ac3-d880-4333-9816-2f076e423844] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 909.309721] env[63515]: DEBUG nova.compute.manager [req-93e88016-27fb-47ed-8299-1b37d49dd8d0 req-ced6455d-c153-4132-97c6-6b583c58ce66 service nova] [instance: 324f7e89-3c17-474d-af1b-6d7a6d042510] Received event network-vif-deleted-c18e8fdc-6f1f-4fd6-8401-cab93bb67126 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 909.309927] env[63515]: INFO nova.compute.manager [req-93e88016-27fb-47ed-8299-1b37d49dd8d0 req-ced6455d-c153-4132-97c6-6b583c58ce66 service nova] [instance: 324f7e89-3c17-474d-af1b-6d7a6d042510] Neutron deleted interface c18e8fdc-6f1f-4fd6-8401-cab93bb67126; detaching it from the instance and deleting it from the info cache [ 909.310212] env[63515]: DEBUG nova.network.neutron [req-93e88016-27fb-47ed-8299-1b37d49dd8d0 req-ced6455d-c153-4132-97c6-6b583c58ce66 service nova] [instance: 324f7e89-3c17-474d-af1b-6d7a6d042510] Updating instance_info_cache with network_info: [{"id": "0ea22959-737c-478d-84d8-17fbb175b186", "address": "fa:16:3e:62:f2:c0", "network": {"id": "a3055a87-44b4-4b02-ae7d-87eb877088ba", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-343503251", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.37", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "26a52fb1944a4f949f6779829e163eaa", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "86a35d07-53d3-46b3-92cb-ae34236c0f41", "external-id": "nsx-vlan-transportzone-811", "segmentation_id": 811, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0ea22959-73", "ovs_interfaceid": "0ea22959-737c-478d-84d8-17fbb175b186", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 909.375232] env[63515]: DEBUG oslo_vmware.api [None req-3173d497-cfe4-4001-b365-7bedc6c45742 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Task: {'id': task-1111533, 'name': PowerOnVM_Task, 'duration_secs': 0.685036} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 909.375505] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-3173d497-cfe4-4001-b365-7bedc6c45742 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: 829e72d7-a56a-451c-b98a-94253e5900d5] Powered on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 909.375699] env[63515]: INFO nova.compute.manager [None req-3173d497-cfe4-4001-b365-7bedc6c45742 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: 829e72d7-a56a-451c-b98a-94253e5900d5] Took 7.75 seconds to spawn the instance on the hypervisor. [ 909.375872] env[63515]: DEBUG nova.compute.manager [None req-3173d497-cfe4-4001-b365-7bedc6c45742 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: 829e72d7-a56a-451c-b98a-94253e5900d5] Checking state {{(pid=63515) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 909.376664] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-490f5cf4-8427-466d-921e-8e58152a8b03 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.390266] env[63515]: DEBUG oslo_concurrency.lockutils [None req-9788214a-e67e-4df5-bfb0-11119b5fe460 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Acquiring lock "2a29f37d-e786-48d3-b126-4854d6e7c67c" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 909.390502] env[63515]: DEBUG oslo_concurrency.lockutils [None req-9788214a-e67e-4df5-bfb0-11119b5fe460 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Lock "2a29f37d-e786-48d3-b126-4854d6e7c67c" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 909.653162] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-ef0f5ab2-1246-4de1-af39-f1efd470cd76 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: ae5d6f34-bac6-4ba0-b0c3-8205982ddc88] Destroying instance {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 909.654196] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b42a2bae-c1ca-4a3e-8ab6-dddbf8ad4ac0 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.662263] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-ef0f5ab2-1246-4de1-af39-f1efd470cd76 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: ae5d6f34-bac6-4ba0-b0c3-8205982ddc88] Unregistering the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 909.662521] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b34e2f77-d7a9-4428-ae39-a483af2caf3a {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.728293] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-ef0f5ab2-1246-4de1-af39-f1efd470cd76 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: ae5d6f34-bac6-4ba0-b0c3-8205982ddc88] Unregistered the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 909.728538] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-ef0f5ab2-1246-4de1-af39-f1efd470cd76 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: ae5d6f34-bac6-4ba0-b0c3-8205982ddc88] Deleting contents of the VM from datastore datastore2 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 909.728649] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-ef0f5ab2-1246-4de1-af39-f1efd470cd76 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Deleting the datastore file [datastore2] ae5d6f34-bac6-4ba0-b0c3-8205982ddc88 {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 909.728911] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7c3e4f00-04e2-43a7-8262-550e9704fa44 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.734832] env[63515]: DEBUG oslo_vmware.api [None req-ef0f5ab2-1246-4de1-af39-f1efd470cd76 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Waiting for the task: (returnval){ [ 909.734832] env[63515]: value = "task-1111535" [ 909.734832] env[63515]: _type = "Task" [ 909.734832] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 909.742292] env[63515]: DEBUG oslo_vmware.api [None req-ef0f5ab2-1246-4de1-af39-f1efd470cd76 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Task: {'id': task-1111535, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.766160] env[63515]: DEBUG nova.network.neutron [None req-29ee5601-80a8-4cbb-90e9-98d09454abfa tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] [instance: 86b1ddcf-585c-41d9-84fe-17b794f1abbc] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 909.788127] env[63515]: DEBUG nova.compute.utils [None req-8c6d15be-1f5e-44a9-a3e4-dab9aa82e26a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Using /dev/sd instead of None {{(pid=63515) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 909.789522] env[63515]: DEBUG nova.compute.manager [None req-8c6d15be-1f5e-44a9-a3e4-dab9aa82e26a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] [instance: 9236607d-47af-4272-93db-79826e918aa1] Allocating IP information in the background. {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 909.789697] env[63515]: DEBUG nova.network.neutron [None req-8c6d15be-1f5e-44a9-a3e4-dab9aa82e26a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] [instance: 9236607d-47af-4272-93db-79826e918aa1] allocate_for_instance() {{(pid=63515) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 909.812790] env[63515]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-8b8979e4-a827-49a8-8783-2862e4124b2b {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.821987] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41cad03e-b5da-4faa-9cd2-223b41ecb62c {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.853462] env[63515]: DEBUG nova.compute.manager [req-93e88016-27fb-47ed-8299-1b37d49dd8d0 req-ced6455d-c153-4132-97c6-6b583c58ce66 service nova] [instance: 324f7e89-3c17-474d-af1b-6d7a6d042510] Detach interface failed, port_id=c18e8fdc-6f1f-4fd6-8401-cab93bb67126, reason: Instance 324f7e89-3c17-474d-af1b-6d7a6d042510 could not be found. {{(pid=63515) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 909.855165] env[63515]: DEBUG nova.policy [None req-8c6d15be-1f5e-44a9-a3e4-dab9aa82e26a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e8b57c2415b44bbd93105c6824a9e322', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3dcf6179891c4841a2901fcffe43cd31', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63515) authorize /opt/stack/nova/nova/policy.py:201}} [ 909.895052] env[63515]: DEBUG nova.compute.utils [None req-9788214a-e67e-4df5-bfb0-11119b5fe460 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Using /dev/sd instead of None {{(pid=63515) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 909.900028] env[63515]: INFO nova.compute.manager [None req-3173d497-cfe4-4001-b365-7bedc6c45742 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: 829e72d7-a56a-451c-b98a-94253e5900d5] Took 22.08 seconds to build instance. [ 909.980215] env[63515]: DEBUG nova.network.neutron [None req-29ee5601-80a8-4cbb-90e9-98d09454abfa tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] [instance: 86b1ddcf-585c-41d9-84fe-17b794f1abbc] Updating instance_info_cache with network_info: [{"id": "43c4b108-fec8-46ea-942f-d3e734fe2be6", "address": "fa:16:3e:04:fd:9b", "network": {"id": "57de0b07-78e5-453d-83ba-b12b7dcaf6ce", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-272799484-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "12a90929c98e40d9a1263929e2b97532", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a1895250-76cc-41f7-b7f8-2e5679494607", "external-id": "nsx-vlan-transportzone-785", "segmentation_id": 785, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap43c4b108-fe", "ovs_interfaceid": "43c4b108-fec8-46ea-942f-d3e734fe2be6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 909.986345] env[63515]: DEBUG nova.network.neutron [-] [instance: 324f7e89-3c17-474d-af1b-6d7a6d042510] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 910.100890] env[63515]: DEBUG nova.network.neutron [None req-8c6d15be-1f5e-44a9-a3e4-dab9aa82e26a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] [instance: 9236607d-47af-4272-93db-79826e918aa1] Successfully created port: ca78220d-deb4-41bc-a20b-b92b8fe2cdba {{(pid=63515) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 910.245131] env[63515]: DEBUG oslo_vmware.api [None req-ef0f5ab2-1246-4de1-af39-f1efd470cd76 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Task: {'id': task-1111535, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.192482} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 910.248084] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-ef0f5ab2-1246-4de1-af39-f1efd470cd76 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Deleted the datastore file {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 910.248084] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-ef0f5ab2-1246-4de1-af39-f1efd470cd76 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: ae5d6f34-bac6-4ba0-b0c3-8205982ddc88] Deleted contents of the VM from datastore datastore2 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 910.248084] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-ef0f5ab2-1246-4de1-af39-f1efd470cd76 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: ae5d6f34-bac6-4ba0-b0c3-8205982ddc88] Instance destroyed {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 910.280223] env[63515]: INFO nova.scheduler.client.report [None req-ef0f5ab2-1246-4de1-af39-f1efd470cd76 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Deleted allocations for instance ae5d6f34-bac6-4ba0-b0c3-8205982ddc88 [ 910.292738] env[63515]: DEBUG nova.compute.manager [None req-8c6d15be-1f5e-44a9-a3e4-dab9aa82e26a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] [instance: 9236607d-47af-4272-93db-79826e918aa1] Start building block device mappings for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 910.399061] env[63515]: DEBUG oslo_concurrency.lockutils [None req-3173d497-cfe4-4001-b365-7bedc6c45742 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Lock "829e72d7-a56a-451c-b98a-94253e5900d5" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 23.594s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 910.399867] env[63515]: DEBUG oslo_concurrency.lockutils [None req-9788214a-e67e-4df5-bfb0-11119b5fe460 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Lock "2a29f37d-e786-48d3-b126-4854d6e7c67c" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.009s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 910.483149] env[63515]: DEBUG oslo_concurrency.lockutils [None req-29ee5601-80a8-4cbb-90e9-98d09454abfa tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Releasing lock "refresh_cache-86b1ddcf-585c-41d9-84fe-17b794f1abbc" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 910.483491] env[63515]: DEBUG nova.compute.manager [None req-29ee5601-80a8-4cbb-90e9-98d09454abfa tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] [instance: 86b1ddcf-585c-41d9-84fe-17b794f1abbc] Instance network_info: |[{"id": "43c4b108-fec8-46ea-942f-d3e734fe2be6", "address": "fa:16:3e:04:fd:9b", "network": {"id": "57de0b07-78e5-453d-83ba-b12b7dcaf6ce", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-272799484-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "12a90929c98e40d9a1263929e2b97532", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a1895250-76cc-41f7-b7f8-2e5679494607", "external-id": "nsx-vlan-transportzone-785", "segmentation_id": 785, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap43c4b108-fe", "ovs_interfaceid": "43c4b108-fec8-46ea-942f-d3e734fe2be6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 910.483931] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-29ee5601-80a8-4cbb-90e9-98d09454abfa tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] [instance: 86b1ddcf-585c-41d9-84fe-17b794f1abbc] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:04:fd:9b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a1895250-76cc-41f7-b7f8-2e5679494607', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '43c4b108-fec8-46ea-942f-d3e734fe2be6', 'vif_model': 'vmxnet3'}] {{(pid=63515) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 910.491726] env[63515]: DEBUG oslo.service.loopingcall [None req-29ee5601-80a8-4cbb-90e9-98d09454abfa tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 910.494805] env[63515]: INFO nova.compute.manager [-] [instance: 324f7e89-3c17-474d-af1b-6d7a6d042510] Took 1.76 seconds to deallocate network for instance. [ 910.495057] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 86b1ddcf-585c-41d9-84fe-17b794f1abbc] Creating VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 910.497197] env[63515]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-94729c27-eed2-42fe-9a78-57bfd638fc79 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.520143] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed689dfe-41ad-40e0-b001-c13d87d6ce6b {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.532626] env[63515]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 910.532626] env[63515]: value = "task-1111536" [ 910.532626] env[63515]: _type = "Task" [ 910.532626] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 910.538972] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5acc9c42-d83f-4fab-adc4-c2e1ebd00f92 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.549253] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1111536, 'name': CreateVM_Task} progress is 6%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.576550] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8898dd4-62bb-42c1-b30c-1664006eb30f {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.585672] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-602ee73f-c75d-4221-bd26-12bd797d804e {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.600179] env[63515]: DEBUG nova.compute.provider_tree [None req-8c6d15be-1f5e-44a9-a3e4-dab9aa82e26a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 910.658381] env[63515]: DEBUG nova.compute.manager [req-2c515b24-9571-4cc8-b80d-05f2d76cb0dd req-2bf53ce2-67f9-4882-a86f-bfd3ffc18b98 service nova] [instance: 86b1ddcf-585c-41d9-84fe-17b794f1abbc] Received event network-changed-43c4b108-fec8-46ea-942f-d3e734fe2be6 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 910.658606] env[63515]: DEBUG nova.compute.manager [req-2c515b24-9571-4cc8-b80d-05f2d76cb0dd req-2bf53ce2-67f9-4882-a86f-bfd3ffc18b98 service nova] [instance: 86b1ddcf-585c-41d9-84fe-17b794f1abbc] Refreshing instance network info cache due to event network-changed-43c4b108-fec8-46ea-942f-d3e734fe2be6. {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 910.658831] env[63515]: DEBUG oslo_concurrency.lockutils [req-2c515b24-9571-4cc8-b80d-05f2d76cb0dd req-2bf53ce2-67f9-4882-a86f-bfd3ffc18b98 service nova] Acquiring lock "refresh_cache-86b1ddcf-585c-41d9-84fe-17b794f1abbc" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 910.659411] env[63515]: DEBUG oslo_concurrency.lockutils [req-2c515b24-9571-4cc8-b80d-05f2d76cb0dd req-2bf53ce2-67f9-4882-a86f-bfd3ffc18b98 service nova] Acquired lock "refresh_cache-86b1ddcf-585c-41d9-84fe-17b794f1abbc" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 910.659608] env[63515]: DEBUG nova.network.neutron [req-2c515b24-9571-4cc8-b80d-05f2d76cb0dd req-2bf53ce2-67f9-4882-a86f-bfd3ffc18b98 service nova] [instance: 86b1ddcf-585c-41d9-84fe-17b794f1abbc] Refreshing network info cache for port 43c4b108-fec8-46ea-942f-d3e734fe2be6 {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 910.784774] env[63515]: DEBUG oslo_concurrency.lockutils [None req-ef0f5ab2-1246-4de1-af39-f1efd470cd76 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 910.945512] env[63515]: DEBUG oslo_concurrency.lockutils [None req-1358b8ba-cf2a-4191-bfd4-730211ac5569 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Acquiring lock "87c468d9-9594-4804-b461-527f01f6118f" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 910.945512] env[63515]: DEBUG oslo_concurrency.lockutils [None req-1358b8ba-cf2a-4191-bfd4-730211ac5569 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Lock "87c468d9-9594-4804-b461-527f01f6118f" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 911.024530] env[63515]: DEBUG oslo_concurrency.lockutils [None req-358377cb-a02d-42b1-aec7-e1ec3bad941d tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 911.044600] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1111536, 'name': CreateVM_Task, 'duration_secs': 0.357985} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.044783] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 86b1ddcf-585c-41d9-84fe-17b794f1abbc] Created VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 911.045482] env[63515]: DEBUG oslo_concurrency.lockutils [None req-29ee5601-80a8-4cbb-90e9-98d09454abfa tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 911.045664] env[63515]: DEBUG oslo_concurrency.lockutils [None req-29ee5601-80a8-4cbb-90e9-98d09454abfa tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 911.045986] env[63515]: DEBUG oslo_concurrency.lockutils [None req-29ee5601-80a8-4cbb-90e9-98d09454abfa tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 911.046279] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c9f29682-2d92-46bf-bc6c-87606992057a {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.050971] env[63515]: DEBUG oslo_vmware.api [None req-29ee5601-80a8-4cbb-90e9-98d09454abfa tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Waiting for the task: (returnval){ [ 911.050971] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]5219290f-1cf8-31a8-c445-048ebc1bae2e" [ 911.050971] env[63515]: _type = "Task" [ 911.050971] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 911.059608] env[63515]: DEBUG oslo_vmware.api [None req-29ee5601-80a8-4cbb-90e9-98d09454abfa tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]5219290f-1cf8-31a8-c445-048ebc1bae2e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.104674] env[63515]: DEBUG nova.scheduler.client.report [None req-8c6d15be-1f5e-44a9-a3e4-dab9aa82e26a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 911.304808] env[63515]: DEBUG nova.compute.manager [None req-8c6d15be-1f5e-44a9-a3e4-dab9aa82e26a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] [instance: 9236607d-47af-4272-93db-79826e918aa1] Start spawning the instance on the hypervisor. {{(pid=63515) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 911.329204] env[63515]: DEBUG nova.virt.hardware [None req-8c6d15be-1f5e-44a9-a3e4-dab9aa82e26a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T02:52:14Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T02:51:56Z,direct_url=,disk_format='vmdk',id=8a120570-cb06-4099-b262-554ca0ad15c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b569255cc43e42e7ae3f2b4ad37c6ef4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T02:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 911.329457] env[63515]: DEBUG nova.virt.hardware [None req-8c6d15be-1f5e-44a9-a3e4-dab9aa82e26a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Flavor limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 911.329619] env[63515]: DEBUG nova.virt.hardware [None req-8c6d15be-1f5e-44a9-a3e4-dab9aa82e26a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Image limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 911.329825] env[63515]: DEBUG nova.virt.hardware [None req-8c6d15be-1f5e-44a9-a3e4-dab9aa82e26a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Flavor pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 911.329941] env[63515]: DEBUG nova.virt.hardware [None req-8c6d15be-1f5e-44a9-a3e4-dab9aa82e26a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Image pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 911.330101] env[63515]: DEBUG nova.virt.hardware [None req-8c6d15be-1f5e-44a9-a3e4-dab9aa82e26a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 911.330312] env[63515]: DEBUG nova.virt.hardware [None req-8c6d15be-1f5e-44a9-a3e4-dab9aa82e26a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 911.330473] env[63515]: DEBUG nova.virt.hardware [None req-8c6d15be-1f5e-44a9-a3e4-dab9aa82e26a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 911.330643] env[63515]: DEBUG nova.virt.hardware [None req-8c6d15be-1f5e-44a9-a3e4-dab9aa82e26a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Got 1 possible topologies {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 911.330807] env[63515]: DEBUG nova.virt.hardware [None req-8c6d15be-1f5e-44a9-a3e4-dab9aa82e26a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 911.330983] env[63515]: DEBUG nova.virt.hardware [None req-8c6d15be-1f5e-44a9-a3e4-dab9aa82e26a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 911.332172] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a4fe0ce-06b1-45b0-b0c5-057c7d1e0044 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.337117] env[63515]: DEBUG nova.compute.manager [req-62b2d190-e039-4726-bb4e-5d7f6a3cb74c req-733ece14-eee2-47ce-ba4d-2efda58f9f52 service nova] [instance: 324f7e89-3c17-474d-af1b-6d7a6d042510] Received event network-vif-deleted-0ea22959-737c-478d-84d8-17fbb175b186 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 911.345136] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e027981-4955-4845-b11f-783f00b527f5 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.386112] env[63515]: DEBUG nova.network.neutron [req-2c515b24-9571-4cc8-b80d-05f2d76cb0dd req-2bf53ce2-67f9-4882-a86f-bfd3ffc18b98 service nova] [instance: 86b1ddcf-585c-41d9-84fe-17b794f1abbc] Updated VIF entry in instance network info cache for port 43c4b108-fec8-46ea-942f-d3e734fe2be6. {{(pid=63515) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 911.386564] env[63515]: DEBUG nova.network.neutron [req-2c515b24-9571-4cc8-b80d-05f2d76cb0dd req-2bf53ce2-67f9-4882-a86f-bfd3ffc18b98 service nova] [instance: 86b1ddcf-585c-41d9-84fe-17b794f1abbc] Updating instance_info_cache with network_info: [{"id": "43c4b108-fec8-46ea-942f-d3e734fe2be6", "address": "fa:16:3e:04:fd:9b", "network": {"id": "57de0b07-78e5-453d-83ba-b12b7dcaf6ce", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-272799484-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "12a90929c98e40d9a1263929e2b97532", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a1895250-76cc-41f7-b7f8-2e5679494607", "external-id": "nsx-vlan-transportzone-785", "segmentation_id": 785, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap43c4b108-fe", "ovs_interfaceid": "43c4b108-fec8-46ea-942f-d3e734fe2be6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 911.448090] env[63515]: INFO nova.compute.manager [None req-1358b8ba-cf2a-4191-bfd4-730211ac5569 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] [instance: 87c468d9-9594-4804-b461-527f01f6118f] Detaching volume 67893a17-820c-4c63-b387-06761ba8e0d5 [ 911.462590] env[63515]: DEBUG oslo_concurrency.lockutils [None req-9788214a-e67e-4df5-bfb0-11119b5fe460 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Acquiring lock "2a29f37d-e786-48d3-b126-4854d6e7c67c" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 911.462848] env[63515]: DEBUG oslo_concurrency.lockutils [None req-9788214a-e67e-4df5-bfb0-11119b5fe460 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Lock "2a29f37d-e786-48d3-b126-4854d6e7c67c" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 911.463092] env[63515]: INFO nova.compute.manager [None req-9788214a-e67e-4df5-bfb0-11119b5fe460 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] [instance: 2a29f37d-e786-48d3-b126-4854d6e7c67c] Attaching volume dea169a3-b753-4313-853a-cbf3a4c05ef8 to /dev/sdb [ 911.491806] env[63515]: INFO nova.virt.block_device [None req-1358b8ba-cf2a-4191-bfd4-730211ac5569 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] [instance: 87c468d9-9594-4804-b461-527f01f6118f] Attempting to driver detach volume 67893a17-820c-4c63-b387-06761ba8e0d5 from mountpoint /dev/sdb [ 911.491806] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-1358b8ba-cf2a-4191-bfd4-730211ac5569 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] [instance: 87c468d9-9594-4804-b461-527f01f6118f] Volume detach. Driver type: vmdk {{(pid=63515) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 911.491806] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-1358b8ba-cf2a-4191-bfd4-730211ac5569 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] [instance: 87c468d9-9594-4804-b461-527f01f6118f] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-243476', 'volume_id': '67893a17-820c-4c63-b387-06761ba8e0d5', 'name': 'volume-67893a17-820c-4c63-b387-06761ba8e0d5', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attached', 'instance': '87c468d9-9594-4804-b461-527f01f6118f', 'attached_at': '', 'detached_at': '', 'volume_id': '67893a17-820c-4c63-b387-06761ba8e0d5', 'serial': '67893a17-820c-4c63-b387-06761ba8e0d5'} {{(pid=63515) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 911.491806] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7119a6db-6ce1-4c54-abc6-995e46e40aa8 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.515427] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41ef7930-184a-4f39-89bc-5fdfb599ce70 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.519027] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-026b2bb2-763a-45ef-85aa-ac0d9b8cbe38 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.526542] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-942353a1-31cb-4f23-b06a-a07440e1774e {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.529503] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1eb4454d-39b9-4ec0-b9f1-42e58715c0b9 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.554723] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6895559-99ec-4565-88a1-bb912def4480 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.558364] env[63515]: DEBUG nova.virt.block_device [None req-9788214a-e67e-4df5-bfb0-11119b5fe460 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] [instance: 2a29f37d-e786-48d3-b126-4854d6e7c67c] Updating existing volume attachment record: 64dda96a-6b80-4246-809f-f71ea639405c {{(pid=63515) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 911.576755] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-1358b8ba-cf2a-4191-bfd4-730211ac5569 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] The volume has not been displaced from its original location: [datastore2] volume-67893a17-820c-4c63-b387-06761ba8e0d5/volume-67893a17-820c-4c63-b387-06761ba8e0d5.vmdk. No consolidation needed. {{(pid=63515) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 911.582219] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-1358b8ba-cf2a-4191-bfd4-730211ac5569 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] [instance: 87c468d9-9594-4804-b461-527f01f6118f] Reconfiguring VM instance instance-00000026 to detach disk 2001 {{(pid=63515) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 911.586714] env[63515]: DEBUG oslo_concurrency.lockutils [None req-bd577155-aa2b-4775-93e5-5bcfdcad6cd6 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Acquiring lock "ae5d6f34-bac6-4ba0-b0c3-8205982ddc88" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 911.586997] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6449263b-e479-414a-b3c6-affab9146a6c {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.602058] env[63515]: DEBUG oslo_vmware.api [None req-29ee5601-80a8-4cbb-90e9-98d09454abfa tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]5219290f-1cf8-31a8-c445-048ebc1bae2e, 'name': SearchDatastore_Task, 'duration_secs': 0.01015} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.602058] env[63515]: DEBUG nova.network.neutron [None req-8c6d15be-1f5e-44a9-a3e4-dab9aa82e26a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] [instance: 9236607d-47af-4272-93db-79826e918aa1] Successfully updated port: ca78220d-deb4-41bc-a20b-b92b8fe2cdba {{(pid=63515) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 911.602594] env[63515]: DEBUG oslo_concurrency.lockutils [None req-29ee5601-80a8-4cbb-90e9-98d09454abfa tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 911.602594] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-29ee5601-80a8-4cbb-90e9-98d09454abfa tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] [instance: 86b1ddcf-585c-41d9-84fe-17b794f1abbc] Processing image 8a120570-cb06-4099-b262-554ca0ad15c5 {{(pid=63515) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 911.602770] env[63515]: DEBUG oslo_concurrency.lockutils [None req-29ee5601-80a8-4cbb-90e9-98d09454abfa tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 911.603022] env[63515]: DEBUG oslo_concurrency.lockutils [None req-29ee5601-80a8-4cbb-90e9-98d09454abfa tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 911.603189] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-29ee5601-80a8-4cbb-90e9-98d09454abfa tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 911.604090] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-7c15bc29-2b24-4c1f-8237-e51a397ec2f9 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.612253] env[63515]: DEBUG oslo_concurrency.lockutils [None req-8c6d15be-1f5e-44a9-a3e4-dab9aa82e26a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.332s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 911.612814] env[63515]: DEBUG nova.compute.manager [None req-8c6d15be-1f5e-44a9-a3e4-dab9aa82e26a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] [instance: ca6a8ac3-d880-4333-9816-2f076e423844] Start building networks asynchronously for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 911.615445] env[63515]: DEBUG oslo_vmware.api [None req-1358b8ba-cf2a-4191-bfd4-730211ac5569 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Waiting for the task: (returnval){ [ 911.615445] env[63515]: value = "task-1111537" [ 911.615445] env[63515]: _type = "Task" [ 911.615445] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 911.616690] env[63515]: DEBUG oslo_concurrency.lockutils [None req-1e59a109-1efb-4ebd-8738-fea0e8fa8e8b tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 7.921s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 911.616930] env[63515]: DEBUG nova.objects.instance [None req-1e59a109-1efb-4ebd-8738-fea0e8fa8e8b tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Lazy-loading 'resources' on Instance uuid b1bf5a0a-42e7-41a6-90fb-a0c311623fdd {{(pid=63515) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 911.618185] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-29ee5601-80a8-4cbb-90e9-98d09454abfa tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 911.618376] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-29ee5601-80a8-4cbb-90e9-98d09454abfa tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63515) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 911.622676] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-168244f4-302b-488d-8db6-be1e7a7e7d8d {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.634636] env[63515]: DEBUG oslo_vmware.api [None req-29ee5601-80a8-4cbb-90e9-98d09454abfa tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Waiting for the task: (returnval){ [ 911.634636] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]52091270-6f7d-406d-697e-0b5491e5580e" [ 911.634636] env[63515]: _type = "Task" [ 911.634636] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 911.634810] env[63515]: DEBUG oslo_vmware.api [None req-1358b8ba-cf2a-4191-bfd4-730211ac5569 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Task: {'id': task-1111537, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.644537] env[63515]: DEBUG oslo_vmware.api [None req-29ee5601-80a8-4cbb-90e9-98d09454abfa tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52091270-6f7d-406d-697e-0b5491e5580e, 'name': SearchDatastore_Task, 'duration_secs': 0.009669} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.645359] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7c386201-fc3f-4109-915f-74f0e1cf34d4 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.650591] env[63515]: DEBUG oslo_vmware.api [None req-29ee5601-80a8-4cbb-90e9-98d09454abfa tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Waiting for the task: (returnval){ [ 911.650591] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]5243e88f-f15c-69ec-46dc-d742446e1d54" [ 911.650591] env[63515]: _type = "Task" [ 911.650591] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 911.658663] env[63515]: DEBUG oslo_vmware.api [None req-29ee5601-80a8-4cbb-90e9-98d09454abfa tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]5243e88f-f15c-69ec-46dc-d742446e1d54, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.840164] env[63515]: DEBUG nova.compute.manager [None req-1d5e76d4-1b01-4bdb-831a-7cba26a22dfc tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: 829e72d7-a56a-451c-b98a-94253e5900d5] Stashing vm_state: active {{(pid=63515) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 911.889030] env[63515]: DEBUG oslo_concurrency.lockutils [req-2c515b24-9571-4cc8-b80d-05f2d76cb0dd req-2bf53ce2-67f9-4882-a86f-bfd3ffc18b98 service nova] Releasing lock "refresh_cache-86b1ddcf-585c-41d9-84fe-17b794f1abbc" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 911.889331] env[63515]: DEBUG nova.compute.manager [req-2c515b24-9571-4cc8-b80d-05f2d76cb0dd req-2bf53ce2-67f9-4882-a86f-bfd3ffc18b98 service nova] [instance: ae5d6f34-bac6-4ba0-b0c3-8205982ddc88] Received event network-vif-unplugged-dbc4bf29-34a5-424b-bf71-8c80ba74be85 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 911.889541] env[63515]: DEBUG oslo_concurrency.lockutils [req-2c515b24-9571-4cc8-b80d-05f2d76cb0dd req-2bf53ce2-67f9-4882-a86f-bfd3ffc18b98 service nova] Acquiring lock "ae5d6f34-bac6-4ba0-b0c3-8205982ddc88-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 911.889754] env[63515]: DEBUG oslo_concurrency.lockutils [req-2c515b24-9571-4cc8-b80d-05f2d76cb0dd req-2bf53ce2-67f9-4882-a86f-bfd3ffc18b98 service nova] Lock "ae5d6f34-bac6-4ba0-b0c3-8205982ddc88-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 911.889924] env[63515]: DEBUG oslo_concurrency.lockutils [req-2c515b24-9571-4cc8-b80d-05f2d76cb0dd req-2bf53ce2-67f9-4882-a86f-bfd3ffc18b98 service nova] Lock "ae5d6f34-bac6-4ba0-b0c3-8205982ddc88-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 911.890114] env[63515]: DEBUG nova.compute.manager [req-2c515b24-9571-4cc8-b80d-05f2d76cb0dd req-2bf53ce2-67f9-4882-a86f-bfd3ffc18b98 service nova] [instance: ae5d6f34-bac6-4ba0-b0c3-8205982ddc88] No waiting events found dispatching network-vif-unplugged-dbc4bf29-34a5-424b-bf71-8c80ba74be85 {{(pid=63515) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 911.890296] env[63515]: WARNING nova.compute.manager [req-2c515b24-9571-4cc8-b80d-05f2d76cb0dd req-2bf53ce2-67f9-4882-a86f-bfd3ffc18b98 service nova] [instance: ae5d6f34-bac6-4ba0-b0c3-8205982ddc88] Received unexpected event network-vif-unplugged-dbc4bf29-34a5-424b-bf71-8c80ba74be85 for instance with vm_state shelved_offloaded and task_state None. [ 911.890463] env[63515]: DEBUG nova.compute.manager [req-2c515b24-9571-4cc8-b80d-05f2d76cb0dd req-2bf53ce2-67f9-4882-a86f-bfd3ffc18b98 service nova] [instance: ae5d6f34-bac6-4ba0-b0c3-8205982ddc88] Received event network-changed-dbc4bf29-34a5-424b-bf71-8c80ba74be85 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 911.890625] env[63515]: DEBUG nova.compute.manager [req-2c515b24-9571-4cc8-b80d-05f2d76cb0dd req-2bf53ce2-67f9-4882-a86f-bfd3ffc18b98 service nova] [instance: ae5d6f34-bac6-4ba0-b0c3-8205982ddc88] Refreshing instance network info cache due to event network-changed-dbc4bf29-34a5-424b-bf71-8c80ba74be85. {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 911.890819] env[63515]: DEBUG oslo_concurrency.lockutils [req-2c515b24-9571-4cc8-b80d-05f2d76cb0dd req-2bf53ce2-67f9-4882-a86f-bfd3ffc18b98 service nova] Acquiring lock "refresh_cache-ae5d6f34-bac6-4ba0-b0c3-8205982ddc88" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 911.890957] env[63515]: DEBUG oslo_concurrency.lockutils [req-2c515b24-9571-4cc8-b80d-05f2d76cb0dd req-2bf53ce2-67f9-4882-a86f-bfd3ffc18b98 service nova] Acquired lock "refresh_cache-ae5d6f34-bac6-4ba0-b0c3-8205982ddc88" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 911.891130] env[63515]: DEBUG nova.network.neutron [req-2c515b24-9571-4cc8-b80d-05f2d76cb0dd req-2bf53ce2-67f9-4882-a86f-bfd3ffc18b98 service nova] [instance: ae5d6f34-bac6-4ba0-b0c3-8205982ddc88] Refreshing network info cache for port dbc4bf29-34a5-424b-bf71-8c80ba74be85 {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 912.107321] env[63515]: DEBUG oslo_concurrency.lockutils [None req-8c6d15be-1f5e-44a9-a3e4-dab9aa82e26a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Acquiring lock "refresh_cache-9236607d-47af-4272-93db-79826e918aa1" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 912.107470] env[63515]: DEBUG oslo_concurrency.lockutils [None req-8c6d15be-1f5e-44a9-a3e4-dab9aa82e26a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Acquired lock "refresh_cache-9236607d-47af-4272-93db-79826e918aa1" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 912.107633] env[63515]: DEBUG nova.network.neutron [None req-8c6d15be-1f5e-44a9-a3e4-dab9aa82e26a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] [instance: 9236607d-47af-4272-93db-79826e918aa1] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 912.126089] env[63515]: DEBUG nova.compute.utils [None req-8c6d15be-1f5e-44a9-a3e4-dab9aa82e26a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Using /dev/sd instead of None {{(pid=63515) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 912.127731] env[63515]: DEBUG nova.compute.manager [None req-8c6d15be-1f5e-44a9-a3e4-dab9aa82e26a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] [instance: ca6a8ac3-d880-4333-9816-2f076e423844] Allocating IP information in the background. {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 912.127935] env[63515]: DEBUG nova.network.neutron [None req-8c6d15be-1f5e-44a9-a3e4-dab9aa82e26a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] [instance: ca6a8ac3-d880-4333-9816-2f076e423844] allocate_for_instance() {{(pid=63515) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 912.135558] env[63515]: DEBUG oslo_vmware.api [None req-1358b8ba-cf2a-4191-bfd4-730211ac5569 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Task: {'id': task-1111537, 'name': ReconfigVM_Task, 'duration_secs': 0.222957} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 912.135865] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-1358b8ba-cf2a-4191-bfd4-730211ac5569 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] [instance: 87c468d9-9594-4804-b461-527f01f6118f] Reconfigured VM instance instance-00000026 to detach disk 2001 {{(pid=63515) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 912.141672] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-61ba6df3-7adc-48e6-a199-a1ea4de3ac2e {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.166933] env[63515]: DEBUG oslo_vmware.api [None req-29ee5601-80a8-4cbb-90e9-98d09454abfa tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]5243e88f-f15c-69ec-46dc-d742446e1d54, 'name': SearchDatastore_Task, 'duration_secs': 0.00852} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 912.168214] env[63515]: DEBUG oslo_concurrency.lockutils [None req-29ee5601-80a8-4cbb-90e9-98d09454abfa tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 912.168418] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-29ee5601-80a8-4cbb-90e9-98d09454abfa tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk to [datastore2] 86b1ddcf-585c-41d9-84fe-17b794f1abbc/86b1ddcf-585c-41d9-84fe-17b794f1abbc.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 912.168735] env[63515]: DEBUG oslo_vmware.api [None req-1358b8ba-cf2a-4191-bfd4-730211ac5569 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Waiting for the task: (returnval){ [ 912.168735] env[63515]: value = "task-1111541" [ 912.168735] env[63515]: _type = "Task" [ 912.168735] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 912.173048] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-4c4222fc-d7e1-41d2-8005-d6fdf508e502 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.179137] env[63515]: DEBUG nova.policy [None req-8c6d15be-1f5e-44a9-a3e4-dab9aa82e26a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e8b57c2415b44bbd93105c6824a9e322', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3dcf6179891c4841a2901fcffe43cd31', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63515) authorize /opt/stack/nova/nova/policy.py:201}} [ 912.189020] env[63515]: DEBUG oslo_vmware.api [None req-29ee5601-80a8-4cbb-90e9-98d09454abfa tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Waiting for the task: (returnval){ [ 912.189020] env[63515]: value = "task-1111542" [ 912.189020] env[63515]: _type = "Task" [ 912.189020] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 912.189344] env[63515]: DEBUG oslo_vmware.api [None req-1358b8ba-cf2a-4191-bfd4-730211ac5569 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Task: {'id': task-1111541, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.198367] env[63515]: DEBUG oslo_vmware.api [None req-29ee5601-80a8-4cbb-90e9-98d09454abfa tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Task: {'id': task-1111542, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.360436] env[63515]: DEBUG oslo_concurrency.lockutils [None req-1d5e76d4-1b01-4bdb-831a-7cba26a22dfc tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 912.422469] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b581192b-d9dd-4e1c-af8a-0fa0b3eb47f9 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.435173] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b138840-a982-4162-9d92-18b84576c16c {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.474076] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35c53306-8f42-4753-af0f-368e2ebde0e2 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.486185] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91bfd25b-e474-4107-8991-4876160b601d {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.511793] env[63515]: DEBUG nova.compute.provider_tree [None req-1e59a109-1efb-4ebd-8738-fea0e8fa8e8b tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 912.553621] env[63515]: DEBUG nova.network.neutron [None req-8c6d15be-1f5e-44a9-a3e4-dab9aa82e26a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] [instance: ca6a8ac3-d880-4333-9816-2f076e423844] Successfully created port: 8dcf0639-35d4-47fc-80ef-cc62d15fe62a {{(pid=63515) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 912.631476] env[63515]: DEBUG nova.compute.manager [None req-8c6d15be-1f5e-44a9-a3e4-dab9aa82e26a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] [instance: ca6a8ac3-d880-4333-9816-2f076e423844] Start building block device mappings for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 912.661848] env[63515]: DEBUG nova.network.neutron [None req-8c6d15be-1f5e-44a9-a3e4-dab9aa82e26a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] [instance: 9236607d-47af-4272-93db-79826e918aa1] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 912.686518] env[63515]: DEBUG oslo_vmware.api [None req-1358b8ba-cf2a-4191-bfd4-730211ac5569 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Task: {'id': task-1111541, 'name': ReconfigVM_Task, 'duration_secs': 0.150985} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 912.689023] env[63515]: DEBUG nova.network.neutron [req-2c515b24-9571-4cc8-b80d-05f2d76cb0dd req-2bf53ce2-67f9-4882-a86f-bfd3ffc18b98 service nova] [instance: ae5d6f34-bac6-4ba0-b0c3-8205982ddc88] Updated VIF entry in instance network info cache for port dbc4bf29-34a5-424b-bf71-8c80ba74be85. {{(pid=63515) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 912.689023] env[63515]: DEBUG nova.network.neutron [req-2c515b24-9571-4cc8-b80d-05f2d76cb0dd req-2bf53ce2-67f9-4882-a86f-bfd3ffc18b98 service nova] [instance: ae5d6f34-bac6-4ba0-b0c3-8205982ddc88] Updating instance_info_cache with network_info: [{"id": "dbc4bf29-34a5-424b-bf71-8c80ba74be85", "address": "fa:16:3e:af:57:b6", "network": {"id": "3e5914c8-0cd4-49ee-81b4-1f3e280440bb", "bridge": null, "label": "tempest-DeleteServersTestJSON-641632277-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ea099ce07bfb4a8da014d9303cf552eb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tapdbc4bf29-34", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 912.691795] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-1358b8ba-cf2a-4191-bfd4-730211ac5569 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] [instance: 87c468d9-9594-4804-b461-527f01f6118f] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-243476', 'volume_id': '67893a17-820c-4c63-b387-06761ba8e0d5', 'name': 'volume-67893a17-820c-4c63-b387-06761ba8e0d5', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attached', 'instance': '87c468d9-9594-4804-b461-527f01f6118f', 'attached_at': '', 'detached_at': '', 'volume_id': '67893a17-820c-4c63-b387-06761ba8e0d5', 'serial': '67893a17-820c-4c63-b387-06761ba8e0d5'} {{(pid=63515) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 912.706111] env[63515]: DEBUG oslo_vmware.api [None req-29ee5601-80a8-4cbb-90e9-98d09454abfa tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Task: {'id': task-1111542, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.464966} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 912.706573] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-29ee5601-80a8-4cbb-90e9-98d09454abfa tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk to [datastore2] 86b1ddcf-585c-41d9-84fe-17b794f1abbc/86b1ddcf-585c-41d9-84fe-17b794f1abbc.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 912.706988] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-29ee5601-80a8-4cbb-90e9-98d09454abfa tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] [instance: 86b1ddcf-585c-41d9-84fe-17b794f1abbc] Extending root virtual disk to 1048576 {{(pid=63515) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 912.707644] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b09e6dbc-70fe-4ce7-a5d3-3c67c7f5b0a3 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.713752] env[63515]: DEBUG oslo_vmware.api [None req-29ee5601-80a8-4cbb-90e9-98d09454abfa tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Waiting for the task: (returnval){ [ 912.713752] env[63515]: value = "task-1111543" [ 912.713752] env[63515]: _type = "Task" [ 912.713752] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 912.723081] env[63515]: DEBUG oslo_vmware.api [None req-29ee5601-80a8-4cbb-90e9-98d09454abfa tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Task: {'id': task-1111543, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.869637] env[63515]: DEBUG nova.network.neutron [None req-8c6d15be-1f5e-44a9-a3e4-dab9aa82e26a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] [instance: 9236607d-47af-4272-93db-79826e918aa1] Updating instance_info_cache with network_info: [{"id": "ca78220d-deb4-41bc-a20b-b92b8fe2cdba", "address": "fa:16:3e:9f:78:a4", "network": {"id": "549c4fb2-0ed6-44ee-bfd1-9e59bb7aed27", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-949461070-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3dcf6179891c4841a2901fcffe43cd31", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8c3e2368-4a35-4aa5-9135-23daedbbf9ef", "external-id": "nsx-vlan-transportzone-125", "segmentation_id": 125, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapca78220d-de", "ovs_interfaceid": "ca78220d-deb4-41bc-a20b-b92b8fe2cdba", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 913.015360] env[63515]: DEBUG nova.scheduler.client.report [None req-1e59a109-1efb-4ebd-8738-fea0e8fa8e8b tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 913.196682] env[63515]: DEBUG oslo_concurrency.lockutils [req-2c515b24-9571-4cc8-b80d-05f2d76cb0dd req-2bf53ce2-67f9-4882-a86f-bfd3ffc18b98 service nova] Releasing lock "refresh_cache-ae5d6f34-bac6-4ba0-b0c3-8205982ddc88" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 913.223493] env[63515]: DEBUG oslo_vmware.api [None req-29ee5601-80a8-4cbb-90e9-98d09454abfa tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Task: {'id': task-1111543, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.233861} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 913.223762] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-29ee5601-80a8-4cbb-90e9-98d09454abfa tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] [instance: 86b1ddcf-585c-41d9-84fe-17b794f1abbc] Extended root virtual disk {{(pid=63515) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 913.224559] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60b80bb9-50a4-4bad-b2c2-f41905659973 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.246327] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-29ee5601-80a8-4cbb-90e9-98d09454abfa tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] [instance: 86b1ddcf-585c-41d9-84fe-17b794f1abbc] Reconfiguring VM instance instance-00000051 to attach disk [datastore2] 86b1ddcf-585c-41d9-84fe-17b794f1abbc/86b1ddcf-585c-41d9-84fe-17b794f1abbc.vmdk or device None with type sparse {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 913.248086] env[63515]: DEBUG nova.objects.instance [None req-1358b8ba-cf2a-4191-bfd4-730211ac5569 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Lazy-loading 'flavor' on Instance uuid 87c468d9-9594-4804-b461-527f01f6118f {{(pid=63515) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 913.249376] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ad6f131d-5c4a-4663-bcd5-dd9373b90ed6 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.273668] env[63515]: DEBUG oslo_vmware.api [None req-29ee5601-80a8-4cbb-90e9-98d09454abfa tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Waiting for the task: (returnval){ [ 913.273668] env[63515]: value = "task-1111544" [ 913.273668] env[63515]: _type = "Task" [ 913.273668] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 913.284186] env[63515]: DEBUG oslo_vmware.api [None req-29ee5601-80a8-4cbb-90e9-98d09454abfa tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Task: {'id': task-1111544, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.361491] env[63515]: DEBUG nova.compute.manager [req-6df96903-958e-4f42-92a7-16f696799186 req-aa67aed8-75c4-4df6-9a26-ea1ee7f44ade service nova] [instance: 9236607d-47af-4272-93db-79826e918aa1] Received event network-vif-plugged-ca78220d-deb4-41bc-a20b-b92b8fe2cdba {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 913.361803] env[63515]: DEBUG oslo_concurrency.lockutils [req-6df96903-958e-4f42-92a7-16f696799186 req-aa67aed8-75c4-4df6-9a26-ea1ee7f44ade service nova] Acquiring lock "9236607d-47af-4272-93db-79826e918aa1-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 913.362033] env[63515]: DEBUG oslo_concurrency.lockutils [req-6df96903-958e-4f42-92a7-16f696799186 req-aa67aed8-75c4-4df6-9a26-ea1ee7f44ade service nova] Lock "9236607d-47af-4272-93db-79826e918aa1-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 913.362205] env[63515]: DEBUG oslo_concurrency.lockutils [req-6df96903-958e-4f42-92a7-16f696799186 req-aa67aed8-75c4-4df6-9a26-ea1ee7f44ade service nova] Lock "9236607d-47af-4272-93db-79826e918aa1-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 913.362385] env[63515]: DEBUG nova.compute.manager [req-6df96903-958e-4f42-92a7-16f696799186 req-aa67aed8-75c4-4df6-9a26-ea1ee7f44ade service nova] [instance: 9236607d-47af-4272-93db-79826e918aa1] No waiting events found dispatching network-vif-plugged-ca78220d-deb4-41bc-a20b-b92b8fe2cdba {{(pid=63515) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 913.362557] env[63515]: WARNING nova.compute.manager [req-6df96903-958e-4f42-92a7-16f696799186 req-aa67aed8-75c4-4df6-9a26-ea1ee7f44ade service nova] [instance: 9236607d-47af-4272-93db-79826e918aa1] Received unexpected event network-vif-plugged-ca78220d-deb4-41bc-a20b-b92b8fe2cdba for instance with vm_state building and task_state spawning. [ 913.362798] env[63515]: DEBUG nova.compute.manager [req-6df96903-958e-4f42-92a7-16f696799186 req-aa67aed8-75c4-4df6-9a26-ea1ee7f44ade service nova] [instance: 9236607d-47af-4272-93db-79826e918aa1] Received event network-changed-ca78220d-deb4-41bc-a20b-b92b8fe2cdba {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 913.363080] env[63515]: DEBUG nova.compute.manager [req-6df96903-958e-4f42-92a7-16f696799186 req-aa67aed8-75c4-4df6-9a26-ea1ee7f44ade service nova] [instance: 9236607d-47af-4272-93db-79826e918aa1] Refreshing instance network info cache due to event network-changed-ca78220d-deb4-41bc-a20b-b92b8fe2cdba. {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 913.363165] env[63515]: DEBUG oslo_concurrency.lockutils [req-6df96903-958e-4f42-92a7-16f696799186 req-aa67aed8-75c4-4df6-9a26-ea1ee7f44ade service nova] Acquiring lock "refresh_cache-9236607d-47af-4272-93db-79826e918aa1" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 913.372812] env[63515]: DEBUG oslo_concurrency.lockutils [None req-8c6d15be-1f5e-44a9-a3e4-dab9aa82e26a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Releasing lock "refresh_cache-9236607d-47af-4272-93db-79826e918aa1" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 913.372907] env[63515]: DEBUG nova.compute.manager [None req-8c6d15be-1f5e-44a9-a3e4-dab9aa82e26a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] [instance: 9236607d-47af-4272-93db-79826e918aa1] Instance network_info: |[{"id": "ca78220d-deb4-41bc-a20b-b92b8fe2cdba", "address": "fa:16:3e:9f:78:a4", "network": {"id": "549c4fb2-0ed6-44ee-bfd1-9e59bb7aed27", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-949461070-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3dcf6179891c4841a2901fcffe43cd31", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8c3e2368-4a35-4aa5-9135-23daedbbf9ef", "external-id": "nsx-vlan-transportzone-125", "segmentation_id": 125, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapca78220d-de", "ovs_interfaceid": "ca78220d-deb4-41bc-a20b-b92b8fe2cdba", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 913.373193] env[63515]: DEBUG oslo_concurrency.lockutils [req-6df96903-958e-4f42-92a7-16f696799186 req-aa67aed8-75c4-4df6-9a26-ea1ee7f44ade service nova] Acquired lock "refresh_cache-9236607d-47af-4272-93db-79826e918aa1" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 913.373382] env[63515]: DEBUG nova.network.neutron [req-6df96903-958e-4f42-92a7-16f696799186 req-aa67aed8-75c4-4df6-9a26-ea1ee7f44ade service nova] [instance: 9236607d-47af-4272-93db-79826e918aa1] Refreshing network info cache for port ca78220d-deb4-41bc-a20b-b92b8fe2cdba {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 913.375069] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-8c6d15be-1f5e-44a9-a3e4-dab9aa82e26a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] [instance: 9236607d-47af-4272-93db-79826e918aa1] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:9f:78:a4', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '8c3e2368-4a35-4aa5-9135-23daedbbf9ef', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ca78220d-deb4-41bc-a20b-b92b8fe2cdba', 'vif_model': 'vmxnet3'}] {{(pid=63515) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 913.382596] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-8c6d15be-1f5e-44a9-a3e4-dab9aa82e26a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Creating folder: Project (3dcf6179891c4841a2901fcffe43cd31). Parent ref: group-v243370. {{(pid=63515) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 913.385607] env[63515]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-90ae1cc1-954d-4d02-a175-8a513780675d {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.396452] env[63515]: INFO nova.virt.vmwareapi.vm_util [None req-8c6d15be-1f5e-44a9-a3e4-dab9aa82e26a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Created folder: Project (3dcf6179891c4841a2901fcffe43cd31) in parent group-v243370. [ 913.396725] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-8c6d15be-1f5e-44a9-a3e4-dab9aa82e26a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Creating folder: Instances. Parent ref: group-v243525. {{(pid=63515) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 913.396870] env[63515]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-722b5a49-ffad-45a5-a46d-2a41f7727f19 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.405517] env[63515]: INFO nova.virt.vmwareapi.vm_util [None req-8c6d15be-1f5e-44a9-a3e4-dab9aa82e26a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Created folder: Instances in parent group-v243525. [ 913.405748] env[63515]: DEBUG oslo.service.loopingcall [None req-8c6d15be-1f5e-44a9-a3e4-dab9aa82e26a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 913.405940] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9236607d-47af-4272-93db-79826e918aa1] Creating VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 913.406172] env[63515]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d38d5d3c-c3d2-4b6a-8222-cb34eb518734 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.427590] env[63515]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 913.427590] env[63515]: value = "task-1111547" [ 913.427590] env[63515]: _type = "Task" [ 913.427590] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 913.435462] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1111547, 'name': CreateVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.520131] env[63515]: DEBUG oslo_concurrency.lockutils [None req-1e59a109-1efb-4ebd-8738-fea0e8fa8e8b tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.903s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 913.523310] env[63515]: DEBUG oslo_concurrency.lockutils [None req-0bb813f4-ce4d-4144-b80d-b179b1c5fd20 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 9.406s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 913.525080] env[63515]: INFO nova.compute.claims [None req-0bb813f4-ce4d-4144-b80d-b179b1c5fd20 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: a2203c5e-66b9-4c7e-8d76-e12a08b57b1d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 913.544739] env[63515]: INFO nova.scheduler.client.report [None req-1e59a109-1efb-4ebd-8738-fea0e8fa8e8b tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Deleted allocations for instance b1bf5a0a-42e7-41a6-90fb-a0c311623fdd [ 913.643698] env[63515]: DEBUG nova.compute.manager [None req-8c6d15be-1f5e-44a9-a3e4-dab9aa82e26a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] [instance: ca6a8ac3-d880-4333-9816-2f076e423844] Start spawning the instance on the hypervisor. {{(pid=63515) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 913.651475] env[63515]: DEBUG nova.network.neutron [req-6df96903-958e-4f42-92a7-16f696799186 req-aa67aed8-75c4-4df6-9a26-ea1ee7f44ade service nova] [instance: 9236607d-47af-4272-93db-79826e918aa1] Updated VIF entry in instance network info cache for port ca78220d-deb4-41bc-a20b-b92b8fe2cdba. {{(pid=63515) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 913.651866] env[63515]: DEBUG nova.network.neutron [req-6df96903-958e-4f42-92a7-16f696799186 req-aa67aed8-75c4-4df6-9a26-ea1ee7f44ade service nova] [instance: 9236607d-47af-4272-93db-79826e918aa1] Updating instance_info_cache with network_info: [{"id": "ca78220d-deb4-41bc-a20b-b92b8fe2cdba", "address": "fa:16:3e:9f:78:a4", "network": {"id": "549c4fb2-0ed6-44ee-bfd1-9e59bb7aed27", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-949461070-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3dcf6179891c4841a2901fcffe43cd31", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8c3e2368-4a35-4aa5-9135-23daedbbf9ef", "external-id": "nsx-vlan-transportzone-125", "segmentation_id": 125, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapca78220d-de", "ovs_interfaceid": "ca78220d-deb4-41bc-a20b-b92b8fe2cdba", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 913.669468] env[63515]: DEBUG nova.virt.hardware [None req-8c6d15be-1f5e-44a9-a3e4-dab9aa82e26a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T02:52:14Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T02:51:56Z,direct_url=,disk_format='vmdk',id=8a120570-cb06-4099-b262-554ca0ad15c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b569255cc43e42e7ae3f2b4ad37c6ef4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T02:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 913.669759] env[63515]: DEBUG nova.virt.hardware [None req-8c6d15be-1f5e-44a9-a3e4-dab9aa82e26a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Flavor limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 913.669946] env[63515]: DEBUG nova.virt.hardware [None req-8c6d15be-1f5e-44a9-a3e4-dab9aa82e26a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Image limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 913.670236] env[63515]: DEBUG nova.virt.hardware [None req-8c6d15be-1f5e-44a9-a3e4-dab9aa82e26a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Flavor pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 913.670359] env[63515]: DEBUG nova.virt.hardware [None req-8c6d15be-1f5e-44a9-a3e4-dab9aa82e26a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Image pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 913.670539] env[63515]: DEBUG nova.virt.hardware [None req-8c6d15be-1f5e-44a9-a3e4-dab9aa82e26a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 913.670784] env[63515]: DEBUG nova.virt.hardware [None req-8c6d15be-1f5e-44a9-a3e4-dab9aa82e26a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 913.670975] env[63515]: DEBUG nova.virt.hardware [None req-8c6d15be-1f5e-44a9-a3e4-dab9aa82e26a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 913.671190] env[63515]: DEBUG nova.virt.hardware [None req-8c6d15be-1f5e-44a9-a3e4-dab9aa82e26a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Got 1 possible topologies {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 913.671388] env[63515]: DEBUG nova.virt.hardware [None req-8c6d15be-1f5e-44a9-a3e4-dab9aa82e26a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 913.671584] env[63515]: DEBUG nova.virt.hardware [None req-8c6d15be-1f5e-44a9-a3e4-dab9aa82e26a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 913.672713] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df51d1ff-15e5-43b7-bb8b-416b539822be {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.681038] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7cc769f9-ff65-477a-ae3e-744628bcfa65 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.768799] env[63515]: DEBUG oslo_concurrency.lockutils [None req-1358b8ba-cf2a-4191-bfd4-730211ac5569 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Lock "87c468d9-9594-4804-b461-527f01f6118f" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 2.824s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 913.785334] env[63515]: DEBUG oslo_vmware.api [None req-29ee5601-80a8-4cbb-90e9-98d09454abfa tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Task: {'id': task-1111544, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.937811] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1111547, 'name': CreateVM_Task, 'duration_secs': 0.484419} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 913.938012] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9236607d-47af-4272-93db-79826e918aa1] Created VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 913.938766] env[63515]: DEBUG oslo_concurrency.lockutils [None req-8c6d15be-1f5e-44a9-a3e4-dab9aa82e26a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 913.938938] env[63515]: DEBUG oslo_concurrency.lockutils [None req-8c6d15be-1f5e-44a9-a3e4-dab9aa82e26a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 913.939278] env[63515]: DEBUG oslo_concurrency.lockutils [None req-8c6d15be-1f5e-44a9-a3e4-dab9aa82e26a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 913.939536] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b3c8d0d8-e672-448f-a901-d55808a80862 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.943834] env[63515]: DEBUG oslo_vmware.api [None req-8c6d15be-1f5e-44a9-a3e4-dab9aa82e26a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Waiting for the task: (returnval){ [ 913.943834] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]528b82ac-baf0-b21b-c37f-785a937ac0be" [ 913.943834] env[63515]: _type = "Task" [ 913.943834] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 913.951668] env[63515]: DEBUG oslo_vmware.api [None req-8c6d15be-1f5e-44a9-a3e4-dab9aa82e26a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]528b82ac-baf0-b21b-c37f-785a937ac0be, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.052101] env[63515]: DEBUG oslo_concurrency.lockutils [None req-1e59a109-1efb-4ebd-8738-fea0e8fa8e8b tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Lock "b1bf5a0a-42e7-41a6-90fb-a0c311623fdd" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 13.220s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 914.080731] env[63515]: DEBUG nova.network.neutron [None req-8c6d15be-1f5e-44a9-a3e4-dab9aa82e26a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] [instance: ca6a8ac3-d880-4333-9816-2f076e423844] Successfully updated port: 8dcf0639-35d4-47fc-80ef-cc62d15fe62a {{(pid=63515) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 914.154902] env[63515]: DEBUG oslo_concurrency.lockutils [req-6df96903-958e-4f42-92a7-16f696799186 req-aa67aed8-75c4-4df6-9a26-ea1ee7f44ade service nova] Releasing lock "refresh_cache-9236607d-47af-4272-93db-79826e918aa1" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 914.287377] env[63515]: DEBUG oslo_vmware.api [None req-29ee5601-80a8-4cbb-90e9-98d09454abfa tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Task: {'id': task-1111544, 'name': ReconfigVM_Task, 'duration_secs': 0.672865} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.287666] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-29ee5601-80a8-4cbb-90e9-98d09454abfa tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] [instance: 86b1ddcf-585c-41d9-84fe-17b794f1abbc] Reconfigured VM instance instance-00000051 to attach disk [datastore2] 86b1ddcf-585c-41d9-84fe-17b794f1abbc/86b1ddcf-585c-41d9-84fe-17b794f1abbc.vmdk or device None with type sparse {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 914.288313] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-163b0030-86a9-45c4-98e8-ca7075b2fa59 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.295489] env[63515]: DEBUG oslo_vmware.api [None req-29ee5601-80a8-4cbb-90e9-98d09454abfa tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Waiting for the task: (returnval){ [ 914.295489] env[63515]: value = "task-1111549" [ 914.295489] env[63515]: _type = "Task" [ 914.295489] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.303582] env[63515]: DEBUG oslo_vmware.api [None req-29ee5601-80a8-4cbb-90e9-98d09454abfa tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Task: {'id': task-1111549, 'name': Rename_Task} progress is 5%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.318537] env[63515]: DEBUG oslo_concurrency.lockutils [None req-44e7f82e-18f0-4009-907e-7846ed9a3683 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Acquiring lock "87c468d9-9594-4804-b461-527f01f6118f" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 914.318802] env[63515]: DEBUG oslo_concurrency.lockutils [None req-44e7f82e-18f0-4009-907e-7846ed9a3683 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Lock "87c468d9-9594-4804-b461-527f01f6118f" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 914.319015] env[63515]: DEBUG oslo_concurrency.lockutils [None req-44e7f82e-18f0-4009-907e-7846ed9a3683 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Acquiring lock "87c468d9-9594-4804-b461-527f01f6118f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 914.319208] env[63515]: DEBUG oslo_concurrency.lockutils [None req-44e7f82e-18f0-4009-907e-7846ed9a3683 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Lock "87c468d9-9594-4804-b461-527f01f6118f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 914.319859] env[63515]: DEBUG oslo_concurrency.lockutils [None req-44e7f82e-18f0-4009-907e-7846ed9a3683 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Lock "87c468d9-9594-4804-b461-527f01f6118f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 914.322856] env[63515]: INFO nova.compute.manager [None req-44e7f82e-18f0-4009-907e-7846ed9a3683 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] [instance: 87c468d9-9594-4804-b461-527f01f6118f] Terminating instance [ 914.325054] env[63515]: DEBUG nova.compute.manager [None req-44e7f82e-18f0-4009-907e-7846ed9a3683 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] [instance: 87c468d9-9594-4804-b461-527f01f6118f] Start destroying the instance on the hypervisor. {{(pid=63515) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 914.325993] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-44e7f82e-18f0-4009-907e-7846ed9a3683 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] [instance: 87c468d9-9594-4804-b461-527f01f6118f] Destroying instance {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 914.326212] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4dba593d-84cd-48bf-bdb1-787396181c33 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.333555] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-44e7f82e-18f0-4009-907e-7846ed9a3683 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] [instance: 87c468d9-9594-4804-b461-527f01f6118f] Powering off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 914.333798] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b22d08dd-42ca-4c07-bb76-b9e645666bfe {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.340172] env[63515]: DEBUG oslo_vmware.api [None req-44e7f82e-18f0-4009-907e-7846ed9a3683 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Waiting for the task: (returnval){ [ 914.340172] env[63515]: value = "task-1111550" [ 914.340172] env[63515]: _type = "Task" [ 914.340172] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.347940] env[63515]: DEBUG oslo_vmware.api [None req-44e7f82e-18f0-4009-907e-7846ed9a3683 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Task: {'id': task-1111550, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.454363] env[63515]: DEBUG oslo_vmware.api [None req-8c6d15be-1f5e-44a9-a3e4-dab9aa82e26a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]528b82ac-baf0-b21b-c37f-785a937ac0be, 'name': SearchDatastore_Task, 'duration_secs': 0.007745} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.454781] env[63515]: DEBUG oslo_concurrency.lockutils [None req-8c6d15be-1f5e-44a9-a3e4-dab9aa82e26a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 914.454939] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-8c6d15be-1f5e-44a9-a3e4-dab9aa82e26a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] [instance: 9236607d-47af-4272-93db-79826e918aa1] Processing image 8a120570-cb06-4099-b262-554ca0ad15c5 {{(pid=63515) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 914.455201] env[63515]: DEBUG oslo_concurrency.lockutils [None req-8c6d15be-1f5e-44a9-a3e4-dab9aa82e26a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 914.455403] env[63515]: DEBUG oslo_concurrency.lockutils [None req-8c6d15be-1f5e-44a9-a3e4-dab9aa82e26a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 914.455541] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-8c6d15be-1f5e-44a9-a3e4-dab9aa82e26a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 914.455848] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-01967a32-1fa7-4b40-b39d-369c075c8505 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.464924] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-8c6d15be-1f5e-44a9-a3e4-dab9aa82e26a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 914.464924] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-8c6d15be-1f5e-44a9-a3e4-dab9aa82e26a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63515) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 914.465247] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b4c121ed-3c82-42f4-8b29-dd348f36d7fb {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.470932] env[63515]: DEBUG oslo_vmware.api [None req-8c6d15be-1f5e-44a9-a3e4-dab9aa82e26a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Waiting for the task: (returnval){ [ 914.470932] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]524bfebd-bdc4-cd0d-1c3e-f266f3535ca5" [ 914.470932] env[63515]: _type = "Task" [ 914.470932] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.479161] env[63515]: DEBUG oslo_vmware.api [None req-8c6d15be-1f5e-44a9-a3e4-dab9aa82e26a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]524bfebd-bdc4-cd0d-1c3e-f266f3535ca5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.587522] env[63515]: DEBUG oslo_concurrency.lockutils [None req-8c6d15be-1f5e-44a9-a3e4-dab9aa82e26a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Acquiring lock "refresh_cache-ca6a8ac3-d880-4333-9816-2f076e423844" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 914.587671] env[63515]: DEBUG oslo_concurrency.lockutils [None req-8c6d15be-1f5e-44a9-a3e4-dab9aa82e26a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Acquired lock "refresh_cache-ca6a8ac3-d880-4333-9816-2f076e423844" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 914.587824] env[63515]: DEBUG nova.network.neutron [None req-8c6d15be-1f5e-44a9-a3e4-dab9aa82e26a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] [instance: ca6a8ac3-d880-4333-9816-2f076e423844] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 914.729926] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc62510d-6b41-4e6b-ad0a-8e1325c8c560 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.737776] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d20b28b-79ac-4de5-9e1c-0991412338d3 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.743208] env[63515]: DEBUG oslo_concurrency.lockutils [None req-872cc7bc-b8ac-4eff-87f3-c2c36aad9630 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Acquiring lock "70df69e5-687b-44fb-b6fc-cdb08e21dda0" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 914.743444] env[63515]: DEBUG oslo_concurrency.lockutils [None req-872cc7bc-b8ac-4eff-87f3-c2c36aad9630 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Lock "70df69e5-687b-44fb-b6fc-cdb08e21dda0" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 914.743659] env[63515]: DEBUG oslo_concurrency.lockutils [None req-872cc7bc-b8ac-4eff-87f3-c2c36aad9630 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Acquiring lock "70df69e5-687b-44fb-b6fc-cdb08e21dda0-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 914.743849] env[63515]: DEBUG oslo_concurrency.lockutils [None req-872cc7bc-b8ac-4eff-87f3-c2c36aad9630 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Lock "70df69e5-687b-44fb-b6fc-cdb08e21dda0-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 914.744062] env[63515]: DEBUG oslo_concurrency.lockutils [None req-872cc7bc-b8ac-4eff-87f3-c2c36aad9630 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Lock "70df69e5-687b-44fb-b6fc-cdb08e21dda0-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 914.768831] env[63515]: INFO nova.compute.manager [None req-872cc7bc-b8ac-4eff-87f3-c2c36aad9630 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: 70df69e5-687b-44fb-b6fc-cdb08e21dda0] Terminating instance [ 914.770777] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09698778-0d3a-4e74-b7f1-d5dda81c1934 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.773487] env[63515]: DEBUG nova.compute.manager [None req-872cc7bc-b8ac-4eff-87f3-c2c36aad9630 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: 70df69e5-687b-44fb-b6fc-cdb08e21dda0] Start destroying the instance on the hypervisor. {{(pid=63515) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 914.773681] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-872cc7bc-b8ac-4eff-87f3-c2c36aad9630 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: 70df69e5-687b-44fb-b6fc-cdb08e21dda0] Destroying instance {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 914.774876] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3d46b44-215a-4759-983f-88c1d6be16a4 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.784463] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25afbed5-4a61-41aa-9169-6f4252fda376 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.787954] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-872cc7bc-b8ac-4eff-87f3-c2c36aad9630 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: 70df69e5-687b-44fb-b6fc-cdb08e21dda0] Powering off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 914.788176] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-cb55c346-6f22-4472-ad18-de242122eaf8 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.798950] env[63515]: DEBUG nova.compute.provider_tree [None req-0bb813f4-ce4d-4144-b80d-b179b1c5fd20 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Updating inventory in ProviderTree for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 914.803314] env[63515]: DEBUG oslo_vmware.api [None req-872cc7bc-b8ac-4eff-87f3-c2c36aad9630 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Waiting for the task: (returnval){ [ 914.803314] env[63515]: value = "task-1111551" [ 914.803314] env[63515]: _type = "Task" [ 914.803314] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.809992] env[63515]: DEBUG oslo_vmware.api [None req-29ee5601-80a8-4cbb-90e9-98d09454abfa tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Task: {'id': task-1111549, 'name': Rename_Task, 'duration_secs': 0.148164} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.810555] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-29ee5601-80a8-4cbb-90e9-98d09454abfa tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] [instance: 86b1ddcf-585c-41d9-84fe-17b794f1abbc] Powering on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 914.810781] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1534e240-ca74-4e07-9f57-aa778206ebf6 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.814978] env[63515]: DEBUG oslo_vmware.api [None req-872cc7bc-b8ac-4eff-87f3-c2c36aad9630 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Task: {'id': task-1111551, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.819263] env[63515]: DEBUG oslo_vmware.api [None req-29ee5601-80a8-4cbb-90e9-98d09454abfa tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Waiting for the task: (returnval){ [ 914.819263] env[63515]: value = "task-1111552" [ 914.819263] env[63515]: _type = "Task" [ 914.819263] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.826493] env[63515]: DEBUG oslo_vmware.api [None req-29ee5601-80a8-4cbb-90e9-98d09454abfa tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Task: {'id': task-1111552, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.848452] env[63515]: DEBUG oslo_vmware.api [None req-44e7f82e-18f0-4009-907e-7846ed9a3683 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Task: {'id': task-1111550, 'name': PowerOffVM_Task, 'duration_secs': 0.188461} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.848715] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-44e7f82e-18f0-4009-907e-7846ed9a3683 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] [instance: 87c468d9-9594-4804-b461-527f01f6118f] Powered off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 914.848885] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-44e7f82e-18f0-4009-907e-7846ed9a3683 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] [instance: 87c468d9-9594-4804-b461-527f01f6118f] Unregistering the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 914.849141] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b2a39b91-5f81-4060-98e1-03789fdcec40 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.912495] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-44e7f82e-18f0-4009-907e-7846ed9a3683 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] [instance: 87c468d9-9594-4804-b461-527f01f6118f] Unregistered the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 914.912753] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-44e7f82e-18f0-4009-907e-7846ed9a3683 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] [instance: 87c468d9-9594-4804-b461-527f01f6118f] Deleting contents of the VM from datastore datastore1 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 914.912938] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-44e7f82e-18f0-4009-907e-7846ed9a3683 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Deleting the datastore file [datastore1] 87c468d9-9594-4804-b461-527f01f6118f {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 914.913297] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-62b44862-afcb-4079-ad93-8495c28cee92 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.920270] env[63515]: DEBUG oslo_vmware.api [None req-44e7f82e-18f0-4009-907e-7846ed9a3683 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Waiting for the task: (returnval){ [ 914.920270] env[63515]: value = "task-1111554" [ 914.920270] env[63515]: _type = "Task" [ 914.920270] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.928761] env[63515]: DEBUG oslo_vmware.api [None req-44e7f82e-18f0-4009-907e-7846ed9a3683 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Task: {'id': task-1111554, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.981775] env[63515]: DEBUG oslo_vmware.api [None req-8c6d15be-1f5e-44a9-a3e4-dab9aa82e26a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]524bfebd-bdc4-cd0d-1c3e-f266f3535ca5, 'name': SearchDatastore_Task, 'duration_secs': 0.008801} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.983049] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6531f60f-c9ad-4594-a45a-a83331a169f7 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.988757] env[63515]: DEBUG oslo_vmware.api [None req-8c6d15be-1f5e-44a9-a3e4-dab9aa82e26a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Waiting for the task: (returnval){ [ 914.988757] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]524d9b0b-868a-0653-3030-af5216a757e9" [ 914.988757] env[63515]: _type = "Task" [ 914.988757] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.998795] env[63515]: DEBUG oslo_vmware.api [None req-8c6d15be-1f5e-44a9-a3e4-dab9aa82e26a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]524d9b0b-868a-0653-3030-af5216a757e9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.134120] env[63515]: DEBUG nova.network.neutron [None req-8c6d15be-1f5e-44a9-a3e4-dab9aa82e26a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] [instance: ca6a8ac3-d880-4333-9816-2f076e423844] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 915.231672] env[63515]: DEBUG oslo_concurrency.lockutils [None req-d63820c0-6e1a-45a6-8579-7bfab43bd25a tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Acquiring lock "2fa098b6-ecc0-491f-8cfc-5ee335bbf943" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 915.231981] env[63515]: DEBUG oslo_concurrency.lockutils [None req-d63820c0-6e1a-45a6-8579-7bfab43bd25a tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Lock "2fa098b6-ecc0-491f-8cfc-5ee335bbf943" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 915.274628] env[63515]: DEBUG nova.network.neutron [None req-8c6d15be-1f5e-44a9-a3e4-dab9aa82e26a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] [instance: ca6a8ac3-d880-4333-9816-2f076e423844] Updating instance_info_cache with network_info: [{"id": "8dcf0639-35d4-47fc-80ef-cc62d15fe62a", "address": "fa:16:3e:56:9c:25", "network": {"id": "549c4fb2-0ed6-44ee-bfd1-9e59bb7aed27", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-949461070-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3dcf6179891c4841a2901fcffe43cd31", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8c3e2368-4a35-4aa5-9135-23daedbbf9ef", "external-id": "nsx-vlan-transportzone-125", "segmentation_id": 125, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8dcf0639-35", "ovs_interfaceid": "8dcf0639-35d4-47fc-80ef-cc62d15fe62a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 915.317310] env[63515]: DEBUG oslo_vmware.api [None req-872cc7bc-b8ac-4eff-87f3-c2c36aad9630 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Task: {'id': task-1111551, 'name': PowerOffVM_Task, 'duration_secs': 0.23344} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.317894] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-872cc7bc-b8ac-4eff-87f3-c2c36aad9630 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: 70df69e5-687b-44fb-b6fc-cdb08e21dda0] Powered off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 915.318267] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-872cc7bc-b8ac-4eff-87f3-c2c36aad9630 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: 70df69e5-687b-44fb-b6fc-cdb08e21dda0] Unregistering the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 915.321020] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a02955b8-3c3a-43e0-a370-c3a89113dd11 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.325389] env[63515]: ERROR nova.scheduler.client.report [None req-0bb813f4-ce4d-4144-b80d-b179b1c5fd20 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [req-731a2384-b9ee-4af9-99bc-f5f1f0793b12] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 2bd86232-2b6f-44d5-9057-1a3a6b27185a. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-731a2384-b9ee-4af9-99bc-f5f1f0793b12"}]} [ 915.335118] env[63515]: DEBUG oslo_vmware.api [None req-29ee5601-80a8-4cbb-90e9-98d09454abfa tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Task: {'id': task-1111552, 'name': PowerOnVM_Task, 'duration_secs': 0.509893} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.335401] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-29ee5601-80a8-4cbb-90e9-98d09454abfa tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] [instance: 86b1ddcf-585c-41d9-84fe-17b794f1abbc] Powered on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 915.335649] env[63515]: INFO nova.compute.manager [None req-29ee5601-80a8-4cbb-90e9-98d09454abfa tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] [instance: 86b1ddcf-585c-41d9-84fe-17b794f1abbc] Took 7.40 seconds to spawn the instance on the hypervisor. [ 915.335868] env[63515]: DEBUG nova.compute.manager [None req-29ee5601-80a8-4cbb-90e9-98d09454abfa tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] [instance: 86b1ddcf-585c-41d9-84fe-17b794f1abbc] Checking state {{(pid=63515) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 915.336710] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0faf497a-090b-49f2-b751-1316a332e9b6 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.342031] env[63515]: DEBUG nova.scheduler.client.report [None req-0bb813f4-ce4d-4144-b80d-b179b1c5fd20 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Refreshing inventories for resource provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 915.363726] env[63515]: DEBUG nova.scheduler.client.report [None req-0bb813f4-ce4d-4144-b80d-b179b1c5fd20 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Updating ProviderTree inventory for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 915.364101] env[63515]: DEBUG nova.compute.provider_tree [None req-0bb813f4-ce4d-4144-b80d-b179b1c5fd20 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Updating inventory in ProviderTree for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 915.379196] env[63515]: DEBUG nova.scheduler.client.report [None req-0bb813f4-ce4d-4144-b80d-b179b1c5fd20 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Refreshing aggregate associations for resource provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a, aggregates: None {{(pid=63515) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 915.403541] env[63515]: DEBUG nova.scheduler.client.report [None req-0bb813f4-ce4d-4144-b80d-b179b1c5fd20 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Refreshing trait associations for resource provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a, traits: COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=63515) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 915.437485] env[63515]: DEBUG oslo_vmware.api [None req-44e7f82e-18f0-4009-907e-7846ed9a3683 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Task: {'id': task-1111554, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.216861} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.437828] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-44e7f82e-18f0-4009-907e-7846ed9a3683 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Deleted the datastore file {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 915.438038] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-44e7f82e-18f0-4009-907e-7846ed9a3683 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] [instance: 87c468d9-9594-4804-b461-527f01f6118f] Deleted contents of the VM from datastore datastore1 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 915.438226] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-44e7f82e-18f0-4009-907e-7846ed9a3683 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] [instance: 87c468d9-9594-4804-b461-527f01f6118f] Instance destroyed {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 915.438409] env[63515]: INFO nova.compute.manager [None req-44e7f82e-18f0-4009-907e-7846ed9a3683 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] [instance: 87c468d9-9594-4804-b461-527f01f6118f] Took 1.11 seconds to destroy the instance on the hypervisor. [ 915.438731] env[63515]: DEBUG oslo.service.loopingcall [None req-44e7f82e-18f0-4009-907e-7846ed9a3683 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 915.438990] env[63515]: DEBUG nova.compute.manager [-] [instance: 87c468d9-9594-4804-b461-527f01f6118f] Deallocating network for instance {{(pid=63515) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 915.439122] env[63515]: DEBUG nova.network.neutron [-] [instance: 87c468d9-9594-4804-b461-527f01f6118f] deallocate_for_instance() {{(pid=63515) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 915.453604] env[63515]: DEBUG nova.compute.manager [req-f76c528c-38e3-4ed1-aaa3-5090e3798ba8 req-84de78fe-4cd3-4464-9c38-f5e9ab681ff1 service nova] [instance: ca6a8ac3-d880-4333-9816-2f076e423844] Received event network-vif-plugged-8dcf0639-35d4-47fc-80ef-cc62d15fe62a {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 915.453889] env[63515]: DEBUG oslo_concurrency.lockutils [req-f76c528c-38e3-4ed1-aaa3-5090e3798ba8 req-84de78fe-4cd3-4464-9c38-f5e9ab681ff1 service nova] Acquiring lock "ca6a8ac3-d880-4333-9816-2f076e423844-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 915.454142] env[63515]: DEBUG oslo_concurrency.lockutils [req-f76c528c-38e3-4ed1-aaa3-5090e3798ba8 req-84de78fe-4cd3-4464-9c38-f5e9ab681ff1 service nova] Lock "ca6a8ac3-d880-4333-9816-2f076e423844-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 915.454315] env[63515]: DEBUG oslo_concurrency.lockutils [req-f76c528c-38e3-4ed1-aaa3-5090e3798ba8 req-84de78fe-4cd3-4464-9c38-f5e9ab681ff1 service nova] Lock "ca6a8ac3-d880-4333-9816-2f076e423844-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 915.454489] env[63515]: DEBUG nova.compute.manager [req-f76c528c-38e3-4ed1-aaa3-5090e3798ba8 req-84de78fe-4cd3-4464-9c38-f5e9ab681ff1 service nova] [instance: ca6a8ac3-d880-4333-9816-2f076e423844] No waiting events found dispatching network-vif-plugged-8dcf0639-35d4-47fc-80ef-cc62d15fe62a {{(pid=63515) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 915.454715] env[63515]: WARNING nova.compute.manager [req-f76c528c-38e3-4ed1-aaa3-5090e3798ba8 req-84de78fe-4cd3-4464-9c38-f5e9ab681ff1 service nova] [instance: ca6a8ac3-d880-4333-9816-2f076e423844] Received unexpected event network-vif-plugged-8dcf0639-35d4-47fc-80ef-cc62d15fe62a for instance with vm_state building and task_state spawning. [ 915.454848] env[63515]: DEBUG nova.compute.manager [req-f76c528c-38e3-4ed1-aaa3-5090e3798ba8 req-84de78fe-4cd3-4464-9c38-f5e9ab681ff1 service nova] [instance: ca6a8ac3-d880-4333-9816-2f076e423844] Received event network-changed-8dcf0639-35d4-47fc-80ef-cc62d15fe62a {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 915.455106] env[63515]: DEBUG nova.compute.manager [req-f76c528c-38e3-4ed1-aaa3-5090e3798ba8 req-84de78fe-4cd3-4464-9c38-f5e9ab681ff1 service nova] [instance: ca6a8ac3-d880-4333-9816-2f076e423844] Refreshing instance network info cache due to event network-changed-8dcf0639-35d4-47fc-80ef-cc62d15fe62a. {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 915.455334] env[63515]: DEBUG oslo_concurrency.lockutils [req-f76c528c-38e3-4ed1-aaa3-5090e3798ba8 req-84de78fe-4cd3-4464-9c38-f5e9ab681ff1 service nova] Acquiring lock "refresh_cache-ca6a8ac3-d880-4333-9816-2f076e423844" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 915.465574] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-872cc7bc-b8ac-4eff-87f3-c2c36aad9630 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: 70df69e5-687b-44fb-b6fc-cdb08e21dda0] Unregistered the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 915.465797] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-872cc7bc-b8ac-4eff-87f3-c2c36aad9630 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: 70df69e5-687b-44fb-b6fc-cdb08e21dda0] Deleting contents of the VM from datastore datastore2 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 915.465988] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-872cc7bc-b8ac-4eff-87f3-c2c36aad9630 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Deleting the datastore file [datastore2] 70df69e5-687b-44fb-b6fc-cdb08e21dda0 {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 915.466333] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-9a3b2805-26b0-423a-8bf0-94ad7b3ad084 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.477019] env[63515]: DEBUG oslo_vmware.api [None req-872cc7bc-b8ac-4eff-87f3-c2c36aad9630 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Waiting for the task: (returnval){ [ 915.477019] env[63515]: value = "task-1111556" [ 915.477019] env[63515]: _type = "Task" [ 915.477019] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.487344] env[63515]: DEBUG oslo_vmware.api [None req-872cc7bc-b8ac-4eff-87f3-c2c36aad9630 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Task: {'id': task-1111556, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.509454] env[63515]: DEBUG oslo_vmware.api [None req-8c6d15be-1f5e-44a9-a3e4-dab9aa82e26a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]524d9b0b-868a-0653-3030-af5216a757e9, 'name': SearchDatastore_Task, 'duration_secs': 0.01035} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.509752] env[63515]: DEBUG oslo_concurrency.lockutils [None req-8c6d15be-1f5e-44a9-a3e4-dab9aa82e26a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 915.510643] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-8c6d15be-1f5e-44a9-a3e4-dab9aa82e26a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk to [datastore2] 9236607d-47af-4272-93db-79826e918aa1/9236607d-47af-4272-93db-79826e918aa1.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 915.510643] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-314ef3e9-c239-4014-9eef-87af2c81d964 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.521334] env[63515]: DEBUG oslo_vmware.api [None req-8c6d15be-1f5e-44a9-a3e4-dab9aa82e26a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Waiting for the task: (returnval){ [ 915.521334] env[63515]: value = "task-1111557" [ 915.521334] env[63515]: _type = "Task" [ 915.521334] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.531696] env[63515]: DEBUG oslo_vmware.api [None req-8c6d15be-1f5e-44a9-a3e4-dab9aa82e26a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Task: {'id': task-1111557, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.692592] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c73ae04c-7df1-4fc4-ace3-1a1f9fea1c28 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.700267] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0505d2b1-ce51-409f-86f3-ccfb69d36b0d {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.732079] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af618bf6-be04-4ec3-8d43-6d5d701d6c17 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.735056] env[63515]: DEBUG nova.compute.manager [None req-d63820c0-6e1a-45a6-8579-7bfab43bd25a tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: 2fa098b6-ecc0-491f-8cfc-5ee335bbf943] Starting instance... {{(pid=63515) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 915.742797] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb99b04e-e30e-4ea8-baf8-e8416ea51831 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.757220] env[63515]: DEBUG nova.compute.provider_tree [None req-0bb813f4-ce4d-4144-b80d-b179b1c5fd20 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Updating inventory in ProviderTree for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 915.778029] env[63515]: DEBUG oslo_concurrency.lockutils [None req-8c6d15be-1f5e-44a9-a3e4-dab9aa82e26a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Releasing lock "refresh_cache-ca6a8ac3-d880-4333-9816-2f076e423844" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 915.778553] env[63515]: DEBUG nova.compute.manager [None req-8c6d15be-1f5e-44a9-a3e4-dab9aa82e26a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] [instance: ca6a8ac3-d880-4333-9816-2f076e423844] Instance network_info: |[{"id": "8dcf0639-35d4-47fc-80ef-cc62d15fe62a", "address": "fa:16:3e:56:9c:25", "network": {"id": "549c4fb2-0ed6-44ee-bfd1-9e59bb7aed27", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-949461070-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3dcf6179891c4841a2901fcffe43cd31", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8c3e2368-4a35-4aa5-9135-23daedbbf9ef", "external-id": "nsx-vlan-transportzone-125", "segmentation_id": 125, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8dcf0639-35", "ovs_interfaceid": "8dcf0639-35d4-47fc-80ef-cc62d15fe62a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 915.779157] env[63515]: DEBUG oslo_concurrency.lockutils [req-f76c528c-38e3-4ed1-aaa3-5090e3798ba8 req-84de78fe-4cd3-4464-9c38-f5e9ab681ff1 service nova] Acquired lock "refresh_cache-ca6a8ac3-d880-4333-9816-2f076e423844" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 915.781023] env[63515]: DEBUG nova.network.neutron [req-f76c528c-38e3-4ed1-aaa3-5090e3798ba8 req-84de78fe-4cd3-4464-9c38-f5e9ab681ff1 service nova] [instance: ca6a8ac3-d880-4333-9816-2f076e423844] Refreshing network info cache for port 8dcf0639-35d4-47fc-80ef-cc62d15fe62a {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 915.781023] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-8c6d15be-1f5e-44a9-a3e4-dab9aa82e26a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] [instance: ca6a8ac3-d880-4333-9816-2f076e423844] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:56:9c:25', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '8c3e2368-4a35-4aa5-9135-23daedbbf9ef', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '8dcf0639-35d4-47fc-80ef-cc62d15fe62a', 'vif_model': 'vmxnet3'}] {{(pid=63515) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 915.790710] env[63515]: DEBUG oslo.service.loopingcall [None req-8c6d15be-1f5e-44a9-a3e4-dab9aa82e26a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 915.791760] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ca6a8ac3-d880-4333-9816-2f076e423844] Creating VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 915.793104] env[63515]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f3af7562-2f36-4bc5-b516-f3612ead6351 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.815951] env[63515]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 915.815951] env[63515]: value = "task-1111558" [ 915.815951] env[63515]: _type = "Task" [ 915.815951] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.828433] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1111558, 'name': CreateVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.860908] env[63515]: INFO nova.compute.manager [None req-29ee5601-80a8-4cbb-90e9-98d09454abfa tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] [instance: 86b1ddcf-585c-41d9-84fe-17b794f1abbc] Took 22.08 seconds to build instance. [ 915.889532] env[63515]: INFO nova.compute.manager [None req-f7b46a02-c5ff-4523-967a-55b60fa40b2d tempest-ServerActionsV293TestJSON-1193370169 tempest-ServerActionsV293TestJSON-1193370169-project-member] [instance: 2ea99c7a-8b61-4718-bee0-f4ce4fc021af] Rebuilding instance [ 915.947250] env[63515]: DEBUG nova.compute.manager [None req-f7b46a02-c5ff-4523-967a-55b60fa40b2d tempest-ServerActionsV293TestJSON-1193370169 tempest-ServerActionsV293TestJSON-1193370169-project-member] [instance: 2ea99c7a-8b61-4718-bee0-f4ce4fc021af] Checking state {{(pid=63515) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 915.947250] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-586245b3-1850-4b25-b37b-1efa4f6bfb73 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.989347] env[63515]: DEBUG oslo_vmware.api [None req-872cc7bc-b8ac-4eff-87f3-c2c36aad9630 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Task: {'id': task-1111556, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.181096} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.989796] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-872cc7bc-b8ac-4eff-87f3-c2c36aad9630 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Deleted the datastore file {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 915.992730] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-872cc7bc-b8ac-4eff-87f3-c2c36aad9630 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: 70df69e5-687b-44fb-b6fc-cdb08e21dda0] Deleted contents of the VM from datastore datastore2 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 915.992730] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-872cc7bc-b8ac-4eff-87f3-c2c36aad9630 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: 70df69e5-687b-44fb-b6fc-cdb08e21dda0] Instance destroyed {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 915.992730] env[63515]: INFO nova.compute.manager [None req-872cc7bc-b8ac-4eff-87f3-c2c36aad9630 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: 70df69e5-687b-44fb-b6fc-cdb08e21dda0] Took 1.22 seconds to destroy the instance on the hypervisor. [ 915.992730] env[63515]: DEBUG oslo.service.loopingcall [None req-872cc7bc-b8ac-4eff-87f3-c2c36aad9630 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 915.992730] env[63515]: DEBUG nova.compute.manager [-] [instance: 70df69e5-687b-44fb-b6fc-cdb08e21dda0] Deallocating network for instance {{(pid=63515) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 915.992730] env[63515]: DEBUG nova.network.neutron [-] [instance: 70df69e5-687b-44fb-b6fc-cdb08e21dda0] deallocate_for_instance() {{(pid=63515) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 915.995112] env[63515]: DEBUG oslo_concurrency.lockutils [None req-87dd6511-0315-48f3-98c4-885fed008155 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Acquiring lock "interface-eab16df9-7bb5-4576-bca0-769a561c5fe9-None" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 915.995362] env[63515]: DEBUG oslo_concurrency.lockutils [None req-87dd6511-0315-48f3-98c4-885fed008155 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Lock "interface-eab16df9-7bb5-4576-bca0-769a561c5fe9-None" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 915.995737] env[63515]: DEBUG nova.objects.instance [None req-87dd6511-0315-48f3-98c4-885fed008155 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Lazy-loading 'flavor' on Instance uuid eab16df9-7bb5-4576-bca0-769a561c5fe9 {{(pid=63515) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 916.032413] env[63515]: DEBUG oslo_vmware.api [None req-8c6d15be-1f5e-44a9-a3e4-dab9aa82e26a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Task: {'id': task-1111557, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.117018] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-9788214a-e67e-4df5-bfb0-11119b5fe460 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] [instance: 2a29f37d-e786-48d3-b126-4854d6e7c67c] Volume attach. Driver type: vmdk {{(pid=63515) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 916.117018] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-9788214a-e67e-4df5-bfb0-11119b5fe460 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] [instance: 2a29f37d-e786-48d3-b126-4854d6e7c67c] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-243524', 'volume_id': 'dea169a3-b753-4313-853a-cbf3a4c05ef8', 'name': 'volume-dea169a3-b753-4313-853a-cbf3a4c05ef8', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '2a29f37d-e786-48d3-b126-4854d6e7c67c', 'attached_at': '', 'detached_at': '', 'volume_id': 'dea169a3-b753-4313-853a-cbf3a4c05ef8', 'serial': 'dea169a3-b753-4313-853a-cbf3a4c05ef8'} {{(pid=63515) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 916.117018] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df8f130e-352b-4a39-a528-a94439b8cf77 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.134824] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af04e9e8-c62d-48c1-adba-b49d85542b53 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.160852] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-9788214a-e67e-4df5-bfb0-11119b5fe460 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] [instance: 2a29f37d-e786-48d3-b126-4854d6e7c67c] Reconfiguring VM instance instance-00000045 to attach disk [datastore2] volume-dea169a3-b753-4313-853a-cbf3a4c05ef8/volume-dea169a3-b753-4313-853a-cbf3a4c05ef8.vmdk or device None with type thin {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 916.161232] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e1b97ac1-0c93-4c32-b98c-70cc8f95fc18 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.180609] env[63515]: DEBUG oslo_vmware.api [None req-9788214a-e67e-4df5-bfb0-11119b5fe460 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Waiting for the task: (returnval){ [ 916.180609] env[63515]: value = "task-1111559" [ 916.180609] env[63515]: _type = "Task" [ 916.180609] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 916.190941] env[63515]: DEBUG oslo_vmware.api [None req-9788214a-e67e-4df5-bfb0-11119b5fe460 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Task: {'id': task-1111559, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.255185] env[63515]: DEBUG oslo_concurrency.lockutils [None req-d63820c0-6e1a-45a6-8579-7bfab43bd25a tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 916.292908] env[63515]: DEBUG nova.scheduler.client.report [None req-0bb813f4-ce4d-4144-b80d-b179b1c5fd20 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Updated inventory for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a with generation 114 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 916.293287] env[63515]: DEBUG nova.compute.provider_tree [None req-0bb813f4-ce4d-4144-b80d-b179b1c5fd20 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Updating resource provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a generation from 114 to 115 during operation: update_inventory {{(pid=63515) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 916.293449] env[63515]: DEBUG nova.compute.provider_tree [None req-0bb813f4-ce4d-4144-b80d-b179b1c5fd20 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Updating inventory in ProviderTree for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 916.328640] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1111558, 'name': CreateVM_Task} progress is 25%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.364763] env[63515]: DEBUG oslo_concurrency.lockutils [None req-29ee5601-80a8-4cbb-90e9-98d09454abfa tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Lock "86b1ddcf-585c-41d9-84fe-17b794f1abbc" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 23.587s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 916.456227] env[63515]: DEBUG nova.network.neutron [-] [instance: 87c468d9-9594-4804-b461-527f01f6118f] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 916.462038] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-f7b46a02-c5ff-4523-967a-55b60fa40b2d tempest-ServerActionsV293TestJSON-1193370169 tempest-ServerActionsV293TestJSON-1193370169-project-member] [instance: 2ea99c7a-8b61-4718-bee0-f4ce4fc021af] Powering off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 916.462670] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-dbfc3e8f-d9a9-46d6-ae92-1924faa7a793 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.470735] env[63515]: DEBUG oslo_vmware.api [None req-f7b46a02-c5ff-4523-967a-55b60fa40b2d tempest-ServerActionsV293TestJSON-1193370169 tempest-ServerActionsV293TestJSON-1193370169-project-member] Waiting for the task: (returnval){ [ 916.470735] env[63515]: value = "task-1111560" [ 916.470735] env[63515]: _type = "Task" [ 916.470735] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 916.483085] env[63515]: DEBUG oslo_vmware.api [None req-f7b46a02-c5ff-4523-967a-55b60fa40b2d tempest-ServerActionsV293TestJSON-1193370169 tempest-ServerActionsV293TestJSON-1193370169-project-member] Task: {'id': task-1111560, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.501926] env[63515]: DEBUG nova.objects.instance [None req-87dd6511-0315-48f3-98c4-885fed008155 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Lazy-loading 'pci_requests' on Instance uuid eab16df9-7bb5-4576-bca0-769a561c5fe9 {{(pid=63515) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 916.536152] env[63515]: DEBUG oslo_vmware.api [None req-8c6d15be-1f5e-44a9-a3e4-dab9aa82e26a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Task: {'id': task-1111557, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.640666} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 916.538255] env[63515]: DEBUG nova.network.neutron [req-f76c528c-38e3-4ed1-aaa3-5090e3798ba8 req-84de78fe-4cd3-4464-9c38-f5e9ab681ff1 service nova] [instance: ca6a8ac3-d880-4333-9816-2f076e423844] Updated VIF entry in instance network info cache for port 8dcf0639-35d4-47fc-80ef-cc62d15fe62a. {{(pid=63515) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 916.538255] env[63515]: DEBUG nova.network.neutron [req-f76c528c-38e3-4ed1-aaa3-5090e3798ba8 req-84de78fe-4cd3-4464-9c38-f5e9ab681ff1 service nova] [instance: ca6a8ac3-d880-4333-9816-2f076e423844] Updating instance_info_cache with network_info: [{"id": "8dcf0639-35d4-47fc-80ef-cc62d15fe62a", "address": "fa:16:3e:56:9c:25", "network": {"id": "549c4fb2-0ed6-44ee-bfd1-9e59bb7aed27", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-949461070-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3dcf6179891c4841a2901fcffe43cd31", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8c3e2368-4a35-4aa5-9135-23daedbbf9ef", "external-id": "nsx-vlan-transportzone-125", "segmentation_id": 125, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8dcf0639-35", "ovs_interfaceid": "8dcf0639-35d4-47fc-80ef-cc62d15fe62a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 916.538681] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-8c6d15be-1f5e-44a9-a3e4-dab9aa82e26a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk to [datastore2] 9236607d-47af-4272-93db-79826e918aa1/9236607d-47af-4272-93db-79826e918aa1.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 916.538907] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-8c6d15be-1f5e-44a9-a3e4-dab9aa82e26a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] [instance: 9236607d-47af-4272-93db-79826e918aa1] Extending root virtual disk to 1048576 {{(pid=63515) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 916.539219] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-34684479-40f0-49e4-99b2-0b85146fee46 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.548025] env[63515]: DEBUG oslo_vmware.api [None req-8c6d15be-1f5e-44a9-a3e4-dab9aa82e26a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Waiting for the task: (returnval){ [ 916.548025] env[63515]: value = "task-1111561" [ 916.548025] env[63515]: _type = "Task" [ 916.548025] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 916.555699] env[63515]: DEBUG oslo_vmware.api [None req-8c6d15be-1f5e-44a9-a3e4-dab9aa82e26a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Task: {'id': task-1111561, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.694836] env[63515]: DEBUG oslo_vmware.api [None req-9788214a-e67e-4df5-bfb0-11119b5fe460 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Task: {'id': task-1111559, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.745242] env[63515]: DEBUG nova.network.neutron [-] [instance: 70df69e5-687b-44fb-b6fc-cdb08e21dda0] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 916.800456] env[63515]: DEBUG oslo_concurrency.lockutils [None req-0bb813f4-ce4d-4144-b80d-b179b1c5fd20 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.277s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 916.801021] env[63515]: DEBUG nova.compute.manager [None req-0bb813f4-ce4d-4144-b80d-b179b1c5fd20 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: a2203c5e-66b9-4c7e-8d76-e12a08b57b1d] Start building networks asynchronously for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 916.804318] env[63515]: DEBUG oslo_concurrency.lockutils [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 8.187s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 916.804534] env[63515]: DEBUG oslo_concurrency.lockutils [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 916.804688] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=63515) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 916.805024] env[63515]: DEBUG oslo_concurrency.lockutils [None req-ef0f5ab2-1246-4de1-af39-f1efd470cd76 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 6.020s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 916.805240] env[63515]: DEBUG nova.objects.instance [None req-ef0f5ab2-1246-4de1-af39-f1efd470cd76 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Lazy-loading 'resources' on Instance uuid ae5d6f34-bac6-4ba0-b0c3-8205982ddc88 {{(pid=63515) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 916.807183] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6b69dff-f067-4fe4-bc2a-fd6a7a00feb1 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.816019] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f49b21ef-53d3-4ee0-8fbf-ca1f297c1689 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.838949] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1111558, 'name': CreateVM_Task, 'duration_secs': 0.697665} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 916.839696] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ca6a8ac3-d880-4333-9816-2f076e423844] Created VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 916.840471] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-182d51ff-3474-4fde-a9f2-9cca19893e13 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.844072] env[63515]: DEBUG oslo_concurrency.lockutils [None req-8c6d15be-1f5e-44a9-a3e4-dab9aa82e26a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 916.844256] env[63515]: DEBUG oslo_concurrency.lockutils [None req-8c6d15be-1f5e-44a9-a3e4-dab9aa82e26a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 916.844578] env[63515]: DEBUG oslo_concurrency.lockutils [None req-8c6d15be-1f5e-44a9-a3e4-dab9aa82e26a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 916.845081] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-666b38cd-71c7-468c-b525-dda3382e68c0 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.852561] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a17fa5a-10f6-4589-b2b5-12b5ce3bae56 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.856068] env[63515]: DEBUG oslo_vmware.api [None req-8c6d15be-1f5e-44a9-a3e4-dab9aa82e26a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Waiting for the task: (returnval){ [ 916.856068] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]52301c86-bf1a-8403-2cf4-79b28bd941c4" [ 916.856068] env[63515]: _type = "Task" [ 916.856068] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 916.884132] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180494MB free_disk=170GB free_vcpus=48 pci_devices=None {{(pid=63515) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 916.884337] env[63515]: DEBUG oslo_concurrency.lockutils [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 916.888771] env[63515]: DEBUG oslo_vmware.api [None req-8c6d15be-1f5e-44a9-a3e4-dab9aa82e26a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52301c86-bf1a-8403-2cf4-79b28bd941c4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.957619] env[63515]: INFO nova.compute.manager [-] [instance: 87c468d9-9594-4804-b461-527f01f6118f] Took 1.52 seconds to deallocate network for instance. [ 916.982189] env[63515]: DEBUG oslo_vmware.api [None req-f7b46a02-c5ff-4523-967a-55b60fa40b2d tempest-ServerActionsV293TestJSON-1193370169 tempest-ServerActionsV293TestJSON-1193370169-project-member] Task: {'id': task-1111560, 'name': PowerOffVM_Task, 'duration_secs': 0.299629} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 916.982274] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-f7b46a02-c5ff-4523-967a-55b60fa40b2d tempest-ServerActionsV293TestJSON-1193370169 tempest-ServerActionsV293TestJSON-1193370169-project-member] [instance: 2ea99c7a-8b61-4718-bee0-f4ce4fc021af] Powered off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 916.982979] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-f7b46a02-c5ff-4523-967a-55b60fa40b2d tempest-ServerActionsV293TestJSON-1193370169 tempest-ServerActionsV293TestJSON-1193370169-project-member] [instance: 2ea99c7a-8b61-4718-bee0-f4ce4fc021af] Powering off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 916.983276] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d8e0dcc4-c62b-4f6e-b3c4-1ddc3ee6d258 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.991237] env[63515]: DEBUG oslo_vmware.api [None req-f7b46a02-c5ff-4523-967a-55b60fa40b2d tempest-ServerActionsV293TestJSON-1193370169 tempest-ServerActionsV293TestJSON-1193370169-project-member] Waiting for the task: (returnval){ [ 916.991237] env[63515]: value = "task-1111562" [ 916.991237] env[63515]: _type = "Task" [ 916.991237] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.003417] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-f7b46a02-c5ff-4523-967a-55b60fa40b2d tempest-ServerActionsV293TestJSON-1193370169 tempest-ServerActionsV293TestJSON-1193370169-project-member] [instance: 2ea99c7a-8b61-4718-bee0-f4ce4fc021af] VM already powered off {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 917.003417] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-f7b46a02-c5ff-4523-967a-55b60fa40b2d tempest-ServerActionsV293TestJSON-1193370169 tempest-ServerActionsV293TestJSON-1193370169-project-member] [instance: 2ea99c7a-8b61-4718-bee0-f4ce4fc021af] Volume detach. Driver type: vmdk {{(pid=63515) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 917.003417] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-f7b46a02-c5ff-4523-967a-55b60fa40b2d tempest-ServerActionsV293TestJSON-1193370169 tempest-ServerActionsV293TestJSON-1193370169-project-member] [instance: 2ea99c7a-8b61-4718-bee0-f4ce4fc021af] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-243483', 'volume_id': '211b1a87-a01e-4d13-aa27-91391bf9cde1', 'name': 'volume-211b1a87-a01e-4d13-aa27-91391bf9cde1', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '2ea99c7a-8b61-4718-bee0-f4ce4fc021af', 'attached_at': '', 'detached_at': '', 'volume_id': '211b1a87-a01e-4d13-aa27-91391bf9cde1', 'serial': '211b1a87-a01e-4d13-aa27-91391bf9cde1'} {{(pid=63515) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 917.003999] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d52d725d-6616-4ce6-894c-ac1fdc04006a {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.007784] env[63515]: DEBUG nova.objects.base [None req-87dd6511-0315-48f3-98c4-885fed008155 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Object Instance lazy-loaded attributes: flavor,pci_requests {{(pid=63515) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 917.007980] env[63515]: DEBUG nova.network.neutron [None req-87dd6511-0315-48f3-98c4-885fed008155 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: eab16df9-7bb5-4576-bca0-769a561c5fe9] allocate_for_instance() {{(pid=63515) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 917.031084] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f8be357-b186-4f1f-9042-fbc063df40b9 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.039947] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c0cce9d-0ab8-45ef-90cc-e759f1862268 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.042986] env[63515]: DEBUG oslo_concurrency.lockutils [req-f76c528c-38e3-4ed1-aaa3-5090e3798ba8 req-84de78fe-4cd3-4464-9c38-f5e9ab681ff1 service nova] Releasing lock "refresh_cache-ca6a8ac3-d880-4333-9816-2f076e423844" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 917.065331] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e3bb5ad-e18c-4e80-95ce-81ba4f2f6d35 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.075244] env[63515]: DEBUG oslo_vmware.api [None req-8c6d15be-1f5e-44a9-a3e4-dab9aa82e26a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Task: {'id': task-1111561, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.233079} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 917.086011] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-8c6d15be-1f5e-44a9-a3e4-dab9aa82e26a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] [instance: 9236607d-47af-4272-93db-79826e918aa1] Extended root virtual disk {{(pid=63515) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 917.086410] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-f7b46a02-c5ff-4523-967a-55b60fa40b2d tempest-ServerActionsV293TestJSON-1193370169 tempest-ServerActionsV293TestJSON-1193370169-project-member] The volume has not been displaced from its original location: [datastore2] volume-211b1a87-a01e-4d13-aa27-91391bf9cde1/volume-211b1a87-a01e-4d13-aa27-91391bf9cde1.vmdk. No consolidation needed. {{(pid=63515) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 917.091628] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-f7b46a02-c5ff-4523-967a-55b60fa40b2d tempest-ServerActionsV293TestJSON-1193370169 tempest-ServerActionsV293TestJSON-1193370169-project-member] [instance: 2ea99c7a-8b61-4718-bee0-f4ce4fc021af] Reconfiguring VM instance instance-00000048 to detach disk 2000 {{(pid=63515) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 917.092504] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4853cb9-72e3-4c1e-93fc-6ee06ce3a166 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.095080] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f4893e61-5d47-4484-a3e3-d3f37a5429e8 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.108629] env[63515]: DEBUG oslo_concurrency.lockutils [None req-87dd6511-0315-48f3-98c4-885fed008155 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Lock "interface-eab16df9-7bb5-4576-bca0-769a561c5fe9-None" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 1.113s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 917.131823] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-8c6d15be-1f5e-44a9-a3e4-dab9aa82e26a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] [instance: 9236607d-47af-4272-93db-79826e918aa1] Reconfiguring VM instance instance-00000052 to attach disk [datastore2] 9236607d-47af-4272-93db-79826e918aa1/9236607d-47af-4272-93db-79826e918aa1.vmdk or device None with type sparse {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 917.135187] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-14d5983a-e56b-4c98-9b8e-b05422bea83c {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.154866] env[63515]: DEBUG oslo_vmware.api [None req-f7b46a02-c5ff-4523-967a-55b60fa40b2d tempest-ServerActionsV293TestJSON-1193370169 tempest-ServerActionsV293TestJSON-1193370169-project-member] Waiting for the task: (returnval){ [ 917.154866] env[63515]: value = "task-1111563" [ 917.154866] env[63515]: _type = "Task" [ 917.154866] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.155031] env[63515]: DEBUG nova.compute.manager [req-bf1778c8-ad03-4305-8552-ef1b4657ca26 req-e28bc8bd-a4a3-4e96-8f4f-c906a03bbc12 service nova] [instance: 86b1ddcf-585c-41d9-84fe-17b794f1abbc] Received event network-changed-43c4b108-fec8-46ea-942f-d3e734fe2be6 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 917.155273] env[63515]: DEBUG nova.compute.manager [req-bf1778c8-ad03-4305-8552-ef1b4657ca26 req-e28bc8bd-a4a3-4e96-8f4f-c906a03bbc12 service nova] [instance: 86b1ddcf-585c-41d9-84fe-17b794f1abbc] Refreshing instance network info cache due to event network-changed-43c4b108-fec8-46ea-942f-d3e734fe2be6. {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 917.155369] env[63515]: DEBUG oslo_concurrency.lockutils [req-bf1778c8-ad03-4305-8552-ef1b4657ca26 req-e28bc8bd-a4a3-4e96-8f4f-c906a03bbc12 service nova] Acquiring lock "refresh_cache-86b1ddcf-585c-41d9-84fe-17b794f1abbc" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 917.155508] env[63515]: DEBUG oslo_concurrency.lockutils [req-bf1778c8-ad03-4305-8552-ef1b4657ca26 req-e28bc8bd-a4a3-4e96-8f4f-c906a03bbc12 service nova] Acquired lock "refresh_cache-86b1ddcf-585c-41d9-84fe-17b794f1abbc" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 917.155664] env[63515]: DEBUG nova.network.neutron [req-bf1778c8-ad03-4305-8552-ef1b4657ca26 req-e28bc8bd-a4a3-4e96-8f4f-c906a03bbc12 service nova] [instance: 86b1ddcf-585c-41d9-84fe-17b794f1abbc] Refreshing network info cache for port 43c4b108-fec8-46ea-942f-d3e734fe2be6 {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 917.163609] env[63515]: DEBUG oslo_vmware.api [None req-8c6d15be-1f5e-44a9-a3e4-dab9aa82e26a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Waiting for the task: (returnval){ [ 917.163609] env[63515]: value = "task-1111564" [ 917.163609] env[63515]: _type = "Task" [ 917.163609] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.170601] env[63515]: DEBUG oslo_vmware.api [None req-f7b46a02-c5ff-4523-967a-55b60fa40b2d tempest-ServerActionsV293TestJSON-1193370169 tempest-ServerActionsV293TestJSON-1193370169-project-member] Task: {'id': task-1111563, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.181947] env[63515]: DEBUG oslo_vmware.api [None req-8c6d15be-1f5e-44a9-a3e4-dab9aa82e26a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Task: {'id': task-1111564, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.193046] env[63515]: DEBUG oslo_vmware.api [None req-9788214a-e67e-4df5-bfb0-11119b5fe460 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Task: {'id': task-1111559, 'name': ReconfigVM_Task, 'duration_secs': 0.897969} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 917.193046] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-9788214a-e67e-4df5-bfb0-11119b5fe460 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] [instance: 2a29f37d-e786-48d3-b126-4854d6e7c67c] Reconfigured VM instance instance-00000045 to attach disk [datastore2] volume-dea169a3-b753-4313-853a-cbf3a4c05ef8/volume-dea169a3-b753-4313-853a-cbf3a4c05ef8.vmdk or device None with type thin {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 917.196218] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-da6951f3-b7aa-4569-b60d-87be0f9a4dac {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.213942] env[63515]: DEBUG oslo_vmware.api [None req-9788214a-e67e-4df5-bfb0-11119b5fe460 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Waiting for the task: (returnval){ [ 917.213942] env[63515]: value = "task-1111565" [ 917.213942] env[63515]: _type = "Task" [ 917.213942] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.222284] env[63515]: DEBUG oslo_vmware.api [None req-9788214a-e67e-4df5-bfb0-11119b5fe460 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Task: {'id': task-1111565, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.247869] env[63515]: INFO nova.compute.manager [-] [instance: 70df69e5-687b-44fb-b6fc-cdb08e21dda0] Took 1.26 seconds to deallocate network for instance. [ 917.308484] env[63515]: DEBUG nova.compute.utils [None req-0bb813f4-ce4d-4144-b80d-b179b1c5fd20 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Using /dev/sd instead of None {{(pid=63515) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 917.309856] env[63515]: DEBUG nova.compute.manager [None req-0bb813f4-ce4d-4144-b80d-b179b1c5fd20 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: a2203c5e-66b9-4c7e-8d76-e12a08b57b1d] Allocating IP information in the background. {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 917.310101] env[63515]: DEBUG nova.network.neutron [None req-0bb813f4-ce4d-4144-b80d-b179b1c5fd20 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: a2203c5e-66b9-4c7e-8d76-e12a08b57b1d] allocate_for_instance() {{(pid=63515) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 917.312902] env[63515]: DEBUG nova.objects.instance [None req-ef0f5ab2-1246-4de1-af39-f1efd470cd76 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Lazy-loading 'numa_topology' on Instance uuid ae5d6f34-bac6-4ba0-b0c3-8205982ddc88 {{(pid=63515) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 917.354478] env[63515]: DEBUG nova.policy [None req-0bb813f4-ce4d-4144-b80d-b179b1c5fd20 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9b19bb2d32d84c019541c3b2e711a202', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '299fdeff647f486390366d5bbf911518', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63515) authorize /opt/stack/nova/nova/policy.py:201}} [ 917.366161] env[63515]: DEBUG oslo_vmware.api [None req-8c6d15be-1f5e-44a9-a3e4-dab9aa82e26a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52301c86-bf1a-8403-2cf4-79b28bd941c4, 'name': SearchDatastore_Task, 'duration_secs': 0.025275} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 917.366527] env[63515]: DEBUG oslo_concurrency.lockutils [None req-8c6d15be-1f5e-44a9-a3e4-dab9aa82e26a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 917.366802] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-8c6d15be-1f5e-44a9-a3e4-dab9aa82e26a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] [instance: ca6a8ac3-d880-4333-9816-2f076e423844] Processing image 8a120570-cb06-4099-b262-554ca0ad15c5 {{(pid=63515) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 917.367096] env[63515]: DEBUG oslo_concurrency.lockutils [None req-8c6d15be-1f5e-44a9-a3e4-dab9aa82e26a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 917.367313] env[63515]: DEBUG oslo_concurrency.lockutils [None req-8c6d15be-1f5e-44a9-a3e4-dab9aa82e26a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 917.367538] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-8c6d15be-1f5e-44a9-a3e4-dab9aa82e26a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 917.367834] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-73f29d9d-1e81-47f9-93d4-b4c894b125c1 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.377257] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-8c6d15be-1f5e-44a9-a3e4-dab9aa82e26a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 917.378985] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-8c6d15be-1f5e-44a9-a3e4-dab9aa82e26a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63515) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 917.380258] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5d1189c6-e907-4635-a065-bd22e1f33820 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.393874] env[63515]: DEBUG oslo_vmware.api [None req-8c6d15be-1f5e-44a9-a3e4-dab9aa82e26a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Waiting for the task: (returnval){ [ 917.393874] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]528f7418-a827-a4c3-75a1-5e552c4b3032" [ 917.393874] env[63515]: _type = "Task" [ 917.393874] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.404315] env[63515]: DEBUG oslo_vmware.api [None req-8c6d15be-1f5e-44a9-a3e4-dab9aa82e26a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]528f7418-a827-a4c3-75a1-5e552c4b3032, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.464352] env[63515]: DEBUG oslo_concurrency.lockutils [None req-44e7f82e-18f0-4009-907e-7846ed9a3683 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 917.671225] env[63515]: DEBUG oslo_vmware.api [None req-f7b46a02-c5ff-4523-967a-55b60fa40b2d tempest-ServerActionsV293TestJSON-1193370169 tempest-ServerActionsV293TestJSON-1193370169-project-member] Task: {'id': task-1111563, 'name': ReconfigVM_Task, 'duration_secs': 0.262602} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 917.671867] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-f7b46a02-c5ff-4523-967a-55b60fa40b2d tempest-ServerActionsV293TestJSON-1193370169 tempest-ServerActionsV293TestJSON-1193370169-project-member] [instance: 2ea99c7a-8b61-4718-bee0-f4ce4fc021af] Reconfigured VM instance instance-00000048 to detach disk 2000 {{(pid=63515) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 917.679533] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a85da0fb-008c-4933-8b64-0cfa5f26831d {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.689717] env[63515]: DEBUG oslo_vmware.api [None req-8c6d15be-1f5e-44a9-a3e4-dab9aa82e26a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Task: {'id': task-1111564, 'name': ReconfigVM_Task, 'duration_secs': 0.327942} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 917.692348] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-8c6d15be-1f5e-44a9-a3e4-dab9aa82e26a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] [instance: 9236607d-47af-4272-93db-79826e918aa1] Reconfigured VM instance instance-00000052 to attach disk [datastore2] 9236607d-47af-4272-93db-79826e918aa1/9236607d-47af-4272-93db-79826e918aa1.vmdk or device None with type sparse {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 917.693717] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b9273fbc-522d-41f3-b9d3-2e3530f991ef {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.698758] env[63515]: DEBUG oslo_vmware.api [None req-f7b46a02-c5ff-4523-967a-55b60fa40b2d tempest-ServerActionsV293TestJSON-1193370169 tempest-ServerActionsV293TestJSON-1193370169-project-member] Waiting for the task: (returnval){ [ 917.698758] env[63515]: value = "task-1111566" [ 917.698758] env[63515]: _type = "Task" [ 917.698758] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.703207] env[63515]: DEBUG oslo_vmware.api [None req-8c6d15be-1f5e-44a9-a3e4-dab9aa82e26a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Waiting for the task: (returnval){ [ 917.703207] env[63515]: value = "task-1111567" [ 917.703207] env[63515]: _type = "Task" [ 917.703207] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.709961] env[63515]: DEBUG oslo_vmware.api [None req-f7b46a02-c5ff-4523-967a-55b60fa40b2d tempest-ServerActionsV293TestJSON-1193370169 tempest-ServerActionsV293TestJSON-1193370169-project-member] Task: {'id': task-1111566, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.714918] env[63515]: DEBUG oslo_vmware.api [None req-8c6d15be-1f5e-44a9-a3e4-dab9aa82e26a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Task: {'id': task-1111567, 'name': Rename_Task} progress is 10%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.720356] env[63515]: DEBUG nova.compute.manager [req-effcc2b3-11a8-417f-8983-1b12e3a2136b req-d1dc44c1-ba5e-474e-bdf6-f566b98353b6 service nova] [instance: 87c468d9-9594-4804-b461-527f01f6118f] Received event network-vif-deleted-d2098868-0475-4b06-bae3-785c608a861e {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 917.720790] env[63515]: DEBUG nova.compute.manager [req-effcc2b3-11a8-417f-8983-1b12e3a2136b req-d1dc44c1-ba5e-474e-bdf6-f566b98353b6 service nova] [instance: 70df69e5-687b-44fb-b6fc-cdb08e21dda0] Received event network-vif-deleted-2639cfaa-8db0-4dcc-ac66-994ec5a84fdc {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 917.727109] env[63515]: DEBUG oslo_vmware.api [None req-9788214a-e67e-4df5-bfb0-11119b5fe460 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Task: {'id': task-1111565, 'name': ReconfigVM_Task, 'duration_secs': 0.175513} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 917.727453] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-9788214a-e67e-4df5-bfb0-11119b5fe460 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] [instance: 2a29f37d-e786-48d3-b126-4854d6e7c67c] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-243524', 'volume_id': 'dea169a3-b753-4313-853a-cbf3a4c05ef8', 'name': 'volume-dea169a3-b753-4313-853a-cbf3a4c05ef8', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '2a29f37d-e786-48d3-b126-4854d6e7c67c', 'attached_at': '', 'detached_at': '', 'volume_id': 'dea169a3-b753-4313-853a-cbf3a4c05ef8', 'serial': 'dea169a3-b753-4313-853a-cbf3a4c05ef8'} {{(pid=63515) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 917.756289] env[63515]: DEBUG oslo_concurrency.lockutils [None req-872cc7bc-b8ac-4eff-87f3-c2c36aad9630 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 917.813608] env[63515]: DEBUG nova.compute.manager [None req-0bb813f4-ce4d-4144-b80d-b179b1c5fd20 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: a2203c5e-66b9-4c7e-8d76-e12a08b57b1d] Start building block device mappings for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 917.819237] env[63515]: DEBUG nova.objects.base [None req-ef0f5ab2-1246-4de1-af39-f1efd470cd76 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Object Instance lazy-loaded attributes: resources,numa_topology {{(pid=63515) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 917.863364] env[63515]: DEBUG nova.network.neutron [None req-0bb813f4-ce4d-4144-b80d-b179b1c5fd20 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: a2203c5e-66b9-4c7e-8d76-e12a08b57b1d] Successfully created port: 8b878aac-8955-4468-893f-63fdbfc0c8cb {{(pid=63515) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 917.906946] env[63515]: DEBUG oslo_vmware.api [None req-8c6d15be-1f5e-44a9-a3e4-dab9aa82e26a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]528f7418-a827-a4c3-75a1-5e552c4b3032, 'name': SearchDatastore_Task, 'duration_secs': 0.012863} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 917.908042] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-29bfdb72-fb9e-48e4-b4dc-c478b616c59e {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.915111] env[63515]: DEBUG oslo_vmware.api [None req-8c6d15be-1f5e-44a9-a3e4-dab9aa82e26a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Waiting for the task: (returnval){ [ 917.915111] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]52a3fa59-b156-be94-874c-182af7c4f3b0" [ 917.915111] env[63515]: _type = "Task" [ 917.915111] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.923013] env[63515]: DEBUG oslo_vmware.api [None req-8c6d15be-1f5e-44a9-a3e4-dab9aa82e26a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52a3fa59-b156-be94-874c-182af7c4f3b0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.981646] env[63515]: DEBUG nova.network.neutron [req-bf1778c8-ad03-4305-8552-ef1b4657ca26 req-e28bc8bd-a4a3-4e96-8f4f-c906a03bbc12 service nova] [instance: 86b1ddcf-585c-41d9-84fe-17b794f1abbc] Updated VIF entry in instance network info cache for port 43c4b108-fec8-46ea-942f-d3e734fe2be6. {{(pid=63515) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 917.982027] env[63515]: DEBUG nova.network.neutron [req-bf1778c8-ad03-4305-8552-ef1b4657ca26 req-e28bc8bd-a4a3-4e96-8f4f-c906a03bbc12 service nova] [instance: 86b1ddcf-585c-41d9-84fe-17b794f1abbc] Updating instance_info_cache with network_info: [{"id": "43c4b108-fec8-46ea-942f-d3e734fe2be6", "address": "fa:16:3e:04:fd:9b", "network": {"id": "57de0b07-78e5-453d-83ba-b12b7dcaf6ce", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-272799484-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.172", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "12a90929c98e40d9a1263929e2b97532", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a1895250-76cc-41f7-b7f8-2e5679494607", "external-id": "nsx-vlan-transportzone-785", "segmentation_id": 785, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap43c4b108-fe", "ovs_interfaceid": "43c4b108-fec8-46ea-942f-d3e734fe2be6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 918.120115] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a7a24b1-5330-4bda-abbf-3372b6beec37 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.131023] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f382329-e2d7-43bb-ab6d-ef1f81e30144 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.165117] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f38b1293-7d0c-49c6-88e9-70c733a19efd {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.173017] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e4d8185-9ffd-4ea0-bc18-925bb3c12760 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.186384] env[63515]: DEBUG nova.compute.provider_tree [None req-ef0f5ab2-1246-4de1-af39-f1efd470cd76 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Updating inventory in ProviderTree for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 918.210782] env[63515]: DEBUG oslo_vmware.api [None req-f7b46a02-c5ff-4523-967a-55b60fa40b2d tempest-ServerActionsV293TestJSON-1193370169 tempest-ServerActionsV293TestJSON-1193370169-project-member] Task: {'id': task-1111566, 'name': ReconfigVM_Task, 'duration_secs': 0.165992} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 918.211484] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-f7b46a02-c5ff-4523-967a-55b60fa40b2d tempest-ServerActionsV293TestJSON-1193370169 tempest-ServerActionsV293TestJSON-1193370169-project-member] [instance: 2ea99c7a-8b61-4718-bee0-f4ce4fc021af] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-243483', 'volume_id': '211b1a87-a01e-4d13-aa27-91391bf9cde1', 'name': 'volume-211b1a87-a01e-4d13-aa27-91391bf9cde1', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '2ea99c7a-8b61-4718-bee0-f4ce4fc021af', 'attached_at': '', 'detached_at': '', 'volume_id': '211b1a87-a01e-4d13-aa27-91391bf9cde1', 'serial': '211b1a87-a01e-4d13-aa27-91391bf9cde1'} {{(pid=63515) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 918.211826] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-f7b46a02-c5ff-4523-967a-55b60fa40b2d tempest-ServerActionsV293TestJSON-1193370169 tempest-ServerActionsV293TestJSON-1193370169-project-member] [instance: 2ea99c7a-8b61-4718-bee0-f4ce4fc021af] Destroying instance {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 918.212828] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a47b5e2-b27e-45b0-902a-ab91865e574b {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.218911] env[63515]: DEBUG oslo_vmware.api [None req-8c6d15be-1f5e-44a9-a3e4-dab9aa82e26a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Task: {'id': task-1111567, 'name': Rename_Task, 'duration_secs': 0.174471} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 918.219519] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-8c6d15be-1f5e-44a9-a3e4-dab9aa82e26a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] [instance: 9236607d-47af-4272-93db-79826e918aa1] Powering on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 918.219753] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-02fd8b02-883f-4a85-8897-bc13b60958a5 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.224122] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-f7b46a02-c5ff-4523-967a-55b60fa40b2d tempest-ServerActionsV293TestJSON-1193370169 tempest-ServerActionsV293TestJSON-1193370169-project-member] [instance: 2ea99c7a-8b61-4718-bee0-f4ce4fc021af] Unregistering the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 918.224858] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ba334f89-4af9-4dde-9837-9aa293acc16d {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.229177] env[63515]: DEBUG oslo_vmware.api [None req-8c6d15be-1f5e-44a9-a3e4-dab9aa82e26a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Waiting for the task: (returnval){ [ 918.229177] env[63515]: value = "task-1111568" [ 918.229177] env[63515]: _type = "Task" [ 918.229177] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 918.240980] env[63515]: DEBUG oslo_vmware.api [None req-8c6d15be-1f5e-44a9-a3e4-dab9aa82e26a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Task: {'id': task-1111568, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.285055] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-f7b46a02-c5ff-4523-967a-55b60fa40b2d tempest-ServerActionsV293TestJSON-1193370169 tempest-ServerActionsV293TestJSON-1193370169-project-member] [instance: 2ea99c7a-8b61-4718-bee0-f4ce4fc021af] Unregistered the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 918.285369] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-f7b46a02-c5ff-4523-967a-55b60fa40b2d tempest-ServerActionsV293TestJSON-1193370169 tempest-ServerActionsV293TestJSON-1193370169-project-member] [instance: 2ea99c7a-8b61-4718-bee0-f4ce4fc021af] Deleting contents of the VM from datastore datastore2 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 918.285847] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-f7b46a02-c5ff-4523-967a-55b60fa40b2d tempest-ServerActionsV293TestJSON-1193370169 tempest-ServerActionsV293TestJSON-1193370169-project-member] Deleting the datastore file [datastore2] 2ea99c7a-8b61-4718-bee0-f4ce4fc021af {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 918.285847] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b133bd68-7c67-481c-8c0c-fa5a18a732de {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.293485] env[63515]: DEBUG oslo_vmware.api [None req-f7b46a02-c5ff-4523-967a-55b60fa40b2d tempest-ServerActionsV293TestJSON-1193370169 tempest-ServerActionsV293TestJSON-1193370169-project-member] Waiting for the task: (returnval){ [ 918.293485] env[63515]: value = "task-1111570" [ 918.293485] env[63515]: _type = "Task" [ 918.293485] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 918.302637] env[63515]: DEBUG oslo_vmware.api [None req-f7b46a02-c5ff-4523-967a-55b60fa40b2d tempest-ServerActionsV293TestJSON-1193370169 tempest-ServerActionsV293TestJSON-1193370169-project-member] Task: {'id': task-1111570, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.426338] env[63515]: DEBUG oslo_vmware.api [None req-8c6d15be-1f5e-44a9-a3e4-dab9aa82e26a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52a3fa59-b156-be94-874c-182af7c4f3b0, 'name': SearchDatastore_Task, 'duration_secs': 0.021887} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 918.426639] env[63515]: DEBUG oslo_concurrency.lockutils [None req-8c6d15be-1f5e-44a9-a3e4-dab9aa82e26a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 918.426929] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-8c6d15be-1f5e-44a9-a3e4-dab9aa82e26a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk to [datastore2] ca6a8ac3-d880-4333-9816-2f076e423844/ca6a8ac3-d880-4333-9816-2f076e423844.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 918.427273] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-10029b51-03d6-4cce-9967-0bdf27b159f4 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.435537] env[63515]: DEBUG oslo_vmware.api [None req-8c6d15be-1f5e-44a9-a3e4-dab9aa82e26a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Waiting for the task: (returnval){ [ 918.435537] env[63515]: value = "task-1111571" [ 918.435537] env[63515]: _type = "Task" [ 918.435537] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 918.445367] env[63515]: DEBUG oslo_vmware.api [None req-8c6d15be-1f5e-44a9-a3e4-dab9aa82e26a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Task: {'id': task-1111571, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.485507] env[63515]: DEBUG oslo_concurrency.lockutils [req-bf1778c8-ad03-4305-8552-ef1b4657ca26 req-e28bc8bd-a4a3-4e96-8f4f-c906a03bbc12 service nova] Releasing lock "refresh_cache-86b1ddcf-585c-41d9-84fe-17b794f1abbc" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 918.731419] env[63515]: DEBUG nova.scheduler.client.report [None req-ef0f5ab2-1246-4de1-af39-f1efd470cd76 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Updated inventory for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a with generation 115 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 918.731861] env[63515]: DEBUG nova.compute.provider_tree [None req-ef0f5ab2-1246-4de1-af39-f1efd470cd76 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Updating resource provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a generation from 115 to 116 during operation: update_inventory {{(pid=63515) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 918.732120] env[63515]: DEBUG nova.compute.provider_tree [None req-ef0f5ab2-1246-4de1-af39-f1efd470cd76 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Updating inventory in ProviderTree for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 918.745728] env[63515]: DEBUG oslo_vmware.api [None req-8c6d15be-1f5e-44a9-a3e4-dab9aa82e26a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Task: {'id': task-1111568, 'name': PowerOnVM_Task, 'duration_secs': 0.499961} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 918.746015] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-8c6d15be-1f5e-44a9-a3e4-dab9aa82e26a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] [instance: 9236607d-47af-4272-93db-79826e918aa1] Powered on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 918.746354] env[63515]: INFO nova.compute.manager [None req-8c6d15be-1f5e-44a9-a3e4-dab9aa82e26a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] [instance: 9236607d-47af-4272-93db-79826e918aa1] Took 7.44 seconds to spawn the instance on the hypervisor. [ 918.746718] env[63515]: DEBUG nova.compute.manager [None req-8c6d15be-1f5e-44a9-a3e4-dab9aa82e26a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] [instance: 9236607d-47af-4272-93db-79826e918aa1] Checking state {{(pid=63515) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 918.747503] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-643094f9-0f51-4504-ab26-c794f0ad2709 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.779754] env[63515]: DEBUG nova.objects.instance [None req-9788214a-e67e-4df5-bfb0-11119b5fe460 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Lazy-loading 'flavor' on Instance uuid 2a29f37d-e786-48d3-b126-4854d6e7c67c {{(pid=63515) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 918.806183] env[63515]: DEBUG oslo_vmware.api [None req-f7b46a02-c5ff-4523-967a-55b60fa40b2d tempest-ServerActionsV293TestJSON-1193370169 tempest-ServerActionsV293TestJSON-1193370169-project-member] Task: {'id': task-1111570, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.12668} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 918.806496] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-f7b46a02-c5ff-4523-967a-55b60fa40b2d tempest-ServerActionsV293TestJSON-1193370169 tempest-ServerActionsV293TestJSON-1193370169-project-member] Deleted the datastore file {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 918.806692] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-f7b46a02-c5ff-4523-967a-55b60fa40b2d tempest-ServerActionsV293TestJSON-1193370169 tempest-ServerActionsV293TestJSON-1193370169-project-member] [instance: 2ea99c7a-8b61-4718-bee0-f4ce4fc021af] Deleted contents of the VM from datastore datastore2 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 918.806890] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-f7b46a02-c5ff-4523-967a-55b60fa40b2d tempest-ServerActionsV293TestJSON-1193370169 tempest-ServerActionsV293TestJSON-1193370169-project-member] [instance: 2ea99c7a-8b61-4718-bee0-f4ce4fc021af] Instance destroyed {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 918.828759] env[63515]: DEBUG nova.compute.manager [None req-0bb813f4-ce4d-4144-b80d-b179b1c5fd20 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: a2203c5e-66b9-4c7e-8d76-e12a08b57b1d] Start spawning the instance on the hypervisor. {{(pid=63515) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 918.857185] env[63515]: DEBUG nova.virt.hardware [None req-0bb813f4-ce4d-4144-b80d-b179b1c5fd20 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T02:52:14Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T02:51:56Z,direct_url=,disk_format='vmdk',id=8a120570-cb06-4099-b262-554ca0ad15c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b569255cc43e42e7ae3f2b4ad37c6ef4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T02:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 918.857185] env[63515]: DEBUG nova.virt.hardware [None req-0bb813f4-ce4d-4144-b80d-b179b1c5fd20 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Flavor limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 918.857185] env[63515]: DEBUG nova.virt.hardware [None req-0bb813f4-ce4d-4144-b80d-b179b1c5fd20 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Image limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 918.857185] env[63515]: DEBUG nova.virt.hardware [None req-0bb813f4-ce4d-4144-b80d-b179b1c5fd20 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Flavor pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 918.857185] env[63515]: DEBUG nova.virt.hardware [None req-0bb813f4-ce4d-4144-b80d-b179b1c5fd20 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Image pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 918.857185] env[63515]: DEBUG nova.virt.hardware [None req-0bb813f4-ce4d-4144-b80d-b179b1c5fd20 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 918.857185] env[63515]: DEBUG nova.virt.hardware [None req-0bb813f4-ce4d-4144-b80d-b179b1c5fd20 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 918.857603] env[63515]: DEBUG nova.virt.hardware [None req-0bb813f4-ce4d-4144-b80d-b179b1c5fd20 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 918.857603] env[63515]: DEBUG nova.virt.hardware [None req-0bb813f4-ce4d-4144-b80d-b179b1c5fd20 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Got 1 possible topologies {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 918.857700] env[63515]: DEBUG nova.virt.hardware [None req-0bb813f4-ce4d-4144-b80d-b179b1c5fd20 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 918.857801] env[63515]: DEBUG nova.virt.hardware [None req-0bb813f4-ce4d-4144-b80d-b179b1c5fd20 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 918.858798] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e31b9575-8d43-4f48-8d08-d2826df0738e {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.868156] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9c0d579-8c3e-4625-bfa5-171682748880 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.883656] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-f7b46a02-c5ff-4523-967a-55b60fa40b2d tempest-ServerActionsV293TestJSON-1193370169 tempest-ServerActionsV293TestJSON-1193370169-project-member] [instance: 2ea99c7a-8b61-4718-bee0-f4ce4fc021af] Volume detach. Driver type: vmdk {{(pid=63515) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 918.883987] env[63515]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-12c0d2d8-8468-4644-91da-8dbf14e31b32 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.893028] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a3b8c3c-3822-4928-9108-82d7f15b7830 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.927385] env[63515]: ERROR nova.compute.manager [None req-f7b46a02-c5ff-4523-967a-55b60fa40b2d tempest-ServerActionsV293TestJSON-1193370169 tempest-ServerActionsV293TestJSON-1193370169-project-member] [instance: 2ea99c7a-8b61-4718-bee0-f4ce4fc021af] Failed to detach volume 211b1a87-a01e-4d13-aa27-91391bf9cde1 from /dev/sda: nova.exception.InstanceNotFound: Instance 2ea99c7a-8b61-4718-bee0-f4ce4fc021af could not be found. [ 918.927385] env[63515]: ERROR nova.compute.manager [instance: 2ea99c7a-8b61-4718-bee0-f4ce4fc021af] Traceback (most recent call last): [ 918.927385] env[63515]: ERROR nova.compute.manager [instance: 2ea99c7a-8b61-4718-bee0-f4ce4fc021af] File "/opt/stack/nova/nova/compute/manager.py", line 4142, in _do_rebuild_instance [ 918.927385] env[63515]: ERROR nova.compute.manager [instance: 2ea99c7a-8b61-4718-bee0-f4ce4fc021af] self.driver.rebuild(**kwargs) [ 918.927385] env[63515]: ERROR nova.compute.manager [instance: 2ea99c7a-8b61-4718-bee0-f4ce4fc021af] File "/opt/stack/nova/nova/virt/driver.py", line 493, in rebuild [ 918.927385] env[63515]: ERROR nova.compute.manager [instance: 2ea99c7a-8b61-4718-bee0-f4ce4fc021af] raise NotImplementedError() [ 918.927385] env[63515]: ERROR nova.compute.manager [instance: 2ea99c7a-8b61-4718-bee0-f4ce4fc021af] NotImplementedError [ 918.927385] env[63515]: ERROR nova.compute.manager [instance: 2ea99c7a-8b61-4718-bee0-f4ce4fc021af] [ 918.927385] env[63515]: ERROR nova.compute.manager [instance: 2ea99c7a-8b61-4718-bee0-f4ce4fc021af] During handling of the above exception, another exception occurred: [ 918.927385] env[63515]: ERROR nova.compute.manager [instance: 2ea99c7a-8b61-4718-bee0-f4ce4fc021af] [ 918.927385] env[63515]: ERROR nova.compute.manager [instance: 2ea99c7a-8b61-4718-bee0-f4ce4fc021af] Traceback (most recent call last): [ 918.927385] env[63515]: ERROR nova.compute.manager [instance: 2ea99c7a-8b61-4718-bee0-f4ce4fc021af] File "/opt/stack/nova/nova/compute/manager.py", line 3565, in _detach_root_volume [ 918.927385] env[63515]: ERROR nova.compute.manager [instance: 2ea99c7a-8b61-4718-bee0-f4ce4fc021af] self.driver.detach_volume(context, old_connection_info, [ 918.927385] env[63515]: ERROR nova.compute.manager [instance: 2ea99c7a-8b61-4718-bee0-f4ce4fc021af] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 553, in detach_volume [ 918.927385] env[63515]: ERROR nova.compute.manager [instance: 2ea99c7a-8b61-4718-bee0-f4ce4fc021af] return self._volumeops.detach_volume(connection_info, instance) [ 918.927385] env[63515]: ERROR nova.compute.manager [instance: 2ea99c7a-8b61-4718-bee0-f4ce4fc021af] File "/opt/stack/nova/nova/virt/vmwareapi/volumeops.py", line 649, in detach_volume [ 918.927385] env[63515]: ERROR nova.compute.manager [instance: 2ea99c7a-8b61-4718-bee0-f4ce4fc021af] self._detach_volume_vmdk(connection_info, instance) [ 918.927385] env[63515]: ERROR nova.compute.manager [instance: 2ea99c7a-8b61-4718-bee0-f4ce4fc021af] File "/opt/stack/nova/nova/virt/vmwareapi/volumeops.py", line 569, in _detach_volume_vmdk [ 918.927385] env[63515]: ERROR nova.compute.manager [instance: 2ea99c7a-8b61-4718-bee0-f4ce4fc021af] vm_ref = vm_util.get_vm_ref(self._session, instance) [ 918.927385] env[63515]: ERROR nova.compute.manager [instance: 2ea99c7a-8b61-4718-bee0-f4ce4fc021af] File "/opt/stack/nova/nova/virt/vmwareapi/vm_util.py", line 1145, in get_vm_ref [ 918.927385] env[63515]: ERROR nova.compute.manager [instance: 2ea99c7a-8b61-4718-bee0-f4ce4fc021af] stable_ref.fetch_moref(session) [ 918.927385] env[63515]: ERROR nova.compute.manager [instance: 2ea99c7a-8b61-4718-bee0-f4ce4fc021af] File "/opt/stack/nova/nova/virt/vmwareapi/vm_util.py", line 1136, in fetch_moref [ 918.927385] env[63515]: ERROR nova.compute.manager [instance: 2ea99c7a-8b61-4718-bee0-f4ce4fc021af] raise exception.InstanceNotFound(instance_id=self._uuid) [ 918.927385] env[63515]: ERROR nova.compute.manager [instance: 2ea99c7a-8b61-4718-bee0-f4ce4fc021af] nova.exception.InstanceNotFound: Instance 2ea99c7a-8b61-4718-bee0-f4ce4fc021af could not be found. [ 918.927385] env[63515]: ERROR nova.compute.manager [instance: 2ea99c7a-8b61-4718-bee0-f4ce4fc021af] [ 918.945375] env[63515]: DEBUG oslo_vmware.api [None req-8c6d15be-1f5e-44a9-a3e4-dab9aa82e26a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Task: {'id': task-1111571, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.072133] env[63515]: DEBUG nova.compute.utils [None req-f7b46a02-c5ff-4523-967a-55b60fa40b2d tempest-ServerActionsV293TestJSON-1193370169 tempest-ServerActionsV293TestJSON-1193370169-project-member] [instance: 2ea99c7a-8b61-4718-bee0-f4ce4fc021af] Build of instance 2ea99c7a-8b61-4718-bee0-f4ce4fc021af aborted: Failed to rebuild volume backed instance. {{(pid=63515) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 919.074021] env[63515]: ERROR nova.compute.manager [None req-f7b46a02-c5ff-4523-967a-55b60fa40b2d tempest-ServerActionsV293TestJSON-1193370169 tempest-ServerActionsV293TestJSON-1193370169-project-member] [instance: 2ea99c7a-8b61-4718-bee0-f4ce4fc021af] Setting instance vm_state to ERROR: nova.exception.BuildAbortException: Build of instance 2ea99c7a-8b61-4718-bee0-f4ce4fc021af aborted: Failed to rebuild volume backed instance. [ 919.074021] env[63515]: ERROR nova.compute.manager [instance: 2ea99c7a-8b61-4718-bee0-f4ce4fc021af] Traceback (most recent call last): [ 919.074021] env[63515]: ERROR nova.compute.manager [instance: 2ea99c7a-8b61-4718-bee0-f4ce4fc021af] File "/opt/stack/nova/nova/compute/manager.py", line 4142, in _do_rebuild_instance [ 919.074021] env[63515]: ERROR nova.compute.manager [instance: 2ea99c7a-8b61-4718-bee0-f4ce4fc021af] self.driver.rebuild(**kwargs) [ 919.074021] env[63515]: ERROR nova.compute.manager [instance: 2ea99c7a-8b61-4718-bee0-f4ce4fc021af] File "/opt/stack/nova/nova/virt/driver.py", line 493, in rebuild [ 919.074021] env[63515]: ERROR nova.compute.manager [instance: 2ea99c7a-8b61-4718-bee0-f4ce4fc021af] raise NotImplementedError() [ 919.074021] env[63515]: ERROR nova.compute.manager [instance: 2ea99c7a-8b61-4718-bee0-f4ce4fc021af] NotImplementedError [ 919.074021] env[63515]: ERROR nova.compute.manager [instance: 2ea99c7a-8b61-4718-bee0-f4ce4fc021af] [ 919.074021] env[63515]: ERROR nova.compute.manager [instance: 2ea99c7a-8b61-4718-bee0-f4ce4fc021af] During handling of the above exception, another exception occurred: [ 919.074021] env[63515]: ERROR nova.compute.manager [instance: 2ea99c7a-8b61-4718-bee0-f4ce4fc021af] [ 919.074021] env[63515]: ERROR nova.compute.manager [instance: 2ea99c7a-8b61-4718-bee0-f4ce4fc021af] Traceback (most recent call last): [ 919.074021] env[63515]: ERROR nova.compute.manager [instance: 2ea99c7a-8b61-4718-bee0-f4ce4fc021af] File "/opt/stack/nova/nova/compute/manager.py", line 3600, in _rebuild_volume_backed_instance [ 919.074021] env[63515]: ERROR nova.compute.manager [instance: 2ea99c7a-8b61-4718-bee0-f4ce4fc021af] self._detach_root_volume(context, instance, root_bdm) [ 919.074021] env[63515]: ERROR nova.compute.manager [instance: 2ea99c7a-8b61-4718-bee0-f4ce4fc021af] File "/opt/stack/nova/nova/compute/manager.py", line 3579, in _detach_root_volume [ 919.074021] env[63515]: ERROR nova.compute.manager [instance: 2ea99c7a-8b61-4718-bee0-f4ce4fc021af] with excutils.save_and_reraise_exception(): [ 919.074021] env[63515]: ERROR nova.compute.manager [instance: 2ea99c7a-8b61-4718-bee0-f4ce4fc021af] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 919.074021] env[63515]: ERROR nova.compute.manager [instance: 2ea99c7a-8b61-4718-bee0-f4ce4fc021af] self.force_reraise() [ 919.074021] env[63515]: ERROR nova.compute.manager [instance: 2ea99c7a-8b61-4718-bee0-f4ce4fc021af] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 919.074021] env[63515]: ERROR nova.compute.manager [instance: 2ea99c7a-8b61-4718-bee0-f4ce4fc021af] raise self.value [ 919.074021] env[63515]: ERROR nova.compute.manager [instance: 2ea99c7a-8b61-4718-bee0-f4ce4fc021af] File "/opt/stack/nova/nova/compute/manager.py", line 3565, in _detach_root_volume [ 919.074021] env[63515]: ERROR nova.compute.manager [instance: 2ea99c7a-8b61-4718-bee0-f4ce4fc021af] self.driver.detach_volume(context, old_connection_info, [ 919.074021] env[63515]: ERROR nova.compute.manager [instance: 2ea99c7a-8b61-4718-bee0-f4ce4fc021af] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 553, in detach_volume [ 919.074021] env[63515]: ERROR nova.compute.manager [instance: 2ea99c7a-8b61-4718-bee0-f4ce4fc021af] return self._volumeops.detach_volume(connection_info, instance) [ 919.074021] env[63515]: ERROR nova.compute.manager [instance: 2ea99c7a-8b61-4718-bee0-f4ce4fc021af] File "/opt/stack/nova/nova/virt/vmwareapi/volumeops.py", line 649, in detach_volume [ 919.074021] env[63515]: ERROR nova.compute.manager [instance: 2ea99c7a-8b61-4718-bee0-f4ce4fc021af] self._detach_volume_vmdk(connection_info, instance) [ 919.074021] env[63515]: ERROR nova.compute.manager [instance: 2ea99c7a-8b61-4718-bee0-f4ce4fc021af] File "/opt/stack/nova/nova/virt/vmwareapi/volumeops.py", line 569, in _detach_volume_vmdk [ 919.074021] env[63515]: ERROR nova.compute.manager [instance: 2ea99c7a-8b61-4718-bee0-f4ce4fc021af] vm_ref = vm_util.get_vm_ref(self._session, instance) [ 919.074021] env[63515]: ERROR nova.compute.manager [instance: 2ea99c7a-8b61-4718-bee0-f4ce4fc021af] File "/opt/stack/nova/nova/virt/vmwareapi/vm_util.py", line 1145, in get_vm_ref [ 919.074021] env[63515]: ERROR nova.compute.manager [instance: 2ea99c7a-8b61-4718-bee0-f4ce4fc021af] stable_ref.fetch_moref(session) [ 919.074021] env[63515]: ERROR nova.compute.manager [instance: 2ea99c7a-8b61-4718-bee0-f4ce4fc021af] File "/opt/stack/nova/nova/virt/vmwareapi/vm_util.py", line 1136, in fetch_moref [ 919.074021] env[63515]: ERROR nova.compute.manager [instance: 2ea99c7a-8b61-4718-bee0-f4ce4fc021af] raise exception.InstanceNotFound(instance_id=self._uuid) [ 919.074021] env[63515]: ERROR nova.compute.manager [instance: 2ea99c7a-8b61-4718-bee0-f4ce4fc021af] nova.exception.InstanceNotFound: Instance 2ea99c7a-8b61-4718-bee0-f4ce4fc021af could not be found. [ 919.074021] env[63515]: ERROR nova.compute.manager [instance: 2ea99c7a-8b61-4718-bee0-f4ce4fc021af] [ 919.074021] env[63515]: ERROR nova.compute.manager [instance: 2ea99c7a-8b61-4718-bee0-f4ce4fc021af] During handling of the above exception, another exception occurred: [ 919.074021] env[63515]: ERROR nova.compute.manager [instance: 2ea99c7a-8b61-4718-bee0-f4ce4fc021af] [ 919.074021] env[63515]: ERROR nova.compute.manager [instance: 2ea99c7a-8b61-4718-bee0-f4ce4fc021af] Traceback (most recent call last): [ 919.074021] env[63515]: ERROR nova.compute.manager [instance: 2ea99c7a-8b61-4718-bee0-f4ce4fc021af] File "/opt/stack/nova/nova/compute/manager.py", line 10863, in _error_out_instance_on_exception [ 919.074021] env[63515]: ERROR nova.compute.manager [instance: 2ea99c7a-8b61-4718-bee0-f4ce4fc021af] yield [ 919.074021] env[63515]: ERROR nova.compute.manager [instance: 2ea99c7a-8b61-4718-bee0-f4ce4fc021af] File "/opt/stack/nova/nova/compute/manager.py", line 3868, in rebuild_instance [ 919.074021] env[63515]: ERROR nova.compute.manager [instance: 2ea99c7a-8b61-4718-bee0-f4ce4fc021af] self._do_rebuild_instance_with_claim( [ 919.075179] env[63515]: ERROR nova.compute.manager [instance: 2ea99c7a-8b61-4718-bee0-f4ce4fc021af] File "/opt/stack/nova/nova/compute/manager.py", line 3954, in _do_rebuild_instance_with_claim [ 919.075179] env[63515]: ERROR nova.compute.manager [instance: 2ea99c7a-8b61-4718-bee0-f4ce4fc021af] self._do_rebuild_instance( [ 919.075179] env[63515]: ERROR nova.compute.manager [instance: 2ea99c7a-8b61-4718-bee0-f4ce4fc021af] File "/opt/stack/nova/nova/compute/manager.py", line 4146, in _do_rebuild_instance [ 919.075179] env[63515]: ERROR nova.compute.manager [instance: 2ea99c7a-8b61-4718-bee0-f4ce4fc021af] self._rebuild_default_impl(**kwargs) [ 919.075179] env[63515]: ERROR nova.compute.manager [instance: 2ea99c7a-8b61-4718-bee0-f4ce4fc021af] File "/opt/stack/nova/nova/compute/manager.py", line 3723, in _rebuild_default_impl [ 919.075179] env[63515]: ERROR nova.compute.manager [instance: 2ea99c7a-8b61-4718-bee0-f4ce4fc021af] self._rebuild_volume_backed_instance( [ 919.075179] env[63515]: ERROR nova.compute.manager [instance: 2ea99c7a-8b61-4718-bee0-f4ce4fc021af] File "/opt/stack/nova/nova/compute/manager.py", line 3615, in _rebuild_volume_backed_instance [ 919.075179] env[63515]: ERROR nova.compute.manager [instance: 2ea99c7a-8b61-4718-bee0-f4ce4fc021af] raise exception.BuildAbortException( [ 919.075179] env[63515]: ERROR nova.compute.manager [instance: 2ea99c7a-8b61-4718-bee0-f4ce4fc021af] nova.exception.BuildAbortException: Build of instance 2ea99c7a-8b61-4718-bee0-f4ce4fc021af aborted: Failed to rebuild volume backed instance. [ 919.075179] env[63515]: ERROR nova.compute.manager [instance: 2ea99c7a-8b61-4718-bee0-f4ce4fc021af] [ 919.101934] env[63515]: INFO nova.compute.manager [None req-094b4e02-a75f-49fa-ba8e-0bf8adbbeb29 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] [instance: 2a29f37d-e786-48d3-b126-4854d6e7c67c] Rescuing [ 919.103033] env[63515]: DEBUG oslo_concurrency.lockutils [None req-094b4e02-a75f-49fa-ba8e-0bf8adbbeb29 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Acquiring lock "refresh_cache-2a29f37d-e786-48d3-b126-4854d6e7c67c" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 919.103033] env[63515]: DEBUG oslo_concurrency.lockutils [None req-094b4e02-a75f-49fa-ba8e-0bf8adbbeb29 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Acquired lock "refresh_cache-2a29f37d-e786-48d3-b126-4854d6e7c67c" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 919.103033] env[63515]: DEBUG nova.network.neutron [None req-094b4e02-a75f-49fa-ba8e-0bf8adbbeb29 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] [instance: 2a29f37d-e786-48d3-b126-4854d6e7c67c] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 919.133539] env[63515]: DEBUG oslo_concurrency.lockutils [None req-525a7c59-1fc2-4bbe-98db-35f4ad5c239b tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Acquiring lock "interface-eab16df9-7bb5-4576-bca0-769a561c5fe9-None" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 919.133819] env[63515]: DEBUG oslo_concurrency.lockutils [None req-525a7c59-1fc2-4bbe-98db-35f4ad5c239b tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Lock "interface-eab16df9-7bb5-4576-bca0-769a561c5fe9-None" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 919.134196] env[63515]: DEBUG nova.objects.instance [None req-525a7c59-1fc2-4bbe-98db-35f4ad5c239b tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Lazy-loading 'flavor' on Instance uuid eab16df9-7bb5-4576-bca0-769a561c5fe9 {{(pid=63515) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 919.241767] env[63515]: DEBUG oslo_concurrency.lockutils [None req-ef0f5ab2-1246-4de1-af39-f1efd470cd76 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.436s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 919.244258] env[63515]: DEBUG oslo_concurrency.lockutils [None req-358377cb-a02d-42b1-aec7-e1ec3bad941d tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 8.220s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 919.244512] env[63515]: DEBUG nova.objects.instance [None req-358377cb-a02d-42b1-aec7-e1ec3bad941d tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] Lazy-loading 'resources' on Instance uuid 324f7e89-3c17-474d-af1b-6d7a6d042510 {{(pid=63515) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 919.267197] env[63515]: INFO nova.compute.manager [None req-8c6d15be-1f5e-44a9-a3e4-dab9aa82e26a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] [instance: 9236607d-47af-4272-93db-79826e918aa1] Took 17.73 seconds to build instance. [ 919.282597] env[63515]: DEBUG oslo_concurrency.lockutils [None req-9788214a-e67e-4df5-bfb0-11119b5fe460 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Lock "2a29f37d-e786-48d3-b126-4854d6e7c67c" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.820s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 919.347908] env[63515]: DEBUG nova.network.neutron [None req-0bb813f4-ce4d-4144-b80d-b179b1c5fd20 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: a2203c5e-66b9-4c7e-8d76-e12a08b57b1d] Successfully updated port: 8b878aac-8955-4468-893f-63fdbfc0c8cb {{(pid=63515) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 919.445955] env[63515]: DEBUG oslo_vmware.api [None req-8c6d15be-1f5e-44a9-a3e4-dab9aa82e26a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Task: {'id': task-1111571, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.658675} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 919.446294] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-8c6d15be-1f5e-44a9-a3e4-dab9aa82e26a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk to [datastore2] ca6a8ac3-d880-4333-9816-2f076e423844/ca6a8ac3-d880-4333-9816-2f076e423844.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 919.446524] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-8c6d15be-1f5e-44a9-a3e4-dab9aa82e26a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] [instance: ca6a8ac3-d880-4333-9816-2f076e423844] Extending root virtual disk to 1048576 {{(pid=63515) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 919.446783] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-98ceae2e-ac68-4eea-9b64-6a3d5879c113 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.453311] env[63515]: DEBUG oslo_vmware.api [None req-8c6d15be-1f5e-44a9-a3e4-dab9aa82e26a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Waiting for the task: (returnval){ [ 919.453311] env[63515]: value = "task-1111572" [ 919.453311] env[63515]: _type = "Task" [ 919.453311] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 919.461111] env[63515]: DEBUG oslo_vmware.api [None req-8c6d15be-1f5e-44a9-a3e4-dab9aa82e26a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Task: {'id': task-1111572, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.738678] env[63515]: DEBUG nova.objects.instance [None req-525a7c59-1fc2-4bbe-98db-35f4ad5c239b tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Lazy-loading 'pci_requests' on Instance uuid eab16df9-7bb5-4576-bca0-769a561c5fe9 {{(pid=63515) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 919.752969] env[63515]: DEBUG oslo_concurrency.lockutils [None req-ef0f5ab2-1246-4de1-af39-f1efd470cd76 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Lock "ae5d6f34-bac6-4ba0-b0c3-8205982ddc88" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 29.384s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 919.753810] env[63515]: DEBUG oslo_concurrency.lockutils [None req-bd577155-aa2b-4775-93e5-5bcfdcad6cd6 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Lock "ae5d6f34-bac6-4ba0-b0c3-8205982ddc88" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 8.167s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 919.754049] env[63515]: DEBUG oslo_concurrency.lockutils [None req-bd577155-aa2b-4775-93e5-5bcfdcad6cd6 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Acquiring lock "ae5d6f34-bac6-4ba0-b0c3-8205982ddc88-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 919.754385] env[63515]: DEBUG oslo_concurrency.lockutils [None req-bd577155-aa2b-4775-93e5-5bcfdcad6cd6 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Lock "ae5d6f34-bac6-4ba0-b0c3-8205982ddc88-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 919.754482] env[63515]: DEBUG oslo_concurrency.lockutils [None req-bd577155-aa2b-4775-93e5-5bcfdcad6cd6 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Lock "ae5d6f34-bac6-4ba0-b0c3-8205982ddc88-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 919.757429] env[63515]: INFO nova.compute.manager [None req-bd577155-aa2b-4775-93e5-5bcfdcad6cd6 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: ae5d6f34-bac6-4ba0-b0c3-8205982ddc88] Terminating instance [ 919.759309] env[63515]: DEBUG nova.compute.manager [None req-bd577155-aa2b-4775-93e5-5bcfdcad6cd6 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: ae5d6f34-bac6-4ba0-b0c3-8205982ddc88] Start destroying the instance on the hypervisor. {{(pid=63515) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 919.759467] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-bd577155-aa2b-4775-93e5-5bcfdcad6cd6 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: ae5d6f34-bac6-4ba0-b0c3-8205982ddc88] Destroying instance {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 919.759723] env[63515]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9b04383e-ac18-4a8d-96af-3db5961ec729 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.769514] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c357d02-112a-41a6-9845-9c1fa1db953d {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.785399] env[63515]: DEBUG oslo_concurrency.lockutils [None req-8c6d15be-1f5e-44a9-a3e4-dab9aa82e26a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Lock "9236607d-47af-4272-93db-79826e918aa1" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 19.259s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 919.795033] env[63515]: DEBUG nova.compute.manager [req-18683797-5fa5-4eb7-95fe-fc7693d41e6f req-a4ce9472-8386-4873-8d7c-920ce5c8e98f service nova] [instance: a2203c5e-66b9-4c7e-8d76-e12a08b57b1d] Received event network-vif-plugged-8b878aac-8955-4468-893f-63fdbfc0c8cb {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 919.795268] env[63515]: DEBUG oslo_concurrency.lockutils [req-18683797-5fa5-4eb7-95fe-fc7693d41e6f req-a4ce9472-8386-4873-8d7c-920ce5c8e98f service nova] Acquiring lock "a2203c5e-66b9-4c7e-8d76-e12a08b57b1d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 919.796345] env[63515]: DEBUG oslo_concurrency.lockutils [req-18683797-5fa5-4eb7-95fe-fc7693d41e6f req-a4ce9472-8386-4873-8d7c-920ce5c8e98f service nova] Lock "a2203c5e-66b9-4c7e-8d76-e12a08b57b1d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 919.796432] env[63515]: DEBUG oslo_concurrency.lockutils [req-18683797-5fa5-4eb7-95fe-fc7693d41e6f req-a4ce9472-8386-4873-8d7c-920ce5c8e98f service nova] Lock "a2203c5e-66b9-4c7e-8d76-e12a08b57b1d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 919.796615] env[63515]: DEBUG nova.compute.manager [req-18683797-5fa5-4eb7-95fe-fc7693d41e6f req-a4ce9472-8386-4873-8d7c-920ce5c8e98f service nova] [instance: a2203c5e-66b9-4c7e-8d76-e12a08b57b1d] No waiting events found dispatching network-vif-plugged-8b878aac-8955-4468-893f-63fdbfc0c8cb {{(pid=63515) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 919.796788] env[63515]: WARNING nova.compute.manager [req-18683797-5fa5-4eb7-95fe-fc7693d41e6f req-a4ce9472-8386-4873-8d7c-920ce5c8e98f service nova] [instance: a2203c5e-66b9-4c7e-8d76-e12a08b57b1d] Received unexpected event network-vif-plugged-8b878aac-8955-4468-893f-63fdbfc0c8cb for instance with vm_state building and task_state spawning. [ 919.796956] env[63515]: DEBUG nova.compute.manager [req-18683797-5fa5-4eb7-95fe-fc7693d41e6f req-a4ce9472-8386-4873-8d7c-920ce5c8e98f service nova] [instance: a2203c5e-66b9-4c7e-8d76-e12a08b57b1d] Received event network-changed-8b878aac-8955-4468-893f-63fdbfc0c8cb {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 919.797133] env[63515]: DEBUG nova.compute.manager [req-18683797-5fa5-4eb7-95fe-fc7693d41e6f req-a4ce9472-8386-4873-8d7c-920ce5c8e98f service nova] [instance: a2203c5e-66b9-4c7e-8d76-e12a08b57b1d] Refreshing instance network info cache due to event network-changed-8b878aac-8955-4468-893f-63fdbfc0c8cb. {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 919.797328] env[63515]: DEBUG oslo_concurrency.lockutils [req-18683797-5fa5-4eb7-95fe-fc7693d41e6f req-a4ce9472-8386-4873-8d7c-920ce5c8e98f service nova] Acquiring lock "refresh_cache-a2203c5e-66b9-4c7e-8d76-e12a08b57b1d" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 919.797469] env[63515]: DEBUG oslo_concurrency.lockutils [req-18683797-5fa5-4eb7-95fe-fc7693d41e6f req-a4ce9472-8386-4873-8d7c-920ce5c8e98f service nova] Acquired lock "refresh_cache-a2203c5e-66b9-4c7e-8d76-e12a08b57b1d" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 919.797629] env[63515]: DEBUG nova.network.neutron [req-18683797-5fa5-4eb7-95fe-fc7693d41e6f req-a4ce9472-8386-4873-8d7c-920ce5c8e98f service nova] [instance: a2203c5e-66b9-4c7e-8d76-e12a08b57b1d] Refreshing network info cache for port 8b878aac-8955-4468-893f-63fdbfc0c8cb {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 919.811473] env[63515]: WARNING nova.virt.vmwareapi.vmops [None req-bd577155-aa2b-4775-93e5-5bcfdcad6cd6 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: ae5d6f34-bac6-4ba0-b0c3-8205982ddc88] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance ae5d6f34-bac6-4ba0-b0c3-8205982ddc88 could not be found. [ 919.811675] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-bd577155-aa2b-4775-93e5-5bcfdcad6cd6 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: ae5d6f34-bac6-4ba0-b0c3-8205982ddc88] Instance destroyed {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 919.811852] env[63515]: INFO nova.compute.manager [None req-bd577155-aa2b-4775-93e5-5bcfdcad6cd6 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: ae5d6f34-bac6-4ba0-b0c3-8205982ddc88] Took 0.05 seconds to destroy the instance on the hypervisor. [ 919.812102] env[63515]: DEBUG oslo.service.loopingcall [None req-bd577155-aa2b-4775-93e5-5bcfdcad6cd6 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 919.814515] env[63515]: DEBUG nova.compute.manager [-] [instance: ae5d6f34-bac6-4ba0-b0c3-8205982ddc88] Deallocating network for instance {{(pid=63515) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 919.815484] env[63515]: DEBUG nova.network.neutron [-] [instance: ae5d6f34-bac6-4ba0-b0c3-8205982ddc88] deallocate_for_instance() {{(pid=63515) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 919.852032] env[63515]: DEBUG oslo_concurrency.lockutils [None req-0bb813f4-ce4d-4144-b80d-b179b1c5fd20 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Acquiring lock "refresh_cache-a2203c5e-66b9-4c7e-8d76-e12a08b57b1d" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 919.857784] env[63515]: DEBUG nova.network.neutron [None req-094b4e02-a75f-49fa-ba8e-0bf8adbbeb29 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] [instance: 2a29f37d-e786-48d3-b126-4854d6e7c67c] Updating instance_info_cache with network_info: [{"id": "956cd361-0d42-4914-85d3-5f75a72a68a1", "address": "fa:16:3e:a3:72:bb", "network": {"id": "850bfa95-3df6-4019-8260-ade20ec58799", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-832000501-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.153", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "08c9aa51b1864410809f8877c7cd49c4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3cf748a8-7ae0-4dca-817d-e727c30d72f4", "external-id": "nsx-vlan-transportzone-853", "segmentation_id": 853, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap956cd361-0d", "ovs_interfaceid": "956cd361-0d42-4914-85d3-5f75a72a68a1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 919.859551] env[63515]: DEBUG nova.network.neutron [req-18683797-5fa5-4eb7-95fe-fc7693d41e6f req-a4ce9472-8386-4873-8d7c-920ce5c8e98f service nova] [instance: a2203c5e-66b9-4c7e-8d76-e12a08b57b1d] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 919.966042] env[63515]: DEBUG oslo_vmware.api [None req-8c6d15be-1f5e-44a9-a3e4-dab9aa82e26a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Task: {'id': task-1111572, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.262882} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 919.968658] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-8c6d15be-1f5e-44a9-a3e4-dab9aa82e26a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] [instance: ca6a8ac3-d880-4333-9816-2f076e423844] Extended root virtual disk {{(pid=63515) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 919.970230] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48e862c7-6d5b-4d7a-9ec8-a9e97be61db5 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.996765] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-8c6d15be-1f5e-44a9-a3e4-dab9aa82e26a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] [instance: ca6a8ac3-d880-4333-9816-2f076e423844] Reconfiguring VM instance instance-00000053 to attach disk [datastore2] ca6a8ac3-d880-4333-9816-2f076e423844/ca6a8ac3-d880-4333-9816-2f076e423844.vmdk or device None with type sparse {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 919.999754] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-93a98417-ce3e-46b7-b272-7b7ab9d7efab {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.020035] env[63515]: DEBUG oslo_vmware.api [None req-8c6d15be-1f5e-44a9-a3e4-dab9aa82e26a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Waiting for the task: (returnval){ [ 920.020035] env[63515]: value = "task-1111573" [ 920.020035] env[63515]: _type = "Task" [ 920.020035] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 920.029989] env[63515]: DEBUG oslo_vmware.api [None req-8c6d15be-1f5e-44a9-a3e4-dab9aa82e26a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Task: {'id': task-1111573, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.038802] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd6e9741-6e53-4bb8-b499-c70423a475f0 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.045483] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-722bfbfd-8e64-4895-8223-80566a32181d {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.049699] env[63515]: DEBUG nova.network.neutron [req-18683797-5fa5-4eb7-95fe-fc7693d41e6f req-a4ce9472-8386-4873-8d7c-920ce5c8e98f service nova] [instance: a2203c5e-66b9-4c7e-8d76-e12a08b57b1d] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 920.076166] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-051a5b57-bf60-4d7f-835d-f9d986c4fd8f {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.086684] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8377f60a-f1ba-4c49-adb2-a2e252492917 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.102015] env[63515]: DEBUG nova.compute.provider_tree [None req-358377cb-a02d-42b1-aec7-e1ec3bad941d tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 920.241789] env[63515]: DEBUG nova.objects.base [None req-525a7c59-1fc2-4bbe-98db-35f4ad5c239b tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Object Instance lazy-loaded attributes: flavor,pci_requests {{(pid=63515) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 920.242062] env[63515]: DEBUG nova.network.neutron [None req-525a7c59-1fc2-4bbe-98db-35f4ad5c239b tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: eab16df9-7bb5-4576-bca0-769a561c5fe9] allocate_for_instance() {{(pid=63515) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 920.304613] env[63515]: DEBUG nova.policy [None req-525a7c59-1fc2-4bbe-98db-35f4ad5c239b tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '16880abf1c5b4341800c94ada2c756c7', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7d4ada9ec35f42b19c6480a9101d21a4', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63515) authorize /opt/stack/nova/nova/policy.py:201}} [ 920.363038] env[63515]: DEBUG oslo_concurrency.lockutils [None req-094b4e02-a75f-49fa-ba8e-0bf8adbbeb29 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Releasing lock "refresh_cache-2a29f37d-e786-48d3-b126-4854d6e7c67c" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 920.532290] env[63515]: DEBUG oslo_vmware.api [None req-8c6d15be-1f5e-44a9-a3e4-dab9aa82e26a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Task: {'id': task-1111573, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.552022] env[63515]: DEBUG oslo_concurrency.lockutils [req-18683797-5fa5-4eb7-95fe-fc7693d41e6f req-a4ce9472-8386-4873-8d7c-920ce5c8e98f service nova] Releasing lock "refresh_cache-a2203c5e-66b9-4c7e-8d76-e12a08b57b1d" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 920.552399] env[63515]: DEBUG oslo_concurrency.lockutils [None req-0bb813f4-ce4d-4144-b80d-b179b1c5fd20 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Acquired lock "refresh_cache-a2203c5e-66b9-4c7e-8d76-e12a08b57b1d" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 920.552564] env[63515]: DEBUG nova.network.neutron [None req-0bb813f4-ce4d-4144-b80d-b179b1c5fd20 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: a2203c5e-66b9-4c7e-8d76-e12a08b57b1d] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 920.606053] env[63515]: DEBUG nova.scheduler.client.report [None req-358377cb-a02d-42b1-aec7-e1ec3bad941d tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 920.653017] env[63515]: DEBUG nova.network.neutron [None req-525a7c59-1fc2-4bbe-98db-35f4ad5c239b tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: eab16df9-7bb5-4576-bca0-769a561c5fe9] Successfully created port: e3c72582-41bf-48d4-8614-b516a96362a7 {{(pid=63515) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 920.668760] env[63515]: DEBUG nova.network.neutron [-] [instance: ae5d6f34-bac6-4ba0-b0c3-8205982ddc88] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 920.904654] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-094b4e02-a75f-49fa-ba8e-0bf8adbbeb29 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] [instance: 2a29f37d-e786-48d3-b126-4854d6e7c67c] Powering off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 920.904940] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f404155e-0f24-448e-a246-ffc4ab51ddaf {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.912167] env[63515]: DEBUG oslo_vmware.api [None req-094b4e02-a75f-49fa-ba8e-0bf8adbbeb29 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Waiting for the task: (returnval){ [ 920.912167] env[63515]: value = "task-1111574" [ 920.912167] env[63515]: _type = "Task" [ 920.912167] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 920.923338] env[63515]: DEBUG oslo_vmware.api [None req-094b4e02-a75f-49fa-ba8e-0bf8adbbeb29 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Task: {'id': task-1111574, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.031440] env[63515]: DEBUG oslo_vmware.api [None req-8c6d15be-1f5e-44a9-a3e4-dab9aa82e26a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Task: {'id': task-1111573, 'name': ReconfigVM_Task, 'duration_secs': 0.562366} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 921.031742] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-8c6d15be-1f5e-44a9-a3e4-dab9aa82e26a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] [instance: ca6a8ac3-d880-4333-9816-2f076e423844] Reconfigured VM instance instance-00000053 to attach disk [datastore2] ca6a8ac3-d880-4333-9816-2f076e423844/ca6a8ac3-d880-4333-9816-2f076e423844.vmdk or device None with type sparse {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 921.032393] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c3957de1-541c-4628-b101-fe5802c36d10 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.039303] env[63515]: DEBUG oslo_vmware.api [None req-8c6d15be-1f5e-44a9-a3e4-dab9aa82e26a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Waiting for the task: (returnval){ [ 921.039303] env[63515]: value = "task-1111575" [ 921.039303] env[63515]: _type = "Task" [ 921.039303] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 921.048101] env[63515]: DEBUG oslo_vmware.api [None req-8c6d15be-1f5e-44a9-a3e4-dab9aa82e26a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Task: {'id': task-1111575, 'name': Rename_Task} progress is 5%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.089839] env[63515]: DEBUG oslo_concurrency.lockutils [None req-f7b46a02-c5ff-4523-967a-55b60fa40b2d tempest-ServerActionsV293TestJSON-1193370169 tempest-ServerActionsV293TestJSON-1193370169-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 921.102747] env[63515]: DEBUG nova.network.neutron [None req-0bb813f4-ce4d-4144-b80d-b179b1c5fd20 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: a2203c5e-66b9-4c7e-8d76-e12a08b57b1d] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 921.111089] env[63515]: DEBUG oslo_concurrency.lockutils [None req-358377cb-a02d-42b1-aec7-e1ec3bad941d tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.867s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 921.114028] env[63515]: DEBUG oslo_concurrency.lockutils [None req-1d5e76d4-1b01-4bdb-831a-7cba26a22dfc tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 8.753s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 921.136500] env[63515]: INFO nova.scheduler.client.report [None req-358377cb-a02d-42b1-aec7-e1ec3bad941d tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] Deleted allocations for instance 324f7e89-3c17-474d-af1b-6d7a6d042510 [ 921.171767] env[63515]: INFO nova.compute.manager [-] [instance: ae5d6f34-bac6-4ba0-b0c3-8205982ddc88] Took 1.36 seconds to deallocate network for instance. [ 921.425483] env[63515]: DEBUG oslo_vmware.api [None req-094b4e02-a75f-49fa-ba8e-0bf8adbbeb29 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Task: {'id': task-1111574, 'name': PowerOffVM_Task, 'duration_secs': 0.228341} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 921.425852] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-094b4e02-a75f-49fa-ba8e-0bf8adbbeb29 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] [instance: 2a29f37d-e786-48d3-b126-4854d6e7c67c] Powered off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 921.426993] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ea3b3dc-102a-40cf-93b6-336d0f5b3ffa {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.454799] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f55e7c77-744d-4886-95de-e4604864ed00 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.486759] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-094b4e02-a75f-49fa-ba8e-0bf8adbbeb29 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] [instance: 2a29f37d-e786-48d3-b126-4854d6e7c67c] Powering off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 921.487122] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-665ec833-af8b-4bd7-b653-5509941944af {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.494128] env[63515]: DEBUG oslo_vmware.api [None req-094b4e02-a75f-49fa-ba8e-0bf8adbbeb29 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Waiting for the task: (returnval){ [ 921.494128] env[63515]: value = "task-1111576" [ 921.494128] env[63515]: _type = "Task" [ 921.494128] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 921.504376] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-094b4e02-a75f-49fa-ba8e-0bf8adbbeb29 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] [instance: 2a29f37d-e786-48d3-b126-4854d6e7c67c] VM already powered off {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 921.504620] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-094b4e02-a75f-49fa-ba8e-0bf8adbbeb29 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] [instance: 2a29f37d-e786-48d3-b126-4854d6e7c67c] Processing image 8a120570-cb06-4099-b262-554ca0ad15c5 {{(pid=63515) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 921.504899] env[63515]: DEBUG oslo_concurrency.lockutils [None req-094b4e02-a75f-49fa-ba8e-0bf8adbbeb29 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 921.505090] env[63515]: DEBUG oslo_concurrency.lockutils [None req-094b4e02-a75f-49fa-ba8e-0bf8adbbeb29 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 921.505303] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-094b4e02-a75f-49fa-ba8e-0bf8adbbeb29 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 921.505577] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-39d30297-3561-4619-968a-1809023764b3 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.514153] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-094b4e02-a75f-49fa-ba8e-0bf8adbbeb29 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 921.514391] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-094b4e02-a75f-49fa-ba8e-0bf8adbbeb29 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63515) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 921.515168] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8407c390-12b4-4d06-b943-1791777fc5b7 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.520517] env[63515]: DEBUG oslo_vmware.api [None req-094b4e02-a75f-49fa-ba8e-0bf8adbbeb29 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Waiting for the task: (returnval){ [ 921.520517] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]520338c0-6823-dcaf-626d-21d0f1ef6d12" [ 921.520517] env[63515]: _type = "Task" [ 921.520517] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 921.528534] env[63515]: DEBUG oslo_vmware.api [None req-094b4e02-a75f-49fa-ba8e-0bf8adbbeb29 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]520338c0-6823-dcaf-626d-21d0f1ef6d12, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.531759] env[63515]: DEBUG nova.network.neutron [None req-0bb813f4-ce4d-4144-b80d-b179b1c5fd20 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: a2203c5e-66b9-4c7e-8d76-e12a08b57b1d] Updating instance_info_cache with network_info: [{"id": "8b878aac-8955-4468-893f-63fdbfc0c8cb", "address": "fa:16:3e:f1:aa:cb", "network": {"id": "581d448e-be1b-4d5b-8a6e-c27326aaffc5", "bridge": "br-int", "label": "tempest-ServersTestJSON-1769610073-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "299fdeff647f486390366d5bbf911518", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fb143ef7-8271-4a8a-a4aa-8eba9a89f6a1", "external-id": "nsx-vlan-transportzone-504", "segmentation_id": 504, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8b878aac-89", "ovs_interfaceid": "8b878aac-8955-4468-893f-63fdbfc0c8cb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 921.548742] env[63515]: DEBUG oslo_vmware.api [None req-8c6d15be-1f5e-44a9-a3e4-dab9aa82e26a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Task: {'id': task-1111575, 'name': Rename_Task, 'duration_secs': 0.165156} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 921.549035] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-8c6d15be-1f5e-44a9-a3e4-dab9aa82e26a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] [instance: ca6a8ac3-d880-4333-9816-2f076e423844] Powering on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 921.549286] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d50d7028-6967-41e4-8154-65c1ecb26c83 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.557028] env[63515]: DEBUG oslo_vmware.api [None req-8c6d15be-1f5e-44a9-a3e4-dab9aa82e26a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Waiting for the task: (returnval){ [ 921.557028] env[63515]: value = "task-1111577" [ 921.557028] env[63515]: _type = "Task" [ 921.557028] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 921.563665] env[63515]: DEBUG oslo_vmware.api [None req-8c6d15be-1f5e-44a9-a3e4-dab9aa82e26a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Task: {'id': task-1111577, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.619525] env[63515]: INFO nova.compute.claims [None req-1d5e76d4-1b01-4bdb-831a-7cba26a22dfc tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: 829e72d7-a56a-451c-b98a-94253e5900d5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 921.644059] env[63515]: DEBUG oslo_concurrency.lockutils [None req-358377cb-a02d-42b1-aec7-e1ec3bad941d tempest-ServersTestMultiNic-1662098323 tempest-ServersTestMultiNic-1662098323-project-member] Lock "324f7e89-3c17-474d-af1b-6d7a6d042510" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 14.132s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 922.031738] env[63515]: DEBUG oslo_vmware.api [None req-094b4e02-a75f-49fa-ba8e-0bf8adbbeb29 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]520338c0-6823-dcaf-626d-21d0f1ef6d12, 'name': SearchDatastore_Task, 'duration_secs': 0.009093} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 922.033146] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1304f1fe-c5bc-4a3c-b651-9d0193d7a7a0 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.035155] env[63515]: DEBUG oslo_concurrency.lockutils [None req-0bb813f4-ce4d-4144-b80d-b179b1c5fd20 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Releasing lock "refresh_cache-a2203c5e-66b9-4c7e-8d76-e12a08b57b1d" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 922.035396] env[63515]: DEBUG nova.compute.manager [None req-0bb813f4-ce4d-4144-b80d-b179b1c5fd20 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: a2203c5e-66b9-4c7e-8d76-e12a08b57b1d] Instance network_info: |[{"id": "8b878aac-8955-4468-893f-63fdbfc0c8cb", "address": "fa:16:3e:f1:aa:cb", "network": {"id": "581d448e-be1b-4d5b-8a6e-c27326aaffc5", "bridge": "br-int", "label": "tempest-ServersTestJSON-1769610073-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "299fdeff647f486390366d5bbf911518", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fb143ef7-8271-4a8a-a4aa-8eba9a89f6a1", "external-id": "nsx-vlan-transportzone-504", "segmentation_id": 504, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8b878aac-89", "ovs_interfaceid": "8b878aac-8955-4468-893f-63fdbfc0c8cb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 922.036398] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-0bb813f4-ce4d-4144-b80d-b179b1c5fd20 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: a2203c5e-66b9-4c7e-8d76-e12a08b57b1d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:f1:aa:cb', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'fb143ef7-8271-4a8a-a4aa-8eba9a89f6a1', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '8b878aac-8955-4468-893f-63fdbfc0c8cb', 'vif_model': 'vmxnet3'}] {{(pid=63515) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 922.043837] env[63515]: DEBUG oslo.service.loopingcall [None req-0bb813f4-ce4d-4144-b80d-b179b1c5fd20 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 922.045288] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a2203c5e-66b9-4c7e-8d76-e12a08b57b1d] Creating VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 922.045461] env[63515]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-30e29ec3-55f7-4a39-88ab-2f8065a5cbf1 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.062022] env[63515]: DEBUG oslo_vmware.api [None req-094b4e02-a75f-49fa-ba8e-0bf8adbbeb29 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Waiting for the task: (returnval){ [ 922.062022] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]52c90399-0fce-8207-d27b-5dd5932c2da3" [ 922.062022] env[63515]: _type = "Task" [ 922.062022] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.070740] env[63515]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 922.070740] env[63515]: value = "task-1111578" [ 922.070740] env[63515]: _type = "Task" [ 922.070740] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.081622] env[63515]: DEBUG oslo_vmware.api [None req-8c6d15be-1f5e-44a9-a3e4-dab9aa82e26a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Task: {'id': task-1111577, 'name': PowerOnVM_Task} progress is 71%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.081877] env[63515]: DEBUG oslo_vmware.api [None req-094b4e02-a75f-49fa-ba8e-0bf8adbbeb29 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52c90399-0fce-8207-d27b-5dd5932c2da3, 'name': SearchDatastore_Task, 'duration_secs': 0.024163} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 922.082552] env[63515]: DEBUG oslo_concurrency.lockutils [None req-094b4e02-a75f-49fa-ba8e-0bf8adbbeb29 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 922.082832] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-094b4e02-a75f-49fa-ba8e-0bf8adbbeb29 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Copying virtual disk from [datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk to [datastore1] 2a29f37d-e786-48d3-b126-4854d6e7c67c/8a120570-cb06-4099-b262-554ca0ad15c5-rescue.vmdk. {{(pid=63515) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 922.083105] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3051c5a8-7a9e-4f09-9be2-50f68fab0fa3 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.088912] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1111578, 'name': CreateVM_Task} progress is 6%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.093155] env[63515]: DEBUG oslo_vmware.api [None req-094b4e02-a75f-49fa-ba8e-0bf8adbbeb29 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Waiting for the task: (returnval){ [ 922.093155] env[63515]: value = "task-1111579" [ 922.093155] env[63515]: _type = "Task" [ 922.093155] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.103313] env[63515]: DEBUG oslo_vmware.api [None req-094b4e02-a75f-49fa-ba8e-0bf8adbbeb29 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Task: {'id': task-1111579, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.126332] env[63515]: INFO nova.compute.resource_tracker [None req-1d5e76d4-1b01-4bdb-831a-7cba26a22dfc tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: 829e72d7-a56a-451c-b98a-94253e5900d5] Updating resource usage from migration c4203df0-81f5-4cda-85fc-996941b25046 [ 922.201172] env[63515]: DEBUG oslo_concurrency.lockutils [None req-bd577155-aa2b-4775-93e5-5bcfdcad6cd6 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Lock "ae5d6f34-bac6-4ba0-b0c3-8205982ddc88" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 2.447s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 922.574203] env[63515]: DEBUG oslo_vmware.api [None req-8c6d15be-1f5e-44a9-a3e4-dab9aa82e26a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Task: {'id': task-1111577, 'name': PowerOnVM_Task} progress is 71%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.588039] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1111578, 'name': CreateVM_Task} progress is 99%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.605432] env[63515]: DEBUG oslo_vmware.api [None req-094b4e02-a75f-49fa-ba8e-0bf8adbbeb29 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Task: {'id': task-1111579, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.656527] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f39900e5-01aa-4adb-b2ed-021909ce2852 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.664810] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be1d1ae6-c4ec-40e1-8cc5-085b5ce26606 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.701009] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53b4f1dd-55da-4250-a98d-4e5d976650a1 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.711714] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83dafbce-1814-4e1b-a16b-0b7f266ae953 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.727253] env[63515]: DEBUG nova.compute.provider_tree [None req-1d5e76d4-1b01-4bdb-831a-7cba26a22dfc tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 922.953793] env[63515]: DEBUG nova.compute.manager [req-d124978c-14d6-4db7-a0fc-5da5909ac424 req-edcee079-0ba9-4e75-9119-34ed588a1d85 service nova] [instance: eab16df9-7bb5-4576-bca0-769a561c5fe9] Received event network-vif-plugged-e3c72582-41bf-48d4-8614-b516a96362a7 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 922.954117] env[63515]: DEBUG oslo_concurrency.lockutils [req-d124978c-14d6-4db7-a0fc-5da5909ac424 req-edcee079-0ba9-4e75-9119-34ed588a1d85 service nova] Acquiring lock "eab16df9-7bb5-4576-bca0-769a561c5fe9-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 922.954443] env[63515]: DEBUG oslo_concurrency.lockutils [req-d124978c-14d6-4db7-a0fc-5da5909ac424 req-edcee079-0ba9-4e75-9119-34ed588a1d85 service nova] Lock "eab16df9-7bb5-4576-bca0-769a561c5fe9-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 922.954571] env[63515]: DEBUG oslo_concurrency.lockutils [req-d124978c-14d6-4db7-a0fc-5da5909ac424 req-edcee079-0ba9-4e75-9119-34ed588a1d85 service nova] Lock "eab16df9-7bb5-4576-bca0-769a561c5fe9-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 922.954744] env[63515]: DEBUG nova.compute.manager [req-d124978c-14d6-4db7-a0fc-5da5909ac424 req-edcee079-0ba9-4e75-9119-34ed588a1d85 service nova] [instance: eab16df9-7bb5-4576-bca0-769a561c5fe9] No waiting events found dispatching network-vif-plugged-e3c72582-41bf-48d4-8614-b516a96362a7 {{(pid=63515) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 922.954985] env[63515]: WARNING nova.compute.manager [req-d124978c-14d6-4db7-a0fc-5da5909ac424 req-edcee079-0ba9-4e75-9119-34ed588a1d85 service nova] [instance: eab16df9-7bb5-4576-bca0-769a561c5fe9] Received unexpected event network-vif-plugged-e3c72582-41bf-48d4-8614-b516a96362a7 for instance with vm_state active and task_state None. [ 923.073417] env[63515]: DEBUG oslo_vmware.api [None req-8c6d15be-1f5e-44a9-a3e4-dab9aa82e26a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Task: {'id': task-1111577, 'name': PowerOnVM_Task, 'duration_secs': 1.387014} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 923.073417] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-8c6d15be-1f5e-44a9-a3e4-dab9aa82e26a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] [instance: ca6a8ac3-d880-4333-9816-2f076e423844] Powered on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 923.073417] env[63515]: INFO nova.compute.manager [None req-8c6d15be-1f5e-44a9-a3e4-dab9aa82e26a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] [instance: ca6a8ac3-d880-4333-9816-2f076e423844] Took 9.43 seconds to spawn the instance on the hypervisor. [ 923.073760] env[63515]: DEBUG nova.compute.manager [None req-8c6d15be-1f5e-44a9-a3e4-dab9aa82e26a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] [instance: ca6a8ac3-d880-4333-9816-2f076e423844] Checking state {{(pid=63515) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 923.074414] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72cb6d72-3a56-451b-b4d8-8d44fe770fb7 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.088717] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1111578, 'name': CreateVM_Task, 'duration_secs': 0.536837} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 923.089274] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a2203c5e-66b9-4c7e-8d76-e12a08b57b1d] Created VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 923.089982] env[63515]: DEBUG oslo_concurrency.lockutils [None req-0bb813f4-ce4d-4144-b80d-b179b1c5fd20 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 923.090132] env[63515]: DEBUG oslo_concurrency.lockutils [None req-0bb813f4-ce4d-4144-b80d-b179b1c5fd20 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 923.090461] env[63515]: DEBUG oslo_concurrency.lockutils [None req-0bb813f4-ce4d-4144-b80d-b179b1c5fd20 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 923.090698] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cab5a9da-47dd-4911-b848-a53fdab05853 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.095286] env[63515]: DEBUG oslo_vmware.api [None req-0bb813f4-ce4d-4144-b80d-b179b1c5fd20 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Waiting for the task: (returnval){ [ 923.095286] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]529645b5-9803-b69a-aeed-d4268095899e" [ 923.095286] env[63515]: _type = "Task" [ 923.095286] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 923.105816] env[63515]: DEBUG oslo_vmware.api [None req-094b4e02-a75f-49fa-ba8e-0bf8adbbeb29 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Task: {'id': task-1111579, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.644911} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 923.108757] env[63515]: INFO nova.virt.vmwareapi.ds_util [None req-094b4e02-a75f-49fa-ba8e-0bf8adbbeb29 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Copied virtual disk from [datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk to [datastore1] 2a29f37d-e786-48d3-b126-4854d6e7c67c/8a120570-cb06-4099-b262-554ca0ad15c5-rescue.vmdk. [ 923.109057] env[63515]: DEBUG oslo_vmware.api [None req-0bb813f4-ce4d-4144-b80d-b179b1c5fd20 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]529645b5-9803-b69a-aeed-d4268095899e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.109782] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e72965d4-cf84-43eb-a94b-fd6733d979fe {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.138182] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-094b4e02-a75f-49fa-ba8e-0bf8adbbeb29 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] [instance: 2a29f37d-e786-48d3-b126-4854d6e7c67c] Reconfiguring VM instance instance-00000045 to attach disk [datastore1] 2a29f37d-e786-48d3-b126-4854d6e7c67c/8a120570-cb06-4099-b262-554ca0ad15c5-rescue.vmdk or device None with type thin {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 923.139223] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c6d69b1e-8c30-47aa-9884-df74dffac05e {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.157055] env[63515]: DEBUG oslo_vmware.api [None req-094b4e02-a75f-49fa-ba8e-0bf8adbbeb29 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Waiting for the task: (returnval){ [ 923.157055] env[63515]: value = "task-1111580" [ 923.157055] env[63515]: _type = "Task" [ 923.157055] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 923.165419] env[63515]: DEBUG oslo_vmware.api [None req-094b4e02-a75f-49fa-ba8e-0bf8adbbeb29 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Task: {'id': task-1111580, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.186687] env[63515]: DEBUG nova.network.neutron [None req-525a7c59-1fc2-4bbe-98db-35f4ad5c239b tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: eab16df9-7bb5-4576-bca0-769a561c5fe9] Successfully updated port: e3c72582-41bf-48d4-8614-b516a96362a7 {{(pid=63515) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 923.230521] env[63515]: DEBUG nova.scheduler.client.report [None req-1d5e76d4-1b01-4bdb-831a-7cba26a22dfc tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 923.424605] env[63515]: DEBUG oslo_concurrency.lockutils [None req-58ca2ad0-b12b-44be-a72b-6a6ba3bc88d1 tempest-ServerActionsV293TestJSON-1193370169 tempest-ServerActionsV293TestJSON-1193370169-project-member] Acquiring lock "2ea99c7a-8b61-4718-bee0-f4ce4fc021af" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 923.424888] env[63515]: DEBUG oslo_concurrency.lockutils [None req-58ca2ad0-b12b-44be-a72b-6a6ba3bc88d1 tempest-ServerActionsV293TestJSON-1193370169 tempest-ServerActionsV293TestJSON-1193370169-project-member] Lock "2ea99c7a-8b61-4718-bee0-f4ce4fc021af" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 923.425121] env[63515]: DEBUG oslo_concurrency.lockutils [None req-58ca2ad0-b12b-44be-a72b-6a6ba3bc88d1 tempest-ServerActionsV293TestJSON-1193370169 tempest-ServerActionsV293TestJSON-1193370169-project-member] Acquiring lock "2ea99c7a-8b61-4718-bee0-f4ce4fc021af-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 923.425323] env[63515]: DEBUG oslo_concurrency.lockutils [None req-58ca2ad0-b12b-44be-a72b-6a6ba3bc88d1 tempest-ServerActionsV293TestJSON-1193370169 tempest-ServerActionsV293TestJSON-1193370169-project-member] Lock "2ea99c7a-8b61-4718-bee0-f4ce4fc021af-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 923.425548] env[63515]: DEBUG oslo_concurrency.lockutils [None req-58ca2ad0-b12b-44be-a72b-6a6ba3bc88d1 tempest-ServerActionsV293TestJSON-1193370169 tempest-ServerActionsV293TestJSON-1193370169-project-member] Lock "2ea99c7a-8b61-4718-bee0-f4ce4fc021af-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 923.427916] env[63515]: INFO nova.compute.manager [None req-58ca2ad0-b12b-44be-a72b-6a6ba3bc88d1 tempest-ServerActionsV293TestJSON-1193370169 tempest-ServerActionsV293TestJSON-1193370169-project-member] [instance: 2ea99c7a-8b61-4718-bee0-f4ce4fc021af] Terminating instance [ 923.430419] env[63515]: DEBUG nova.compute.manager [None req-58ca2ad0-b12b-44be-a72b-6a6ba3bc88d1 tempest-ServerActionsV293TestJSON-1193370169 tempest-ServerActionsV293TestJSON-1193370169-project-member] [instance: 2ea99c7a-8b61-4718-bee0-f4ce4fc021af] Start destroying the instance on the hypervisor. {{(pid=63515) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 923.430802] env[63515]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a4bb4680-cadd-42e8-acb9-e39bcd5a8dd3 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.439803] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2885a87b-3258-4db5-81b9-6b09aa7eb45d {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.472419] env[63515]: WARNING nova.virt.vmwareapi.driver [None req-58ca2ad0-b12b-44be-a72b-6a6ba3bc88d1 tempest-ServerActionsV293TestJSON-1193370169 tempest-ServerActionsV293TestJSON-1193370169-project-member] [instance: 2ea99c7a-8b61-4718-bee0-f4ce4fc021af] Instance does not exists. Proceeding to delete instance properties on datastore: nova.exception.InstanceNotFound: Instance 2ea99c7a-8b61-4718-bee0-f4ce4fc021af could not be found. [ 923.472654] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-58ca2ad0-b12b-44be-a72b-6a6ba3bc88d1 tempest-ServerActionsV293TestJSON-1193370169 tempest-ServerActionsV293TestJSON-1193370169-project-member] [instance: 2ea99c7a-8b61-4718-bee0-f4ce4fc021af] Destroying instance {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 923.474118] env[63515]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-41a81f50-274e-44ec-9861-e0adeb04d3fe {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.488333] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df84d87c-8566-46b1-93ed-5d74cb7b59d9 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.519895] env[63515]: WARNING nova.virt.vmwareapi.vmops [None req-58ca2ad0-b12b-44be-a72b-6a6ba3bc88d1 tempest-ServerActionsV293TestJSON-1193370169 tempest-ServerActionsV293TestJSON-1193370169-project-member] [instance: 2ea99c7a-8b61-4718-bee0-f4ce4fc021af] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 2ea99c7a-8b61-4718-bee0-f4ce4fc021af could not be found. [ 923.520129] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-58ca2ad0-b12b-44be-a72b-6a6ba3bc88d1 tempest-ServerActionsV293TestJSON-1193370169 tempest-ServerActionsV293TestJSON-1193370169-project-member] [instance: 2ea99c7a-8b61-4718-bee0-f4ce4fc021af] Instance destroyed {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 923.520306] env[63515]: INFO nova.compute.manager [None req-58ca2ad0-b12b-44be-a72b-6a6ba3bc88d1 tempest-ServerActionsV293TestJSON-1193370169 tempest-ServerActionsV293TestJSON-1193370169-project-member] [instance: 2ea99c7a-8b61-4718-bee0-f4ce4fc021af] Took 0.09 seconds to destroy the instance on the hypervisor. [ 923.520566] env[63515]: DEBUG oslo.service.loopingcall [None req-58ca2ad0-b12b-44be-a72b-6a6ba3bc88d1 tempest-ServerActionsV293TestJSON-1193370169 tempest-ServerActionsV293TestJSON-1193370169-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 923.520822] env[63515]: DEBUG nova.compute.manager [-] [instance: 2ea99c7a-8b61-4718-bee0-f4ce4fc021af] Deallocating network for instance {{(pid=63515) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 923.520887] env[63515]: DEBUG nova.network.neutron [-] [instance: 2ea99c7a-8b61-4718-bee0-f4ce4fc021af] deallocate_for_instance() {{(pid=63515) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 923.602801] env[63515]: INFO nova.compute.manager [None req-8c6d15be-1f5e-44a9-a3e4-dab9aa82e26a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] [instance: ca6a8ac3-d880-4333-9816-2f076e423844] Took 21.99 seconds to build instance. [ 923.609700] env[63515]: DEBUG oslo_vmware.api [None req-0bb813f4-ce4d-4144-b80d-b179b1c5fd20 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]529645b5-9803-b69a-aeed-d4268095899e, 'name': SearchDatastore_Task, 'duration_secs': 0.0203} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 923.609997] env[63515]: DEBUG oslo_concurrency.lockutils [None req-0bb813f4-ce4d-4144-b80d-b179b1c5fd20 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 923.610244] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-0bb813f4-ce4d-4144-b80d-b179b1c5fd20 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: a2203c5e-66b9-4c7e-8d76-e12a08b57b1d] Processing image 8a120570-cb06-4099-b262-554ca0ad15c5 {{(pid=63515) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 923.610484] env[63515]: DEBUG oslo_concurrency.lockutils [None req-0bb813f4-ce4d-4144-b80d-b179b1c5fd20 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 923.610632] env[63515]: DEBUG oslo_concurrency.lockutils [None req-0bb813f4-ce4d-4144-b80d-b179b1c5fd20 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 923.610832] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-0bb813f4-ce4d-4144-b80d-b179b1c5fd20 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 923.611108] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-825fa827-05b9-4681-a1e3-d4137c848508 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.620854] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-0bb813f4-ce4d-4144-b80d-b179b1c5fd20 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 923.622412] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-0bb813f4-ce4d-4144-b80d-b179b1c5fd20 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63515) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 923.622412] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-94870c7f-26b8-43fd-a72f-342732f3100f {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.627948] env[63515]: DEBUG oslo_vmware.api [None req-0bb813f4-ce4d-4144-b80d-b179b1c5fd20 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Waiting for the task: (returnval){ [ 923.627948] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]5246cdb4-e4e3-8162-7c72-19f29e63aa24" [ 923.627948] env[63515]: _type = "Task" [ 923.627948] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 923.635788] env[63515]: DEBUG oslo_vmware.api [None req-0bb813f4-ce4d-4144-b80d-b179b1c5fd20 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]5246cdb4-e4e3-8162-7c72-19f29e63aa24, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.666829] env[63515]: DEBUG oslo_vmware.api [None req-094b4e02-a75f-49fa-ba8e-0bf8adbbeb29 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Task: {'id': task-1111580, 'name': ReconfigVM_Task, 'duration_secs': 0.332582} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 923.667315] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-094b4e02-a75f-49fa-ba8e-0bf8adbbeb29 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] [instance: 2a29f37d-e786-48d3-b126-4854d6e7c67c] Reconfigured VM instance instance-00000045 to attach disk [datastore1] 2a29f37d-e786-48d3-b126-4854d6e7c67c/8a120570-cb06-4099-b262-554ca0ad15c5-rescue.vmdk or device None with type thin {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 923.668249] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33d465f6-c96d-4114-8563-b041162a79ae {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.693266] env[63515]: DEBUG oslo_concurrency.lockutils [None req-525a7c59-1fc2-4bbe-98db-35f4ad5c239b tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Acquiring lock "refresh_cache-eab16df9-7bb5-4576-bca0-769a561c5fe9" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 923.693477] env[63515]: DEBUG oslo_concurrency.lockutils [None req-525a7c59-1fc2-4bbe-98db-35f4ad5c239b tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Acquired lock "refresh_cache-eab16df9-7bb5-4576-bca0-769a561c5fe9" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 923.693640] env[63515]: DEBUG nova.network.neutron [None req-525a7c59-1fc2-4bbe-98db-35f4ad5c239b tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: eab16df9-7bb5-4576-bca0-769a561c5fe9] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 923.701415] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6e21be48-4503-4e7e-a5b5-c997ed37090a {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.718067] env[63515]: DEBUG oslo_vmware.api [None req-094b4e02-a75f-49fa-ba8e-0bf8adbbeb29 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Waiting for the task: (returnval){ [ 923.718067] env[63515]: value = "task-1111581" [ 923.718067] env[63515]: _type = "Task" [ 923.718067] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 923.726733] env[63515]: DEBUG oslo_vmware.api [None req-094b4e02-a75f-49fa-ba8e-0bf8adbbeb29 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Task: {'id': task-1111581, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.735961] env[63515]: DEBUG oslo_concurrency.lockutils [None req-1d5e76d4-1b01-4bdb-831a-7cba26a22dfc tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.622s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 923.736237] env[63515]: INFO nova.compute.manager [None req-1d5e76d4-1b01-4bdb-831a-7cba26a22dfc tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: 829e72d7-a56a-451c-b98a-94253e5900d5] Migrating [ 923.746346] env[63515]: DEBUG oslo_concurrency.lockutils [None req-d63820c0-6e1a-45a6-8579-7bfab43bd25a tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 7.491s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 923.747878] env[63515]: INFO nova.compute.claims [None req-d63820c0-6e1a-45a6-8579-7bfab43bd25a tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: 2fa098b6-ecc0-491f-8cfc-5ee335bbf943] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 923.979911] env[63515]: DEBUG nova.compute.manager [None req-b95e91a2-83f6-4cd7-9448-65181d09829a tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 68766cd4-84be-475b-8494-d7ab43a9e969] Checking state {{(pid=63515) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 923.980898] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58dbaec1-93c8-419c-906b-63a445267eaf {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.108714] env[63515]: DEBUG oslo_concurrency.lockutils [None req-8c6d15be-1f5e-44a9-a3e4-dab9aa82e26a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Lock "ca6a8ac3-d880-4333-9816-2f076e423844" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 23.507s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 924.143359] env[63515]: DEBUG oslo_vmware.api [None req-0bb813f4-ce4d-4144-b80d-b179b1c5fd20 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]5246cdb4-e4e3-8162-7c72-19f29e63aa24, 'name': SearchDatastore_Task, 'duration_secs': 0.03946} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 924.144052] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-12325bee-bc6c-43dd-bdf0-c4d9be53b7fc {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.153235] env[63515]: DEBUG oslo_vmware.api [None req-0bb813f4-ce4d-4144-b80d-b179b1c5fd20 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Waiting for the task: (returnval){ [ 924.153235] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]524149d1-3ccb-0279-5e38-db1361456804" [ 924.153235] env[63515]: _type = "Task" [ 924.153235] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 924.158892] env[63515]: DEBUG nova.compute.manager [req-39d93c87-c26a-494e-8c52-5a129e1bb256 req-c736a420-f791-4a16-91db-06eb2097c5fa service nova] [instance: eab16df9-7bb5-4576-bca0-769a561c5fe9] Received event network-changed-e3c72582-41bf-48d4-8614-b516a96362a7 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 924.159149] env[63515]: DEBUG nova.compute.manager [req-39d93c87-c26a-494e-8c52-5a129e1bb256 req-c736a420-f791-4a16-91db-06eb2097c5fa service nova] [instance: eab16df9-7bb5-4576-bca0-769a561c5fe9] Refreshing instance network info cache due to event network-changed-e3c72582-41bf-48d4-8614-b516a96362a7. {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 924.159370] env[63515]: DEBUG oslo_concurrency.lockutils [req-39d93c87-c26a-494e-8c52-5a129e1bb256 req-c736a420-f791-4a16-91db-06eb2097c5fa service nova] Acquiring lock "refresh_cache-eab16df9-7bb5-4576-bca0-769a561c5fe9" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 924.165247] env[63515]: DEBUG oslo_vmware.api [None req-0bb813f4-ce4d-4144-b80d-b179b1c5fd20 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]524149d1-3ccb-0279-5e38-db1361456804, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.229750] env[63515]: DEBUG oslo_vmware.api [None req-094b4e02-a75f-49fa-ba8e-0bf8adbbeb29 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Task: {'id': task-1111581, 'name': ReconfigVM_Task, 'duration_secs': 0.186182} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 924.230289] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-094b4e02-a75f-49fa-ba8e-0bf8adbbeb29 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] [instance: 2a29f37d-e786-48d3-b126-4854d6e7c67c] Powering on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 924.230707] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9848fcfd-a22c-4324-8716-fb7bc6680f27 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.236785] env[63515]: DEBUG oslo_vmware.api [None req-094b4e02-a75f-49fa-ba8e-0bf8adbbeb29 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Waiting for the task: (returnval){ [ 924.236785] env[63515]: value = "task-1111582" [ 924.236785] env[63515]: _type = "Task" [ 924.236785] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 924.242047] env[63515]: WARNING nova.network.neutron [None req-525a7c59-1fc2-4bbe-98db-35f4ad5c239b tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: eab16df9-7bb5-4576-bca0-769a561c5fe9] 88dfb29d-41e0-4df7-8f70-e52cda05620c already exists in list: networks containing: ['88dfb29d-41e0-4df7-8f70-e52cda05620c']. ignoring it [ 924.247660] env[63515]: DEBUG oslo_vmware.api [None req-094b4e02-a75f-49fa-ba8e-0bf8adbbeb29 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Task: {'id': task-1111582, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.259478] env[63515]: DEBUG oslo_concurrency.lockutils [None req-1d5e76d4-1b01-4bdb-831a-7cba26a22dfc tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Acquiring lock "refresh_cache-829e72d7-a56a-451c-b98a-94253e5900d5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 924.259920] env[63515]: DEBUG oslo_concurrency.lockutils [None req-1d5e76d4-1b01-4bdb-831a-7cba26a22dfc tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Acquired lock "refresh_cache-829e72d7-a56a-451c-b98a-94253e5900d5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 924.260267] env[63515]: DEBUG nova.network.neutron [None req-1d5e76d4-1b01-4bdb-831a-7cba26a22dfc tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: 829e72d7-a56a-451c-b98a-94253e5900d5] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 924.491171] env[63515]: INFO nova.compute.manager [None req-b95e91a2-83f6-4cd7-9448-65181d09829a tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 68766cd4-84be-475b-8494-d7ab43a9e969] instance snapshotting [ 924.491791] env[63515]: DEBUG nova.objects.instance [None req-b95e91a2-83f6-4cd7-9448-65181d09829a tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Lazy-loading 'flavor' on Instance uuid 68766cd4-84be-475b-8494-d7ab43a9e969 {{(pid=63515) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 924.653024] env[63515]: DEBUG nova.network.neutron [-] [instance: 2ea99c7a-8b61-4718-bee0-f4ce4fc021af] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 924.659962] env[63515]: DEBUG nova.network.neutron [None req-525a7c59-1fc2-4bbe-98db-35f4ad5c239b tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: eab16df9-7bb5-4576-bca0-769a561c5fe9] Updating instance_info_cache with network_info: [{"id": "4168f5a5-c282-4f13-9c97-0b0236bd659b", "address": "fa:16:3e:a0:6a:50", "network": {"id": "88dfb29d-41e0-4df7-8f70-e52cda05620c", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1355578400-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.203", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7d4ada9ec35f42b19c6480a9101d21a4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "357d2811-e990-4985-9f9e-b158d10d3699", "external-id": "nsx-vlan-transportzone-641", "segmentation_id": 641, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4168f5a5-c2", "ovs_interfaceid": "4168f5a5-c282-4f13-9c97-0b0236bd659b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "e3c72582-41bf-48d4-8614-b516a96362a7", "address": "fa:16:3e:c0:fd:b9", "network": {"id": "88dfb29d-41e0-4df7-8f70-e52cda05620c", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1355578400-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7d4ada9ec35f42b19c6480a9101d21a4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "357d2811-e990-4985-9f9e-b158d10d3699", "external-id": "nsx-vlan-transportzone-641", "segmentation_id": 641, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape3c72582-41", "ovs_interfaceid": "e3c72582-41bf-48d4-8614-b516a96362a7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 924.665427] env[63515]: DEBUG oslo_vmware.api [None req-0bb813f4-ce4d-4144-b80d-b179b1c5fd20 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]524149d1-3ccb-0279-5e38-db1361456804, 'name': SearchDatastore_Task, 'duration_secs': 0.05901} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 924.666148] env[63515]: DEBUG oslo_concurrency.lockutils [None req-0bb813f4-ce4d-4144-b80d-b179b1c5fd20 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 924.666558] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-0bb813f4-ce4d-4144-b80d-b179b1c5fd20 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk to [datastore2] a2203c5e-66b9-4c7e-8d76-e12a08b57b1d/a2203c5e-66b9-4c7e-8d76-e12a08b57b1d.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 924.666994] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f641bc16-a75b-40eb-ac7a-35a722002178 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.673921] env[63515]: DEBUG oslo_vmware.api [None req-0bb813f4-ce4d-4144-b80d-b179b1c5fd20 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Waiting for the task: (returnval){ [ 924.673921] env[63515]: value = "task-1111583" [ 924.673921] env[63515]: _type = "Task" [ 924.673921] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 924.685432] env[63515]: DEBUG oslo_vmware.api [None req-0bb813f4-ce4d-4144-b80d-b179b1c5fd20 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Task: {'id': task-1111583, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.723648] env[63515]: DEBUG oslo_concurrency.lockutils [None req-be7cccf6-3f5b-4895-b678-cf60de2c827a tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Acquiring lock "014720fe-87b9-4e55-b9ba-295b009c7ff6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 924.723888] env[63515]: DEBUG oslo_concurrency.lockutils [None req-be7cccf6-3f5b-4895-b678-cf60de2c827a tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Lock "014720fe-87b9-4e55-b9ba-295b009c7ff6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 924.747037] env[63515]: DEBUG oslo_vmware.api [None req-094b4e02-a75f-49fa-ba8e-0bf8adbbeb29 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Task: {'id': task-1111582, 'name': PowerOnVM_Task, 'duration_secs': 0.423901} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 924.747349] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-094b4e02-a75f-49fa-ba8e-0bf8adbbeb29 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] [instance: 2a29f37d-e786-48d3-b126-4854d6e7c67c] Powered on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 924.752357] env[63515]: DEBUG nova.compute.manager [None req-094b4e02-a75f-49fa-ba8e-0bf8adbbeb29 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] [instance: 2a29f37d-e786-48d3-b126-4854d6e7c67c] Checking state {{(pid=63515) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 924.753194] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60dd3aea-76e3-44f9-8407-b064ad199a74 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.997230] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50306e31-cd23-4da5-9c75-2a91027e8518 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.020212] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c183ee47-a0fa-4da8-bf27-6a041e03e283 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.023597] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65e5646b-9bd6-4b90-adad-280b0004858e {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.036323] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28d5513c-3b9c-4f41-ab0b-8072c624c1a4 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.069962] env[63515]: DEBUG nova.network.neutron [None req-1d5e76d4-1b01-4bdb-831a-7cba26a22dfc tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: 829e72d7-a56a-451c-b98a-94253e5900d5] Updating instance_info_cache with network_info: [{"id": "856de677-b1c8-479b-91d9-34d31a703a63", "address": "fa:16:3e:13:27:b4", "network": {"id": "15fdaabe-6793-400f-9737-2cdff8f07238", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1175204866-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f9c98e9f6020475490aaa7e76d907ab7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7d377d75-3add-4a15-8691-74b2eb010924", "external-id": "nsx-vlan-transportzone-71", "segmentation_id": 71, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap856de677-b1", "ovs_interfaceid": "856de677-b1c8-479b-91d9-34d31a703a63", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 925.071807] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41cd6f8e-b3f3-430f-9d72-c35bbde0ad68 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.080683] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3923e6f0-bc9a-45e8-8c7a-42de822d8ea4 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.095807] env[63515]: DEBUG nova.compute.provider_tree [None req-d63820c0-6e1a-45a6-8579-7bfab43bd25a tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 925.159105] env[63515]: INFO nova.compute.manager [-] [instance: 2ea99c7a-8b61-4718-bee0-f4ce4fc021af] Took 1.64 seconds to deallocate network for instance. [ 925.167065] env[63515]: DEBUG oslo_concurrency.lockutils [None req-525a7c59-1fc2-4bbe-98db-35f4ad5c239b tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Releasing lock "refresh_cache-eab16df9-7bb5-4576-bca0-769a561c5fe9" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 925.167886] env[63515]: DEBUG oslo_concurrency.lockutils [None req-525a7c59-1fc2-4bbe-98db-35f4ad5c239b tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Acquiring lock "eab16df9-7bb5-4576-bca0-769a561c5fe9" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 925.168113] env[63515]: DEBUG oslo_concurrency.lockutils [None req-525a7c59-1fc2-4bbe-98db-35f4ad5c239b tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Acquired lock "eab16df9-7bb5-4576-bca0-769a561c5fe9" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 925.168495] env[63515]: DEBUG oslo_concurrency.lockutils [req-39d93c87-c26a-494e-8c52-5a129e1bb256 req-c736a420-f791-4a16-91db-06eb2097c5fa service nova] Acquired lock "refresh_cache-eab16df9-7bb5-4576-bca0-769a561c5fe9" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 925.168759] env[63515]: DEBUG nova.network.neutron [req-39d93c87-c26a-494e-8c52-5a129e1bb256 req-c736a420-f791-4a16-91db-06eb2097c5fa service nova] [instance: eab16df9-7bb5-4576-bca0-769a561c5fe9] Refreshing network info cache for port e3c72582-41bf-48d4-8614-b516a96362a7 {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 925.170851] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7f444be-9bae-4f8b-9323-adba1e15b1c5 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.189621] env[63515]: DEBUG nova.virt.hardware [None req-525a7c59-1fc2-4bbe-98db-35f4ad5c239b tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T02:52:14Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 925.189868] env[63515]: DEBUG nova.virt.hardware [None req-525a7c59-1fc2-4bbe-98db-35f4ad5c239b tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Flavor limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 925.190141] env[63515]: DEBUG nova.virt.hardware [None req-525a7c59-1fc2-4bbe-98db-35f4ad5c239b tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Image limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 925.190871] env[63515]: DEBUG nova.virt.hardware [None req-525a7c59-1fc2-4bbe-98db-35f4ad5c239b tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Flavor pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 925.190871] env[63515]: DEBUG nova.virt.hardware [None req-525a7c59-1fc2-4bbe-98db-35f4ad5c239b tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Image pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 925.190871] env[63515]: DEBUG nova.virt.hardware [None req-525a7c59-1fc2-4bbe-98db-35f4ad5c239b tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 925.190871] env[63515]: DEBUG nova.virt.hardware [None req-525a7c59-1fc2-4bbe-98db-35f4ad5c239b tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 925.191170] env[63515]: DEBUG nova.virt.hardware [None req-525a7c59-1fc2-4bbe-98db-35f4ad5c239b tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 925.191170] env[63515]: DEBUG nova.virt.hardware [None req-525a7c59-1fc2-4bbe-98db-35f4ad5c239b tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Got 1 possible topologies {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 925.191252] env[63515]: DEBUG nova.virt.hardware [None req-525a7c59-1fc2-4bbe-98db-35f4ad5c239b tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 925.191416] env[63515]: DEBUG nova.virt.hardware [None req-525a7c59-1fc2-4bbe-98db-35f4ad5c239b tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 925.197788] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-525a7c59-1fc2-4bbe-98db-35f4ad5c239b tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: eab16df9-7bb5-4576-bca0-769a561c5fe9] Reconfiguring VM to attach interface {{(pid=63515) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1923}} [ 925.198869] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7b47fa7a-d2ac-446a-960b-dd89e80f042b {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.214204] env[63515]: DEBUG oslo_vmware.api [None req-0bb813f4-ce4d-4144-b80d-b179b1c5fd20 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Task: {'id': task-1111583, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.221068] env[63515]: DEBUG oslo_vmware.api [None req-525a7c59-1fc2-4bbe-98db-35f4ad5c239b tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Waiting for the task: (returnval){ [ 925.221068] env[63515]: value = "task-1111584" [ 925.221068] env[63515]: _type = "Task" [ 925.221068] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.226554] env[63515]: DEBUG nova.compute.manager [None req-be7cccf6-3f5b-4895-b678-cf60de2c827a tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: 014720fe-87b9-4e55-b9ba-295b009c7ff6] Starting instance... {{(pid=63515) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 925.233364] env[63515]: DEBUG oslo_vmware.api [None req-525a7c59-1fc2-4bbe-98db-35f4ad5c239b tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Task: {'id': task-1111584, 'name': ReconfigVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.537609] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-b95e91a2-83f6-4cd7-9448-65181d09829a tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 68766cd4-84be-475b-8494-d7ab43a9e969] Creating Snapshot of the VM instance {{(pid=63515) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 925.538524] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-114cfb10-1c87-4a1d-b30e-02b577973798 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.549575] env[63515]: DEBUG oslo_vmware.api [None req-b95e91a2-83f6-4cd7-9448-65181d09829a tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Waiting for the task: (returnval){ [ 925.549575] env[63515]: value = "task-1111585" [ 925.549575] env[63515]: _type = "Task" [ 925.549575] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.560352] env[63515]: DEBUG oslo_vmware.api [None req-b95e91a2-83f6-4cd7-9448-65181d09829a tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Task: {'id': task-1111585, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.579813] env[63515]: DEBUG oslo_concurrency.lockutils [None req-1d5e76d4-1b01-4bdb-831a-7cba26a22dfc tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Releasing lock "refresh_cache-829e72d7-a56a-451c-b98a-94253e5900d5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 925.599367] env[63515]: DEBUG nova.scheduler.client.report [None req-d63820c0-6e1a-45a6-8579-7bfab43bd25a tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 925.691172] env[63515]: DEBUG oslo_vmware.api [None req-0bb813f4-ce4d-4144-b80d-b179b1c5fd20 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Task: {'id': task-1111583, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.883392} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.691668] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-0bb813f4-ce4d-4144-b80d-b179b1c5fd20 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk to [datastore2] a2203c5e-66b9-4c7e-8d76-e12a08b57b1d/a2203c5e-66b9-4c7e-8d76-e12a08b57b1d.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 925.692140] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-0bb813f4-ce4d-4144-b80d-b179b1c5fd20 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: a2203c5e-66b9-4c7e-8d76-e12a08b57b1d] Extending root virtual disk to 1048576 {{(pid=63515) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 925.692535] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-9d36f424-3adb-4fae-b870-37a81926fd03 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.701058] env[63515]: DEBUG oslo_vmware.api [None req-0bb813f4-ce4d-4144-b80d-b179b1c5fd20 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Waiting for the task: (returnval){ [ 925.701058] env[63515]: value = "task-1111586" [ 925.701058] env[63515]: _type = "Task" [ 925.701058] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.710662] env[63515]: DEBUG oslo_vmware.api [None req-0bb813f4-ce4d-4144-b80d-b179b1c5fd20 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Task: {'id': task-1111586, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.725032] env[63515]: DEBUG oslo_concurrency.lockutils [None req-f7e007a5-60f7-44d2-a529-05d489daa7aa tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Acquiring lock "9236607d-47af-4272-93db-79826e918aa1" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 925.725032] env[63515]: DEBUG oslo_concurrency.lockutils [None req-f7e007a5-60f7-44d2-a529-05d489daa7aa tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Lock "9236607d-47af-4272-93db-79826e918aa1" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 925.725032] env[63515]: DEBUG oslo_concurrency.lockutils [None req-f7e007a5-60f7-44d2-a529-05d489daa7aa tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Acquiring lock "9236607d-47af-4272-93db-79826e918aa1-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 925.725032] env[63515]: DEBUG oslo_concurrency.lockutils [None req-f7e007a5-60f7-44d2-a529-05d489daa7aa tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Lock "9236607d-47af-4272-93db-79826e918aa1-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 925.725032] env[63515]: DEBUG oslo_concurrency.lockutils [None req-f7e007a5-60f7-44d2-a529-05d489daa7aa tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Lock "9236607d-47af-4272-93db-79826e918aa1-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 925.726809] env[63515]: INFO nova.compute.manager [None req-58ca2ad0-b12b-44be-a72b-6a6ba3bc88d1 tempest-ServerActionsV293TestJSON-1193370169 tempest-ServerActionsV293TestJSON-1193370169-project-member] [instance: 2ea99c7a-8b61-4718-bee0-f4ce4fc021af] Took 0.57 seconds to detach 1 volumes for instance. [ 925.728885] env[63515]: INFO nova.compute.manager [None req-f7e007a5-60f7-44d2-a529-05d489daa7aa tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] [instance: 9236607d-47af-4272-93db-79826e918aa1] Terminating instance [ 925.735639] env[63515]: DEBUG nova.compute.manager [None req-58ca2ad0-b12b-44be-a72b-6a6ba3bc88d1 tempest-ServerActionsV293TestJSON-1193370169 tempest-ServerActionsV293TestJSON-1193370169-project-member] [instance: 2ea99c7a-8b61-4718-bee0-f4ce4fc021af] Deleting volume: 211b1a87-a01e-4d13-aa27-91391bf9cde1 {{(pid=63515) _cleanup_volumes /opt/stack/nova/nova/compute/manager.py:3247}} [ 925.738935] env[63515]: DEBUG nova.compute.manager [None req-f7e007a5-60f7-44d2-a529-05d489daa7aa tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] [instance: 9236607d-47af-4272-93db-79826e918aa1] Start destroying the instance on the hypervisor. {{(pid=63515) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 925.739310] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-f7e007a5-60f7-44d2-a529-05d489daa7aa tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] [instance: 9236607d-47af-4272-93db-79826e918aa1] Destroying instance {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 925.742485] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28d83203-bbe7-4957-96ac-010c9ceef16b {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.760436] env[63515]: DEBUG oslo_vmware.api [None req-525a7c59-1fc2-4bbe-98db-35f4ad5c239b tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Task: {'id': task-1111584, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.763840] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-f7e007a5-60f7-44d2-a529-05d489daa7aa tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] [instance: 9236607d-47af-4272-93db-79826e918aa1] Powering off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 925.764363] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-dcca7d18-0ca5-4692-81d0-0dafd88ac43d {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.771738] env[63515]: DEBUG oslo_concurrency.lockutils [None req-be7cccf6-3f5b-4895-b678-cf60de2c827a tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 925.773436] env[63515]: DEBUG oslo_vmware.api [None req-f7e007a5-60f7-44d2-a529-05d489daa7aa tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Waiting for the task: (returnval){ [ 925.773436] env[63515]: value = "task-1111587" [ 925.773436] env[63515]: _type = "Task" [ 925.773436] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.784854] env[63515]: DEBUG oslo_vmware.api [None req-f7e007a5-60f7-44d2-a529-05d489daa7aa tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Task: {'id': task-1111587, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.869024] env[63515]: DEBUG oslo_concurrency.lockutils [None req-9e4db499-5494-499b-b448-2d861439cb14 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Acquiring lock "ca6a8ac3-d880-4333-9816-2f076e423844" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 925.869024] env[63515]: DEBUG oslo_concurrency.lockutils [None req-9e4db499-5494-499b-b448-2d861439cb14 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Lock "ca6a8ac3-d880-4333-9816-2f076e423844" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 925.869024] env[63515]: DEBUG oslo_concurrency.lockutils [None req-9e4db499-5494-499b-b448-2d861439cb14 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Acquiring lock "ca6a8ac3-d880-4333-9816-2f076e423844-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 925.869024] env[63515]: DEBUG oslo_concurrency.lockutils [None req-9e4db499-5494-499b-b448-2d861439cb14 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Lock "ca6a8ac3-d880-4333-9816-2f076e423844-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 925.869024] env[63515]: DEBUG oslo_concurrency.lockutils [None req-9e4db499-5494-499b-b448-2d861439cb14 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Lock "ca6a8ac3-d880-4333-9816-2f076e423844-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 925.872761] env[63515]: INFO nova.compute.manager [None req-9e4db499-5494-499b-b448-2d861439cb14 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] [instance: ca6a8ac3-d880-4333-9816-2f076e423844] Terminating instance [ 925.875286] env[63515]: DEBUG nova.compute.manager [None req-9e4db499-5494-499b-b448-2d861439cb14 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] [instance: ca6a8ac3-d880-4333-9816-2f076e423844] Start destroying the instance on the hypervisor. {{(pid=63515) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 925.876069] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-9e4db499-5494-499b-b448-2d861439cb14 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] [instance: ca6a8ac3-d880-4333-9816-2f076e423844] Destroying instance {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 925.877179] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e4923b9-4339-4702-8b27-a4fb5046180f {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.887781] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-9e4db499-5494-499b-b448-2d861439cb14 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] [instance: ca6a8ac3-d880-4333-9816-2f076e423844] Powering off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 925.888462] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8630156e-35f1-4872-aabf-27ea5e59f930 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.899852] env[63515]: DEBUG oslo_vmware.api [None req-9e4db499-5494-499b-b448-2d861439cb14 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Waiting for the task: (returnval){ [ 925.899852] env[63515]: value = "task-1111589" [ 925.899852] env[63515]: _type = "Task" [ 925.899852] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.909811] env[63515]: DEBUG oslo_vmware.api [None req-9e4db499-5494-499b-b448-2d861439cb14 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Task: {'id': task-1111589, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.073341] env[63515]: DEBUG oslo_vmware.api [None req-b95e91a2-83f6-4cd7-9448-65181d09829a tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Task: {'id': task-1111585, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.106300] env[63515]: DEBUG oslo_concurrency.lockutils [None req-d63820c0-6e1a-45a6-8579-7bfab43bd25a tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.360s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 926.106938] env[63515]: DEBUG nova.compute.manager [None req-d63820c0-6e1a-45a6-8579-7bfab43bd25a tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: 2fa098b6-ecc0-491f-8cfc-5ee335bbf943] Start building networks asynchronously for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 926.110193] env[63515]: DEBUG oslo_concurrency.lockutils [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 9.226s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 926.147430] env[63515]: DEBUG nova.network.neutron [req-39d93c87-c26a-494e-8c52-5a129e1bb256 req-c736a420-f791-4a16-91db-06eb2097c5fa service nova] [instance: eab16df9-7bb5-4576-bca0-769a561c5fe9] Updated VIF entry in instance network info cache for port e3c72582-41bf-48d4-8614-b516a96362a7. {{(pid=63515) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 926.147430] env[63515]: DEBUG nova.network.neutron [req-39d93c87-c26a-494e-8c52-5a129e1bb256 req-c736a420-f791-4a16-91db-06eb2097c5fa service nova] [instance: eab16df9-7bb5-4576-bca0-769a561c5fe9] Updating instance_info_cache with network_info: [{"id": "4168f5a5-c282-4f13-9c97-0b0236bd659b", "address": "fa:16:3e:a0:6a:50", "network": {"id": "88dfb29d-41e0-4df7-8f70-e52cda05620c", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1355578400-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.203", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7d4ada9ec35f42b19c6480a9101d21a4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "357d2811-e990-4985-9f9e-b158d10d3699", "external-id": "nsx-vlan-transportzone-641", "segmentation_id": 641, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4168f5a5-c2", "ovs_interfaceid": "4168f5a5-c282-4f13-9c97-0b0236bd659b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "e3c72582-41bf-48d4-8614-b516a96362a7", "address": "fa:16:3e:c0:fd:b9", "network": {"id": "88dfb29d-41e0-4df7-8f70-e52cda05620c", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1355578400-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7d4ada9ec35f42b19c6480a9101d21a4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "357d2811-e990-4985-9f9e-b158d10d3699", "external-id": "nsx-vlan-transportzone-641", "segmentation_id": 641, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape3c72582-41", "ovs_interfaceid": "e3c72582-41bf-48d4-8614-b516a96362a7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 926.214453] env[63515]: DEBUG oslo_vmware.api [None req-0bb813f4-ce4d-4144-b80d-b179b1c5fd20 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Task: {'id': task-1111586, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.070811} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 926.214770] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-0bb813f4-ce4d-4144-b80d-b179b1c5fd20 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: a2203c5e-66b9-4c7e-8d76-e12a08b57b1d] Extended root virtual disk {{(pid=63515) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 926.215644] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca0d7d89-67f3-4f57-b4a7-e2d63cb71068 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.244795] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-0bb813f4-ce4d-4144-b80d-b179b1c5fd20 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: a2203c5e-66b9-4c7e-8d76-e12a08b57b1d] Reconfiguring VM instance instance-00000054 to attach disk [datastore2] a2203c5e-66b9-4c7e-8d76-e12a08b57b1d/a2203c5e-66b9-4c7e-8d76-e12a08b57b1d.vmdk or device None with type sparse {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 926.248968] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bd35165f-a38b-4a4f-964d-2bafffd1d6ac {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.270315] env[63515]: DEBUG oslo_vmware.api [None req-525a7c59-1fc2-4bbe-98db-35f4ad5c239b tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Task: {'id': task-1111584, 'name': ReconfigVM_Task, 'duration_secs': 0.819117} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 926.271965] env[63515]: DEBUG oslo_concurrency.lockutils [None req-525a7c59-1fc2-4bbe-98db-35f4ad5c239b tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Releasing lock "eab16df9-7bb5-4576-bca0-769a561c5fe9" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 926.272219] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-525a7c59-1fc2-4bbe-98db-35f4ad5c239b tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: eab16df9-7bb5-4576-bca0-769a561c5fe9] Reconfigured VM to attach interface {{(pid=63515) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1937}} [ 926.275258] env[63515]: DEBUG oslo_vmware.api [None req-0bb813f4-ce4d-4144-b80d-b179b1c5fd20 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Waiting for the task: (returnval){ [ 926.275258] env[63515]: value = "task-1111590" [ 926.275258] env[63515]: _type = "Task" [ 926.275258] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 926.287745] env[63515]: DEBUG oslo_vmware.api [None req-f7e007a5-60f7-44d2-a529-05d489daa7aa tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Task: {'id': task-1111587, 'name': PowerOffVM_Task, 'duration_secs': 0.188848} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 926.290937] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-f7e007a5-60f7-44d2-a529-05d489daa7aa tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] [instance: 9236607d-47af-4272-93db-79826e918aa1] Powered off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 926.291150] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-f7e007a5-60f7-44d2-a529-05d489daa7aa tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] [instance: 9236607d-47af-4272-93db-79826e918aa1] Unregistering the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 926.291639] env[63515]: DEBUG oslo_vmware.api [None req-0bb813f4-ce4d-4144-b80d-b179b1c5fd20 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Task: {'id': task-1111590, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.292969] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d208ee65-b7b0-46cd-93f3-c693e47f7551 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.296068] env[63515]: DEBUG oslo_concurrency.lockutils [None req-58ca2ad0-b12b-44be-a72b-6a6ba3bc88d1 tempest-ServerActionsV293TestJSON-1193370169 tempest-ServerActionsV293TestJSON-1193370169-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 926.361203] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-f7e007a5-60f7-44d2-a529-05d489daa7aa tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] [instance: 9236607d-47af-4272-93db-79826e918aa1] Unregistered the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 926.361496] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-f7e007a5-60f7-44d2-a529-05d489daa7aa tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] [instance: 9236607d-47af-4272-93db-79826e918aa1] Deleting contents of the VM from datastore datastore2 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 926.361910] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-f7e007a5-60f7-44d2-a529-05d489daa7aa tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Deleting the datastore file [datastore2] 9236607d-47af-4272-93db-79826e918aa1 {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 926.361910] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ce9d35c0-cefc-4b67-8807-6b491edb3e20 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.368505] env[63515]: DEBUG oslo_vmware.api [None req-f7e007a5-60f7-44d2-a529-05d489daa7aa tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Waiting for the task: (returnval){ [ 926.368505] env[63515]: value = "task-1111592" [ 926.368505] env[63515]: _type = "Task" [ 926.368505] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 926.376422] env[63515]: DEBUG oslo_vmware.api [None req-f7e007a5-60f7-44d2-a529-05d489daa7aa tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Task: {'id': task-1111592, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.384542] env[63515]: DEBUG nova.compute.manager [req-81746d00-f124-45db-9d62-30ecb0f7f4d2 req-9817ec92-5dc7-40ea-86df-53e974f5f907 service nova] [instance: 2ea99c7a-8b61-4718-bee0-f4ce4fc021af] Received event network-vif-deleted-1ddac51a-f040-4ab9-b3bd-3f75ed5e1f5b {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 926.409027] env[63515]: DEBUG oslo_vmware.api [None req-9e4db499-5494-499b-b448-2d861439cb14 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Task: {'id': task-1111589, 'name': PowerOffVM_Task, 'duration_secs': 0.328222} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 926.409344] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-9e4db499-5494-499b-b448-2d861439cb14 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] [instance: ca6a8ac3-d880-4333-9816-2f076e423844] Powered off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 926.409529] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-9e4db499-5494-499b-b448-2d861439cb14 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] [instance: ca6a8ac3-d880-4333-9816-2f076e423844] Unregistering the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 926.409797] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-569e0eda-bfe3-4cc3-82ff-2da4ef7a37dd {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.498864] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-9e4db499-5494-499b-b448-2d861439cb14 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] [instance: ca6a8ac3-d880-4333-9816-2f076e423844] Unregistered the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 926.499108] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-9e4db499-5494-499b-b448-2d861439cb14 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] [instance: ca6a8ac3-d880-4333-9816-2f076e423844] Deleting contents of the VM from datastore datastore2 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 926.499300] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-9e4db499-5494-499b-b448-2d861439cb14 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Deleting the datastore file [datastore2] ca6a8ac3-d880-4333-9816-2f076e423844 {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 926.499574] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e5bd0d25-267a-4aab-9016-42f68178c942 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.507458] env[63515]: DEBUG oslo_vmware.api [None req-9e4db499-5494-499b-b448-2d861439cb14 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Waiting for the task: (returnval){ [ 926.507458] env[63515]: value = "task-1111594" [ 926.507458] env[63515]: _type = "Task" [ 926.507458] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 926.518651] env[63515]: DEBUG oslo_vmware.api [None req-9e4db499-5494-499b-b448-2d861439cb14 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Task: {'id': task-1111594, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.561426] env[63515]: DEBUG oslo_vmware.api [None req-b95e91a2-83f6-4cd7-9448-65181d09829a tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Task: {'id': task-1111585, 'name': CreateSnapshot_Task, 'duration_secs': 0.903511} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 926.562259] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-b95e91a2-83f6-4cd7-9448-65181d09829a tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 68766cd4-84be-475b-8494-d7ab43a9e969] Created Snapshot of the VM instance {{(pid=63515) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 926.563236] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dbd72673-2a07-4030-80d2-e03aa83422c2 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.616300] env[63515]: DEBUG nova.compute.utils [None req-d63820c0-6e1a-45a6-8579-7bfab43bd25a tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Using /dev/sd instead of None {{(pid=63515) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 926.627525] env[63515]: DEBUG nova.compute.manager [None req-d63820c0-6e1a-45a6-8579-7bfab43bd25a tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: 2fa098b6-ecc0-491f-8cfc-5ee335bbf943] Allocating IP information in the background. {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 926.627698] env[63515]: DEBUG nova.network.neutron [None req-d63820c0-6e1a-45a6-8579-7bfab43bd25a tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: 2fa098b6-ecc0-491f-8cfc-5ee335bbf943] allocate_for_instance() {{(pid=63515) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 926.653268] env[63515]: DEBUG oslo_concurrency.lockutils [req-39d93c87-c26a-494e-8c52-5a129e1bb256 req-c736a420-f791-4a16-91db-06eb2097c5fa service nova] Releasing lock "refresh_cache-eab16df9-7bb5-4576-bca0-769a561c5fe9" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 926.694859] env[63515]: DEBUG nova.policy [None req-d63820c0-6e1a-45a6-8579-7bfab43bd25a tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '002c1a9364184b82b649fc18d017d97f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2259667ddc0c42639ea1af3b6e6fa0e8', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63515) authorize /opt/stack/nova/nova/policy.py:201}} [ 926.778332] env[63515]: DEBUG oslo_concurrency.lockutils [None req-525a7c59-1fc2-4bbe-98db-35f4ad5c239b tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Lock "interface-eab16df9-7bb5-4576-bca0-769a561c5fe9-None" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 7.644s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 926.790270] env[63515]: DEBUG oslo_vmware.api [None req-0bb813f4-ce4d-4144-b80d-b179b1c5fd20 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Task: {'id': task-1111590, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.880361] env[63515]: DEBUG oslo_vmware.api [None req-f7e007a5-60f7-44d2-a529-05d489daa7aa tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Task: {'id': task-1111592, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.236916} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 926.880694] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-f7e007a5-60f7-44d2-a529-05d489daa7aa tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Deleted the datastore file {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 926.880928] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-f7e007a5-60f7-44d2-a529-05d489daa7aa tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] [instance: 9236607d-47af-4272-93db-79826e918aa1] Deleted contents of the VM from datastore datastore2 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 926.881134] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-f7e007a5-60f7-44d2-a529-05d489daa7aa tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] [instance: 9236607d-47af-4272-93db-79826e918aa1] Instance destroyed {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 926.881873] env[63515]: INFO nova.compute.manager [None req-f7e007a5-60f7-44d2-a529-05d489daa7aa tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] [instance: 9236607d-47af-4272-93db-79826e918aa1] Took 1.14 seconds to destroy the instance on the hypervisor. [ 926.882201] env[63515]: DEBUG oslo.service.loopingcall [None req-f7e007a5-60f7-44d2-a529-05d489daa7aa tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 926.882426] env[63515]: DEBUG nova.compute.manager [-] [instance: 9236607d-47af-4272-93db-79826e918aa1] Deallocating network for instance {{(pid=63515) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 926.882519] env[63515]: DEBUG nova.network.neutron [-] [instance: 9236607d-47af-4272-93db-79826e918aa1] deallocate_for_instance() {{(pid=63515) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 926.945850] env[63515]: INFO nova.compute.manager [None req-575abd55-8fd4-45a9-87be-30608bf1eab5 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] [instance: 2a29f37d-e786-48d3-b126-4854d6e7c67c] Unrescuing [ 926.946171] env[63515]: DEBUG oslo_concurrency.lockutils [None req-575abd55-8fd4-45a9-87be-30608bf1eab5 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Acquiring lock "refresh_cache-2a29f37d-e786-48d3-b126-4854d6e7c67c" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 926.946354] env[63515]: DEBUG oslo_concurrency.lockutils [None req-575abd55-8fd4-45a9-87be-30608bf1eab5 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Acquired lock "refresh_cache-2a29f37d-e786-48d3-b126-4854d6e7c67c" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 926.946525] env[63515]: DEBUG nova.network.neutron [None req-575abd55-8fd4-45a9-87be-30608bf1eab5 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] [instance: 2a29f37d-e786-48d3-b126-4854d6e7c67c] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 927.019427] env[63515]: DEBUG oslo_vmware.api [None req-9e4db499-5494-499b-b448-2d861439cb14 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Task: {'id': task-1111594, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.161261} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 927.019619] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-9e4db499-5494-499b-b448-2d861439cb14 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Deleted the datastore file {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 927.019823] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-9e4db499-5494-499b-b448-2d861439cb14 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] [instance: ca6a8ac3-d880-4333-9816-2f076e423844] Deleted contents of the VM from datastore datastore2 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 927.020011] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-9e4db499-5494-499b-b448-2d861439cb14 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] [instance: ca6a8ac3-d880-4333-9816-2f076e423844] Instance destroyed {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 927.020205] env[63515]: INFO nova.compute.manager [None req-9e4db499-5494-499b-b448-2d861439cb14 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] [instance: ca6a8ac3-d880-4333-9816-2f076e423844] Took 1.14 seconds to destroy the instance on the hypervisor. [ 927.020493] env[63515]: DEBUG oslo.service.loopingcall [None req-9e4db499-5494-499b-b448-2d861439cb14 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 927.020740] env[63515]: DEBUG nova.compute.manager [-] [instance: ca6a8ac3-d880-4333-9816-2f076e423844] Deallocating network for instance {{(pid=63515) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 927.020857] env[63515]: DEBUG nova.network.neutron [-] [instance: ca6a8ac3-d880-4333-9816-2f076e423844] deallocate_for_instance() {{(pid=63515) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 927.084123] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-b95e91a2-83f6-4cd7-9448-65181d09829a tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 68766cd4-84be-475b-8494-d7ab43a9e969] Creating linked-clone VM from snapshot {{(pid=63515) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 927.084123] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-ff3ea024-3e56-4fec-8562-1b8650b3aec3 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.094326] env[63515]: DEBUG oslo_vmware.api [None req-b95e91a2-83f6-4cd7-9448-65181d09829a tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Waiting for the task: (returnval){ [ 927.094326] env[63515]: value = "task-1111595" [ 927.094326] env[63515]: _type = "Task" [ 927.094326] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 927.101675] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80b005e5-939d-4492-8888-7c341c1e8883 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.109939] env[63515]: DEBUG oslo_vmware.api [None req-b95e91a2-83f6-4cd7-9448-65181d09829a tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Task: {'id': task-1111595, 'name': CloneVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.127152] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Applying migration context for instance 829e72d7-a56a-451c-b98a-94253e5900d5 as it has an incoming, in-progress migration c4203df0-81f5-4cda-85fc-996941b25046. Migration status is migrating {{(pid=63515) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1016}} [ 927.128952] env[63515]: INFO nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] [instance: 829e72d7-a56a-451c-b98a-94253e5900d5] Updating resource usage from migration c4203df0-81f5-4cda-85fc-996941b25046 [ 927.131539] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-1d5e76d4-1b01-4bdb-831a-7cba26a22dfc tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: 829e72d7-a56a-451c-b98a-94253e5900d5] Updating instance '829e72d7-a56a-451c-b98a-94253e5900d5' progress to 0 {{(pid=63515) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 927.140156] env[63515]: DEBUG nova.compute.manager [None req-d63820c0-6e1a-45a6-8579-7bfab43bd25a tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: 2fa098b6-ecc0-491f-8cfc-5ee335bbf943] Start building block device mappings for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 927.159677] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Instance 5cbce760-0163-4b27-8ae3-e46c926c8916 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63515) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 927.159823] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Instance 63ce797d-7180-4209-ac2c-81978bf7607a actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63515) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 927.159963] env[63515]: WARNING nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Instance 87c468d9-9594-4804-b461-527f01f6118f is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 927.160107] env[63515]: WARNING nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Instance 70df69e5-687b-44fb-b6fc-cdb08e21dda0 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 927.162463] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Instance 19475235-dbec-40a5-bd8f-8070616b7ddb actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63515) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 927.162463] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Instance 2a29f37d-e786-48d3-b126-4854d6e7c67c actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63515) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 927.162463] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Instance eab16df9-7bb5-4576-bca0-769a561c5fe9 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63515) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 927.162463] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Instance 68766cd4-84be-475b-8494-d7ab43a9e969 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63515) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 927.162463] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Instance e13da90c-28e6-43d2-99b7-19c5095954ca actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63515) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 927.162463] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Instance 86b1ddcf-585c-41d9-84fe-17b794f1abbc actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63515) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 927.162463] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Instance 9236607d-47af-4272-93db-79826e918aa1 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63515) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 927.162463] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Instance ca6a8ac3-d880-4333-9816-2f076e423844 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63515) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 927.162463] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Instance a2203c5e-66b9-4c7e-8d76-e12a08b57b1d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63515) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 927.162463] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Migration c4203df0-81f5-4cda-85fc-996941b25046 is active on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63515) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1712}} [ 927.162463] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Instance 829e72d7-a56a-451c-b98a-94253e5900d5 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=63515) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 927.162463] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Instance 2fa098b6-ecc0-491f-8cfc-5ee335bbf943 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63515) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 927.232959] env[63515]: DEBUG nova.network.neutron [None req-d63820c0-6e1a-45a6-8579-7bfab43bd25a tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: 2fa098b6-ecc0-491f-8cfc-5ee335bbf943] Successfully created port: e136c744-91c7-433e-979f-0a202f915a28 {{(pid=63515) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 927.295045] env[63515]: DEBUG oslo_vmware.api [None req-0bb813f4-ce4d-4144-b80d-b179b1c5fd20 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Task: {'id': task-1111590, 'name': ReconfigVM_Task, 'duration_secs': 0.650116} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 927.295840] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-0bb813f4-ce4d-4144-b80d-b179b1c5fd20 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: a2203c5e-66b9-4c7e-8d76-e12a08b57b1d] Reconfigured VM instance instance-00000054 to attach disk [datastore2] a2203c5e-66b9-4c7e-8d76-e12a08b57b1d/a2203c5e-66b9-4c7e-8d76-e12a08b57b1d.vmdk or device None with type sparse {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 927.296803] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-9e8bc804-259e-4834-a151-2eb0e97a716d {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.306018] env[63515]: DEBUG oslo_vmware.api [None req-0bb813f4-ce4d-4144-b80d-b179b1c5fd20 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Waiting for the task: (returnval){ [ 927.306018] env[63515]: value = "task-1111596" [ 927.306018] env[63515]: _type = "Task" [ 927.306018] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 927.316890] env[63515]: DEBUG oslo_vmware.api [None req-0bb813f4-ce4d-4144-b80d-b179b1c5fd20 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Task: {'id': task-1111596, 'name': Rename_Task} progress is 5%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.610270] env[63515]: DEBUG oslo_vmware.api [None req-b95e91a2-83f6-4cd7-9448-65181d09829a tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Task: {'id': task-1111595, 'name': CloneVM_Task} progress is 94%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.614170] env[63515]: DEBUG nova.compute.manager [req-fca655c0-d1e2-49d2-9db7-eda76f93faba req-2d3323bf-1a2f-4167-b751-bd9afcc158ed service nova] [instance: ca6a8ac3-d880-4333-9816-2f076e423844] Received event network-vif-deleted-8dcf0639-35d4-47fc-80ef-cc62d15fe62a {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 927.614370] env[63515]: INFO nova.compute.manager [req-fca655c0-d1e2-49d2-9db7-eda76f93faba req-2d3323bf-1a2f-4167-b751-bd9afcc158ed service nova] [instance: ca6a8ac3-d880-4333-9816-2f076e423844] Neutron deleted interface 8dcf0639-35d4-47fc-80ef-cc62d15fe62a; detaching it from the instance and deleting it from the info cache [ 927.614881] env[63515]: DEBUG nova.network.neutron [req-fca655c0-d1e2-49d2-9db7-eda76f93faba req-2d3323bf-1a2f-4167-b751-bd9afcc158ed service nova] [instance: ca6a8ac3-d880-4333-9816-2f076e423844] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 927.645179] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-1d5e76d4-1b01-4bdb-831a-7cba26a22dfc tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: 829e72d7-a56a-451c-b98a-94253e5900d5] Powering off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 927.650554] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-752cffc7-4874-4e5e-a4cb-d76af063b5e5 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.660572] env[63515]: DEBUG oslo_vmware.api [None req-1d5e76d4-1b01-4bdb-831a-7cba26a22dfc tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Waiting for the task: (returnval){ [ 927.660572] env[63515]: value = "task-1111597" [ 927.660572] env[63515]: _type = "Task" [ 927.660572] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 927.664763] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Instance 014720fe-87b9-4e55-b9ba-295b009c7ff6 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63515) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 927.666708] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Total usable vcpus: 48, total allocated vcpus: 14 {{(pid=63515) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 927.667356] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=3264MB phys_disk=200GB used_disk=14GB total_vcpus=48 used_vcpus=14 pci_stats=[] {{(pid=63515) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 927.675806] env[63515]: DEBUG oslo_vmware.api [None req-1d5e76d4-1b01-4bdb-831a-7cba26a22dfc tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Task: {'id': task-1111597, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.816404] env[63515]: DEBUG oslo_vmware.api [None req-0bb813f4-ce4d-4144-b80d-b179b1c5fd20 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Task: {'id': task-1111596, 'name': Rename_Task} progress is 99%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.830441] env[63515]: DEBUG nova.network.neutron [-] [instance: 9236607d-47af-4272-93db-79826e918aa1] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 927.936136] env[63515]: DEBUG nova.network.neutron [None req-575abd55-8fd4-45a9-87be-30608bf1eab5 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] [instance: 2a29f37d-e786-48d3-b126-4854d6e7c67c] Updating instance_info_cache with network_info: [{"id": "956cd361-0d42-4914-85d3-5f75a72a68a1", "address": "fa:16:3e:a3:72:bb", "network": {"id": "850bfa95-3df6-4019-8260-ade20ec58799", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-832000501-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.153", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "08c9aa51b1864410809f8877c7cd49c4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3cf748a8-7ae0-4dca-817d-e727c30d72f4", "external-id": "nsx-vlan-transportzone-853", "segmentation_id": 853, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap956cd361-0d", "ovs_interfaceid": "956cd361-0d42-4914-85d3-5f75a72a68a1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 927.980207] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1f7a001-88fa-4376-bd55-83c6d76bf7ef {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.990880] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f62e9a08-be2e-470e-b7a4-6d00354376d7 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.023047] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9c6108b-8981-463b-9d87-9e4dd24a5731 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.031068] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89fbe054-bbf3-4b6e-b584-385e4a2458c9 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.036840] env[63515]: DEBUG nova.network.neutron [-] [instance: ca6a8ac3-d880-4333-9816-2f076e423844] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 928.052667] env[63515]: DEBUG nova.compute.provider_tree [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Updating inventory in ProviderTree for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 928.105727] env[63515]: DEBUG oslo_vmware.api [None req-b95e91a2-83f6-4cd7-9448-65181d09829a tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Task: {'id': task-1111595, 'name': CloneVM_Task} progress is 94%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.120722] env[63515]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b482d3d6-8a9b-421f-b50e-81476db7c596 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.133024] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7eda48b-7121-4bd2-be72-fd2e85a04baf {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.154052] env[63515]: DEBUG nova.compute.manager [None req-d63820c0-6e1a-45a6-8579-7bfab43bd25a tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: 2fa098b6-ecc0-491f-8cfc-5ee335bbf943] Start spawning the instance on the hypervisor. {{(pid=63515) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 928.163972] env[63515]: DEBUG nova.compute.manager [req-fca655c0-d1e2-49d2-9db7-eda76f93faba req-2d3323bf-1a2f-4167-b751-bd9afcc158ed service nova] [instance: ca6a8ac3-d880-4333-9816-2f076e423844] Detach interface failed, port_id=8dcf0639-35d4-47fc-80ef-cc62d15fe62a, reason: Instance ca6a8ac3-d880-4333-9816-2f076e423844 could not be found. {{(pid=63515) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 928.174027] env[63515]: DEBUG oslo_vmware.api [None req-1d5e76d4-1b01-4bdb-831a-7cba26a22dfc tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Task: {'id': task-1111597, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.194133] env[63515]: DEBUG nova.virt.hardware [None req-d63820c0-6e1a-45a6-8579-7bfab43bd25a tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T02:52:14Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T02:51:56Z,direct_url=,disk_format='vmdk',id=8a120570-cb06-4099-b262-554ca0ad15c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b569255cc43e42e7ae3f2b4ad37c6ef4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T02:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 928.194133] env[63515]: DEBUG nova.virt.hardware [None req-d63820c0-6e1a-45a6-8579-7bfab43bd25a tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Flavor limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 928.194133] env[63515]: DEBUG nova.virt.hardware [None req-d63820c0-6e1a-45a6-8579-7bfab43bd25a tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Image limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 928.194133] env[63515]: DEBUG nova.virt.hardware [None req-d63820c0-6e1a-45a6-8579-7bfab43bd25a tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Flavor pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 928.194133] env[63515]: DEBUG nova.virt.hardware [None req-d63820c0-6e1a-45a6-8579-7bfab43bd25a tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Image pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 928.194133] env[63515]: DEBUG nova.virt.hardware [None req-d63820c0-6e1a-45a6-8579-7bfab43bd25a tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 928.194133] env[63515]: DEBUG nova.virt.hardware [None req-d63820c0-6e1a-45a6-8579-7bfab43bd25a tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 928.194505] env[63515]: DEBUG nova.virt.hardware [None req-d63820c0-6e1a-45a6-8579-7bfab43bd25a tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 928.194836] env[63515]: DEBUG nova.virt.hardware [None req-d63820c0-6e1a-45a6-8579-7bfab43bd25a tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Got 1 possible topologies {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 928.195384] env[63515]: DEBUG nova.virt.hardware [None req-d63820c0-6e1a-45a6-8579-7bfab43bd25a tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 928.198052] env[63515]: DEBUG nova.virt.hardware [None req-d63820c0-6e1a-45a6-8579-7bfab43bd25a tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 928.198052] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9331f9a5-0953-47f7-b27b-6e64c20b388c {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.206567] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-068c0df2-b5e8-4dc3-982c-5e8d53ffed48 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.316650] env[63515]: DEBUG oslo_vmware.api [None req-0bb813f4-ce4d-4144-b80d-b179b1c5fd20 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Task: {'id': task-1111596, 'name': Rename_Task} progress is 99%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.334773] env[63515]: INFO nova.compute.manager [-] [instance: 9236607d-47af-4272-93db-79826e918aa1] Took 1.45 seconds to deallocate network for instance. [ 928.448048] env[63515]: DEBUG oslo_concurrency.lockutils [None req-575abd55-8fd4-45a9-87be-30608bf1eab5 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Releasing lock "refresh_cache-2a29f37d-e786-48d3-b126-4854d6e7c67c" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 928.448048] env[63515]: DEBUG nova.objects.instance [None req-575abd55-8fd4-45a9-87be-30608bf1eab5 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Lazy-loading 'flavor' on Instance uuid 2a29f37d-e786-48d3-b126-4854d6e7c67c {{(pid=63515) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 928.541810] env[63515]: INFO nova.compute.manager [-] [instance: ca6a8ac3-d880-4333-9816-2f076e423844] Took 1.52 seconds to deallocate network for instance. [ 928.584898] env[63515]: DEBUG nova.compute.manager [req-007934dc-115b-47bc-9feb-ee21f35ddc34 req-e5c4fab3-6470-4659-bbac-ef1fc1809dcc service nova] [instance: 9236607d-47af-4272-93db-79826e918aa1] Received event network-vif-deleted-ca78220d-deb4-41bc-a20b-b92b8fe2cdba {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 928.586379] env[63515]: ERROR nova.scheduler.client.report [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] [req-c3106137-c80b-4508-8651-a237b2232578] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 2bd86232-2b6f-44d5-9057-1a3a6b27185a. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-c3106137-c80b-4508-8651-a237b2232578"}]} [ 928.610081] env[63515]: DEBUG oslo_vmware.api [None req-b95e91a2-83f6-4cd7-9448-65181d09829a tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Task: {'id': task-1111595, 'name': CloneVM_Task, 'duration_secs': 1.472317} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 928.611856] env[63515]: DEBUG nova.scheduler.client.report [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Refreshing inventories for resource provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 928.613716] env[63515]: INFO nova.virt.vmwareapi.vmops [None req-b95e91a2-83f6-4cd7-9448-65181d09829a tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 68766cd4-84be-475b-8494-d7ab43a9e969] Created linked-clone VM from snapshot [ 928.615804] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5987778d-be96-4633-a7a3-7fecfdf4e7be {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.628429] env[63515]: DEBUG nova.virt.vmwareapi.images [None req-b95e91a2-83f6-4cd7-9448-65181d09829a tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 68766cd4-84be-475b-8494-d7ab43a9e969] Uploading image ca8a2595-00a8-4a1d-8bdc-e37e8a7f93e5 {{(pid=63515) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 928.635514] env[63515]: DEBUG nova.scheduler.client.report [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Updating ProviderTree inventory for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 928.635645] env[63515]: DEBUG nova.compute.provider_tree [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Updating inventory in ProviderTree for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 928.651691] env[63515]: DEBUG nova.scheduler.client.report [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Refreshing aggregate associations for resource provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a, aggregates: None {{(pid=63515) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 928.666057] env[63515]: DEBUG oslo_vmware.rw_handles [None req-b95e91a2-83f6-4cd7-9448-65181d09829a tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 928.666057] env[63515]: value = "vm-243531" [ 928.666057] env[63515]: _type = "VirtualMachine" [ 928.666057] env[63515]: }. {{(pid=63515) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 928.666538] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-7e8ccc14-f998-447b-8bd2-7ae4b88d3774 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.677951] env[63515]: DEBUG nova.scheduler.client.report [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Refreshing trait associations for resource provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a, traits: COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=63515) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 928.685843] env[63515]: DEBUG oslo_vmware.rw_handles [None req-b95e91a2-83f6-4cd7-9448-65181d09829a tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Lease: (returnval){ [ 928.685843] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]525ef456-73fa-4a2b-6cc8-11a73988e681" [ 928.685843] env[63515]: _type = "HttpNfcLease" [ 928.685843] env[63515]: } obtained for exporting VM: (result){ [ 928.685843] env[63515]: value = "vm-243531" [ 928.685843] env[63515]: _type = "VirtualMachine" [ 928.685843] env[63515]: }. {{(pid=63515) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 928.686329] env[63515]: DEBUG oslo_vmware.api [None req-b95e91a2-83f6-4cd7-9448-65181d09829a tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Waiting for the lease: (returnval){ [ 928.686329] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]525ef456-73fa-4a2b-6cc8-11a73988e681" [ 928.686329] env[63515]: _type = "HttpNfcLease" [ 928.686329] env[63515]: } to be ready. {{(pid=63515) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 928.686625] env[63515]: DEBUG oslo_vmware.api [None req-1d5e76d4-1b01-4bdb-831a-7cba26a22dfc tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Task: {'id': task-1111597, 'name': PowerOffVM_Task} progress is 100%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.697023] env[63515]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 928.697023] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]525ef456-73fa-4a2b-6cc8-11a73988e681" [ 928.697023] env[63515]: _type = "HttpNfcLease" [ 928.697023] env[63515]: } is ready. {{(pid=63515) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 928.697023] env[63515]: DEBUG oslo_vmware.rw_handles [None req-b95e91a2-83f6-4cd7-9448-65181d09829a tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 928.697023] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]525ef456-73fa-4a2b-6cc8-11a73988e681" [ 928.697023] env[63515]: _type = "HttpNfcLease" [ 928.697023] env[63515]: }. {{(pid=63515) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 928.697023] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6d9def8-c033-40ca-9ba5-5329dedda09d {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.703746] env[63515]: DEBUG oslo_vmware.rw_handles [None req-b95e91a2-83f6-4cd7-9448-65181d09829a tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52467804-bf2c-9afd-896b-b66372e45b11/disk-0.vmdk from lease info. {{(pid=63515) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 928.703746] env[63515]: DEBUG oslo_vmware.rw_handles [None req-b95e91a2-83f6-4cd7-9448-65181d09829a tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52467804-bf2c-9afd-896b-b66372e45b11/disk-0.vmdk for reading. {{(pid=63515) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 928.799330] env[63515]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-57f63945-4a3a-458e-b0db-724e7eb4c0d1 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.817836] env[63515]: DEBUG oslo_vmware.api [None req-0bb813f4-ce4d-4144-b80d-b179b1c5fd20 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Task: {'id': task-1111596, 'name': Rename_Task, 'duration_secs': 1.212004} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 928.817836] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-0bb813f4-ce4d-4144-b80d-b179b1c5fd20 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: a2203c5e-66b9-4c7e-8d76-e12a08b57b1d] Powering on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 928.818051] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-720fad97-598d-4ac3-b409-57bb249de2c0 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.829218] env[63515]: DEBUG oslo_vmware.api [None req-0bb813f4-ce4d-4144-b80d-b179b1c5fd20 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Waiting for the task: (returnval){ [ 928.829218] env[63515]: value = "task-1111599" [ 928.829218] env[63515]: _type = "Task" [ 928.829218] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 928.842564] env[63515]: DEBUG oslo_concurrency.lockutils [None req-f7e007a5-60f7-44d2-a529-05d489daa7aa tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 928.848022] env[63515]: DEBUG oslo_vmware.api [None req-0bb813f4-ce4d-4144-b80d-b179b1c5fd20 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Task: {'id': task-1111599, 'name': PowerOnVM_Task} progress is 33%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.950924] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9831b91b-3946-4e99-b660-8029b2cb6596 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.974730] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-575abd55-8fd4-45a9-87be-30608bf1eab5 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] [instance: 2a29f37d-e786-48d3-b126-4854d6e7c67c] Powering off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 928.980229] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1ac8ae87-76ea-4aa6-a6a0-4660d5d47267 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.985704] env[63515]: DEBUG oslo_vmware.api [None req-575abd55-8fd4-45a9-87be-30608bf1eab5 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Waiting for the task: (returnval){ [ 928.985704] env[63515]: value = "task-1111600" [ 928.985704] env[63515]: _type = "Task" [ 928.985704] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 928.995646] env[63515]: DEBUG oslo_vmware.api [None req-575abd55-8fd4-45a9-87be-30608bf1eab5 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Task: {'id': task-1111600, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.030354] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b91f5f0-f0be-4776-a381-d7bff69f91fd {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.044561] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-319c81fe-090f-46e8-9c56-7ae0fcf7ffca {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.053766] env[63515]: DEBUG oslo_concurrency.lockutils [None req-9e4db499-5494-499b-b448-2d861439cb14 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 929.085126] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1cff0fa2-b385-44ea-835e-ef57e84e3582 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.094677] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3472b355-f3d2-4239-93e0-cd3e6d31ed97 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.111744] env[63515]: DEBUG nova.compute.provider_tree [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Updating inventory in ProviderTree for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 929.138805] env[63515]: DEBUG nova.network.neutron [None req-d63820c0-6e1a-45a6-8579-7bfab43bd25a tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: 2fa098b6-ecc0-491f-8cfc-5ee335bbf943] Successfully updated port: e136c744-91c7-433e-979f-0a202f915a28 {{(pid=63515) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 929.179085] env[63515]: DEBUG oslo_vmware.api [None req-1d5e76d4-1b01-4bdb-831a-7cba26a22dfc tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Task: {'id': task-1111597, 'name': PowerOffVM_Task, 'duration_secs': 1.086008} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 929.179440] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-1d5e76d4-1b01-4bdb-831a-7cba26a22dfc tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: 829e72d7-a56a-451c-b98a-94253e5900d5] Powered off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 929.180866] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-1d5e76d4-1b01-4bdb-831a-7cba26a22dfc tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: 829e72d7-a56a-451c-b98a-94253e5900d5] Updating instance '829e72d7-a56a-451c-b98a-94253e5900d5' progress to 17 {{(pid=63515) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 929.342382] env[63515]: DEBUG oslo_vmware.api [None req-0bb813f4-ce4d-4144-b80d-b179b1c5fd20 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Task: {'id': task-1111599, 'name': PowerOnVM_Task} progress is 89%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.366214] env[63515]: DEBUG oslo_concurrency.lockutils [None req-bd214878-de3c-49e1-b706-bcb83d66311c tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Acquiring lock "interface-eab16df9-7bb5-4576-bca0-769a561c5fe9-ff4043f6-7576-4093-80ed-d36240b37f2f" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 929.366796] env[63515]: DEBUG oslo_concurrency.lockutils [None req-bd214878-de3c-49e1-b706-bcb83d66311c tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Lock "interface-eab16df9-7bb5-4576-bca0-769a561c5fe9-ff4043f6-7576-4093-80ed-d36240b37f2f" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.001s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 929.367374] env[63515]: DEBUG nova.objects.instance [None req-bd214878-de3c-49e1-b706-bcb83d66311c tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Lazy-loading 'flavor' on Instance uuid eab16df9-7bb5-4576-bca0-769a561c5fe9 {{(pid=63515) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 929.496293] env[63515]: DEBUG oslo_vmware.api [None req-575abd55-8fd4-45a9-87be-30608bf1eab5 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Task: {'id': task-1111600, 'name': PowerOffVM_Task, 'duration_secs': 0.299807} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 929.496707] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-575abd55-8fd4-45a9-87be-30608bf1eab5 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] [instance: 2a29f37d-e786-48d3-b126-4854d6e7c67c] Powered off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 929.502546] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-575abd55-8fd4-45a9-87be-30608bf1eab5 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] [instance: 2a29f37d-e786-48d3-b126-4854d6e7c67c] Reconfiguring VM instance instance-00000045 to detach disk 2002 {{(pid=63515) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 929.503829] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-49fefd84-405b-4554-8969-5a8f4c85e531 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.529384] env[63515]: DEBUG oslo_vmware.api [None req-575abd55-8fd4-45a9-87be-30608bf1eab5 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Waiting for the task: (returnval){ [ 929.529384] env[63515]: value = "task-1111601" [ 929.529384] env[63515]: _type = "Task" [ 929.529384] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 929.537743] env[63515]: DEBUG oslo_vmware.api [None req-575abd55-8fd4-45a9-87be-30608bf1eab5 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Task: {'id': task-1111601, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.641840] env[63515]: DEBUG oslo_concurrency.lockutils [None req-d63820c0-6e1a-45a6-8579-7bfab43bd25a tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Acquiring lock "refresh_cache-2fa098b6-ecc0-491f-8cfc-5ee335bbf943" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 929.641996] env[63515]: DEBUG oslo_concurrency.lockutils [None req-d63820c0-6e1a-45a6-8579-7bfab43bd25a tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Acquired lock "refresh_cache-2fa098b6-ecc0-491f-8cfc-5ee335bbf943" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 929.642169] env[63515]: DEBUG nova.network.neutron [None req-d63820c0-6e1a-45a6-8579-7bfab43bd25a tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: 2fa098b6-ecc0-491f-8cfc-5ee335bbf943] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 929.648210] env[63515]: DEBUG nova.scheduler.client.report [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Updated inventory for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a with generation 118 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 929.648687] env[63515]: DEBUG nova.compute.provider_tree [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Updating resource provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a generation from 118 to 119 during operation: update_inventory {{(pid=63515) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 929.649252] env[63515]: DEBUG nova.compute.provider_tree [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Updating inventory in ProviderTree for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 929.690235] env[63515]: DEBUG nova.virt.hardware [None req-1d5e76d4-1b01-4bdb-831a-7cba26a22dfc tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T02:52:16Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=8a120570-cb06-4099-b262-554ca0ad15c5,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 929.690586] env[63515]: DEBUG nova.virt.hardware [None req-1d5e76d4-1b01-4bdb-831a-7cba26a22dfc tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Flavor limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 929.690970] env[63515]: DEBUG nova.virt.hardware [None req-1d5e76d4-1b01-4bdb-831a-7cba26a22dfc tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Image limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 929.691213] env[63515]: DEBUG nova.virt.hardware [None req-1d5e76d4-1b01-4bdb-831a-7cba26a22dfc tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Flavor pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 929.691404] env[63515]: DEBUG nova.virt.hardware [None req-1d5e76d4-1b01-4bdb-831a-7cba26a22dfc tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Image pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 929.691706] env[63515]: DEBUG nova.virt.hardware [None req-1d5e76d4-1b01-4bdb-831a-7cba26a22dfc tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 929.692039] env[63515]: DEBUG nova.virt.hardware [None req-1d5e76d4-1b01-4bdb-831a-7cba26a22dfc tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 929.692252] env[63515]: DEBUG nova.virt.hardware [None req-1d5e76d4-1b01-4bdb-831a-7cba26a22dfc tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 929.692527] env[63515]: DEBUG nova.virt.hardware [None req-1d5e76d4-1b01-4bdb-831a-7cba26a22dfc tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Got 1 possible topologies {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 929.692707] env[63515]: DEBUG nova.virt.hardware [None req-1d5e76d4-1b01-4bdb-831a-7cba26a22dfc tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 929.692956] env[63515]: DEBUG nova.virt.hardware [None req-1d5e76d4-1b01-4bdb-831a-7cba26a22dfc tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 929.699049] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-249451d8-0644-408d-ba7c-9dc60ffb9b52 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.718954] env[63515]: DEBUG oslo_vmware.api [None req-1d5e76d4-1b01-4bdb-831a-7cba26a22dfc tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Waiting for the task: (returnval){ [ 929.718954] env[63515]: value = "task-1111602" [ 929.718954] env[63515]: _type = "Task" [ 929.718954] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 929.728863] env[63515]: DEBUG oslo_vmware.api [None req-1d5e76d4-1b01-4bdb-831a-7cba26a22dfc tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Task: {'id': task-1111602, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.846843] env[63515]: DEBUG oslo_vmware.api [None req-0bb813f4-ce4d-4144-b80d-b179b1c5fd20 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Task: {'id': task-1111599, 'name': PowerOnVM_Task, 'duration_secs': 0.594831} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 929.847344] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-0bb813f4-ce4d-4144-b80d-b179b1c5fd20 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: a2203c5e-66b9-4c7e-8d76-e12a08b57b1d] Powered on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 929.847963] env[63515]: INFO nova.compute.manager [None req-0bb813f4-ce4d-4144-b80d-b179b1c5fd20 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: a2203c5e-66b9-4c7e-8d76-e12a08b57b1d] Took 11.02 seconds to spawn the instance on the hypervisor. [ 929.848350] env[63515]: DEBUG nova.compute.manager [None req-0bb813f4-ce4d-4144-b80d-b179b1c5fd20 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: a2203c5e-66b9-4c7e-8d76-e12a08b57b1d] Checking state {{(pid=63515) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 929.849611] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15711d78-dc45-4371-bfa3-ba81562f6264 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.985888] env[63515]: DEBUG nova.objects.instance [None req-bd214878-de3c-49e1-b706-bcb83d66311c tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Lazy-loading 'pci_requests' on Instance uuid eab16df9-7bb5-4576-bca0-769a561c5fe9 {{(pid=63515) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 930.041517] env[63515]: DEBUG oslo_vmware.api [None req-575abd55-8fd4-45a9-87be-30608bf1eab5 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Task: {'id': task-1111601, 'name': ReconfigVM_Task} progress is 99%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.156158] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=63515) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 930.156424] env[63515]: DEBUG oslo_concurrency.lockutils [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 4.046s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 930.156783] env[63515]: DEBUG oslo_concurrency.lockutils [None req-44e7f82e-18f0-4009-907e-7846ed9a3683 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 12.693s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 930.157079] env[63515]: DEBUG oslo_concurrency.lockutils [None req-44e7f82e-18f0-4009-907e-7846ed9a3683 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 930.159172] env[63515]: DEBUG oslo_concurrency.lockutils [None req-872cc7bc-b8ac-4eff-87f3-c2c36aad9630 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 12.403s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 930.159467] env[63515]: DEBUG oslo_concurrency.lockutils [None req-872cc7bc-b8ac-4eff-87f3-c2c36aad9630 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 930.161613] env[63515]: DEBUG oslo_concurrency.lockutils [None req-f7b46a02-c5ff-4523-967a-55b60fa40b2d tempest-ServerActionsV293TestJSON-1193370169 tempest-ServerActionsV293TestJSON-1193370169-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 9.072s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 930.161917] env[63515]: DEBUG oslo_concurrency.lockutils [None req-f7b46a02-c5ff-4523-967a-55b60fa40b2d tempest-ServerActionsV293TestJSON-1193370169 tempest-ServerActionsV293TestJSON-1193370169-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 930.162219] env[63515]: INFO nova.compute.manager [None req-f7b46a02-c5ff-4523-967a-55b60fa40b2d tempest-ServerActionsV293TestJSON-1193370169 tempest-ServerActionsV293TestJSON-1193370169-project-member] [instance: 2ea99c7a-8b61-4718-bee0-f4ce4fc021af] Successfully reverted task state from rebuilding on failure for instance. [ 930.168882] env[63515]: DEBUG oslo_concurrency.lockutils [None req-be7cccf6-3f5b-4895-b678-cf60de2c827a tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.397s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 930.173019] env[63515]: INFO nova.compute.claims [None req-be7cccf6-3f5b-4895-b678-cf60de2c827a tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: 014720fe-87b9-4e55-b9ba-295b009c7ff6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 930.185236] env[63515]: DEBUG nova.network.neutron [None req-d63820c0-6e1a-45a6-8579-7bfab43bd25a tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: 2fa098b6-ecc0-491f-8cfc-5ee335bbf943] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 930.209152] env[63515]: INFO nova.scheduler.client.report [None req-872cc7bc-b8ac-4eff-87f3-c2c36aad9630 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Deleted allocations for instance 70df69e5-687b-44fb-b6fc-cdb08e21dda0 [ 930.212373] env[63515]: INFO nova.scheduler.client.report [None req-44e7f82e-18f0-4009-907e-7846ed9a3683 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Deleted allocations for instance 87c468d9-9594-4804-b461-527f01f6118f [ 930.229522] env[63515]: DEBUG oslo_vmware.api [None req-1d5e76d4-1b01-4bdb-831a-7cba26a22dfc tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Task: {'id': task-1111602, 'name': ReconfigVM_Task, 'duration_secs': 0.434714} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 930.231684] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-1d5e76d4-1b01-4bdb-831a-7cba26a22dfc tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: 829e72d7-a56a-451c-b98a-94253e5900d5] Updating instance '829e72d7-a56a-451c-b98a-94253e5900d5' progress to 33 {{(pid=63515) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 930.363503] env[63515]: DEBUG nova.network.neutron [None req-d63820c0-6e1a-45a6-8579-7bfab43bd25a tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: 2fa098b6-ecc0-491f-8cfc-5ee335bbf943] Updating instance_info_cache with network_info: [{"id": "e136c744-91c7-433e-979f-0a202f915a28", "address": "fa:16:3e:28:2c:4f", "network": {"id": "320921bd-a6c9-4726-a801-19009d70d16e", "bridge": "br-int", "label": "tempest-ImagesTestJSON-304158429-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2259667ddc0c42639ea1af3b6e6fa0e8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2d94740a-bce8-4103-8ecf-230d02ec0a44", "external-id": "nsx-vlan-transportzone-149", "segmentation_id": 149, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape136c744-91", "ovs_interfaceid": "e136c744-91c7-433e-979f-0a202f915a28", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 930.381196] env[63515]: INFO nova.compute.manager [None req-0bb813f4-ce4d-4144-b80d-b179b1c5fd20 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: a2203c5e-66b9-4c7e-8d76-e12a08b57b1d] Took 26.29 seconds to build instance. [ 930.488698] env[63515]: DEBUG nova.objects.base [None req-bd214878-de3c-49e1-b706-bcb83d66311c tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Object Instance lazy-loaded attributes: flavor,pci_requests {{(pid=63515) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 930.488942] env[63515]: DEBUG nova.network.neutron [None req-bd214878-de3c-49e1-b706-bcb83d66311c tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: eab16df9-7bb5-4576-bca0-769a561c5fe9] allocate_for_instance() {{(pid=63515) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 930.540364] env[63515]: DEBUG oslo_vmware.api [None req-575abd55-8fd4-45a9-87be-30608bf1eab5 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Task: {'id': task-1111601, 'name': ReconfigVM_Task, 'duration_secs': 0.645944} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 930.540739] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-575abd55-8fd4-45a9-87be-30608bf1eab5 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] [instance: 2a29f37d-e786-48d3-b126-4854d6e7c67c] Reconfigured VM instance instance-00000045 to detach disk 2002 {{(pid=63515) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 930.541844] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-575abd55-8fd4-45a9-87be-30608bf1eab5 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] [instance: 2a29f37d-e786-48d3-b126-4854d6e7c67c] Powering on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 930.542712] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-21b48068-024f-47ff-a20a-8bbf23ca2cbd {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.550888] env[63515]: DEBUG oslo_vmware.api [None req-575abd55-8fd4-45a9-87be-30608bf1eab5 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Waiting for the task: (returnval){ [ 930.550888] env[63515]: value = "task-1111603" [ 930.550888] env[63515]: _type = "Task" [ 930.550888] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 930.564813] env[63515]: DEBUG oslo_vmware.api [None req-575abd55-8fd4-45a9-87be-30608bf1eab5 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Task: {'id': task-1111603, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.693960] env[63515]: DEBUG nova.policy [None req-bd214878-de3c-49e1-b706-bcb83d66311c tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '16880abf1c5b4341800c94ada2c756c7', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7d4ada9ec35f42b19c6480a9101d21a4', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63515) authorize /opt/stack/nova/nova/policy.py:201}} [ 930.724567] env[63515]: DEBUG oslo_concurrency.lockutils [None req-872cc7bc-b8ac-4eff-87f3-c2c36aad9630 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Lock "70df69e5-687b-44fb-b6fc-cdb08e21dda0" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 15.980s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 930.726718] env[63515]: DEBUG oslo_concurrency.lockutils [None req-44e7f82e-18f0-4009-907e-7846ed9a3683 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Lock "87c468d9-9594-4804-b461-527f01f6118f" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 16.408s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 930.737219] env[63515]: DEBUG nova.virt.hardware [None req-1d5e76d4-1b01-4bdb-831a-7cba26a22dfc tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T02:52:14Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=8a120570-cb06-4099-b262-554ca0ad15c5,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 930.740893] env[63515]: DEBUG nova.virt.hardware [None req-1d5e76d4-1b01-4bdb-831a-7cba26a22dfc tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Flavor limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 930.740893] env[63515]: DEBUG nova.virt.hardware [None req-1d5e76d4-1b01-4bdb-831a-7cba26a22dfc tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Image limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 930.740893] env[63515]: DEBUG nova.virt.hardware [None req-1d5e76d4-1b01-4bdb-831a-7cba26a22dfc tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Flavor pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 930.740893] env[63515]: DEBUG nova.virt.hardware [None req-1d5e76d4-1b01-4bdb-831a-7cba26a22dfc tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Image pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 930.740893] env[63515]: DEBUG nova.virt.hardware [None req-1d5e76d4-1b01-4bdb-831a-7cba26a22dfc tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 930.740893] env[63515]: DEBUG nova.virt.hardware [None req-1d5e76d4-1b01-4bdb-831a-7cba26a22dfc tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 930.740893] env[63515]: DEBUG nova.virt.hardware [None req-1d5e76d4-1b01-4bdb-831a-7cba26a22dfc tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 930.740893] env[63515]: DEBUG nova.virt.hardware [None req-1d5e76d4-1b01-4bdb-831a-7cba26a22dfc tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Got 1 possible topologies {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 930.740893] env[63515]: DEBUG nova.virt.hardware [None req-1d5e76d4-1b01-4bdb-831a-7cba26a22dfc tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 930.740893] env[63515]: DEBUG nova.virt.hardware [None req-1d5e76d4-1b01-4bdb-831a-7cba26a22dfc tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 930.747853] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-1d5e76d4-1b01-4bdb-831a-7cba26a22dfc tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: 829e72d7-a56a-451c-b98a-94253e5900d5] Reconfiguring VM instance instance-00000050 to detach disk 2000 {{(pid=63515) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 930.748222] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0081aee9-b631-452c-8164-b91d4715a344 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.770666] env[63515]: DEBUG nova.compute.manager [req-9de1fa68-443c-45da-8b98-20419f380537 req-cd2e6312-0bd4-4691-8f7f-4cde309fc987 service nova] [instance: 2fa098b6-ecc0-491f-8cfc-5ee335bbf943] Received event network-vif-plugged-e136c744-91c7-433e-979f-0a202f915a28 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 930.771043] env[63515]: DEBUG oslo_concurrency.lockutils [req-9de1fa68-443c-45da-8b98-20419f380537 req-cd2e6312-0bd4-4691-8f7f-4cde309fc987 service nova] Acquiring lock "2fa098b6-ecc0-491f-8cfc-5ee335bbf943-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 930.771355] env[63515]: DEBUG oslo_concurrency.lockutils [req-9de1fa68-443c-45da-8b98-20419f380537 req-cd2e6312-0bd4-4691-8f7f-4cde309fc987 service nova] Lock "2fa098b6-ecc0-491f-8cfc-5ee335bbf943-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 930.771697] env[63515]: DEBUG oslo_concurrency.lockutils [req-9de1fa68-443c-45da-8b98-20419f380537 req-cd2e6312-0bd4-4691-8f7f-4cde309fc987 service nova] Lock "2fa098b6-ecc0-491f-8cfc-5ee335bbf943-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 930.771868] env[63515]: DEBUG nova.compute.manager [req-9de1fa68-443c-45da-8b98-20419f380537 req-cd2e6312-0bd4-4691-8f7f-4cde309fc987 service nova] [instance: 2fa098b6-ecc0-491f-8cfc-5ee335bbf943] No waiting events found dispatching network-vif-plugged-e136c744-91c7-433e-979f-0a202f915a28 {{(pid=63515) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 930.772111] env[63515]: WARNING nova.compute.manager [req-9de1fa68-443c-45da-8b98-20419f380537 req-cd2e6312-0bd4-4691-8f7f-4cde309fc987 service nova] [instance: 2fa098b6-ecc0-491f-8cfc-5ee335bbf943] Received unexpected event network-vif-plugged-e136c744-91c7-433e-979f-0a202f915a28 for instance with vm_state building and task_state spawning. [ 930.772394] env[63515]: DEBUG nova.compute.manager [req-9de1fa68-443c-45da-8b98-20419f380537 req-cd2e6312-0bd4-4691-8f7f-4cde309fc987 service nova] [instance: 2fa098b6-ecc0-491f-8cfc-5ee335bbf943] Received event network-changed-e136c744-91c7-433e-979f-0a202f915a28 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 930.772681] env[63515]: DEBUG nova.compute.manager [req-9de1fa68-443c-45da-8b98-20419f380537 req-cd2e6312-0bd4-4691-8f7f-4cde309fc987 service nova] [instance: 2fa098b6-ecc0-491f-8cfc-5ee335bbf943] Refreshing instance network info cache due to event network-changed-e136c744-91c7-433e-979f-0a202f915a28. {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 930.772875] env[63515]: DEBUG oslo_concurrency.lockutils [req-9de1fa68-443c-45da-8b98-20419f380537 req-cd2e6312-0bd4-4691-8f7f-4cde309fc987 service nova] Acquiring lock "refresh_cache-2fa098b6-ecc0-491f-8cfc-5ee335bbf943" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 930.780334] env[63515]: DEBUG oslo_vmware.api [None req-1d5e76d4-1b01-4bdb-831a-7cba26a22dfc tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Waiting for the task: (returnval){ [ 930.780334] env[63515]: value = "task-1111604" [ 930.780334] env[63515]: _type = "Task" [ 930.780334] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 930.791928] env[63515]: DEBUG oslo_vmware.api [None req-1d5e76d4-1b01-4bdb-831a-7cba26a22dfc tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Task: {'id': task-1111604, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.866521] env[63515]: DEBUG oslo_concurrency.lockutils [None req-d63820c0-6e1a-45a6-8579-7bfab43bd25a tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Releasing lock "refresh_cache-2fa098b6-ecc0-491f-8cfc-5ee335bbf943" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 930.866906] env[63515]: DEBUG nova.compute.manager [None req-d63820c0-6e1a-45a6-8579-7bfab43bd25a tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: 2fa098b6-ecc0-491f-8cfc-5ee335bbf943] Instance network_info: |[{"id": "e136c744-91c7-433e-979f-0a202f915a28", "address": "fa:16:3e:28:2c:4f", "network": {"id": "320921bd-a6c9-4726-a801-19009d70d16e", "bridge": "br-int", "label": "tempest-ImagesTestJSON-304158429-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2259667ddc0c42639ea1af3b6e6fa0e8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2d94740a-bce8-4103-8ecf-230d02ec0a44", "external-id": "nsx-vlan-transportzone-149", "segmentation_id": 149, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape136c744-91", "ovs_interfaceid": "e136c744-91c7-433e-979f-0a202f915a28", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 930.867284] env[63515]: DEBUG oslo_concurrency.lockutils [req-9de1fa68-443c-45da-8b98-20419f380537 req-cd2e6312-0bd4-4691-8f7f-4cde309fc987 service nova] Acquired lock "refresh_cache-2fa098b6-ecc0-491f-8cfc-5ee335bbf943" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 930.867494] env[63515]: DEBUG nova.network.neutron [req-9de1fa68-443c-45da-8b98-20419f380537 req-cd2e6312-0bd4-4691-8f7f-4cde309fc987 service nova] [instance: 2fa098b6-ecc0-491f-8cfc-5ee335bbf943] Refreshing network info cache for port e136c744-91c7-433e-979f-0a202f915a28 {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 930.868993] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-d63820c0-6e1a-45a6-8579-7bfab43bd25a tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: 2fa098b6-ecc0-491f-8cfc-5ee335bbf943] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:28:2c:4f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '2d94740a-bce8-4103-8ecf-230d02ec0a44', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'e136c744-91c7-433e-979f-0a202f915a28', 'vif_model': 'vmxnet3'}] {{(pid=63515) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 930.882267] env[63515]: DEBUG oslo.service.loopingcall [None req-d63820c0-6e1a-45a6-8579-7bfab43bd25a tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 930.888032] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2fa098b6-ecc0-491f-8cfc-5ee335bbf943] Creating VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 930.888729] env[63515]: DEBUG oslo_concurrency.lockutils [None req-0bb813f4-ce4d-4144-b80d-b179b1c5fd20 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Lock "a2203c5e-66b9-4c7e-8d76-e12a08b57b1d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 27.807s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 930.889070] env[63515]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f51b23f9-6318-4f3e-8d9e-6645553d6665 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.914511] env[63515]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 930.914511] env[63515]: value = "task-1111605" [ 930.914511] env[63515]: _type = "Task" [ 930.914511] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 930.923611] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1111605, 'name': CreateVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.062920] env[63515]: DEBUG oslo_vmware.api [None req-575abd55-8fd4-45a9-87be-30608bf1eab5 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Task: {'id': task-1111603, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.187805] env[63515]: DEBUG nova.network.neutron [req-9de1fa68-443c-45da-8b98-20419f380537 req-cd2e6312-0bd4-4691-8f7f-4cde309fc987 service nova] [instance: 2fa098b6-ecc0-491f-8cfc-5ee335bbf943] Updated VIF entry in instance network info cache for port e136c744-91c7-433e-979f-0a202f915a28. {{(pid=63515) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 931.187805] env[63515]: DEBUG nova.network.neutron [req-9de1fa68-443c-45da-8b98-20419f380537 req-cd2e6312-0bd4-4691-8f7f-4cde309fc987 service nova] [instance: 2fa098b6-ecc0-491f-8cfc-5ee335bbf943] Updating instance_info_cache with network_info: [{"id": "e136c744-91c7-433e-979f-0a202f915a28", "address": "fa:16:3e:28:2c:4f", "network": {"id": "320921bd-a6c9-4726-a801-19009d70d16e", "bridge": "br-int", "label": "tempest-ImagesTestJSON-304158429-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2259667ddc0c42639ea1af3b6e6fa0e8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2d94740a-bce8-4103-8ecf-230d02ec0a44", "external-id": "nsx-vlan-transportzone-149", "segmentation_id": 149, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape136c744-91", "ovs_interfaceid": "e136c744-91c7-433e-979f-0a202f915a28", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 931.290121] env[63515]: DEBUG oslo_vmware.api [None req-1d5e76d4-1b01-4bdb-831a-7cba26a22dfc tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Task: {'id': task-1111604, 'name': ReconfigVM_Task, 'duration_secs': 0.23367} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 931.293455] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-1d5e76d4-1b01-4bdb-831a-7cba26a22dfc tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: 829e72d7-a56a-451c-b98a-94253e5900d5] Reconfigured VM instance instance-00000050 to detach disk 2000 {{(pid=63515) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 931.294201] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21e39fdb-52bd-4a7c-b9cb-812b7880dce7 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.318307] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-1d5e76d4-1b01-4bdb-831a-7cba26a22dfc tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: 829e72d7-a56a-451c-b98a-94253e5900d5] Reconfiguring VM instance instance-00000050 to attach disk [datastore1] 829e72d7-a56a-451c-b98a-94253e5900d5/829e72d7-a56a-451c-b98a-94253e5900d5.vmdk or device None with type thin {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 931.321218] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-de16b474-c37f-43ab-9c69-764b90c1a79a {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.342031] env[63515]: DEBUG oslo_vmware.api [None req-1d5e76d4-1b01-4bdb-831a-7cba26a22dfc tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Waiting for the task: (returnval){ [ 931.342031] env[63515]: value = "task-1111606" [ 931.342031] env[63515]: _type = "Task" [ 931.342031] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 931.350225] env[63515]: DEBUG oslo_vmware.api [None req-1d5e76d4-1b01-4bdb-831a-7cba26a22dfc tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Task: {'id': task-1111606, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.427344] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1111605, 'name': CreateVM_Task, 'duration_secs': 0.399321} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 931.427649] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2fa098b6-ecc0-491f-8cfc-5ee335bbf943] Created VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 931.428861] env[63515]: DEBUG oslo_concurrency.lockutils [None req-d63820c0-6e1a-45a6-8579-7bfab43bd25a tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 931.428975] env[63515]: DEBUG oslo_concurrency.lockutils [None req-d63820c0-6e1a-45a6-8579-7bfab43bd25a tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 931.429329] env[63515]: DEBUG oslo_concurrency.lockutils [None req-d63820c0-6e1a-45a6-8579-7bfab43bd25a tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 931.429662] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-60b74846-1a5c-4211-b62a-af8703c6a479 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.434163] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41049106-ee86-4d58-bdcb-7768902bb2c1 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.438461] env[63515]: DEBUG oslo_vmware.api [None req-d63820c0-6e1a-45a6-8579-7bfab43bd25a tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Waiting for the task: (returnval){ [ 931.438461] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]52774a8b-8de0-253e-1e93-2254917f088e" [ 931.438461] env[63515]: _type = "Task" [ 931.438461] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 931.444509] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f86e5ab-764d-49ad-9600-96022ce09f49 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.453268] env[63515]: DEBUG oslo_vmware.api [None req-d63820c0-6e1a-45a6-8579-7bfab43bd25a tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52774a8b-8de0-253e-1e93-2254917f088e, 'name': SearchDatastore_Task, 'duration_secs': 0.011113} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 931.478786] env[63515]: DEBUG oslo_concurrency.lockutils [None req-d63820c0-6e1a-45a6-8579-7bfab43bd25a tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 931.478786] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-d63820c0-6e1a-45a6-8579-7bfab43bd25a tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: 2fa098b6-ecc0-491f-8cfc-5ee335bbf943] Processing image 8a120570-cb06-4099-b262-554ca0ad15c5 {{(pid=63515) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 931.478970] env[63515]: DEBUG oslo_concurrency.lockutils [None req-d63820c0-6e1a-45a6-8579-7bfab43bd25a tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 931.479032] env[63515]: DEBUG oslo_concurrency.lockutils [None req-d63820c0-6e1a-45a6-8579-7bfab43bd25a tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 931.479225] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-d63820c0-6e1a-45a6-8579-7bfab43bd25a tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 931.480103] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-934265af-37a6-4c28-babb-8cf48cf31fe4 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.482619] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d218261a-c8e8-4db1-b11f-68e7955a5c8d {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.490908] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb55aa21-14cf-453c-bd66-41e7d3036636 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.496024] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-d63820c0-6e1a-45a6-8579-7bfab43bd25a tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 931.496200] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-d63820c0-6e1a-45a6-8579-7bfab43bd25a tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63515) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 931.497333] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a73e4788-df80-4bb0-9de0-20b15e7b0f6a {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.509238] env[63515]: DEBUG nova.compute.provider_tree [None req-be7cccf6-3f5b-4895-b678-cf60de2c827a tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 931.515318] env[63515]: DEBUG oslo_vmware.api [None req-d63820c0-6e1a-45a6-8579-7bfab43bd25a tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Waiting for the task: (returnval){ [ 931.515318] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]52b50cc4-5162-058f-02db-4c28f584692a" [ 931.515318] env[63515]: _type = "Task" [ 931.515318] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 931.525730] env[63515]: DEBUG oslo_vmware.api [None req-d63820c0-6e1a-45a6-8579-7bfab43bd25a tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52b50cc4-5162-058f-02db-4c28f584692a, 'name': SearchDatastore_Task, 'duration_secs': 0.009706} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 931.526869] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fc9e0654-b39a-4dbb-bf0d-93edd59a9ff2 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.532835] env[63515]: DEBUG oslo_vmware.api [None req-d63820c0-6e1a-45a6-8579-7bfab43bd25a tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Waiting for the task: (returnval){ [ 931.532835] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]52bce7be-78a7-59ca-3440-b08008d5063c" [ 931.532835] env[63515]: _type = "Task" [ 931.532835] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 931.545066] env[63515]: DEBUG oslo_vmware.api [None req-d63820c0-6e1a-45a6-8579-7bfab43bd25a tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52bce7be-78a7-59ca-3440-b08008d5063c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.560779] env[63515]: DEBUG oslo_vmware.api [None req-575abd55-8fd4-45a9-87be-30608bf1eab5 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Task: {'id': task-1111603, 'name': PowerOnVM_Task, 'duration_secs': 0.567376} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 931.561558] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-575abd55-8fd4-45a9-87be-30608bf1eab5 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] [instance: 2a29f37d-e786-48d3-b126-4854d6e7c67c] Powered on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 931.561763] env[63515]: DEBUG nova.compute.manager [None req-575abd55-8fd4-45a9-87be-30608bf1eab5 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] [instance: 2a29f37d-e786-48d3-b126-4854d6e7c67c] Checking state {{(pid=63515) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 931.562677] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6d1cc86-1420-415d-81d6-f80d67ab325f {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.666686] env[63515]: DEBUG oslo_concurrency.lockutils [None req-56611604-1203-4b20-9871-2ec521700550 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Acquiring lock "a2203c5e-66b9-4c7e-8d76-e12a08b57b1d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 931.666820] env[63515]: DEBUG oslo_concurrency.lockutils [None req-56611604-1203-4b20-9871-2ec521700550 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Lock "a2203c5e-66b9-4c7e-8d76-e12a08b57b1d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 931.667027] env[63515]: DEBUG oslo_concurrency.lockutils [None req-56611604-1203-4b20-9871-2ec521700550 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Acquiring lock "a2203c5e-66b9-4c7e-8d76-e12a08b57b1d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 931.667442] env[63515]: DEBUG oslo_concurrency.lockutils [None req-56611604-1203-4b20-9871-2ec521700550 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Lock "a2203c5e-66b9-4c7e-8d76-e12a08b57b1d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 931.667595] env[63515]: DEBUG oslo_concurrency.lockutils [None req-56611604-1203-4b20-9871-2ec521700550 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Lock "a2203c5e-66b9-4c7e-8d76-e12a08b57b1d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 931.669940] env[63515]: INFO nova.compute.manager [None req-56611604-1203-4b20-9871-2ec521700550 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: a2203c5e-66b9-4c7e-8d76-e12a08b57b1d] Terminating instance [ 931.671739] env[63515]: DEBUG nova.compute.manager [None req-56611604-1203-4b20-9871-2ec521700550 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: a2203c5e-66b9-4c7e-8d76-e12a08b57b1d] Start destroying the instance on the hypervisor. {{(pid=63515) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 931.674135] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-56611604-1203-4b20-9871-2ec521700550 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: a2203c5e-66b9-4c7e-8d76-e12a08b57b1d] Destroying instance {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 931.674135] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ec56042-46cc-4fff-96e9-de98755c8909 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.683098] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-56611604-1203-4b20-9871-2ec521700550 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: a2203c5e-66b9-4c7e-8d76-e12a08b57b1d] Powering off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 931.683098] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f08b1e67-07cf-4b73-a41c-cb2e0a646627 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.689532] env[63515]: DEBUG oslo_vmware.api [None req-56611604-1203-4b20-9871-2ec521700550 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Waiting for the task: (returnval){ [ 931.689532] env[63515]: value = "task-1111607" [ 931.689532] env[63515]: _type = "Task" [ 931.689532] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 931.691223] env[63515]: DEBUG oslo_concurrency.lockutils [req-9de1fa68-443c-45da-8b98-20419f380537 req-cd2e6312-0bd4-4691-8f7f-4cde309fc987 service nova] Releasing lock "refresh_cache-2fa098b6-ecc0-491f-8cfc-5ee335bbf943" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 931.699149] env[63515]: DEBUG oslo_vmware.api [None req-56611604-1203-4b20-9871-2ec521700550 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Task: {'id': task-1111607, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.852133] env[63515]: DEBUG oslo_vmware.api [None req-1d5e76d4-1b01-4bdb-831a-7cba26a22dfc tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Task: {'id': task-1111606, 'name': ReconfigVM_Task, 'duration_secs': 0.444675} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 931.852133] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-1d5e76d4-1b01-4bdb-831a-7cba26a22dfc tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: 829e72d7-a56a-451c-b98a-94253e5900d5] Reconfigured VM instance instance-00000050 to attach disk [datastore1] 829e72d7-a56a-451c-b98a-94253e5900d5/829e72d7-a56a-451c-b98a-94253e5900d5.vmdk or device None with type thin {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 931.852133] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-1d5e76d4-1b01-4bdb-831a-7cba26a22dfc tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: 829e72d7-a56a-451c-b98a-94253e5900d5] Updating instance '829e72d7-a56a-451c-b98a-94253e5900d5' progress to 50 {{(pid=63515) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 932.011929] env[63515]: DEBUG nova.scheduler.client.report [None req-be7cccf6-3f5b-4895-b678-cf60de2c827a tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 932.044025] env[63515]: DEBUG oslo_vmware.api [None req-d63820c0-6e1a-45a6-8579-7bfab43bd25a tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52bce7be-78a7-59ca-3440-b08008d5063c, 'name': SearchDatastore_Task, 'duration_secs': 0.010117} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 932.044344] env[63515]: DEBUG oslo_concurrency.lockutils [None req-d63820c0-6e1a-45a6-8579-7bfab43bd25a tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 932.044637] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-d63820c0-6e1a-45a6-8579-7bfab43bd25a tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk to [datastore2] 2fa098b6-ecc0-491f-8cfc-5ee335bbf943/2fa098b6-ecc0-491f-8cfc-5ee335bbf943.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 932.044931] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5539c314-4a10-477c-804b-02fc3c6bdd24 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.051484] env[63515]: DEBUG oslo_vmware.api [None req-d63820c0-6e1a-45a6-8579-7bfab43bd25a tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Waiting for the task: (returnval){ [ 932.051484] env[63515]: value = "task-1111608" [ 932.051484] env[63515]: _type = "Task" [ 932.051484] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 932.059667] env[63515]: DEBUG oslo_vmware.api [None req-d63820c0-6e1a-45a6-8579-7bfab43bd25a tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Task: {'id': task-1111608, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.202313] env[63515]: DEBUG oslo_vmware.api [None req-56611604-1203-4b20-9871-2ec521700550 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Task: {'id': task-1111607, 'name': PowerOffVM_Task, 'duration_secs': 0.268281} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 932.202621] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-56611604-1203-4b20-9871-2ec521700550 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: a2203c5e-66b9-4c7e-8d76-e12a08b57b1d] Powered off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 932.202793] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-56611604-1203-4b20-9871-2ec521700550 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: a2203c5e-66b9-4c7e-8d76-e12a08b57b1d] Unregistering the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 932.203091] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ad428cd9-c2bf-4e6e-b5b4-ed8a17745506 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.217537] env[63515]: DEBUG oslo_concurrency.lockutils [None req-8fdfb203-5786-4efd-8927-4ee465e68e44 tempest-ServerMetadataTestJSON-1034558863 tempest-ServerMetadataTestJSON-1034558863-project-member] Acquiring lock "b25bf2ba-cf08-44ac-a524-aa87cf785575" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 932.217791] env[63515]: DEBUG oslo_concurrency.lockutils [None req-8fdfb203-5786-4efd-8927-4ee465e68e44 tempest-ServerMetadataTestJSON-1034558863 tempest-ServerMetadataTestJSON-1034558863-project-member] Lock "b25bf2ba-cf08-44ac-a524-aa87cf785575" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 932.284978] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-56611604-1203-4b20-9871-2ec521700550 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: a2203c5e-66b9-4c7e-8d76-e12a08b57b1d] Unregistered the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 932.285339] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-56611604-1203-4b20-9871-2ec521700550 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: a2203c5e-66b9-4c7e-8d76-e12a08b57b1d] Deleting contents of the VM from datastore datastore2 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 932.285603] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-56611604-1203-4b20-9871-2ec521700550 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Deleting the datastore file [datastore2] a2203c5e-66b9-4c7e-8d76-e12a08b57b1d {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 932.285924] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-aebd17f9-054f-409c-ac98-208a38a55336 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.292936] env[63515]: DEBUG oslo_vmware.api [None req-56611604-1203-4b20-9871-2ec521700550 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Waiting for the task: (returnval){ [ 932.292936] env[63515]: value = "task-1111610" [ 932.292936] env[63515]: _type = "Task" [ 932.292936] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 932.302356] env[63515]: DEBUG oslo_vmware.api [None req-56611604-1203-4b20-9871-2ec521700550 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Task: {'id': task-1111610, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.363647] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de73e3c7-2f94-4338-8795-aa1b746d8dde {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.389142] env[63515]: DEBUG nova.network.neutron [None req-bd214878-de3c-49e1-b706-bcb83d66311c tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: eab16df9-7bb5-4576-bca0-769a561c5fe9] Successfully updated port: ff4043f6-7576-4093-80ed-d36240b37f2f {{(pid=63515) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 932.390981] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e3246a7-7814-44ab-b147-7268efa2c973 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.412908] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-1d5e76d4-1b01-4bdb-831a-7cba26a22dfc tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: 829e72d7-a56a-451c-b98a-94253e5900d5] Updating instance '829e72d7-a56a-451c-b98a-94253e5900d5' progress to 67 {{(pid=63515) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 932.517310] env[63515]: DEBUG oslo_concurrency.lockutils [None req-be7cccf6-3f5b-4895-b678-cf60de2c827a tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.348s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 932.518235] env[63515]: DEBUG nova.compute.manager [None req-be7cccf6-3f5b-4895-b678-cf60de2c827a tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: 014720fe-87b9-4e55-b9ba-295b009c7ff6] Start building networks asynchronously for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 932.521287] env[63515]: DEBUG oslo_concurrency.lockutils [None req-58ca2ad0-b12b-44be-a72b-6a6ba3bc88d1 tempest-ServerActionsV293TestJSON-1193370169 tempest-ServerActionsV293TestJSON-1193370169-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 6.225s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 932.521556] env[63515]: DEBUG oslo_concurrency.lockutils [None req-58ca2ad0-b12b-44be-a72b-6a6ba3bc88d1 tempest-ServerActionsV293TestJSON-1193370169 tempest-ServerActionsV293TestJSON-1193370169-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 932.526444] env[63515]: DEBUG oslo_concurrency.lockutils [None req-f7e007a5-60f7-44d2-a529-05d489daa7aa tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.682s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 932.526444] env[63515]: DEBUG nova.objects.instance [None req-f7e007a5-60f7-44d2-a529-05d489daa7aa tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Lazy-loading 'resources' on Instance uuid 9236607d-47af-4272-93db-79826e918aa1 {{(pid=63515) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 932.564451] env[63515]: DEBUG oslo_vmware.api [None req-d63820c0-6e1a-45a6-8579-7bfab43bd25a tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Task: {'id': task-1111608, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.468871} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 932.564771] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-d63820c0-6e1a-45a6-8579-7bfab43bd25a tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk to [datastore2] 2fa098b6-ecc0-491f-8cfc-5ee335bbf943/2fa098b6-ecc0-491f-8cfc-5ee335bbf943.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 932.564995] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-d63820c0-6e1a-45a6-8579-7bfab43bd25a tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: 2fa098b6-ecc0-491f-8cfc-5ee335bbf943] Extending root virtual disk to 1048576 {{(pid=63515) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 932.565295] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c8b9392a-8937-4e22-a968-9254daebaf6d {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.572752] env[63515]: DEBUG oslo_vmware.api [None req-d63820c0-6e1a-45a6-8579-7bfab43bd25a tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Waiting for the task: (returnval){ [ 932.572752] env[63515]: value = "task-1111611" [ 932.572752] env[63515]: _type = "Task" [ 932.572752] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 932.582460] env[63515]: DEBUG oslo_vmware.api [None req-d63820c0-6e1a-45a6-8579-7bfab43bd25a tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Task: {'id': task-1111611, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.720628] env[63515]: DEBUG nova.compute.manager [None req-8fdfb203-5786-4efd-8927-4ee465e68e44 tempest-ServerMetadataTestJSON-1034558863 tempest-ServerMetadataTestJSON-1034558863-project-member] [instance: b25bf2ba-cf08-44ac-a524-aa87cf785575] Starting instance... {{(pid=63515) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 932.791205] env[63515]: DEBUG nova.compute.manager [req-01aa6168-0e97-4f98-a197-cc7496ce58f2 req-d7c0bfee-5845-477d-80fe-001a05d1a8dd service nova] [instance: eab16df9-7bb5-4576-bca0-769a561c5fe9] Received event network-vif-plugged-ff4043f6-7576-4093-80ed-d36240b37f2f {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 932.791458] env[63515]: DEBUG oslo_concurrency.lockutils [req-01aa6168-0e97-4f98-a197-cc7496ce58f2 req-d7c0bfee-5845-477d-80fe-001a05d1a8dd service nova] Acquiring lock "eab16df9-7bb5-4576-bca0-769a561c5fe9-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 932.791674] env[63515]: DEBUG oslo_concurrency.lockutils [req-01aa6168-0e97-4f98-a197-cc7496ce58f2 req-d7c0bfee-5845-477d-80fe-001a05d1a8dd service nova] Lock "eab16df9-7bb5-4576-bca0-769a561c5fe9-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 932.791898] env[63515]: DEBUG oslo_concurrency.lockutils [req-01aa6168-0e97-4f98-a197-cc7496ce58f2 req-d7c0bfee-5845-477d-80fe-001a05d1a8dd service nova] Lock "eab16df9-7bb5-4576-bca0-769a561c5fe9-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 932.795410] env[63515]: DEBUG nova.compute.manager [req-01aa6168-0e97-4f98-a197-cc7496ce58f2 req-d7c0bfee-5845-477d-80fe-001a05d1a8dd service nova] [instance: eab16df9-7bb5-4576-bca0-769a561c5fe9] No waiting events found dispatching network-vif-plugged-ff4043f6-7576-4093-80ed-d36240b37f2f {{(pid=63515) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 932.795682] env[63515]: WARNING nova.compute.manager [req-01aa6168-0e97-4f98-a197-cc7496ce58f2 req-d7c0bfee-5845-477d-80fe-001a05d1a8dd service nova] [instance: eab16df9-7bb5-4576-bca0-769a561c5fe9] Received unexpected event network-vif-plugged-ff4043f6-7576-4093-80ed-d36240b37f2f for instance with vm_state active and task_state None. [ 932.795936] env[63515]: DEBUG nova.compute.manager [req-01aa6168-0e97-4f98-a197-cc7496ce58f2 req-d7c0bfee-5845-477d-80fe-001a05d1a8dd service nova] [instance: eab16df9-7bb5-4576-bca0-769a561c5fe9] Received event network-changed-ff4043f6-7576-4093-80ed-d36240b37f2f {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 932.796194] env[63515]: DEBUG nova.compute.manager [req-01aa6168-0e97-4f98-a197-cc7496ce58f2 req-d7c0bfee-5845-477d-80fe-001a05d1a8dd service nova] [instance: eab16df9-7bb5-4576-bca0-769a561c5fe9] Refreshing instance network info cache due to event network-changed-ff4043f6-7576-4093-80ed-d36240b37f2f. {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 932.796504] env[63515]: DEBUG oslo_concurrency.lockutils [req-01aa6168-0e97-4f98-a197-cc7496ce58f2 req-d7c0bfee-5845-477d-80fe-001a05d1a8dd service nova] Acquiring lock "refresh_cache-eab16df9-7bb5-4576-bca0-769a561c5fe9" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 932.796717] env[63515]: DEBUG oslo_concurrency.lockutils [req-01aa6168-0e97-4f98-a197-cc7496ce58f2 req-d7c0bfee-5845-477d-80fe-001a05d1a8dd service nova] Acquired lock "refresh_cache-eab16df9-7bb5-4576-bca0-769a561c5fe9" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 932.797269] env[63515]: DEBUG nova.network.neutron [req-01aa6168-0e97-4f98-a197-cc7496ce58f2 req-d7c0bfee-5845-477d-80fe-001a05d1a8dd service nova] [instance: eab16df9-7bb5-4576-bca0-769a561c5fe9] Refreshing network info cache for port ff4043f6-7576-4093-80ed-d36240b37f2f {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 932.811845] env[63515]: DEBUG oslo_vmware.api [None req-56611604-1203-4b20-9871-2ec521700550 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Task: {'id': task-1111610, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.350689} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 932.812799] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-56611604-1203-4b20-9871-2ec521700550 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Deleted the datastore file {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 932.812972] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-56611604-1203-4b20-9871-2ec521700550 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: a2203c5e-66b9-4c7e-8d76-e12a08b57b1d] Deleted contents of the VM from datastore datastore2 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 932.813106] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-56611604-1203-4b20-9871-2ec521700550 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: a2203c5e-66b9-4c7e-8d76-e12a08b57b1d] Instance destroyed {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 932.813291] env[63515]: INFO nova.compute.manager [None req-56611604-1203-4b20-9871-2ec521700550 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: a2203c5e-66b9-4c7e-8d76-e12a08b57b1d] Took 1.14 seconds to destroy the instance on the hypervisor. [ 932.813683] env[63515]: DEBUG oslo.service.loopingcall [None req-56611604-1203-4b20-9871-2ec521700550 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 932.813823] env[63515]: DEBUG nova.compute.manager [-] [instance: a2203c5e-66b9-4c7e-8d76-e12a08b57b1d] Deallocating network for instance {{(pid=63515) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 932.813823] env[63515]: DEBUG nova.network.neutron [-] [instance: a2203c5e-66b9-4c7e-8d76-e12a08b57b1d] deallocate_for_instance() {{(pid=63515) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 932.896280] env[63515]: DEBUG oslo_concurrency.lockutils [None req-bd214878-de3c-49e1-b706-bcb83d66311c tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Acquiring lock "refresh_cache-eab16df9-7bb5-4576-bca0-769a561c5fe9" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 933.020628] env[63515]: DEBUG nova.network.neutron [None req-1d5e76d4-1b01-4bdb-831a-7cba26a22dfc tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: 829e72d7-a56a-451c-b98a-94253e5900d5] Port 856de677-b1c8-479b-91d9-34d31a703a63 binding to destination host cpu-1 is already ACTIVE {{(pid=63515) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 933.030880] env[63515]: DEBUG nova.compute.utils [None req-be7cccf6-3f5b-4895-b678-cf60de2c827a tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Using /dev/sd instead of None {{(pid=63515) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 933.037667] env[63515]: DEBUG nova.compute.manager [None req-be7cccf6-3f5b-4895-b678-cf60de2c827a tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: 014720fe-87b9-4e55-b9ba-295b009c7ff6] Allocating IP information in the background. {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 933.037667] env[63515]: DEBUG nova.network.neutron [None req-be7cccf6-3f5b-4895-b678-cf60de2c827a tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: 014720fe-87b9-4e55-b9ba-295b009c7ff6] allocate_for_instance() {{(pid=63515) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 933.049234] env[63515]: DEBUG oslo_concurrency.lockutils [None req-58ca2ad0-b12b-44be-a72b-6a6ba3bc88d1 tempest-ServerActionsV293TestJSON-1193370169 tempest-ServerActionsV293TestJSON-1193370169-project-member] Lock "2ea99c7a-8b61-4718-bee0-f4ce4fc021af" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 9.624s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 933.091405] env[63515]: DEBUG oslo_vmware.api [None req-d63820c0-6e1a-45a6-8579-7bfab43bd25a tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Task: {'id': task-1111611, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.107579} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 933.091887] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-d63820c0-6e1a-45a6-8579-7bfab43bd25a tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: 2fa098b6-ecc0-491f-8cfc-5ee335bbf943] Extended root virtual disk {{(pid=63515) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 933.093132] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35e7fa31-9a91-471d-884f-b96a612ea02b {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.123458] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-d63820c0-6e1a-45a6-8579-7bfab43bd25a tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: 2fa098b6-ecc0-491f-8cfc-5ee335bbf943] Reconfiguring VM instance instance-00000055 to attach disk [datastore2] 2fa098b6-ecc0-491f-8cfc-5ee335bbf943/2fa098b6-ecc0-491f-8cfc-5ee335bbf943.vmdk or device None with type sparse {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 933.126765] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c02be666-ab2e-4525-ab8e-4dfcbfd3ffb3 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.146711] env[63515]: DEBUG nova.compute.manager [req-98df7774-dee6-4990-a782-fffdc2cbe48c req-4546f3d7-198b-49ee-8b11-58d6da0d9763 service nova] [instance: 2a29f37d-e786-48d3-b126-4854d6e7c67c] Received event network-changed-956cd361-0d42-4914-85d3-5f75a72a68a1 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 933.146794] env[63515]: DEBUG nova.compute.manager [req-98df7774-dee6-4990-a782-fffdc2cbe48c req-4546f3d7-198b-49ee-8b11-58d6da0d9763 service nova] [instance: 2a29f37d-e786-48d3-b126-4854d6e7c67c] Refreshing instance network info cache due to event network-changed-956cd361-0d42-4914-85d3-5f75a72a68a1. {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 933.148526] env[63515]: DEBUG oslo_concurrency.lockutils [req-98df7774-dee6-4990-a782-fffdc2cbe48c req-4546f3d7-198b-49ee-8b11-58d6da0d9763 service nova] Acquiring lock "refresh_cache-2a29f37d-e786-48d3-b126-4854d6e7c67c" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 933.148526] env[63515]: DEBUG oslo_concurrency.lockutils [req-98df7774-dee6-4990-a782-fffdc2cbe48c req-4546f3d7-198b-49ee-8b11-58d6da0d9763 service nova] Acquired lock "refresh_cache-2a29f37d-e786-48d3-b126-4854d6e7c67c" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 933.148526] env[63515]: DEBUG nova.network.neutron [req-98df7774-dee6-4990-a782-fffdc2cbe48c req-4546f3d7-198b-49ee-8b11-58d6da0d9763 service nova] [instance: 2a29f37d-e786-48d3-b126-4854d6e7c67c] Refreshing network info cache for port 956cd361-0d42-4914-85d3-5f75a72a68a1 {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 933.150604] env[63515]: DEBUG oslo_vmware.api [None req-d63820c0-6e1a-45a6-8579-7bfab43bd25a tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Waiting for the task: (returnval){ [ 933.150604] env[63515]: value = "task-1111612" [ 933.150604] env[63515]: _type = "Task" [ 933.150604] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 933.152114] env[63515]: DEBUG nova.policy [None req-be7cccf6-3f5b-4895-b678-cf60de2c827a tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f58ac812970845dc825202b842feb5d0', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ea099ce07bfb4a8da014d9303cf552eb', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63515) authorize /opt/stack/nova/nova/policy.py:201}} [ 933.165495] env[63515]: DEBUG oslo_vmware.api [None req-d63820c0-6e1a-45a6-8579-7bfab43bd25a tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Task: {'id': task-1111612, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.247991] env[63515]: DEBUG oslo_concurrency.lockutils [None req-8fdfb203-5786-4efd-8927-4ee465e68e44 tempest-ServerMetadataTestJSON-1034558863 tempest-ServerMetadataTestJSON-1034558863-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 933.373267] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc433cd4-307c-4419-8000-f22d2d42c195 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.384666] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5a5d518-5cd6-4119-a6a4-01ada8fbd611 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.424369] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c6834cf-3dfa-40a5-ad9c-8549c8441d60 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.432691] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b33e1106-14d7-4c2e-adee-397219138f40 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.451241] env[63515]: DEBUG nova.compute.provider_tree [None req-f7e007a5-60f7-44d2-a529-05d489daa7aa tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Updating inventory in ProviderTree for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 933.538497] env[63515]: DEBUG nova.compute.manager [None req-be7cccf6-3f5b-4895-b678-cf60de2c827a tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: 014720fe-87b9-4e55-b9ba-295b009c7ff6] Start building block device mappings for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 933.666047] env[63515]: DEBUG oslo_vmware.api [None req-d63820c0-6e1a-45a6-8579-7bfab43bd25a tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Task: {'id': task-1111612, 'name': ReconfigVM_Task} progress is 99%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.688027] env[63515]: DEBUG nova.network.neutron [None req-be7cccf6-3f5b-4895-b678-cf60de2c827a tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: 014720fe-87b9-4e55-b9ba-295b009c7ff6] Successfully created port: 406b4b98-709d-4013-9ce0-c4232abb4b1d {{(pid=63515) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 933.718419] env[63515]: DEBUG nova.network.neutron [req-01aa6168-0e97-4f98-a197-cc7496ce58f2 req-d7c0bfee-5845-477d-80fe-001a05d1a8dd service nova] [instance: eab16df9-7bb5-4576-bca0-769a561c5fe9] Added VIF to instance network info cache for port ff4043f6-7576-4093-80ed-d36240b37f2f. {{(pid=63515) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3489}} [ 933.719023] env[63515]: DEBUG nova.network.neutron [req-01aa6168-0e97-4f98-a197-cc7496ce58f2 req-d7c0bfee-5845-477d-80fe-001a05d1a8dd service nova] [instance: eab16df9-7bb5-4576-bca0-769a561c5fe9] Updating instance_info_cache with network_info: [{"id": "4168f5a5-c282-4f13-9c97-0b0236bd659b", "address": "fa:16:3e:a0:6a:50", "network": {"id": "88dfb29d-41e0-4df7-8f70-e52cda05620c", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1355578400-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.203", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7d4ada9ec35f42b19c6480a9101d21a4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "357d2811-e990-4985-9f9e-b158d10d3699", "external-id": "nsx-vlan-transportzone-641", "segmentation_id": 641, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4168f5a5-c2", "ovs_interfaceid": "4168f5a5-c282-4f13-9c97-0b0236bd659b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "e3c72582-41bf-48d4-8614-b516a96362a7", "address": "fa:16:3e:c0:fd:b9", "network": {"id": "88dfb29d-41e0-4df7-8f70-e52cda05620c", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1355578400-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7d4ada9ec35f42b19c6480a9101d21a4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "357d2811-e990-4985-9f9e-b158d10d3699", "external-id": "nsx-vlan-transportzone-641", "segmentation_id": 641, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape3c72582-41", "ovs_interfaceid": "e3c72582-41bf-48d4-8614-b516a96362a7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "ff4043f6-7576-4093-80ed-d36240b37f2f", "address": "fa:16:3e:47:af:a2", "network": {"id": "88dfb29d-41e0-4df7-8f70-e52cda05620c", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1355578400-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7d4ada9ec35f42b19c6480a9101d21a4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "357d2811-e990-4985-9f9e-b158d10d3699", "external-id": "nsx-vlan-transportzone-641", "segmentation_id": 641, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapff4043f6-75", "ovs_interfaceid": "ff4043f6-7576-4093-80ed-d36240b37f2f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 933.776447] env[63515]: DEBUG nova.network.neutron [-] [instance: a2203c5e-66b9-4c7e-8d76-e12a08b57b1d] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 933.974272] env[63515]: ERROR nova.scheduler.client.report [None req-f7e007a5-60f7-44d2-a529-05d489daa7aa tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] [req-74928ab1-d3f6-4629-91b3-6986614b5963] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 2bd86232-2b6f-44d5-9057-1a3a6b27185a. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-74928ab1-d3f6-4629-91b3-6986614b5963"}]} [ 933.993590] env[63515]: DEBUG nova.scheduler.client.report [None req-f7e007a5-60f7-44d2-a529-05d489daa7aa tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Refreshing inventories for resource provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 934.010368] env[63515]: DEBUG nova.scheduler.client.report [None req-f7e007a5-60f7-44d2-a529-05d489daa7aa tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Updating ProviderTree inventory for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 934.010608] env[63515]: DEBUG nova.compute.provider_tree [None req-f7e007a5-60f7-44d2-a529-05d489daa7aa tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Updating inventory in ProviderTree for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 934.013221] env[63515]: DEBUG nova.network.neutron [req-98df7774-dee6-4990-a782-fffdc2cbe48c req-4546f3d7-198b-49ee-8b11-58d6da0d9763 service nova] [instance: 2a29f37d-e786-48d3-b126-4854d6e7c67c] Updated VIF entry in instance network info cache for port 956cd361-0d42-4914-85d3-5f75a72a68a1. {{(pid=63515) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 934.013560] env[63515]: DEBUG nova.network.neutron [req-98df7774-dee6-4990-a782-fffdc2cbe48c req-4546f3d7-198b-49ee-8b11-58d6da0d9763 service nova] [instance: 2a29f37d-e786-48d3-b126-4854d6e7c67c] Updating instance_info_cache with network_info: [{"id": "956cd361-0d42-4914-85d3-5f75a72a68a1", "address": "fa:16:3e:a3:72:bb", "network": {"id": "850bfa95-3df6-4019-8260-ade20ec58799", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-832000501-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.153", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "08c9aa51b1864410809f8877c7cd49c4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3cf748a8-7ae0-4dca-817d-e727c30d72f4", "external-id": "nsx-vlan-transportzone-853", "segmentation_id": 853, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap956cd361-0d", "ovs_interfaceid": "956cd361-0d42-4914-85d3-5f75a72a68a1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 934.025321] env[63515]: DEBUG nova.scheduler.client.report [None req-f7e007a5-60f7-44d2-a529-05d489daa7aa tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Refreshing aggregate associations for resource provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a, aggregates: None {{(pid=63515) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 934.052840] env[63515]: DEBUG oslo_concurrency.lockutils [None req-1d5e76d4-1b01-4bdb-831a-7cba26a22dfc tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Acquiring lock "829e72d7-a56a-451c-b98a-94253e5900d5-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 934.053100] env[63515]: DEBUG oslo_concurrency.lockutils [None req-1d5e76d4-1b01-4bdb-831a-7cba26a22dfc tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Lock "829e72d7-a56a-451c-b98a-94253e5900d5-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 934.053283] env[63515]: DEBUG oslo_concurrency.lockutils [None req-1d5e76d4-1b01-4bdb-831a-7cba26a22dfc tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Lock "829e72d7-a56a-451c-b98a-94253e5900d5-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 934.059095] env[63515]: DEBUG nova.scheduler.client.report [None req-f7e007a5-60f7-44d2-a529-05d489daa7aa tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Refreshing trait associations for resource provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a, traits: COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=63515) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 934.169247] env[63515]: DEBUG oslo_vmware.api [None req-d63820c0-6e1a-45a6-8579-7bfab43bd25a tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Task: {'id': task-1111612, 'name': ReconfigVM_Task} progress is 99%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.222544] env[63515]: DEBUG oslo_concurrency.lockutils [req-01aa6168-0e97-4f98-a197-cc7496ce58f2 req-d7c0bfee-5845-477d-80fe-001a05d1a8dd service nova] Releasing lock "refresh_cache-eab16df9-7bb5-4576-bca0-769a561c5fe9" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 934.222979] env[63515]: DEBUG oslo_concurrency.lockutils [None req-bd214878-de3c-49e1-b706-bcb83d66311c tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Acquired lock "refresh_cache-eab16df9-7bb5-4576-bca0-769a561c5fe9" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 934.223181] env[63515]: DEBUG nova.network.neutron [None req-bd214878-de3c-49e1-b706-bcb83d66311c tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: eab16df9-7bb5-4576-bca0-769a561c5fe9] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 934.280664] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da0c6bf4-72ce-4037-aff4-987a4481125e {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.285289] env[63515]: INFO nova.compute.manager [-] [instance: a2203c5e-66b9-4c7e-8d76-e12a08b57b1d] Took 1.47 seconds to deallocate network for instance. [ 934.298050] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2880ce05-0a2f-4bee-ae4b-a49f5cee157f {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.334175] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a5993a8-35d0-4f2f-863c-e01439c255c1 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.342993] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b0ddfae-6db1-49c3-9f18-541bfb717350 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.357803] env[63515]: DEBUG nova.compute.provider_tree [None req-f7e007a5-60f7-44d2-a529-05d489daa7aa tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Updating inventory in ProviderTree for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 934.518799] env[63515]: DEBUG oslo_concurrency.lockutils [req-98df7774-dee6-4990-a782-fffdc2cbe48c req-4546f3d7-198b-49ee-8b11-58d6da0d9763 service nova] Releasing lock "refresh_cache-2a29f37d-e786-48d3-b126-4854d6e7c67c" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 934.554513] env[63515]: DEBUG nova.compute.manager [None req-be7cccf6-3f5b-4895-b678-cf60de2c827a tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: 014720fe-87b9-4e55-b9ba-295b009c7ff6] Start spawning the instance on the hypervisor. {{(pid=63515) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 934.584453] env[63515]: DEBUG nova.virt.hardware [None req-be7cccf6-3f5b-4895-b678-cf60de2c827a tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T02:52:14Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T02:51:56Z,direct_url=,disk_format='vmdk',id=8a120570-cb06-4099-b262-554ca0ad15c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b569255cc43e42e7ae3f2b4ad37c6ef4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T02:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 934.585399] env[63515]: DEBUG nova.virt.hardware [None req-be7cccf6-3f5b-4895-b678-cf60de2c827a tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Flavor limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 934.585828] env[63515]: DEBUG nova.virt.hardware [None req-be7cccf6-3f5b-4895-b678-cf60de2c827a tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Image limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 934.586285] env[63515]: DEBUG nova.virt.hardware [None req-be7cccf6-3f5b-4895-b678-cf60de2c827a tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Flavor pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 934.586662] env[63515]: DEBUG nova.virt.hardware [None req-be7cccf6-3f5b-4895-b678-cf60de2c827a tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Image pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 934.587022] env[63515]: DEBUG nova.virt.hardware [None req-be7cccf6-3f5b-4895-b678-cf60de2c827a tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 934.587461] env[63515]: DEBUG nova.virt.hardware [None req-be7cccf6-3f5b-4895-b678-cf60de2c827a tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 934.587850] env[63515]: DEBUG nova.virt.hardware [None req-be7cccf6-3f5b-4895-b678-cf60de2c827a tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 934.588433] env[63515]: DEBUG nova.virt.hardware [None req-be7cccf6-3f5b-4895-b678-cf60de2c827a tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Got 1 possible topologies {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 934.589717] env[63515]: DEBUG nova.virt.hardware [None req-be7cccf6-3f5b-4895-b678-cf60de2c827a tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 934.589717] env[63515]: DEBUG nova.virt.hardware [None req-be7cccf6-3f5b-4895-b678-cf60de2c827a tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 934.591915] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1abdec5-95c0-42e6-9e98-041baa3e0c64 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.602376] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d15533f-cc80-4705-9044-96fa1da674c5 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.666651] env[63515]: DEBUG oslo_vmware.api [None req-d63820c0-6e1a-45a6-8579-7bfab43bd25a tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Task: {'id': task-1111612, 'name': ReconfigVM_Task, 'duration_secs': 1.296795} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 934.666953] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-d63820c0-6e1a-45a6-8579-7bfab43bd25a tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: 2fa098b6-ecc0-491f-8cfc-5ee335bbf943] Reconfigured VM instance instance-00000055 to attach disk [datastore2] 2fa098b6-ecc0-491f-8cfc-5ee335bbf943/2fa098b6-ecc0-491f-8cfc-5ee335bbf943.vmdk or device None with type sparse {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 934.667682] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-be294b2b-f8be-43f7-bb36-33a1fb345307 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.673861] env[63515]: DEBUG oslo_vmware.api [None req-d63820c0-6e1a-45a6-8579-7bfab43bd25a tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Waiting for the task: (returnval){ [ 934.673861] env[63515]: value = "task-1111613" [ 934.673861] env[63515]: _type = "Task" [ 934.673861] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 934.682256] env[63515]: DEBUG oslo_vmware.api [None req-d63820c0-6e1a-45a6-8579-7bfab43bd25a tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Task: {'id': task-1111613, 'name': Rename_Task} progress is 5%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.734515] env[63515]: DEBUG nova.network.neutron [None req-be7cccf6-3f5b-4895-b678-cf60de2c827a tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: 014720fe-87b9-4e55-b9ba-295b009c7ff6] Successfully updated port: 406b4b98-709d-4013-9ce0-c4232abb4b1d {{(pid=63515) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 934.780711] env[63515]: WARNING nova.network.neutron [None req-bd214878-de3c-49e1-b706-bcb83d66311c tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: eab16df9-7bb5-4576-bca0-769a561c5fe9] 88dfb29d-41e0-4df7-8f70-e52cda05620c already exists in list: networks containing: ['88dfb29d-41e0-4df7-8f70-e52cda05620c']. ignoring it [ 934.780950] env[63515]: WARNING nova.network.neutron [None req-bd214878-de3c-49e1-b706-bcb83d66311c tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: eab16df9-7bb5-4576-bca0-769a561c5fe9] 88dfb29d-41e0-4df7-8f70-e52cda05620c already exists in list: networks containing: ['88dfb29d-41e0-4df7-8f70-e52cda05620c']. ignoring it [ 934.781187] env[63515]: WARNING nova.network.neutron [None req-bd214878-de3c-49e1-b706-bcb83d66311c tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: eab16df9-7bb5-4576-bca0-769a561c5fe9] 88dfb29d-41e0-4df7-8f70-e52cda05620c already exists in list: networks containing: ['88dfb29d-41e0-4df7-8f70-e52cda05620c']. ignoring it [ 934.781522] env[63515]: WARNING nova.network.neutron [None req-bd214878-de3c-49e1-b706-bcb83d66311c tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: eab16df9-7bb5-4576-bca0-769a561c5fe9] ff4043f6-7576-4093-80ed-d36240b37f2f already exists in list: port_ids containing: ['ff4043f6-7576-4093-80ed-d36240b37f2f']. ignoring it [ 934.796009] env[63515]: DEBUG oslo_concurrency.lockutils [None req-56611604-1203-4b20-9871-2ec521700550 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 934.878812] env[63515]: ERROR nova.scheduler.client.report [None req-f7e007a5-60f7-44d2-a529-05d489daa7aa tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] [req-cf881709-75ab-4c3f-b1e2-215bb4aa3d1a] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 2bd86232-2b6f-44d5-9057-1a3a6b27185a. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-cf881709-75ab-4c3f-b1e2-215bb4aa3d1a"}]} [ 934.898457] env[63515]: DEBUG nova.scheduler.client.report [None req-f7e007a5-60f7-44d2-a529-05d489daa7aa tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Refreshing inventories for resource provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 934.917761] env[63515]: DEBUG nova.scheduler.client.report [None req-f7e007a5-60f7-44d2-a529-05d489daa7aa tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Updating ProviderTree inventory for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 934.918330] env[63515]: DEBUG nova.compute.provider_tree [None req-f7e007a5-60f7-44d2-a529-05d489daa7aa tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Updating inventory in ProviderTree for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 934.931253] env[63515]: DEBUG nova.scheduler.client.report [None req-f7e007a5-60f7-44d2-a529-05d489daa7aa tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Refreshing aggregate associations for resource provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a, aggregates: None {{(pid=63515) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 934.933636] env[63515]: DEBUG oslo_concurrency.lockutils [None req-127af73c-afc8-4436-9075-a886f5e9c7d5 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Acquiring lock "7d572ac0-f6af-4622-96cc-e75983420222" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 934.933937] env[63515]: DEBUG oslo_concurrency.lockutils [None req-127af73c-afc8-4436-9075-a886f5e9c7d5 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Lock "7d572ac0-f6af-4622-96cc-e75983420222" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 934.961758] env[63515]: DEBUG nova.scheduler.client.report [None req-f7e007a5-60f7-44d2-a529-05d489daa7aa tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Refreshing trait associations for resource provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a, traits: COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=63515) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 935.096152] env[63515]: DEBUG oslo_concurrency.lockutils [None req-1d5e76d4-1b01-4bdb-831a-7cba26a22dfc tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Acquiring lock "refresh_cache-829e72d7-a56a-451c-b98a-94253e5900d5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 935.096376] env[63515]: DEBUG oslo_concurrency.lockutils [None req-1d5e76d4-1b01-4bdb-831a-7cba26a22dfc tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Acquired lock "refresh_cache-829e72d7-a56a-451c-b98a-94253e5900d5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 935.096561] env[63515]: DEBUG nova.network.neutron [None req-1d5e76d4-1b01-4bdb-831a-7cba26a22dfc tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: 829e72d7-a56a-451c-b98a-94253e5900d5] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 935.198132] env[63515]: DEBUG oslo_vmware.api [None req-d63820c0-6e1a-45a6-8579-7bfab43bd25a tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Task: {'id': task-1111613, 'name': Rename_Task, 'duration_secs': 0.148419} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 935.204919] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-d63820c0-6e1a-45a6-8579-7bfab43bd25a tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: 2fa098b6-ecc0-491f-8cfc-5ee335bbf943] Powering on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 935.205283] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6e194d6f-98b0-4a8a-a6a0-5a110457a909 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.215462] env[63515]: DEBUG nova.compute.manager [req-51bfb957-3fd2-4eaf-aa8c-c6a959a160ea req-d7fa06c5-6ec8-40af-b8a8-0f8dd88b70ca service nova] [instance: a2203c5e-66b9-4c7e-8d76-e12a08b57b1d] Received event network-vif-deleted-8b878aac-8955-4468-893f-63fdbfc0c8cb {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 935.215620] env[63515]: DEBUG nova.compute.manager [req-51bfb957-3fd2-4eaf-aa8c-c6a959a160ea req-d7fa06c5-6ec8-40af-b8a8-0f8dd88b70ca service nova] [instance: 014720fe-87b9-4e55-b9ba-295b009c7ff6] Received event network-vif-plugged-406b4b98-709d-4013-9ce0-c4232abb4b1d {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 935.215840] env[63515]: DEBUG oslo_concurrency.lockutils [req-51bfb957-3fd2-4eaf-aa8c-c6a959a160ea req-d7fa06c5-6ec8-40af-b8a8-0f8dd88b70ca service nova] Acquiring lock "014720fe-87b9-4e55-b9ba-295b009c7ff6-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 935.216055] env[63515]: DEBUG oslo_concurrency.lockutils [req-51bfb957-3fd2-4eaf-aa8c-c6a959a160ea req-d7fa06c5-6ec8-40af-b8a8-0f8dd88b70ca service nova] Lock "014720fe-87b9-4e55-b9ba-295b009c7ff6-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 935.216238] env[63515]: DEBUG oslo_concurrency.lockutils [req-51bfb957-3fd2-4eaf-aa8c-c6a959a160ea req-d7fa06c5-6ec8-40af-b8a8-0f8dd88b70ca service nova] Lock "014720fe-87b9-4e55-b9ba-295b009c7ff6-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 935.216796] env[63515]: DEBUG nova.compute.manager [req-51bfb957-3fd2-4eaf-aa8c-c6a959a160ea req-d7fa06c5-6ec8-40af-b8a8-0f8dd88b70ca service nova] [instance: 014720fe-87b9-4e55-b9ba-295b009c7ff6] No waiting events found dispatching network-vif-plugged-406b4b98-709d-4013-9ce0-c4232abb4b1d {{(pid=63515) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 935.216796] env[63515]: WARNING nova.compute.manager [req-51bfb957-3fd2-4eaf-aa8c-c6a959a160ea req-d7fa06c5-6ec8-40af-b8a8-0f8dd88b70ca service nova] [instance: 014720fe-87b9-4e55-b9ba-295b009c7ff6] Received unexpected event network-vif-plugged-406b4b98-709d-4013-9ce0-c4232abb4b1d for instance with vm_state building and task_state spawning. [ 935.216796] env[63515]: DEBUG nova.compute.manager [req-51bfb957-3fd2-4eaf-aa8c-c6a959a160ea req-d7fa06c5-6ec8-40af-b8a8-0f8dd88b70ca service nova] [instance: 014720fe-87b9-4e55-b9ba-295b009c7ff6] Received event network-changed-406b4b98-709d-4013-9ce0-c4232abb4b1d {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 935.217127] env[63515]: DEBUG nova.compute.manager [req-51bfb957-3fd2-4eaf-aa8c-c6a959a160ea req-d7fa06c5-6ec8-40af-b8a8-0f8dd88b70ca service nova] [instance: 014720fe-87b9-4e55-b9ba-295b009c7ff6] Refreshing instance network info cache due to event network-changed-406b4b98-709d-4013-9ce0-c4232abb4b1d. {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 935.217127] env[63515]: DEBUG oslo_concurrency.lockutils [req-51bfb957-3fd2-4eaf-aa8c-c6a959a160ea req-d7fa06c5-6ec8-40af-b8a8-0f8dd88b70ca service nova] Acquiring lock "refresh_cache-014720fe-87b9-4e55-b9ba-295b009c7ff6" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 935.217268] env[63515]: DEBUG oslo_concurrency.lockutils [req-51bfb957-3fd2-4eaf-aa8c-c6a959a160ea req-d7fa06c5-6ec8-40af-b8a8-0f8dd88b70ca service nova] Acquired lock "refresh_cache-014720fe-87b9-4e55-b9ba-295b009c7ff6" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 935.218219] env[63515]: DEBUG nova.network.neutron [req-51bfb957-3fd2-4eaf-aa8c-c6a959a160ea req-d7fa06c5-6ec8-40af-b8a8-0f8dd88b70ca service nova] [instance: 014720fe-87b9-4e55-b9ba-295b009c7ff6] Refreshing network info cache for port 406b4b98-709d-4013-9ce0-c4232abb4b1d {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 935.223622] env[63515]: DEBUG oslo_vmware.api [None req-d63820c0-6e1a-45a6-8579-7bfab43bd25a tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Waiting for the task: (returnval){ [ 935.223622] env[63515]: value = "task-1111614" [ 935.223622] env[63515]: _type = "Task" [ 935.223622] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 935.233181] env[63515]: DEBUG nova.compute.manager [req-563ca71f-6989-4c61-8a42-bb2c6bfc93a3 req-5d3ff12c-1989-4052-937a-60c07e3b96c9 service nova] [instance: 2a29f37d-e786-48d3-b126-4854d6e7c67c] Received event network-changed-956cd361-0d42-4914-85d3-5f75a72a68a1 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 935.233181] env[63515]: DEBUG nova.compute.manager [req-563ca71f-6989-4c61-8a42-bb2c6bfc93a3 req-5d3ff12c-1989-4052-937a-60c07e3b96c9 service nova] [instance: 2a29f37d-e786-48d3-b126-4854d6e7c67c] Refreshing instance network info cache due to event network-changed-956cd361-0d42-4914-85d3-5f75a72a68a1. {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 935.233181] env[63515]: DEBUG oslo_concurrency.lockutils [req-563ca71f-6989-4c61-8a42-bb2c6bfc93a3 req-5d3ff12c-1989-4052-937a-60c07e3b96c9 service nova] Acquiring lock "refresh_cache-2a29f37d-e786-48d3-b126-4854d6e7c67c" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 935.233181] env[63515]: DEBUG oslo_concurrency.lockutils [req-563ca71f-6989-4c61-8a42-bb2c6bfc93a3 req-5d3ff12c-1989-4052-937a-60c07e3b96c9 service nova] Acquired lock "refresh_cache-2a29f37d-e786-48d3-b126-4854d6e7c67c" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 935.233181] env[63515]: DEBUG nova.network.neutron [req-563ca71f-6989-4c61-8a42-bb2c6bfc93a3 req-5d3ff12c-1989-4052-937a-60c07e3b96c9 service nova] [instance: 2a29f37d-e786-48d3-b126-4854d6e7c67c] Refreshing network info cache for port 956cd361-0d42-4914-85d3-5f75a72a68a1 {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 935.236694] env[63515]: DEBUG oslo_concurrency.lockutils [None req-be7cccf6-3f5b-4895-b678-cf60de2c827a tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Acquiring lock "refresh_cache-014720fe-87b9-4e55-b9ba-295b009c7ff6" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 935.240485] env[63515]: DEBUG oslo_vmware.api [None req-d63820c0-6e1a-45a6-8579-7bfab43bd25a tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Task: {'id': task-1111614, 'name': PowerOnVM_Task} progress is 33%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.319345] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dace8c49-2c24-455e-a2ec-3da89675491d {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.329931] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6672fc4e-abf8-4996-a34f-39d75b13bc02 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.360557] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14dda01d-963e-4f6e-a9d5-880e06adf5ee {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.370498] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85d4365f-abe6-4344-8686-2de57888a24a {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.386585] env[63515]: DEBUG nova.compute.provider_tree [None req-f7e007a5-60f7-44d2-a529-05d489daa7aa tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Updating inventory in ProviderTree for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 935.436506] env[63515]: DEBUG nova.compute.manager [None req-127af73c-afc8-4436-9075-a886f5e9c7d5 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] [instance: 7d572ac0-f6af-4622-96cc-e75983420222] Starting instance... {{(pid=63515) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 935.573859] env[63515]: DEBUG nova.network.neutron [None req-bd214878-de3c-49e1-b706-bcb83d66311c tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: eab16df9-7bb5-4576-bca0-769a561c5fe9] Updating instance_info_cache with network_info: [{"id": "4168f5a5-c282-4f13-9c97-0b0236bd659b", "address": "fa:16:3e:a0:6a:50", "network": {"id": "88dfb29d-41e0-4df7-8f70-e52cda05620c", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1355578400-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.203", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7d4ada9ec35f42b19c6480a9101d21a4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "357d2811-e990-4985-9f9e-b158d10d3699", "external-id": "nsx-vlan-transportzone-641", "segmentation_id": 641, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4168f5a5-c2", "ovs_interfaceid": "4168f5a5-c282-4f13-9c97-0b0236bd659b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "e3c72582-41bf-48d4-8614-b516a96362a7", "address": "fa:16:3e:c0:fd:b9", "network": {"id": "88dfb29d-41e0-4df7-8f70-e52cda05620c", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1355578400-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7d4ada9ec35f42b19c6480a9101d21a4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "357d2811-e990-4985-9f9e-b158d10d3699", "external-id": "nsx-vlan-transportzone-641", "segmentation_id": 641, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape3c72582-41", "ovs_interfaceid": "e3c72582-41bf-48d4-8614-b516a96362a7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "ff4043f6-7576-4093-80ed-d36240b37f2f", "address": "fa:16:3e:47:af:a2", "network": {"id": "88dfb29d-41e0-4df7-8f70-e52cda05620c", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1355578400-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7d4ada9ec35f42b19c6480a9101d21a4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "357d2811-e990-4985-9f9e-b158d10d3699", "external-id": "nsx-vlan-transportzone-641", "segmentation_id": 641, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapff4043f6-75", "ovs_interfaceid": "ff4043f6-7576-4093-80ed-d36240b37f2f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 935.740273] env[63515]: DEBUG oslo_vmware.api [None req-d63820c0-6e1a-45a6-8579-7bfab43bd25a tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Task: {'id': task-1111614, 'name': PowerOnVM_Task, 'duration_secs': 0.481069} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 935.740577] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-d63820c0-6e1a-45a6-8579-7bfab43bd25a tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: 2fa098b6-ecc0-491f-8cfc-5ee335bbf943] Powered on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 935.740792] env[63515]: INFO nova.compute.manager [None req-d63820c0-6e1a-45a6-8579-7bfab43bd25a tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: 2fa098b6-ecc0-491f-8cfc-5ee335bbf943] Took 7.59 seconds to spawn the instance on the hypervisor. [ 935.741014] env[63515]: DEBUG nova.compute.manager [None req-d63820c0-6e1a-45a6-8579-7bfab43bd25a tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: 2fa098b6-ecc0-491f-8cfc-5ee335bbf943] Checking state {{(pid=63515) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 935.742144] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85f9bc73-0794-4f9d-b563-e964e84f7d0a {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.770761] env[63515]: DEBUG nova.network.neutron [req-51bfb957-3fd2-4eaf-aa8c-c6a959a160ea req-d7fa06c5-6ec8-40af-b8a8-0f8dd88b70ca service nova] [instance: 014720fe-87b9-4e55-b9ba-295b009c7ff6] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 935.924021] env[63515]: DEBUG nova.network.neutron [req-51bfb957-3fd2-4eaf-aa8c-c6a959a160ea req-d7fa06c5-6ec8-40af-b8a8-0f8dd88b70ca service nova] [instance: 014720fe-87b9-4e55-b9ba-295b009c7ff6] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 935.926781] env[63515]: DEBUG nova.scheduler.client.report [None req-f7e007a5-60f7-44d2-a529-05d489daa7aa tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Updated inventory for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a with generation 121 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 935.927319] env[63515]: DEBUG nova.compute.provider_tree [None req-f7e007a5-60f7-44d2-a529-05d489daa7aa tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Updating resource provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a generation from 121 to 122 during operation: update_inventory {{(pid=63515) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 935.927680] env[63515]: DEBUG nova.compute.provider_tree [None req-f7e007a5-60f7-44d2-a529-05d489daa7aa tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Updating inventory in ProviderTree for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 935.932698] env[63515]: DEBUG nova.network.neutron [None req-1d5e76d4-1b01-4bdb-831a-7cba26a22dfc tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: 829e72d7-a56a-451c-b98a-94253e5900d5] Updating instance_info_cache with network_info: [{"id": "856de677-b1c8-479b-91d9-34d31a703a63", "address": "fa:16:3e:13:27:b4", "network": {"id": "15fdaabe-6793-400f-9737-2cdff8f07238", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1175204866-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f9c98e9f6020475490aaa7e76d907ab7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7d377d75-3add-4a15-8691-74b2eb010924", "external-id": "nsx-vlan-transportzone-71", "segmentation_id": 71, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap856de677-b1", "ovs_interfaceid": "856de677-b1c8-479b-91d9-34d31a703a63", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 935.969325] env[63515]: DEBUG oslo_concurrency.lockutils [None req-127af73c-afc8-4436-9075-a886f5e9c7d5 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 936.064111] env[63515]: DEBUG nova.network.neutron [req-563ca71f-6989-4c61-8a42-bb2c6bfc93a3 req-5d3ff12c-1989-4052-937a-60c07e3b96c9 service nova] [instance: 2a29f37d-e786-48d3-b126-4854d6e7c67c] Updated VIF entry in instance network info cache for port 956cd361-0d42-4914-85d3-5f75a72a68a1. {{(pid=63515) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 936.064513] env[63515]: DEBUG nova.network.neutron [req-563ca71f-6989-4c61-8a42-bb2c6bfc93a3 req-5d3ff12c-1989-4052-937a-60c07e3b96c9 service nova] [instance: 2a29f37d-e786-48d3-b126-4854d6e7c67c] Updating instance_info_cache with network_info: [{"id": "956cd361-0d42-4914-85d3-5f75a72a68a1", "address": "fa:16:3e:a3:72:bb", "network": {"id": "850bfa95-3df6-4019-8260-ade20ec58799", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-832000501-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.153", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "08c9aa51b1864410809f8877c7cd49c4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3cf748a8-7ae0-4dca-817d-e727c30d72f4", "external-id": "nsx-vlan-transportzone-853", "segmentation_id": 853, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap956cd361-0d", "ovs_interfaceid": "956cd361-0d42-4914-85d3-5f75a72a68a1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 936.077028] env[63515]: DEBUG oslo_concurrency.lockutils [None req-bd214878-de3c-49e1-b706-bcb83d66311c tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Releasing lock "refresh_cache-eab16df9-7bb5-4576-bca0-769a561c5fe9" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 936.077792] env[63515]: DEBUG oslo_concurrency.lockutils [None req-bd214878-de3c-49e1-b706-bcb83d66311c tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Acquiring lock "eab16df9-7bb5-4576-bca0-769a561c5fe9" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 936.077995] env[63515]: DEBUG oslo_concurrency.lockutils [None req-bd214878-de3c-49e1-b706-bcb83d66311c tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Acquired lock "eab16df9-7bb5-4576-bca0-769a561c5fe9" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 936.079343] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0d09ebe-148e-44a6-821e-29dd1c6d46ac {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.097884] env[63515]: DEBUG nova.virt.hardware [None req-bd214878-de3c-49e1-b706-bcb83d66311c tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T02:52:14Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 936.098182] env[63515]: DEBUG nova.virt.hardware [None req-bd214878-de3c-49e1-b706-bcb83d66311c tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Flavor limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 936.098321] env[63515]: DEBUG nova.virt.hardware [None req-bd214878-de3c-49e1-b706-bcb83d66311c tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Image limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 936.098533] env[63515]: DEBUG nova.virt.hardware [None req-bd214878-de3c-49e1-b706-bcb83d66311c tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Flavor pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 936.098695] env[63515]: DEBUG nova.virt.hardware [None req-bd214878-de3c-49e1-b706-bcb83d66311c tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Image pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 936.098824] env[63515]: DEBUG nova.virt.hardware [None req-bd214878-de3c-49e1-b706-bcb83d66311c tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 936.099040] env[63515]: DEBUG nova.virt.hardware [None req-bd214878-de3c-49e1-b706-bcb83d66311c tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 936.099230] env[63515]: DEBUG nova.virt.hardware [None req-bd214878-de3c-49e1-b706-bcb83d66311c tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 936.099410] env[63515]: DEBUG nova.virt.hardware [None req-bd214878-de3c-49e1-b706-bcb83d66311c tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Got 1 possible topologies {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 936.099577] env[63515]: DEBUG nova.virt.hardware [None req-bd214878-de3c-49e1-b706-bcb83d66311c tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 936.099754] env[63515]: DEBUG nova.virt.hardware [None req-bd214878-de3c-49e1-b706-bcb83d66311c tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 936.106369] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-bd214878-de3c-49e1-b706-bcb83d66311c tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: eab16df9-7bb5-4576-bca0-769a561c5fe9] Reconfiguring VM to attach interface {{(pid=63515) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1923}} [ 936.106762] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-33a7bda3-879e-498e-8925-3d8412c958f4 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.125296] env[63515]: DEBUG oslo_vmware.api [None req-bd214878-de3c-49e1-b706-bcb83d66311c tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Waiting for the task: (returnval){ [ 936.125296] env[63515]: value = "task-1111615" [ 936.125296] env[63515]: _type = "Task" [ 936.125296] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 936.133298] env[63515]: DEBUG oslo_vmware.api [None req-bd214878-de3c-49e1-b706-bcb83d66311c tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Task: {'id': task-1111615, 'name': ReconfigVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.259714] env[63515]: INFO nova.compute.manager [None req-d63820c0-6e1a-45a6-8579-7bfab43bd25a tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: 2fa098b6-ecc0-491f-8cfc-5ee335bbf943] Took 20.02 seconds to build instance. [ 936.427594] env[63515]: DEBUG oslo_concurrency.lockutils [req-51bfb957-3fd2-4eaf-aa8c-c6a959a160ea req-d7fa06c5-6ec8-40af-b8a8-0f8dd88b70ca service nova] Releasing lock "refresh_cache-014720fe-87b9-4e55-b9ba-295b009c7ff6" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 936.428775] env[63515]: DEBUG oslo_concurrency.lockutils [None req-be7cccf6-3f5b-4895-b678-cf60de2c827a tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Acquired lock "refresh_cache-014720fe-87b9-4e55-b9ba-295b009c7ff6" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 936.428775] env[63515]: DEBUG nova.network.neutron [None req-be7cccf6-3f5b-4895-b678-cf60de2c827a tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: 014720fe-87b9-4e55-b9ba-295b009c7ff6] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 936.433387] env[63515]: DEBUG oslo_concurrency.lockutils [None req-f7e007a5-60f7-44d2-a529-05d489daa7aa tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 3.909s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 936.436758] env[63515]: DEBUG oslo_concurrency.lockutils [None req-9e4db499-5494-499b-b448-2d861439cb14 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 7.383s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 936.437089] env[63515]: DEBUG nova.objects.instance [None req-9e4db499-5494-499b-b448-2d861439cb14 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Lazy-loading 'resources' on Instance uuid ca6a8ac3-d880-4333-9816-2f076e423844 {{(pid=63515) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 936.438330] env[63515]: DEBUG oslo_concurrency.lockutils [None req-1d5e76d4-1b01-4bdb-831a-7cba26a22dfc tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Releasing lock "refresh_cache-829e72d7-a56a-451c-b98a-94253e5900d5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 936.456938] env[63515]: INFO nova.scheduler.client.report [None req-f7e007a5-60f7-44d2-a529-05d489daa7aa tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Deleted allocations for instance 9236607d-47af-4272-93db-79826e918aa1 [ 936.567673] env[63515]: DEBUG oslo_concurrency.lockutils [req-563ca71f-6989-4c61-8a42-bb2c6bfc93a3 req-5d3ff12c-1989-4052-937a-60c07e3b96c9 service nova] Releasing lock "refresh_cache-2a29f37d-e786-48d3-b126-4854d6e7c67c" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 936.637869] env[63515]: DEBUG oslo_vmware.api [None req-bd214878-de3c-49e1-b706-bcb83d66311c tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Task: {'id': task-1111615, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.761888] env[63515]: DEBUG oslo_concurrency.lockutils [None req-d63820c0-6e1a-45a6-8579-7bfab43bd25a tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Lock "2fa098b6-ecc0-491f-8cfc-5ee335bbf943" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 21.530s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 936.971527] env[63515]: DEBUG nova.network.neutron [None req-be7cccf6-3f5b-4895-b678-cf60de2c827a tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: 014720fe-87b9-4e55-b9ba-295b009c7ff6] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 936.974321] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-231bdd1c-d4c5-4e81-9e9f-3a29c04dcddb {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.978524] env[63515]: DEBUG oslo_concurrency.lockutils [None req-f7e007a5-60f7-44d2-a529-05d489daa7aa tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Lock "9236607d-47af-4272-93db-79826e918aa1" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 11.256s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 937.005786] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be56a934-5d3b-40ca-9b78-5e476821c5df {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.015384] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-1d5e76d4-1b01-4bdb-831a-7cba26a22dfc tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: 829e72d7-a56a-451c-b98a-94253e5900d5] Updating instance '829e72d7-a56a-451c-b98a-94253e5900d5' progress to 83 {{(pid=63515) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 937.139985] env[63515]: DEBUG oslo_vmware.api [None req-bd214878-de3c-49e1-b706-bcb83d66311c tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Task: {'id': task-1111615, 'name': ReconfigVM_Task, 'duration_secs': 0.851553} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 937.140712] env[63515]: DEBUG oslo_concurrency.lockutils [None req-bd214878-de3c-49e1-b706-bcb83d66311c tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Releasing lock "eab16df9-7bb5-4576-bca0-769a561c5fe9" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 937.140960] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-bd214878-de3c-49e1-b706-bcb83d66311c tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: eab16df9-7bb5-4576-bca0-769a561c5fe9] Reconfigured VM to attach interface {{(pid=63515) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1937}} [ 937.149103] env[63515]: DEBUG nova.network.neutron [None req-be7cccf6-3f5b-4895-b678-cf60de2c827a tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: 014720fe-87b9-4e55-b9ba-295b009c7ff6] Updating instance_info_cache with network_info: [{"id": "406b4b98-709d-4013-9ce0-c4232abb4b1d", "address": "fa:16:3e:36:42:96", "network": {"id": "3e5914c8-0cd4-49ee-81b4-1f3e280440bb", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-641632277-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ea099ce07bfb4a8da014d9303cf552eb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "225b6979-9329-403b-91fa-138bd41f6e83", "external-id": "nsx-vlan-transportzone-38", "segmentation_id": 38, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap406b4b98-70", "ovs_interfaceid": "406b4b98-709d-4013-9ce0-c4232abb4b1d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 937.292941] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04e8cc83-d625-4225-a1f8-46f2cf85f369 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.301020] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb2d26f5-9f46-4500-90e7-6d1a2aa156f7 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.089595] env[63515]: DEBUG nova.compute.manager [None req-9550b400-f3b1-456b-9e7a-27ee304f516f tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: 2fa098b6-ecc0-491f-8cfc-5ee335bbf943] Checking state {{(pid=63515) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 938.091970] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-1d5e76d4-1b01-4bdb-831a-7cba26a22dfc tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: 829e72d7-a56a-451c-b98a-94253e5900d5] Powering on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 938.096021] env[63515]: DEBUG oslo_concurrency.lockutils [None req-bd214878-de3c-49e1-b706-bcb83d66311c tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Lock "interface-eab16df9-7bb5-4576-bca0-769a561c5fe9-ff4043f6-7576-4093-80ed-d36240b37f2f" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 8.726s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 938.096021] env[63515]: DEBUG oslo_concurrency.lockutils [None req-be7cccf6-3f5b-4895-b678-cf60de2c827a tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Releasing lock "refresh_cache-014720fe-87b9-4e55-b9ba-295b009c7ff6" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 938.096021] env[63515]: DEBUG nova.compute.manager [None req-be7cccf6-3f5b-4895-b678-cf60de2c827a tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: 014720fe-87b9-4e55-b9ba-295b009c7ff6] Instance network_info: |[{"id": "406b4b98-709d-4013-9ce0-c4232abb4b1d", "address": "fa:16:3e:36:42:96", "network": {"id": "3e5914c8-0cd4-49ee-81b4-1f3e280440bb", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-641632277-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ea099ce07bfb4a8da014d9303cf552eb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "225b6979-9329-403b-91fa-138bd41f6e83", "external-id": "nsx-vlan-transportzone-38", "segmentation_id": 38, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap406b4b98-70", "ovs_interfaceid": "406b4b98-709d-4013-9ce0-c4232abb4b1d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 938.096021] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e392a5df-c5ac-4849-bab9-fd292596166a {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.097964] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-22924d50-e92a-4c7c-91bf-daee698729f1 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.099596] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-be7cccf6-3f5b-4895-b678-cf60de2c827a tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: 014720fe-87b9-4e55-b9ba-295b009c7ff6] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:36:42:96', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '225b6979-9329-403b-91fa-138bd41f6e83', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '406b4b98-709d-4013-9ce0-c4232abb4b1d', 'vif_model': 'vmxnet3'}] {{(pid=63515) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 938.107024] env[63515]: DEBUG oslo.service.loopingcall [None req-be7cccf6-3f5b-4895-b678-cf60de2c827a tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 938.107773] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6782d1f-a35d-45ee-8706-8576fdc84315 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.111926] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 014720fe-87b9-4e55-b9ba-295b009c7ff6] Creating VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 938.111926] env[63515]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d08ea0af-492d-400c-b2b9-bcc390a4843b {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.129857] env[63515]: DEBUG oslo_vmware.api [None req-1d5e76d4-1b01-4bdb-831a-7cba26a22dfc tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Waiting for the task: (returnval){ [ 938.129857] env[63515]: value = "task-1111616" [ 938.129857] env[63515]: _type = "Task" [ 938.129857] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 938.143187] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-140c9c16-1d69-4f40-bf52-5cd002014a5c {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.152216] env[63515]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 938.152216] env[63515]: value = "task-1111617" [ 938.152216] env[63515]: _type = "Task" [ 938.152216] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 938.166343] env[63515]: DEBUG nova.compute.provider_tree [None req-9e4db499-5494-499b-b448-2d861439cb14 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 938.167593] env[63515]: DEBUG oslo_vmware.api [None req-1d5e76d4-1b01-4bdb-831a-7cba26a22dfc tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Task: {'id': task-1111616, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.173554] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1111617, 'name': CreateVM_Task} progress is 10%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.216812] env[63515]: DEBUG oslo_vmware.rw_handles [None req-b95e91a2-83f6-4cd7-9448-65181d09829a tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52467804-bf2c-9afd-896b-b66372e45b11/disk-0.vmdk. {{(pid=63515) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 938.217800] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-629808ec-5e62-40ba-9468-5ea6d8eedf7f {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.224453] env[63515]: DEBUG oslo_vmware.rw_handles [None req-b95e91a2-83f6-4cd7-9448-65181d09829a tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52467804-bf2c-9afd-896b-b66372e45b11/disk-0.vmdk is in state: ready. {{(pid=63515) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 938.224638] env[63515]: ERROR oslo_vmware.rw_handles [None req-b95e91a2-83f6-4cd7-9448-65181d09829a tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52467804-bf2c-9afd-896b-b66372e45b11/disk-0.vmdk due to incomplete transfer. [ 938.224870] env[63515]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-23348e04-5d48-46dc-bb31-c263e2a958dd {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.233311] env[63515]: DEBUG oslo_concurrency.lockutils [None req-857b4940-0702-491c-a977-f4c63c5dd5f5 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Acquiring lock "918d5b35-c46d-47c5-b59f-175831aed9d6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 938.233543] env[63515]: DEBUG oslo_concurrency.lockutils [None req-857b4940-0702-491c-a977-f4c63c5dd5f5 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Lock "918d5b35-c46d-47c5-b59f-175831aed9d6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 938.234647] env[63515]: DEBUG oslo_vmware.rw_handles [None req-b95e91a2-83f6-4cd7-9448-65181d09829a tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52467804-bf2c-9afd-896b-b66372e45b11/disk-0.vmdk. {{(pid=63515) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 938.234835] env[63515]: DEBUG nova.virt.vmwareapi.images [None req-b95e91a2-83f6-4cd7-9448-65181d09829a tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 68766cd4-84be-475b-8494-d7ab43a9e969] Uploaded image ca8a2595-00a8-4a1d-8bdc-e37e8a7f93e5 to the Glance image server {{(pid=63515) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 938.237391] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-b95e91a2-83f6-4cd7-9448-65181d09829a tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 68766cd4-84be-475b-8494-d7ab43a9e969] Destroying the VM {{(pid=63515) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 938.238048] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-688e5aee-ebeb-4de4-a6b7-cda93ac218f1 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.244774] env[63515]: DEBUG oslo_vmware.api [None req-b95e91a2-83f6-4cd7-9448-65181d09829a tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Waiting for the task: (returnval){ [ 938.244774] env[63515]: value = "task-1111618" [ 938.244774] env[63515]: _type = "Task" [ 938.244774] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 938.253178] env[63515]: DEBUG oslo_vmware.api [None req-b95e91a2-83f6-4cd7-9448-65181d09829a tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Task: {'id': task-1111618, 'name': Destroy_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.639157] env[63515]: INFO nova.compute.manager [None req-9550b400-f3b1-456b-9e7a-27ee304f516f tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: 2fa098b6-ecc0-491f-8cfc-5ee335bbf943] instance snapshotting [ 938.642525] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c09a8e87-f332-42db-ac75-30f50223faad {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.650375] env[63515]: DEBUG oslo_vmware.api [None req-1d5e76d4-1b01-4bdb-831a-7cba26a22dfc tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Task: {'id': task-1111616, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.670473] env[63515]: DEBUG nova.scheduler.client.report [None req-9e4db499-5494-499b-b448-2d861439cb14 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 938.674919] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57ab51e2-f7b8-44e6-9e64-795e41a7f141 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.680985] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1111617, 'name': CreateVM_Task, 'duration_secs': 0.347781} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 938.681532] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 014720fe-87b9-4e55-b9ba-295b009c7ff6] Created VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 938.682251] env[63515]: DEBUG oslo_concurrency.lockutils [None req-be7cccf6-3f5b-4895-b678-cf60de2c827a tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 938.682440] env[63515]: DEBUG oslo_concurrency.lockutils [None req-be7cccf6-3f5b-4895-b678-cf60de2c827a tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 938.682762] env[63515]: DEBUG oslo_concurrency.lockutils [None req-be7cccf6-3f5b-4895-b678-cf60de2c827a tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 938.683056] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b0ac2ed9-f26f-4973-9549-b83f7abecbfe {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.692150] env[63515]: DEBUG oslo_vmware.api [None req-be7cccf6-3f5b-4895-b678-cf60de2c827a tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Waiting for the task: (returnval){ [ 938.692150] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]52750c69-8b19-f5e0-2612-3f2e4a5669d3" [ 938.692150] env[63515]: _type = "Task" [ 938.692150] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 938.700289] env[63515]: DEBUG oslo_vmware.api [None req-be7cccf6-3f5b-4895-b678-cf60de2c827a tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52750c69-8b19-f5e0-2612-3f2e4a5669d3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.738167] env[63515]: DEBUG nova.compute.manager [None req-857b4940-0702-491c-a977-f4c63c5dd5f5 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: 918d5b35-c46d-47c5-b59f-175831aed9d6] Starting instance... {{(pid=63515) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 938.755394] env[63515]: DEBUG oslo_vmware.api [None req-b95e91a2-83f6-4cd7-9448-65181d09829a tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Task: {'id': task-1111618, 'name': Destroy_Task, 'duration_secs': 0.341208} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 938.756027] env[63515]: INFO nova.virt.vmwareapi.vm_util [None req-b95e91a2-83f6-4cd7-9448-65181d09829a tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 68766cd4-84be-475b-8494-d7ab43a9e969] Destroyed the VM [ 938.756072] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-b95e91a2-83f6-4cd7-9448-65181d09829a tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 68766cd4-84be-475b-8494-d7ab43a9e969] Deleting Snapshot of the VM instance {{(pid=63515) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 938.756357] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-a1013df9-a918-4c3e-a2a9-7474de3a14e2 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.763097] env[63515]: DEBUG oslo_vmware.api [None req-b95e91a2-83f6-4cd7-9448-65181d09829a tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Waiting for the task: (returnval){ [ 938.763097] env[63515]: value = "task-1111619" [ 938.763097] env[63515]: _type = "Task" [ 938.763097] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 938.771538] env[63515]: DEBUG oslo_vmware.api [None req-b95e91a2-83f6-4cd7-9448-65181d09829a tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Task: {'id': task-1111619, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.144241] env[63515]: DEBUG oslo_vmware.api [None req-1d5e76d4-1b01-4bdb-831a-7cba26a22dfc tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Task: {'id': task-1111616, 'name': PowerOnVM_Task, 'duration_secs': 0.593522} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 939.144532] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-1d5e76d4-1b01-4bdb-831a-7cba26a22dfc tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: 829e72d7-a56a-451c-b98a-94253e5900d5] Powered on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 939.144680] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-1d5e76d4-1b01-4bdb-831a-7cba26a22dfc tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: 829e72d7-a56a-451c-b98a-94253e5900d5] Updating instance '829e72d7-a56a-451c-b98a-94253e5900d5' progress to 100 {{(pid=63515) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 939.182839] env[63515]: DEBUG oslo_concurrency.lockutils [None req-9e4db499-5494-499b-b448-2d861439cb14 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.746s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 939.185051] env[63515]: DEBUG oslo_concurrency.lockutils [None req-8fdfb203-5786-4efd-8927-4ee465e68e44 tempest-ServerMetadataTestJSON-1034558863 tempest-ServerMetadataTestJSON-1034558863-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.938s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 939.186560] env[63515]: INFO nova.compute.claims [None req-8fdfb203-5786-4efd-8927-4ee465e68e44 tempest-ServerMetadataTestJSON-1034558863 tempest-ServerMetadataTestJSON-1034558863-project-member] [instance: b25bf2ba-cf08-44ac-a524-aa87cf785575] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 939.191252] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-9550b400-f3b1-456b-9e7a-27ee304f516f tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: 2fa098b6-ecc0-491f-8cfc-5ee335bbf943] Creating Snapshot of the VM instance {{(pid=63515) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 939.191717] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-12e6830e-808c-4b80-b4d6-a3c178d553f5 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.210018] env[63515]: DEBUG oslo_vmware.api [None req-be7cccf6-3f5b-4895-b678-cf60de2c827a tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52750c69-8b19-f5e0-2612-3f2e4a5669d3, 'name': SearchDatastore_Task, 'duration_secs': 0.019555} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 939.210018] env[63515]: DEBUG oslo_concurrency.lockutils [None req-be7cccf6-3f5b-4895-b678-cf60de2c827a tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 939.210395] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-be7cccf6-3f5b-4895-b678-cf60de2c827a tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: 014720fe-87b9-4e55-b9ba-295b009c7ff6] Processing image 8a120570-cb06-4099-b262-554ca0ad15c5 {{(pid=63515) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 939.210861] env[63515]: DEBUG oslo_concurrency.lockutils [None req-be7cccf6-3f5b-4895-b678-cf60de2c827a tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 939.211259] env[63515]: DEBUG oslo_concurrency.lockutils [None req-be7cccf6-3f5b-4895-b678-cf60de2c827a tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 939.214019] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-be7cccf6-3f5b-4895-b678-cf60de2c827a tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 939.214019] env[63515]: DEBUG oslo_vmware.api [None req-9550b400-f3b1-456b-9e7a-27ee304f516f tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Waiting for the task: (returnval){ [ 939.214019] env[63515]: value = "task-1111620" [ 939.214019] env[63515]: _type = "Task" [ 939.214019] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 939.214019] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-fc114968-c08a-4dab-82b0-b293eacb30e9 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.215947] env[63515]: INFO nova.scheduler.client.report [None req-9e4db499-5494-499b-b448-2d861439cb14 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Deleted allocations for instance ca6a8ac3-d880-4333-9816-2f076e423844 [ 939.228594] env[63515]: DEBUG oslo_vmware.api [None req-9550b400-f3b1-456b-9e7a-27ee304f516f tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Task: {'id': task-1111620, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.229766] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-be7cccf6-3f5b-4895-b678-cf60de2c827a tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 939.230036] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-be7cccf6-3f5b-4895-b678-cf60de2c827a tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63515) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 939.230833] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-67a8b9fd-0ab0-4d2f-a1d0-1ad9488eead6 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.236275] env[63515]: DEBUG oslo_vmware.api [None req-be7cccf6-3f5b-4895-b678-cf60de2c827a tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Waiting for the task: (returnval){ [ 939.236275] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]52904164-bb05-adf0-8881-6473c1274912" [ 939.236275] env[63515]: _type = "Task" [ 939.236275] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 939.248395] env[63515]: DEBUG oslo_vmware.api [None req-be7cccf6-3f5b-4895-b678-cf60de2c827a tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52904164-bb05-adf0-8881-6473c1274912, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.261138] env[63515]: DEBUG oslo_concurrency.lockutils [None req-857b4940-0702-491c-a977-f4c63c5dd5f5 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 939.273654] env[63515]: DEBUG oslo_vmware.api [None req-b95e91a2-83f6-4cd7-9448-65181d09829a tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Task: {'id': task-1111619, 'name': RemoveSnapshot_Task} progress is 36%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.727011] env[63515]: DEBUG oslo_vmware.api [None req-9550b400-f3b1-456b-9e7a-27ee304f516f tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Task: {'id': task-1111620, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.727479] env[63515]: DEBUG oslo_concurrency.lockutils [None req-9e4db499-5494-499b-b448-2d861439cb14 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Lock "ca6a8ac3-d880-4333-9816-2f076e423844" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 13.860s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 939.746291] env[63515]: DEBUG oslo_vmware.api [None req-be7cccf6-3f5b-4895-b678-cf60de2c827a tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52904164-bb05-adf0-8881-6473c1274912, 'name': SearchDatastore_Task, 'duration_secs': 0.00913} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 939.747118] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-57cb59c5-8817-493c-aab7-2c61c1dd2a78 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.752368] env[63515]: DEBUG oslo_vmware.api [None req-be7cccf6-3f5b-4895-b678-cf60de2c827a tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Waiting for the task: (returnval){ [ 939.752368] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]52171972-dc2d-0f55-7c6f-ef4abab5d1db" [ 939.752368] env[63515]: _type = "Task" [ 939.752368] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 939.760111] env[63515]: DEBUG oslo_vmware.api [None req-be7cccf6-3f5b-4895-b678-cf60de2c827a tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52171972-dc2d-0f55-7c6f-ef4abab5d1db, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.771859] env[63515]: DEBUG oslo_vmware.api [None req-b95e91a2-83f6-4cd7-9448-65181d09829a tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Task: {'id': task-1111619, 'name': RemoveSnapshot_Task, 'duration_secs': 0.550882} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 939.772122] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-b95e91a2-83f6-4cd7-9448-65181d09829a tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 68766cd4-84be-475b-8494-d7ab43a9e969] Deleted Snapshot of the VM instance {{(pid=63515) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 939.772351] env[63515]: INFO nova.compute.manager [None req-b95e91a2-83f6-4cd7-9448-65181d09829a tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 68766cd4-84be-475b-8494-d7ab43a9e969] Took 14.78 seconds to snapshot the instance on the hypervisor. [ 939.856729] env[63515]: DEBUG oslo_concurrency.lockutils [None req-7db48089-ba7f-4a2a-9098-e95b53997aff tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Acquiring lock "interface-eab16df9-7bb5-4576-bca0-769a561c5fe9-e3c72582-41bf-48d4-8614-b516a96362a7" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 939.857016] env[63515]: DEBUG oslo_concurrency.lockutils [None req-7db48089-ba7f-4a2a-9098-e95b53997aff tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Lock "interface-eab16df9-7bb5-4576-bca0-769a561c5fe9-e3c72582-41bf-48d4-8614-b516a96362a7" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 940.228392] env[63515]: DEBUG oslo_vmware.api [None req-9550b400-f3b1-456b-9e7a-27ee304f516f tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Task: {'id': task-1111620, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.264450] env[63515]: DEBUG oslo_vmware.api [None req-be7cccf6-3f5b-4895-b678-cf60de2c827a tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52171972-dc2d-0f55-7c6f-ef4abab5d1db, 'name': SearchDatastore_Task, 'duration_secs': 0.02047} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 940.267799] env[63515]: DEBUG oslo_concurrency.lockutils [None req-be7cccf6-3f5b-4895-b678-cf60de2c827a tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 940.268178] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-be7cccf6-3f5b-4895-b678-cf60de2c827a tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk to [datastore2] 014720fe-87b9-4e55-b9ba-295b009c7ff6/014720fe-87b9-4e55-b9ba-295b009c7ff6.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 940.269096] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-08cd55db-9b86-4744-a95e-73b71ea086e5 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.278853] env[63515]: DEBUG oslo_vmware.api [None req-be7cccf6-3f5b-4895-b678-cf60de2c827a tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Waiting for the task: (returnval){ [ 940.278853] env[63515]: value = "task-1111621" [ 940.278853] env[63515]: _type = "Task" [ 940.278853] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 940.297026] env[63515]: DEBUG oslo_vmware.api [None req-be7cccf6-3f5b-4895-b678-cf60de2c827a tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Task: {'id': task-1111621, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.322709] env[63515]: DEBUG nova.compute.manager [None req-b95e91a2-83f6-4cd7-9448-65181d09829a tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 68766cd4-84be-475b-8494-d7ab43a9e969] Found 1 images (rotation: 2) {{(pid=63515) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4554}} [ 940.359731] env[63515]: DEBUG oslo_concurrency.lockutils [None req-7db48089-ba7f-4a2a-9098-e95b53997aff tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Acquiring lock "eab16df9-7bb5-4576-bca0-769a561c5fe9" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 940.360349] env[63515]: DEBUG oslo_concurrency.lockutils [None req-7db48089-ba7f-4a2a-9098-e95b53997aff tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Acquired lock "eab16df9-7bb5-4576-bca0-769a561c5fe9" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 940.362656] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e6e58fc-9b8e-44b4-81cc-adf522d85012 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.384960] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8c67b4c-f908-4295-a492-ba2ac955cfd7 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.413769] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-7db48089-ba7f-4a2a-9098-e95b53997aff tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: eab16df9-7bb5-4576-bca0-769a561c5fe9] Reconfiguring VM to detach interface {{(pid=63515) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1968}} [ 940.418131] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e66a4bf3-91d0-4f43-a28a-9e35f9288139 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.434158] env[63515]: DEBUG oslo_concurrency.lockutils [None req-3f7a26fd-cc84-46d9-9d28-feb378f8b5e6 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Acquiring lock "75f4dc79-4b73-4ddc-be03-3653a5a0797e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 940.434433] env[63515]: DEBUG oslo_concurrency.lockutils [None req-3f7a26fd-cc84-46d9-9d28-feb378f8b5e6 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Lock "75f4dc79-4b73-4ddc-be03-3653a5a0797e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 940.439822] env[63515]: DEBUG oslo_concurrency.lockutils [None req-3f7a26fd-cc84-46d9-9d28-feb378f8b5e6 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Acquiring lock "d438dd30-b816-48a6-94c6-11e6e9bd5b71" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 940.440080] env[63515]: DEBUG oslo_concurrency.lockutils [None req-3f7a26fd-cc84-46d9-9d28-feb378f8b5e6 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Lock "d438dd30-b816-48a6-94c6-11e6e9bd5b71" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 940.452061] env[63515]: DEBUG oslo_vmware.api [None req-7db48089-ba7f-4a2a-9098-e95b53997aff tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Waiting for the task: (returnval){ [ 940.452061] env[63515]: value = "task-1111622" [ 940.452061] env[63515]: _type = "Task" [ 940.452061] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 940.462623] env[63515]: DEBUG oslo_vmware.api [None req-7db48089-ba7f-4a2a-9098-e95b53997aff tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Task: {'id': task-1111622, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.516969] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92f58a09-2023-4efe-b2c5-502853239762 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.527238] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27da46d9-e13f-4695-b17c-fefb09fc7f09 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.561783] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3afbc758-1252-45e6-be35-c88b835195d1 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.570026] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e693e56-79e7-4c55-94f8-7841920c3f5a {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.583767] env[63515]: DEBUG nova.compute.provider_tree [None req-8fdfb203-5786-4efd-8927-4ee465e68e44 tempest-ServerMetadataTestJSON-1034558863 tempest-ServerMetadataTestJSON-1034558863-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 940.731582] env[63515]: DEBUG oslo_vmware.api [None req-9550b400-f3b1-456b-9e7a-27ee304f516f tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Task: {'id': task-1111620, 'name': CreateSnapshot_Task, 'duration_secs': 1.158388} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 940.731874] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-9550b400-f3b1-456b-9e7a-27ee304f516f tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: 2fa098b6-ecc0-491f-8cfc-5ee335bbf943] Created Snapshot of the VM instance {{(pid=63515) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 940.732708] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a251aa44-2500-4afe-bd27-8d1965964c60 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.794329] env[63515]: DEBUG oslo_vmware.api [None req-be7cccf6-3f5b-4895-b678-cf60de2c827a tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Task: {'id': task-1111621, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.936940] env[63515]: DEBUG nova.compute.manager [None req-3f7a26fd-cc84-46d9-9d28-feb378f8b5e6 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] [instance: 75f4dc79-4b73-4ddc-be03-3653a5a0797e] Starting instance... {{(pid=63515) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 940.942500] env[63515]: DEBUG nova.compute.manager [None req-3f7a26fd-cc84-46d9-9d28-feb378f8b5e6 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] [instance: d438dd30-b816-48a6-94c6-11e6e9bd5b71] Starting instance... {{(pid=63515) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 940.962219] env[63515]: DEBUG oslo_vmware.api [None req-7db48089-ba7f-4a2a-9098-e95b53997aff tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Task: {'id': task-1111622, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.994580] env[63515]: DEBUG oslo_concurrency.lockutils [None req-b94389d0-616d-4e7e-a415-7b0c3a16a683 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Acquiring lock "829e72d7-a56a-451c-b98a-94253e5900d5" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 940.994901] env[63515]: DEBUG oslo_concurrency.lockutils [None req-b94389d0-616d-4e7e-a415-7b0c3a16a683 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Lock "829e72d7-a56a-451c-b98a-94253e5900d5" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 940.995158] env[63515]: DEBUG nova.compute.manager [None req-b94389d0-616d-4e7e-a415-7b0c3a16a683 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: 829e72d7-a56a-451c-b98a-94253e5900d5] Going to confirm migration 3 {{(pid=63515) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 941.090214] env[63515]: DEBUG nova.scheduler.client.report [None req-8fdfb203-5786-4efd-8927-4ee465e68e44 tempest-ServerMetadataTestJSON-1034558863 tempest-ServerMetadataTestJSON-1034558863-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 941.092982] env[63515]: DEBUG nova.compute.manager [None req-8e6475b3-95eb-4b44-a2f3-5ddd5db39f52 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 68766cd4-84be-475b-8494-d7ab43a9e969] Checking state {{(pid=63515) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 941.094687] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55e982d8-4fc7-4695-8d31-2e46d4631a4b {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.251306] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-9550b400-f3b1-456b-9e7a-27ee304f516f tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: 2fa098b6-ecc0-491f-8cfc-5ee335bbf943] Creating linked-clone VM from snapshot {{(pid=63515) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 941.251306] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-8d7e05b6-ffaa-4a68-af58-45082b0a68ea {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.260433] env[63515]: DEBUG oslo_vmware.api [None req-9550b400-f3b1-456b-9e7a-27ee304f516f tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Waiting for the task: (returnval){ [ 941.260433] env[63515]: value = "task-1111623" [ 941.260433] env[63515]: _type = "Task" [ 941.260433] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 941.268556] env[63515]: DEBUG oslo_vmware.api [None req-9550b400-f3b1-456b-9e7a-27ee304f516f tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Task: {'id': task-1111623, 'name': CloneVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.293019] env[63515]: DEBUG oslo_vmware.api [None req-be7cccf6-3f5b-4895-b678-cf60de2c827a tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Task: {'id': task-1111621, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.547628} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 941.293019] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-be7cccf6-3f5b-4895-b678-cf60de2c827a tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk to [datastore2] 014720fe-87b9-4e55-b9ba-295b009c7ff6/014720fe-87b9-4e55-b9ba-295b009c7ff6.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 941.293019] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-be7cccf6-3f5b-4895-b678-cf60de2c827a tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: 014720fe-87b9-4e55-b9ba-295b009c7ff6] Extending root virtual disk to 1048576 {{(pid=63515) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 941.293252] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-4415cc4a-a51a-4fb5-aa11-ad9390bab804 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.299993] env[63515]: DEBUG oslo_vmware.api [None req-be7cccf6-3f5b-4895-b678-cf60de2c827a tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Waiting for the task: (returnval){ [ 941.299993] env[63515]: value = "task-1111624" [ 941.299993] env[63515]: _type = "Task" [ 941.299993] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 941.308742] env[63515]: DEBUG oslo_vmware.api [None req-be7cccf6-3f5b-4895-b678-cf60de2c827a tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Task: {'id': task-1111624, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.459993] env[63515]: DEBUG oslo_concurrency.lockutils [None req-3f7a26fd-cc84-46d9-9d28-feb378f8b5e6 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 941.463686] env[63515]: DEBUG oslo_vmware.api [None req-7db48089-ba7f-4a2a-9098-e95b53997aff tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Task: {'id': task-1111622, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.464572] env[63515]: DEBUG oslo_concurrency.lockutils [None req-3f7a26fd-cc84-46d9-9d28-feb378f8b5e6 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 941.534443] env[63515]: DEBUG oslo_concurrency.lockutils [None req-b94389d0-616d-4e7e-a415-7b0c3a16a683 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Acquiring lock "refresh_cache-829e72d7-a56a-451c-b98a-94253e5900d5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 941.534692] env[63515]: DEBUG oslo_concurrency.lockutils [None req-b94389d0-616d-4e7e-a415-7b0c3a16a683 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Acquired lock "refresh_cache-829e72d7-a56a-451c-b98a-94253e5900d5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 941.534843] env[63515]: DEBUG nova.network.neutron [None req-b94389d0-616d-4e7e-a415-7b0c3a16a683 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: 829e72d7-a56a-451c-b98a-94253e5900d5] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 941.535049] env[63515]: DEBUG nova.objects.instance [None req-b94389d0-616d-4e7e-a415-7b0c3a16a683 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Lazy-loading 'info_cache' on Instance uuid 829e72d7-a56a-451c-b98a-94253e5900d5 {{(pid=63515) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 941.594826] env[63515]: DEBUG oslo_concurrency.lockutils [None req-8fdfb203-5786-4efd-8927-4ee465e68e44 tempest-ServerMetadataTestJSON-1034558863 tempest-ServerMetadataTestJSON-1034558863-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.410s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 941.595392] env[63515]: DEBUG nova.compute.manager [None req-8fdfb203-5786-4efd-8927-4ee465e68e44 tempest-ServerMetadataTestJSON-1034558863 tempest-ServerMetadataTestJSON-1034558863-project-member] [instance: b25bf2ba-cf08-44ac-a524-aa87cf785575] Start building networks asynchronously for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 941.598110] env[63515]: DEBUG oslo_concurrency.lockutils [None req-56611604-1203-4b20-9871-2ec521700550 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 6.802s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 941.598345] env[63515]: DEBUG nova.objects.instance [None req-56611604-1203-4b20-9871-2ec521700550 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Lazy-loading 'resources' on Instance uuid a2203c5e-66b9-4c7e-8d76-e12a08b57b1d {{(pid=63515) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 941.606929] env[63515]: INFO nova.compute.manager [None req-8e6475b3-95eb-4b44-a2f3-5ddd5db39f52 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 68766cd4-84be-475b-8494-d7ab43a9e969] instance snapshotting [ 941.607554] env[63515]: DEBUG nova.objects.instance [None req-8e6475b3-95eb-4b44-a2f3-5ddd5db39f52 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Lazy-loading 'flavor' on Instance uuid 68766cd4-84be-475b-8494-d7ab43a9e969 {{(pid=63515) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 941.774627] env[63515]: DEBUG oslo_vmware.api [None req-9550b400-f3b1-456b-9e7a-27ee304f516f tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Task: {'id': task-1111623, 'name': CloneVM_Task} progress is 94%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.810679] env[63515]: DEBUG oslo_vmware.api [None req-be7cccf6-3f5b-4895-b678-cf60de2c827a tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Task: {'id': task-1111624, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.07128} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 941.810954] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-be7cccf6-3f5b-4895-b678-cf60de2c827a tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: 014720fe-87b9-4e55-b9ba-295b009c7ff6] Extended root virtual disk {{(pid=63515) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 941.811755] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7fc839f1-6cd4-418d-a8ce-b4410471b288 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.833420] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-be7cccf6-3f5b-4895-b678-cf60de2c827a tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: 014720fe-87b9-4e55-b9ba-295b009c7ff6] Reconfiguring VM instance instance-00000056 to attach disk [datastore2] 014720fe-87b9-4e55-b9ba-295b009c7ff6/014720fe-87b9-4e55-b9ba-295b009c7ff6.vmdk or device None with type sparse {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 941.833705] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-530abf6d-4ba5-48fc-9b0e-8708033b8b45 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.852691] env[63515]: DEBUG oslo_vmware.api [None req-be7cccf6-3f5b-4895-b678-cf60de2c827a tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Waiting for the task: (returnval){ [ 941.852691] env[63515]: value = "task-1111625" [ 941.852691] env[63515]: _type = "Task" [ 941.852691] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 941.860517] env[63515]: DEBUG oslo_vmware.api [None req-be7cccf6-3f5b-4895-b678-cf60de2c827a tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Task: {'id': task-1111625, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.965485] env[63515]: DEBUG oslo_vmware.api [None req-7db48089-ba7f-4a2a-9098-e95b53997aff tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Task: {'id': task-1111622, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.102228] env[63515]: DEBUG nova.compute.utils [None req-8fdfb203-5786-4efd-8927-4ee465e68e44 tempest-ServerMetadataTestJSON-1034558863 tempest-ServerMetadataTestJSON-1034558863-project-member] Using /dev/sd instead of None {{(pid=63515) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 942.106679] env[63515]: DEBUG nova.compute.manager [None req-8fdfb203-5786-4efd-8927-4ee465e68e44 tempest-ServerMetadataTestJSON-1034558863 tempest-ServerMetadataTestJSON-1034558863-project-member] [instance: b25bf2ba-cf08-44ac-a524-aa87cf785575] Allocating IP information in the background. {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 942.107046] env[63515]: DEBUG nova.network.neutron [None req-8fdfb203-5786-4efd-8927-4ee465e68e44 tempest-ServerMetadataTestJSON-1034558863 tempest-ServerMetadataTestJSON-1034558863-project-member] [instance: b25bf2ba-cf08-44ac-a524-aa87cf785575] allocate_for_instance() {{(pid=63515) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 942.113437] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eebe1f04-448e-4d4f-8a92-55abf3378b4e {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.136350] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3f76417-4665-4ccb-9577-9cc7265ec03e {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.154943] env[63515]: DEBUG nova.policy [None req-8fdfb203-5786-4efd-8927-4ee465e68e44 tempest-ServerMetadataTestJSON-1034558863 tempest-ServerMetadataTestJSON-1034558863-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c30692ed40234685b258020a3413859b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '49ee798463ab46c2b84738523ba7d096', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63515) authorize /opt/stack/nova/nova/policy.py:201}} [ 942.270799] env[63515]: DEBUG oslo_vmware.api [None req-9550b400-f3b1-456b-9e7a-27ee304f516f tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Task: {'id': task-1111623, 'name': CloneVM_Task} progress is 94%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.351023] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-961fdd48-e966-4e2a-aa80-7892a97fb7d6 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.359170] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e71a05f4-a837-462e-bf83-60203b3f4e49 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.367497] env[63515]: DEBUG oslo_vmware.api [None req-be7cccf6-3f5b-4895-b678-cf60de2c827a tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Task: {'id': task-1111625, 'name': ReconfigVM_Task, 'duration_secs': 0.283373} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 942.368283] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-be7cccf6-3f5b-4895-b678-cf60de2c827a tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: 014720fe-87b9-4e55-b9ba-295b009c7ff6] Reconfigured VM instance instance-00000056 to attach disk [datastore2] 014720fe-87b9-4e55-b9ba-295b009c7ff6/014720fe-87b9-4e55-b9ba-295b009c7ff6.vmdk or device None with type sparse {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 942.391352] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ba2415a6-0d63-430a-ac9f-2d208941275a {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.393803] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-503607c3-e98d-49de-bd60-8f3934d07737 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.404286] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae9cc67f-38ca-44b0-bd48-796db6033d6d {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.413159] env[63515]: DEBUG oslo_vmware.api [None req-be7cccf6-3f5b-4895-b678-cf60de2c827a tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Waiting for the task: (returnval){ [ 942.413159] env[63515]: value = "task-1111626" [ 942.413159] env[63515]: _type = "Task" [ 942.413159] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 942.426484] env[63515]: DEBUG nova.compute.provider_tree [None req-56611604-1203-4b20-9871-2ec521700550 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Updating inventory in ProviderTree for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 942.434436] env[63515]: DEBUG oslo_vmware.api [None req-be7cccf6-3f5b-4895-b678-cf60de2c827a tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Task: {'id': task-1111626, 'name': Rename_Task} progress is 14%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.451158] env[63515]: DEBUG nova.network.neutron [None req-8fdfb203-5786-4efd-8927-4ee465e68e44 tempest-ServerMetadataTestJSON-1034558863 tempest-ServerMetadataTestJSON-1034558863-project-member] [instance: b25bf2ba-cf08-44ac-a524-aa87cf785575] Successfully created port: 5d351a64-ccc8-4d4c-9057-7ef8d2c3edbd {{(pid=63515) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 942.467109] env[63515]: DEBUG oslo_vmware.api [None req-7db48089-ba7f-4a2a-9098-e95b53997aff tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Task: {'id': task-1111622, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.607536] env[63515]: DEBUG nova.compute.manager [None req-8fdfb203-5786-4efd-8927-4ee465e68e44 tempest-ServerMetadataTestJSON-1034558863 tempest-ServerMetadataTestJSON-1034558863-project-member] [instance: b25bf2ba-cf08-44ac-a524-aa87cf785575] Start building block device mappings for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 942.647188] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-8e6475b3-95eb-4b44-a2f3-5ddd5db39f52 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 68766cd4-84be-475b-8494-d7ab43a9e969] Creating Snapshot of the VM instance {{(pid=63515) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 942.647516] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-5578160d-1806-4df1-8044-2b5a84690b1b {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.656044] env[63515]: DEBUG oslo_vmware.api [None req-8e6475b3-95eb-4b44-a2f3-5ddd5db39f52 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Waiting for the task: (returnval){ [ 942.656044] env[63515]: value = "task-1111627" [ 942.656044] env[63515]: _type = "Task" [ 942.656044] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 942.663709] env[63515]: DEBUG oslo_vmware.api [None req-8e6475b3-95eb-4b44-a2f3-5ddd5db39f52 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Task: {'id': task-1111627, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.762421] env[63515]: DEBUG nova.network.neutron [None req-b94389d0-616d-4e7e-a415-7b0c3a16a683 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: 829e72d7-a56a-451c-b98a-94253e5900d5] Updating instance_info_cache with network_info: [{"id": "856de677-b1c8-479b-91d9-34d31a703a63", "address": "fa:16:3e:13:27:b4", "network": {"id": "15fdaabe-6793-400f-9737-2cdff8f07238", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1175204866-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f9c98e9f6020475490aaa7e76d907ab7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7d377d75-3add-4a15-8691-74b2eb010924", "external-id": "nsx-vlan-transportzone-71", "segmentation_id": 71, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap856de677-b1", "ovs_interfaceid": "856de677-b1c8-479b-91d9-34d31a703a63", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 942.773502] env[63515]: DEBUG oslo_vmware.api [None req-9550b400-f3b1-456b-9e7a-27ee304f516f tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Task: {'id': task-1111623, 'name': CloneVM_Task} progress is 94%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.921351] env[63515]: DEBUG oslo_vmware.api [None req-be7cccf6-3f5b-4895-b678-cf60de2c827a tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Task: {'id': task-1111626, 'name': Rename_Task, 'duration_secs': 0.137671} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 942.921636] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-be7cccf6-3f5b-4895-b678-cf60de2c827a tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: 014720fe-87b9-4e55-b9ba-295b009c7ff6] Powering on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 942.921892] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d9863367-5fa2-42e6-a64d-245e0dbc43ed {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.931048] env[63515]: DEBUG oslo_vmware.api [None req-be7cccf6-3f5b-4895-b678-cf60de2c827a tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Waiting for the task: (returnval){ [ 942.931048] env[63515]: value = "task-1111628" [ 942.931048] env[63515]: _type = "Task" [ 942.931048] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 942.938820] env[63515]: DEBUG oslo_vmware.api [None req-be7cccf6-3f5b-4895-b678-cf60de2c827a tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Task: {'id': task-1111628, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.947811] env[63515]: ERROR nova.scheduler.client.report [None req-56611604-1203-4b20-9871-2ec521700550 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [req-f6472bbe-551e-41c2-a239-13942b9eb7e8] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 2bd86232-2b6f-44d5-9057-1a3a6b27185a. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-f6472bbe-551e-41c2-a239-13942b9eb7e8"}]} [ 942.966182] env[63515]: DEBUG oslo_vmware.api [None req-7db48089-ba7f-4a2a-9098-e95b53997aff tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Task: {'id': task-1111622, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.970912] env[63515]: DEBUG nova.scheduler.client.report [None req-56611604-1203-4b20-9871-2ec521700550 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Refreshing inventories for resource provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 942.990872] env[63515]: DEBUG nova.scheduler.client.report [None req-56611604-1203-4b20-9871-2ec521700550 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Updating ProviderTree inventory for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 942.991175] env[63515]: DEBUG nova.compute.provider_tree [None req-56611604-1203-4b20-9871-2ec521700550 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Updating inventory in ProviderTree for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 943.003935] env[63515]: DEBUG nova.scheduler.client.report [None req-56611604-1203-4b20-9871-2ec521700550 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Refreshing aggregate associations for resource provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a, aggregates: None {{(pid=63515) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 943.023695] env[63515]: DEBUG nova.scheduler.client.report [None req-56611604-1203-4b20-9871-2ec521700550 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Refreshing trait associations for resource provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a, traits: COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=63515) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 943.167710] env[63515]: DEBUG oslo_vmware.api [None req-8e6475b3-95eb-4b44-a2f3-5ddd5db39f52 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Task: {'id': task-1111627, 'name': CreateSnapshot_Task, 'duration_secs': 0.491221} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 943.167897] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-8e6475b3-95eb-4b44-a2f3-5ddd5db39f52 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 68766cd4-84be-475b-8494-d7ab43a9e969] Created Snapshot of the VM instance {{(pid=63515) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 943.168663] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ec35676-615e-423f-a5be-f87d48ca7c8a {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.263983] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7185a8b-9832-4b18-af64-c1f98bc443d0 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.271553] env[63515]: DEBUG oslo_concurrency.lockutils [None req-b94389d0-616d-4e7e-a415-7b0c3a16a683 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Releasing lock "refresh_cache-829e72d7-a56a-451c-b98a-94253e5900d5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 943.271856] env[63515]: DEBUG nova.objects.instance [None req-b94389d0-616d-4e7e-a415-7b0c3a16a683 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Lazy-loading 'migration_context' on Instance uuid 829e72d7-a56a-451c-b98a-94253e5900d5 {{(pid=63515) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 943.279430] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8b3133b-6c12-4250-813c-a1ddc6bd93d4 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.282528] env[63515]: DEBUG oslo_vmware.api [None req-9550b400-f3b1-456b-9e7a-27ee304f516f tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Task: {'id': task-1111623, 'name': CloneVM_Task, 'duration_secs': 1.892729} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 943.282528] env[63515]: INFO nova.virt.vmwareapi.vmops [None req-9550b400-f3b1-456b-9e7a-27ee304f516f tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: 2fa098b6-ecc0-491f-8cfc-5ee335bbf943] Created linked-clone VM from snapshot [ 943.283339] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f214805-fd73-49ff-9120-72f5500be8e1 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.314561] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f881f3d-f1e1-462b-a0eb-c0752312d6e9 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.317486] env[63515]: DEBUG nova.virt.vmwareapi.images [None req-9550b400-f3b1-456b-9e7a-27ee304f516f tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: 2fa098b6-ecc0-491f-8cfc-5ee335bbf943] Uploading image e8441460-0adf-484c-ab99-e706ea197728 {{(pid=63515) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 943.325069] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9c0fed4-ea62-4fa3-8bfa-f7b07435002b {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.330284] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-9550b400-f3b1-456b-9e7a-27ee304f516f tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: 2fa098b6-ecc0-491f-8cfc-5ee335bbf943] Destroying the VM {{(pid=63515) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 943.330559] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-61c7b928-6ca7-420c-a67e-3e1780b3d9e4 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.340185] env[63515]: DEBUG nova.compute.provider_tree [None req-56611604-1203-4b20-9871-2ec521700550 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Updating inventory in ProviderTree for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 943.346681] env[63515]: DEBUG oslo_vmware.api [None req-9550b400-f3b1-456b-9e7a-27ee304f516f tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Waiting for the task: (returnval){ [ 943.346681] env[63515]: value = "task-1111629" [ 943.346681] env[63515]: _type = "Task" [ 943.346681] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 943.354796] env[63515]: DEBUG oslo_vmware.api [None req-9550b400-f3b1-456b-9e7a-27ee304f516f tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Task: {'id': task-1111629, 'name': Destroy_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.441265] env[63515]: DEBUG oslo_vmware.api [None req-be7cccf6-3f5b-4895-b678-cf60de2c827a tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Task: {'id': task-1111628, 'name': PowerOnVM_Task} progress is 99%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.465751] env[63515]: DEBUG oslo_vmware.api [None req-7db48089-ba7f-4a2a-9098-e95b53997aff tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Task: {'id': task-1111622, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.618592] env[63515]: DEBUG nova.compute.manager [None req-8fdfb203-5786-4efd-8927-4ee465e68e44 tempest-ServerMetadataTestJSON-1034558863 tempest-ServerMetadataTestJSON-1034558863-project-member] [instance: b25bf2ba-cf08-44ac-a524-aa87cf785575] Start spawning the instance on the hypervisor. {{(pid=63515) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 943.644221] env[63515]: DEBUG nova.virt.hardware [None req-8fdfb203-5786-4efd-8927-4ee465e68e44 tempest-ServerMetadataTestJSON-1034558863 tempest-ServerMetadataTestJSON-1034558863-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T02:52:14Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T02:51:56Z,direct_url=,disk_format='vmdk',id=8a120570-cb06-4099-b262-554ca0ad15c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b569255cc43e42e7ae3f2b4ad37c6ef4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T02:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 943.644600] env[63515]: DEBUG nova.virt.hardware [None req-8fdfb203-5786-4efd-8927-4ee465e68e44 tempest-ServerMetadataTestJSON-1034558863 tempest-ServerMetadataTestJSON-1034558863-project-member] Flavor limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 943.644771] env[63515]: DEBUG nova.virt.hardware [None req-8fdfb203-5786-4efd-8927-4ee465e68e44 tempest-ServerMetadataTestJSON-1034558863 tempest-ServerMetadataTestJSON-1034558863-project-member] Image limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 943.644981] env[63515]: DEBUG nova.virt.hardware [None req-8fdfb203-5786-4efd-8927-4ee465e68e44 tempest-ServerMetadataTestJSON-1034558863 tempest-ServerMetadataTestJSON-1034558863-project-member] Flavor pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 943.645238] env[63515]: DEBUG nova.virt.hardware [None req-8fdfb203-5786-4efd-8927-4ee465e68e44 tempest-ServerMetadataTestJSON-1034558863 tempest-ServerMetadataTestJSON-1034558863-project-member] Image pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 943.645451] env[63515]: DEBUG nova.virt.hardware [None req-8fdfb203-5786-4efd-8927-4ee465e68e44 tempest-ServerMetadataTestJSON-1034558863 tempest-ServerMetadataTestJSON-1034558863-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 943.645757] env[63515]: DEBUG nova.virt.hardware [None req-8fdfb203-5786-4efd-8927-4ee465e68e44 tempest-ServerMetadataTestJSON-1034558863 tempest-ServerMetadataTestJSON-1034558863-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 943.645991] env[63515]: DEBUG nova.virt.hardware [None req-8fdfb203-5786-4efd-8927-4ee465e68e44 tempest-ServerMetadataTestJSON-1034558863 tempest-ServerMetadataTestJSON-1034558863-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 943.646245] env[63515]: DEBUG nova.virt.hardware [None req-8fdfb203-5786-4efd-8927-4ee465e68e44 tempest-ServerMetadataTestJSON-1034558863 tempest-ServerMetadataTestJSON-1034558863-project-member] Got 1 possible topologies {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 943.646471] env[63515]: DEBUG nova.virt.hardware [None req-8fdfb203-5786-4efd-8927-4ee465e68e44 tempest-ServerMetadataTestJSON-1034558863 tempest-ServerMetadataTestJSON-1034558863-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 943.646653] env[63515]: DEBUG nova.virt.hardware [None req-8fdfb203-5786-4efd-8927-4ee465e68e44 tempest-ServerMetadataTestJSON-1034558863 tempest-ServerMetadataTestJSON-1034558863-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 943.647565] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e082330-a39c-4a06-85d9-855b78738d51 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.655419] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-546712a5-e5a4-4adb-b166-9cc32180345d {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.687254] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-8e6475b3-95eb-4b44-a2f3-5ddd5db39f52 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 68766cd4-84be-475b-8494-d7ab43a9e969] Creating linked-clone VM from snapshot {{(pid=63515) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 943.687603] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-c9276a86-c607-4f06-8cd7-5aad52c97cb5 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.696856] env[63515]: DEBUG oslo_vmware.api [None req-8e6475b3-95eb-4b44-a2f3-5ddd5db39f52 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Waiting for the task: (returnval){ [ 943.696856] env[63515]: value = "task-1111630" [ 943.696856] env[63515]: _type = "Task" [ 943.696856] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 943.704978] env[63515]: DEBUG oslo_vmware.api [None req-8e6475b3-95eb-4b44-a2f3-5ddd5db39f52 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Task: {'id': task-1111630, 'name': CloneVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.775401] env[63515]: DEBUG nova.objects.base [None req-b94389d0-616d-4e7e-a415-7b0c3a16a683 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Object Instance<829e72d7-a56a-451c-b98a-94253e5900d5> lazy-loaded attributes: info_cache,migration_context {{(pid=63515) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 943.776648] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-beb9b094-904e-4ad4-ab23-74ea0137299a {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.796717] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a5ab83df-5d24-4151-81ef-3775fb7a8205 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.802789] env[63515]: DEBUG oslo_vmware.api [None req-b94389d0-616d-4e7e-a415-7b0c3a16a683 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Waiting for the task: (returnval){ [ 943.802789] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]5266e565-c259-b01a-2a72-8a748d7b3a34" [ 943.802789] env[63515]: _type = "Task" [ 943.802789] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 943.811083] env[63515]: DEBUG oslo_vmware.api [None req-b94389d0-616d-4e7e-a415-7b0c3a16a683 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]5266e565-c259-b01a-2a72-8a748d7b3a34, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.857156] env[63515]: DEBUG oslo_vmware.api [None req-9550b400-f3b1-456b-9e7a-27ee304f516f tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Task: {'id': task-1111629, 'name': Destroy_Task} progress is 33%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.878052] env[63515]: DEBUG nova.scheduler.client.report [None req-56611604-1203-4b20-9871-2ec521700550 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Updated inventory for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a with generation 125 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 943.878052] env[63515]: DEBUG nova.compute.provider_tree [None req-56611604-1203-4b20-9871-2ec521700550 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Updating resource provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a generation from 125 to 126 during operation: update_inventory {{(pid=63515) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 943.878052] env[63515]: DEBUG nova.compute.provider_tree [None req-56611604-1203-4b20-9871-2ec521700550 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Updating inventory in ProviderTree for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 943.942443] env[63515]: DEBUG oslo_vmware.api [None req-be7cccf6-3f5b-4895-b678-cf60de2c827a tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Task: {'id': task-1111628, 'name': PowerOnVM_Task, 'duration_secs': 0.688593} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 943.942784] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-be7cccf6-3f5b-4895-b678-cf60de2c827a tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: 014720fe-87b9-4e55-b9ba-295b009c7ff6] Powered on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 943.943075] env[63515]: INFO nova.compute.manager [None req-be7cccf6-3f5b-4895-b678-cf60de2c827a tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: 014720fe-87b9-4e55-b9ba-295b009c7ff6] Took 9.39 seconds to spawn the instance on the hypervisor. [ 943.943334] env[63515]: DEBUG nova.compute.manager [None req-be7cccf6-3f5b-4895-b678-cf60de2c827a tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: 014720fe-87b9-4e55-b9ba-295b009c7ff6] Checking state {{(pid=63515) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 943.945246] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b82152d8-ca3a-4d53-8c8d-b33ca8fb2281 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.948986] env[63515]: DEBUG nova.compute.manager [req-64b3dd7e-2fec-42d8-8e1f-3c07e8b6943f req-e42c9c2c-0869-488a-aeb7-72c574c455ba service nova] [instance: b25bf2ba-cf08-44ac-a524-aa87cf785575] Received event network-vif-plugged-5d351a64-ccc8-4d4c-9057-7ef8d2c3edbd {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 943.949252] env[63515]: DEBUG oslo_concurrency.lockutils [req-64b3dd7e-2fec-42d8-8e1f-3c07e8b6943f req-e42c9c2c-0869-488a-aeb7-72c574c455ba service nova] Acquiring lock "b25bf2ba-cf08-44ac-a524-aa87cf785575-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 943.949530] env[63515]: DEBUG oslo_concurrency.lockutils [req-64b3dd7e-2fec-42d8-8e1f-3c07e8b6943f req-e42c9c2c-0869-488a-aeb7-72c574c455ba service nova] Lock "b25bf2ba-cf08-44ac-a524-aa87cf785575-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 943.949749] env[63515]: DEBUG oslo_concurrency.lockutils [req-64b3dd7e-2fec-42d8-8e1f-3c07e8b6943f req-e42c9c2c-0869-488a-aeb7-72c574c455ba service nova] Lock "b25bf2ba-cf08-44ac-a524-aa87cf785575-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 943.949954] env[63515]: DEBUG nova.compute.manager [req-64b3dd7e-2fec-42d8-8e1f-3c07e8b6943f req-e42c9c2c-0869-488a-aeb7-72c574c455ba service nova] [instance: b25bf2ba-cf08-44ac-a524-aa87cf785575] No waiting events found dispatching network-vif-plugged-5d351a64-ccc8-4d4c-9057-7ef8d2c3edbd {{(pid=63515) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 943.950475] env[63515]: WARNING nova.compute.manager [req-64b3dd7e-2fec-42d8-8e1f-3c07e8b6943f req-e42c9c2c-0869-488a-aeb7-72c574c455ba service nova] [instance: b25bf2ba-cf08-44ac-a524-aa87cf785575] Received unexpected event network-vif-plugged-5d351a64-ccc8-4d4c-9057-7ef8d2c3edbd for instance with vm_state building and task_state spawning. [ 943.965397] env[63515]: DEBUG oslo_vmware.api [None req-7db48089-ba7f-4a2a-9098-e95b53997aff tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Task: {'id': task-1111622, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.043337] env[63515]: DEBUG nova.network.neutron [None req-8fdfb203-5786-4efd-8927-4ee465e68e44 tempest-ServerMetadataTestJSON-1034558863 tempest-ServerMetadataTestJSON-1034558863-project-member] [instance: b25bf2ba-cf08-44ac-a524-aa87cf785575] Successfully updated port: 5d351a64-ccc8-4d4c-9057-7ef8d2c3edbd {{(pid=63515) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 944.207022] env[63515]: DEBUG oslo_vmware.api [None req-8e6475b3-95eb-4b44-a2f3-5ddd5db39f52 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Task: {'id': task-1111630, 'name': CloneVM_Task} progress is 94%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.312943] env[63515]: DEBUG oslo_vmware.api [None req-b94389d0-616d-4e7e-a415-7b0c3a16a683 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]5266e565-c259-b01a-2a72-8a748d7b3a34, 'name': SearchDatastore_Task, 'duration_secs': 0.007706} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 944.313601] env[63515]: DEBUG oslo_concurrency.lockutils [None req-b94389d0-616d-4e7e-a415-7b0c3a16a683 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 944.359501] env[63515]: DEBUG oslo_vmware.api [None req-9550b400-f3b1-456b-9e7a-27ee304f516f tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Task: {'id': task-1111629, 'name': Destroy_Task, 'duration_secs': 0.538801} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 944.359783] env[63515]: INFO nova.virt.vmwareapi.vm_util [None req-9550b400-f3b1-456b-9e7a-27ee304f516f tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: 2fa098b6-ecc0-491f-8cfc-5ee335bbf943] Destroyed the VM [ 944.360046] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-9550b400-f3b1-456b-9e7a-27ee304f516f tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: 2fa098b6-ecc0-491f-8cfc-5ee335bbf943] Deleting Snapshot of the VM instance {{(pid=63515) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 944.360321] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-ec10eb91-630e-4976-a2b9-dd41ba41883c {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.366223] env[63515]: DEBUG oslo_vmware.api [None req-9550b400-f3b1-456b-9e7a-27ee304f516f tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Waiting for the task: (returnval){ [ 944.366223] env[63515]: value = "task-1111631" [ 944.366223] env[63515]: _type = "Task" [ 944.366223] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 944.374744] env[63515]: DEBUG oslo_vmware.api [None req-9550b400-f3b1-456b-9e7a-27ee304f516f tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Task: {'id': task-1111631, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.381771] env[63515]: DEBUG oslo_concurrency.lockutils [None req-56611604-1203-4b20-9871-2ec521700550 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.784s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 944.384408] env[63515]: DEBUG oslo_concurrency.lockutils [None req-127af73c-afc8-4436-9075-a886f5e9c7d5 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 8.415s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 944.386067] env[63515]: INFO nova.compute.claims [None req-127af73c-afc8-4436-9075-a886f5e9c7d5 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] [instance: 7d572ac0-f6af-4622-96cc-e75983420222] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 944.404203] env[63515]: INFO nova.scheduler.client.report [None req-56611604-1203-4b20-9871-2ec521700550 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Deleted allocations for instance a2203c5e-66b9-4c7e-8d76-e12a08b57b1d [ 944.467165] env[63515]: INFO nova.compute.manager [None req-be7cccf6-3f5b-4895-b678-cf60de2c827a tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: 014720fe-87b9-4e55-b9ba-295b009c7ff6] Took 18.73 seconds to build instance. [ 944.475463] env[63515]: DEBUG oslo_vmware.api [None req-7db48089-ba7f-4a2a-9098-e95b53997aff tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Task: {'id': task-1111622, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.543691] env[63515]: DEBUG oslo_concurrency.lockutils [None req-8fdfb203-5786-4efd-8927-4ee465e68e44 tempest-ServerMetadataTestJSON-1034558863 tempest-ServerMetadataTestJSON-1034558863-project-member] Acquiring lock "refresh_cache-b25bf2ba-cf08-44ac-a524-aa87cf785575" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 944.543928] env[63515]: DEBUG oslo_concurrency.lockutils [None req-8fdfb203-5786-4efd-8927-4ee465e68e44 tempest-ServerMetadataTestJSON-1034558863 tempest-ServerMetadataTestJSON-1034558863-project-member] Acquired lock "refresh_cache-b25bf2ba-cf08-44ac-a524-aa87cf785575" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 944.544075] env[63515]: DEBUG nova.network.neutron [None req-8fdfb203-5786-4efd-8927-4ee465e68e44 tempest-ServerMetadataTestJSON-1034558863 tempest-ServerMetadataTestJSON-1034558863-project-member] [instance: b25bf2ba-cf08-44ac-a524-aa87cf785575] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 944.708439] env[63515]: DEBUG oslo_vmware.api [None req-8e6475b3-95eb-4b44-a2f3-5ddd5db39f52 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Task: {'id': task-1111630, 'name': CloneVM_Task} progress is 94%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.875823] env[63515]: DEBUG oslo_vmware.api [None req-9550b400-f3b1-456b-9e7a-27ee304f516f tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Task: {'id': task-1111631, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.911439] env[63515]: DEBUG oslo_concurrency.lockutils [None req-56611604-1203-4b20-9871-2ec521700550 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Lock "a2203c5e-66b9-4c7e-8d76-e12a08b57b1d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 13.245s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 944.968844] env[63515]: DEBUG oslo_vmware.api [None req-7db48089-ba7f-4a2a-9098-e95b53997aff tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Task: {'id': task-1111622, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.969271] env[63515]: DEBUG oslo_concurrency.lockutils [None req-be7cccf6-3f5b-4895-b678-cf60de2c827a tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Lock "014720fe-87b9-4e55-b9ba-295b009c7ff6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 20.245s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 945.015590] env[63515]: DEBUG oslo_concurrency.lockutils [None req-ce4a187a-30cc-42fa-8b30-3e62a1b3d589 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Acquiring lock "014720fe-87b9-4e55-b9ba-295b009c7ff6" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 945.015910] env[63515]: DEBUG oslo_concurrency.lockutils [None req-ce4a187a-30cc-42fa-8b30-3e62a1b3d589 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Lock "014720fe-87b9-4e55-b9ba-295b009c7ff6" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 945.016186] env[63515]: DEBUG nova.compute.manager [None req-ce4a187a-30cc-42fa-8b30-3e62a1b3d589 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: 014720fe-87b9-4e55-b9ba-295b009c7ff6] Checking state {{(pid=63515) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 945.017113] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f588828-36ec-429a-b414-c405e2b5aac5 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.023748] env[63515]: DEBUG nova.compute.manager [None req-ce4a187a-30cc-42fa-8b30-3e62a1b3d589 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: 014720fe-87b9-4e55-b9ba-295b009c7ff6] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=63515) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 945.024382] env[63515]: DEBUG nova.objects.instance [None req-ce4a187a-30cc-42fa-8b30-3e62a1b3d589 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Lazy-loading 'flavor' on Instance uuid 014720fe-87b9-4e55-b9ba-295b009c7ff6 {{(pid=63515) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 945.079960] env[63515]: DEBUG nova.network.neutron [None req-8fdfb203-5786-4efd-8927-4ee465e68e44 tempest-ServerMetadataTestJSON-1034558863 tempest-ServerMetadataTestJSON-1034558863-project-member] [instance: b25bf2ba-cf08-44ac-a524-aa87cf785575] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 945.210917] env[63515]: DEBUG oslo_vmware.api [None req-8e6475b3-95eb-4b44-a2f3-5ddd5db39f52 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Task: {'id': task-1111630, 'name': CloneVM_Task} progress is 95%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.239185] env[63515]: DEBUG nova.network.neutron [None req-8fdfb203-5786-4efd-8927-4ee465e68e44 tempest-ServerMetadataTestJSON-1034558863 tempest-ServerMetadataTestJSON-1034558863-project-member] [instance: b25bf2ba-cf08-44ac-a524-aa87cf785575] Updating instance_info_cache with network_info: [{"id": "5d351a64-ccc8-4d4c-9057-7ef8d2c3edbd", "address": "fa:16:3e:57:0b:0b", "network": {"id": "7812a19b-9968-4a98-b3f9-2827425949d5", "bridge": "br-int", "label": "tempest-ServerMetadataTestJSON-846086881-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "49ee798463ab46c2b84738523ba7d096", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bd8c6be9-575e-4605-b779-98606281a3bf", "external-id": "nsx-vlan-transportzone-273", "segmentation_id": 273, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5d351a64-cc", "ovs_interfaceid": "5d351a64-ccc8-4d4c-9057-7ef8d2c3edbd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 945.375960] env[63515]: DEBUG oslo_vmware.api [None req-9550b400-f3b1-456b-9e7a-27ee304f516f tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Task: {'id': task-1111631, 'name': RemoveSnapshot_Task} progress is 97%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.468796] env[63515]: DEBUG oslo_vmware.api [None req-7db48089-ba7f-4a2a-9098-e95b53997aff tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Task: {'id': task-1111622, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.528938] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-ce4a187a-30cc-42fa-8b30-3e62a1b3d589 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: 014720fe-87b9-4e55-b9ba-295b009c7ff6] Powering off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 945.529230] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4375058b-aa11-478d-b023-33297dbd549a {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.538072] env[63515]: DEBUG oslo_vmware.api [None req-ce4a187a-30cc-42fa-8b30-3e62a1b3d589 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Waiting for the task: (returnval){ [ 945.538072] env[63515]: value = "task-1111632" [ 945.538072] env[63515]: _type = "Task" [ 945.538072] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 945.547819] env[63515]: DEBUG oslo_vmware.api [None req-ce4a187a-30cc-42fa-8b30-3e62a1b3d589 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Task: {'id': task-1111632, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.633521] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd2c055a-57f2-46d7-af06-c99e4a998c83 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.641940] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-427e19a0-36fc-4c14-953f-03b44c019a96 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.679272] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78b9e3c1-37a6-4f09-a2eb-1dad6431d6a8 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.689953] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52b34151-d109-4973-9d3d-d3f6d24ecf13 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.703593] env[63515]: DEBUG nova.compute.provider_tree [None req-127af73c-afc8-4436-9075-a886f5e9c7d5 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Updating inventory in ProviderTree for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 945.712745] env[63515]: DEBUG oslo_vmware.api [None req-8e6475b3-95eb-4b44-a2f3-5ddd5db39f52 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Task: {'id': task-1111630, 'name': CloneVM_Task, 'duration_secs': 1.82684} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 945.713709] env[63515]: INFO nova.virt.vmwareapi.vmops [None req-8e6475b3-95eb-4b44-a2f3-5ddd5db39f52 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 68766cd4-84be-475b-8494-d7ab43a9e969] Created linked-clone VM from snapshot [ 945.714568] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6f1fb8f-f497-4d82-85d6-0f6711bdfebf {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.722494] env[63515]: DEBUG nova.virt.vmwareapi.images [None req-8e6475b3-95eb-4b44-a2f3-5ddd5db39f52 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 68766cd4-84be-475b-8494-d7ab43a9e969] Uploading image 4978423f-d581-473b-85c6-34574f05d1d5 {{(pid=63515) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 945.742052] env[63515]: DEBUG oslo_concurrency.lockutils [None req-8fdfb203-5786-4efd-8927-4ee465e68e44 tempest-ServerMetadataTestJSON-1034558863 tempest-ServerMetadataTestJSON-1034558863-project-member] Releasing lock "refresh_cache-b25bf2ba-cf08-44ac-a524-aa87cf785575" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 945.742396] env[63515]: DEBUG nova.compute.manager [None req-8fdfb203-5786-4efd-8927-4ee465e68e44 tempest-ServerMetadataTestJSON-1034558863 tempest-ServerMetadataTestJSON-1034558863-project-member] [instance: b25bf2ba-cf08-44ac-a524-aa87cf785575] Instance network_info: |[{"id": "5d351a64-ccc8-4d4c-9057-7ef8d2c3edbd", "address": "fa:16:3e:57:0b:0b", "network": {"id": "7812a19b-9968-4a98-b3f9-2827425949d5", "bridge": "br-int", "label": "tempest-ServerMetadataTestJSON-846086881-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "49ee798463ab46c2b84738523ba7d096", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bd8c6be9-575e-4605-b779-98606281a3bf", "external-id": "nsx-vlan-transportzone-273", "segmentation_id": 273, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5d351a64-cc", "ovs_interfaceid": "5d351a64-ccc8-4d4c-9057-7ef8d2c3edbd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 945.744356] env[63515]: DEBUG oslo_vmware.rw_handles [None req-8e6475b3-95eb-4b44-a2f3-5ddd5db39f52 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 945.744356] env[63515]: value = "vm-243537" [ 945.744356] env[63515]: _type = "VirtualMachine" [ 945.744356] env[63515]: }. {{(pid=63515) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 945.744768] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-8fdfb203-5786-4efd-8927-4ee465e68e44 tempest-ServerMetadataTestJSON-1034558863 tempest-ServerMetadataTestJSON-1034558863-project-member] [instance: b25bf2ba-cf08-44ac-a524-aa87cf785575] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:57:0b:0b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'bd8c6be9-575e-4605-b779-98606281a3bf', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '5d351a64-ccc8-4d4c-9057-7ef8d2c3edbd', 'vif_model': 'vmxnet3'}] {{(pid=63515) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 945.752178] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-8fdfb203-5786-4efd-8927-4ee465e68e44 tempest-ServerMetadataTestJSON-1034558863 tempest-ServerMetadataTestJSON-1034558863-project-member] Creating folder: Project (49ee798463ab46c2b84738523ba7d096). Parent ref: group-v243370. {{(pid=63515) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 945.752439] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-86c248da-2b50-4f10-9421-4174cf67798d {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.754103] env[63515]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-cd79dd1b-35a2-4229-8229-fe27c155b09c {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.761135] env[63515]: DEBUG oslo_vmware.rw_handles [None req-8e6475b3-95eb-4b44-a2f3-5ddd5db39f52 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Lease: (returnval){ [ 945.761135] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]52167ef8-b9d2-569b-42af-bc52c2dc7abe" [ 945.761135] env[63515]: _type = "HttpNfcLease" [ 945.761135] env[63515]: } obtained for exporting VM: (result){ [ 945.761135] env[63515]: value = "vm-243537" [ 945.761135] env[63515]: _type = "VirtualMachine" [ 945.761135] env[63515]: }. {{(pid=63515) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 945.761402] env[63515]: DEBUG oslo_vmware.api [None req-8e6475b3-95eb-4b44-a2f3-5ddd5db39f52 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Waiting for the lease: (returnval){ [ 945.761402] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]52167ef8-b9d2-569b-42af-bc52c2dc7abe" [ 945.761402] env[63515]: _type = "HttpNfcLease" [ 945.761402] env[63515]: } to be ready. {{(pid=63515) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 945.766344] env[63515]: INFO nova.virt.vmwareapi.vm_util [None req-8fdfb203-5786-4efd-8927-4ee465e68e44 tempest-ServerMetadataTestJSON-1034558863 tempest-ServerMetadataTestJSON-1034558863-project-member] Created folder: Project (49ee798463ab46c2b84738523ba7d096) in parent group-v243370. [ 945.766552] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-8fdfb203-5786-4efd-8927-4ee465e68e44 tempest-ServerMetadataTestJSON-1034558863 tempest-ServerMetadataTestJSON-1034558863-project-member] Creating folder: Instances. Parent ref: group-v243538. {{(pid=63515) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 945.768029] env[63515]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a0442bf3-2522-432a-9c26-2ce12535c68b {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.769464] env[63515]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 945.769464] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]52167ef8-b9d2-569b-42af-bc52c2dc7abe" [ 945.769464] env[63515]: _type = "HttpNfcLease" [ 945.769464] env[63515]: } is initializing. {{(pid=63515) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 945.776793] env[63515]: INFO nova.virt.vmwareapi.vm_util [None req-8fdfb203-5786-4efd-8927-4ee465e68e44 tempest-ServerMetadataTestJSON-1034558863 tempest-ServerMetadataTestJSON-1034558863-project-member] Created folder: Instances in parent group-v243538. [ 945.777045] env[63515]: DEBUG oslo.service.loopingcall [None req-8fdfb203-5786-4efd-8927-4ee465e68e44 tempest-ServerMetadataTestJSON-1034558863 tempest-ServerMetadataTestJSON-1034558863-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 945.777563] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b25bf2ba-cf08-44ac-a524-aa87cf785575] Creating VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 945.777563] env[63515]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3fd2c356-0aaf-4f39-b886-091103f336d3 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.796189] env[63515]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 945.796189] env[63515]: value = "task-1111636" [ 945.796189] env[63515]: _type = "Task" [ 945.796189] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 945.805149] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1111636, 'name': CreateVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.879566] env[63515]: DEBUG oslo_vmware.api [None req-9550b400-f3b1-456b-9e7a-27ee304f516f tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Task: {'id': task-1111631, 'name': RemoveSnapshot_Task, 'duration_secs': 1.296805} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 945.880065] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-9550b400-f3b1-456b-9e7a-27ee304f516f tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: 2fa098b6-ecc0-491f-8cfc-5ee335bbf943] Deleted Snapshot of the VM instance {{(pid=63515) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 945.972288] env[63515]: DEBUG oslo_vmware.api [None req-7db48089-ba7f-4a2a-9098-e95b53997aff tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Task: {'id': task-1111622, 'name': ReconfigVM_Task} progress is 18%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.031267] env[63515]: DEBUG nova.compute.manager [req-fc506143-afa4-4dae-ae3d-13d2c21900b7 req-beea6dca-e014-437d-8945-d78db15ce696 service nova] [instance: b25bf2ba-cf08-44ac-a524-aa87cf785575] Received event network-changed-5d351a64-ccc8-4d4c-9057-7ef8d2c3edbd {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 946.031376] env[63515]: DEBUG nova.compute.manager [req-fc506143-afa4-4dae-ae3d-13d2c21900b7 req-beea6dca-e014-437d-8945-d78db15ce696 service nova] [instance: b25bf2ba-cf08-44ac-a524-aa87cf785575] Refreshing instance network info cache due to event network-changed-5d351a64-ccc8-4d4c-9057-7ef8d2c3edbd. {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 946.031533] env[63515]: DEBUG oslo_concurrency.lockutils [req-fc506143-afa4-4dae-ae3d-13d2c21900b7 req-beea6dca-e014-437d-8945-d78db15ce696 service nova] Acquiring lock "refresh_cache-b25bf2ba-cf08-44ac-a524-aa87cf785575" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 946.031678] env[63515]: DEBUG oslo_concurrency.lockutils [req-fc506143-afa4-4dae-ae3d-13d2c21900b7 req-beea6dca-e014-437d-8945-d78db15ce696 service nova] Acquired lock "refresh_cache-b25bf2ba-cf08-44ac-a524-aa87cf785575" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 946.031841] env[63515]: DEBUG nova.network.neutron [req-fc506143-afa4-4dae-ae3d-13d2c21900b7 req-beea6dca-e014-437d-8945-d78db15ce696 service nova] [instance: b25bf2ba-cf08-44ac-a524-aa87cf785575] Refreshing network info cache for port 5d351a64-ccc8-4d4c-9057-7ef8d2c3edbd {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 946.048141] env[63515]: DEBUG oslo_vmware.api [None req-ce4a187a-30cc-42fa-8b30-3e62a1b3d589 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Task: {'id': task-1111632, 'name': PowerOffVM_Task, 'duration_secs': 0.188306} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 946.048424] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-ce4a187a-30cc-42fa-8b30-3e62a1b3d589 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: 014720fe-87b9-4e55-b9ba-295b009c7ff6] Powered off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 946.048772] env[63515]: DEBUG nova.compute.manager [None req-ce4a187a-30cc-42fa-8b30-3e62a1b3d589 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: 014720fe-87b9-4e55-b9ba-295b009c7ff6] Checking state {{(pid=63515) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 946.049612] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7584a6bb-c0b9-4c37-a454-7f7f47f15b66 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.228919] env[63515]: ERROR nova.scheduler.client.report [None req-127af73c-afc8-4436-9075-a886f5e9c7d5 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] [req-f78a4757-87e2-4372-b184-428eaa02af6d] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 2bd86232-2b6f-44d5-9057-1a3a6b27185a. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-f78a4757-87e2-4372-b184-428eaa02af6d"}]} [ 946.246091] env[63515]: DEBUG nova.scheduler.client.report [None req-127af73c-afc8-4436-9075-a886f5e9c7d5 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Refreshing inventories for resource provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 946.260247] env[63515]: DEBUG nova.scheduler.client.report [None req-127af73c-afc8-4436-9075-a886f5e9c7d5 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Updating ProviderTree inventory for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 946.260529] env[63515]: DEBUG nova.compute.provider_tree [None req-127af73c-afc8-4436-9075-a886f5e9c7d5 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Updating inventory in ProviderTree for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 946.269258] env[63515]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 946.269258] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]52167ef8-b9d2-569b-42af-bc52c2dc7abe" [ 946.269258] env[63515]: _type = "HttpNfcLease" [ 946.269258] env[63515]: } is ready. {{(pid=63515) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 946.269558] env[63515]: DEBUG oslo_vmware.rw_handles [None req-8e6475b3-95eb-4b44-a2f3-5ddd5db39f52 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 946.269558] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]52167ef8-b9d2-569b-42af-bc52c2dc7abe" [ 946.269558] env[63515]: _type = "HttpNfcLease" [ 946.269558] env[63515]: }. {{(pid=63515) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 946.270375] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03b5d8a4-b35c-4f74-b11f-c8cb9d537531 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.273427] env[63515]: DEBUG nova.scheduler.client.report [None req-127af73c-afc8-4436-9075-a886f5e9c7d5 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Refreshing aggregate associations for resource provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a, aggregates: None {{(pid=63515) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 946.280433] env[63515]: DEBUG oslo_vmware.rw_handles [None req-8e6475b3-95eb-4b44-a2f3-5ddd5db39f52 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52c2c74c-f165-626c-5ca1-1950d805bd0e/disk-0.vmdk from lease info. {{(pid=63515) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 946.280625] env[63515]: DEBUG oslo_vmware.rw_handles [None req-8e6475b3-95eb-4b44-a2f3-5ddd5db39f52 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52c2c74c-f165-626c-5ca1-1950d805bd0e/disk-0.vmdk for reading. {{(pid=63515) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 946.339277] env[63515]: DEBUG nova.scheduler.client.report [None req-127af73c-afc8-4436-9075-a886f5e9c7d5 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Refreshing trait associations for resource provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a, traits: COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=63515) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 946.351257] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1111636, 'name': CreateVM_Task, 'duration_secs': 0.430755} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 946.351452] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b25bf2ba-cf08-44ac-a524-aa87cf785575] Created VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 946.352171] env[63515]: DEBUG oslo_concurrency.lockutils [None req-8fdfb203-5786-4efd-8927-4ee465e68e44 tempest-ServerMetadataTestJSON-1034558863 tempest-ServerMetadataTestJSON-1034558863-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 946.352353] env[63515]: DEBUG oslo_concurrency.lockutils [None req-8fdfb203-5786-4efd-8927-4ee465e68e44 tempest-ServerMetadataTestJSON-1034558863 tempest-ServerMetadataTestJSON-1034558863-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 946.352687] env[63515]: DEBUG oslo_concurrency.lockutils [None req-8fdfb203-5786-4efd-8927-4ee465e68e44 tempest-ServerMetadataTestJSON-1034558863 tempest-ServerMetadataTestJSON-1034558863-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 946.352960] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c921173b-8094-482f-889b-4c8b76d41395 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.357733] env[63515]: DEBUG oslo_vmware.api [None req-8fdfb203-5786-4efd-8927-4ee465e68e44 tempest-ServerMetadataTestJSON-1034558863 tempest-ServerMetadataTestJSON-1034558863-project-member] Waiting for the task: (returnval){ [ 946.357733] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]524d41f3-2f83-d66f-588e-473a758227b5" [ 946.357733] env[63515]: _type = "Task" [ 946.357733] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 946.365613] env[63515]: DEBUG oslo_vmware.api [None req-8fdfb203-5786-4efd-8927-4ee465e68e44 tempest-ServerMetadataTestJSON-1034558863 tempest-ServerMetadataTestJSON-1034558863-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]524d41f3-2f83-d66f-588e-473a758227b5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.374535] env[63515]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-8d50ab87-1eb7-4262-911a-c694e9f7955d {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.386266] env[63515]: WARNING nova.compute.manager [None req-9550b400-f3b1-456b-9e7a-27ee304f516f tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: 2fa098b6-ecc0-491f-8cfc-5ee335bbf943] Image not found during snapshot: nova.exception.ImageNotFound: Image e8441460-0adf-484c-ab99-e706ea197728 could not be found. [ 946.471674] env[63515]: DEBUG oslo_vmware.api [None req-7db48089-ba7f-4a2a-9098-e95b53997aff tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Task: {'id': task-1111622, 'name': ReconfigVM_Task, 'duration_secs': 5.811961} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 946.471933] env[63515]: DEBUG oslo_concurrency.lockutils [None req-7db48089-ba7f-4a2a-9098-e95b53997aff tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Releasing lock "eab16df9-7bb5-4576-bca0-769a561c5fe9" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 946.472168] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-7db48089-ba7f-4a2a-9098-e95b53997aff tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: eab16df9-7bb5-4576-bca0-769a561c5fe9] Reconfigured VM to detach interface {{(pid=63515) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1978}} [ 946.498734] env[63515]: DEBUG oslo_concurrency.lockutils [None req-165bb947-b1a3-4f80-91f4-6990bed3b114 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Acquiring lock "d024b6b2-dd10-4112-89e8-ced57efe8208" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 946.498892] env[63515]: DEBUG oslo_concurrency.lockutils [None req-165bb947-b1a3-4f80-91f4-6990bed3b114 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Lock "d024b6b2-dd10-4112-89e8-ced57efe8208" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 946.562111] env[63515]: DEBUG oslo_concurrency.lockutils [None req-ce4a187a-30cc-42fa-8b30-3e62a1b3d589 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Lock "014720fe-87b9-4e55-b9ba-295b009c7ff6" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.546s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 946.593386] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d3a8e60-6044-4840-b6b3-486b2c616e35 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.601645] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e0c30e7-6ee9-4aa8-a7f0-5cd3eb213b88 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.634589] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a109fc8-02c1-4762-bff9-0bd893d4c15b {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.642638] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8d8fdbc-2482-4ba4-8948-aabd891fa230 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.656316] env[63515]: DEBUG nova.compute.provider_tree [None req-127af73c-afc8-4436-9075-a886f5e9c7d5 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Updating inventory in ProviderTree for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 946.771286] env[63515]: DEBUG nova.network.neutron [req-fc506143-afa4-4dae-ae3d-13d2c21900b7 req-beea6dca-e014-437d-8945-d78db15ce696 service nova] [instance: b25bf2ba-cf08-44ac-a524-aa87cf785575] Updated VIF entry in instance network info cache for port 5d351a64-ccc8-4d4c-9057-7ef8d2c3edbd. {{(pid=63515) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 946.771796] env[63515]: DEBUG nova.network.neutron [req-fc506143-afa4-4dae-ae3d-13d2c21900b7 req-beea6dca-e014-437d-8945-d78db15ce696 service nova] [instance: b25bf2ba-cf08-44ac-a524-aa87cf785575] Updating instance_info_cache with network_info: [{"id": "5d351a64-ccc8-4d4c-9057-7ef8d2c3edbd", "address": "fa:16:3e:57:0b:0b", "network": {"id": "7812a19b-9968-4a98-b3f9-2827425949d5", "bridge": "br-int", "label": "tempest-ServerMetadataTestJSON-846086881-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "49ee798463ab46c2b84738523ba7d096", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bd8c6be9-575e-4605-b779-98606281a3bf", "external-id": "nsx-vlan-transportzone-273", "segmentation_id": 273, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5d351a64-cc", "ovs_interfaceid": "5d351a64-ccc8-4d4c-9057-7ef8d2c3edbd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 946.870300] env[63515]: DEBUG oslo_vmware.api [None req-8fdfb203-5786-4efd-8927-4ee465e68e44 tempest-ServerMetadataTestJSON-1034558863 tempest-ServerMetadataTestJSON-1034558863-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]524d41f3-2f83-d66f-588e-473a758227b5, 'name': SearchDatastore_Task, 'duration_secs': 0.010345} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 946.870765] env[63515]: DEBUG oslo_concurrency.lockutils [None req-8fdfb203-5786-4efd-8927-4ee465e68e44 tempest-ServerMetadataTestJSON-1034558863 tempest-ServerMetadataTestJSON-1034558863-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 946.871039] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-8fdfb203-5786-4efd-8927-4ee465e68e44 tempest-ServerMetadataTestJSON-1034558863 tempest-ServerMetadataTestJSON-1034558863-project-member] [instance: b25bf2ba-cf08-44ac-a524-aa87cf785575] Processing image 8a120570-cb06-4099-b262-554ca0ad15c5 {{(pid=63515) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 946.871371] env[63515]: DEBUG oslo_concurrency.lockutils [None req-8fdfb203-5786-4efd-8927-4ee465e68e44 tempest-ServerMetadataTestJSON-1034558863 tempest-ServerMetadataTestJSON-1034558863-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 946.871593] env[63515]: DEBUG oslo_concurrency.lockutils [None req-8fdfb203-5786-4efd-8927-4ee465e68e44 tempest-ServerMetadataTestJSON-1034558863 tempest-ServerMetadataTestJSON-1034558863-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 946.871863] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-8fdfb203-5786-4efd-8927-4ee465e68e44 tempest-ServerMetadataTestJSON-1034558863 tempest-ServerMetadataTestJSON-1034558863-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 946.872208] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-af6bab71-defa-49b0-947f-eec6509d69d1 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.881322] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-8fdfb203-5786-4efd-8927-4ee465e68e44 tempest-ServerMetadataTestJSON-1034558863 tempest-ServerMetadataTestJSON-1034558863-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 946.881557] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-8fdfb203-5786-4efd-8927-4ee465e68e44 tempest-ServerMetadataTestJSON-1034558863 tempest-ServerMetadataTestJSON-1034558863-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63515) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 946.882440] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0751f7a3-37bb-448f-9d83-3093c6317fa3 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.888697] env[63515]: DEBUG oslo_vmware.api [None req-8fdfb203-5786-4efd-8927-4ee465e68e44 tempest-ServerMetadataTestJSON-1034558863 tempest-ServerMetadataTestJSON-1034558863-project-member] Waiting for the task: (returnval){ [ 946.888697] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]5228592c-a091-d822-8502-ae26f4c950be" [ 946.888697] env[63515]: _type = "Task" [ 946.888697] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 946.896728] env[63515]: DEBUG oslo_vmware.api [None req-8fdfb203-5786-4efd-8927-4ee465e68e44 tempest-ServerMetadataTestJSON-1034558863 tempest-ServerMetadataTestJSON-1034558863-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]5228592c-a091-d822-8502-ae26f4c950be, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.005203] env[63515]: DEBUG nova.compute.manager [None req-165bb947-b1a3-4f80-91f4-6990bed3b114 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: d024b6b2-dd10-4112-89e8-ced57efe8208] Starting instance... {{(pid=63515) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 947.194047] env[63515]: DEBUG nova.scheduler.client.report [None req-127af73c-afc8-4436-9075-a886f5e9c7d5 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Updated inventory for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a with generation 127 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 947.194457] env[63515]: DEBUG nova.compute.provider_tree [None req-127af73c-afc8-4436-9075-a886f5e9c7d5 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Updating resource provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a generation from 127 to 128 during operation: update_inventory {{(pid=63515) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 947.194720] env[63515]: DEBUG nova.compute.provider_tree [None req-127af73c-afc8-4436-9075-a886f5e9c7d5 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Updating inventory in ProviderTree for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 947.274594] env[63515]: DEBUG oslo_concurrency.lockutils [req-fc506143-afa4-4dae-ae3d-13d2c21900b7 req-beea6dca-e014-437d-8945-d78db15ce696 service nova] Releasing lock "refresh_cache-b25bf2ba-cf08-44ac-a524-aa87cf785575" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 947.400952] env[63515]: DEBUG oslo_vmware.api [None req-8fdfb203-5786-4efd-8927-4ee465e68e44 tempest-ServerMetadataTestJSON-1034558863 tempest-ServerMetadataTestJSON-1034558863-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]5228592c-a091-d822-8502-ae26f4c950be, 'name': SearchDatastore_Task, 'duration_secs': 0.014377} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 947.401853] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3382fa2a-f3af-4c39-bd63-e96676d53dc8 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.408410] env[63515]: DEBUG oslo_vmware.api [None req-8fdfb203-5786-4efd-8927-4ee465e68e44 tempest-ServerMetadataTestJSON-1034558863 tempest-ServerMetadataTestJSON-1034558863-project-member] Waiting for the task: (returnval){ [ 947.408410] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]52b844b7-5899-8a72-ec8a-9ae026a5a768" [ 947.408410] env[63515]: _type = "Task" [ 947.408410] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 947.416421] env[63515]: DEBUG oslo_vmware.api [None req-8fdfb203-5786-4efd-8927-4ee465e68e44 tempest-ServerMetadataTestJSON-1034558863 tempest-ServerMetadataTestJSON-1034558863-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52b844b7-5899-8a72-ec8a-9ae026a5a768, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.519859] env[63515]: DEBUG oslo_concurrency.lockutils [None req-0773e564-f19d-40eb-a88a-6e28c35f232c tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Acquiring lock "014720fe-87b9-4e55-b9ba-295b009c7ff6" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 947.520204] env[63515]: DEBUG oslo_concurrency.lockutils [None req-0773e564-f19d-40eb-a88a-6e28c35f232c tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Lock "014720fe-87b9-4e55-b9ba-295b009c7ff6" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 947.522989] env[63515]: DEBUG oslo_concurrency.lockutils [None req-0773e564-f19d-40eb-a88a-6e28c35f232c tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Acquiring lock "014720fe-87b9-4e55-b9ba-295b009c7ff6-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 947.523624] env[63515]: DEBUG oslo_concurrency.lockutils [None req-0773e564-f19d-40eb-a88a-6e28c35f232c tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Lock "014720fe-87b9-4e55-b9ba-295b009c7ff6-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 947.523818] env[63515]: DEBUG oslo_concurrency.lockutils [None req-0773e564-f19d-40eb-a88a-6e28c35f232c tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Lock "014720fe-87b9-4e55-b9ba-295b009c7ff6-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 947.526346] env[63515]: INFO nova.compute.manager [None req-0773e564-f19d-40eb-a88a-6e28c35f232c tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: 014720fe-87b9-4e55-b9ba-295b009c7ff6] Terminating instance [ 947.528514] env[63515]: DEBUG nova.compute.manager [None req-0773e564-f19d-40eb-a88a-6e28c35f232c tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: 014720fe-87b9-4e55-b9ba-295b009c7ff6] Start destroying the instance on the hypervisor. {{(pid=63515) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 947.528732] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-0773e564-f19d-40eb-a88a-6e28c35f232c tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: 014720fe-87b9-4e55-b9ba-295b009c7ff6] Destroying instance {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 947.530066] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1121afb1-2ef5-48d4-9585-d1f273fffdf5 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.538693] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-0773e564-f19d-40eb-a88a-6e28c35f232c tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: 014720fe-87b9-4e55-b9ba-295b009c7ff6] Unregistering the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 947.539822] env[63515]: DEBUG oslo_concurrency.lockutils [None req-165bb947-b1a3-4f80-91f4-6990bed3b114 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 947.540135] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-0b57c9ae-7bb8-4fea-adc5-a4789ed499b8 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.552263] env[63515]: DEBUG nova.compute.manager [req-ba527675-754d-400d-b509-eff7b0dc8952 req-1b7b37f8-bc15-4b25-a524-f350142909bf service nova] [instance: eab16df9-7bb5-4576-bca0-769a561c5fe9] Received event network-vif-deleted-ff4043f6-7576-4093-80ed-d36240b37f2f {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 947.552607] env[63515]: INFO nova.compute.manager [req-ba527675-754d-400d-b509-eff7b0dc8952 req-1b7b37f8-bc15-4b25-a524-f350142909bf service nova] [instance: eab16df9-7bb5-4576-bca0-769a561c5fe9] Neutron deleted interface ff4043f6-7576-4093-80ed-d36240b37f2f; detaching it from the instance and deleting it from the info cache [ 947.552871] env[63515]: DEBUG nova.network.neutron [req-ba527675-754d-400d-b509-eff7b0dc8952 req-1b7b37f8-bc15-4b25-a524-f350142909bf service nova] [instance: eab16df9-7bb5-4576-bca0-769a561c5fe9] Updating instance_info_cache with network_info: [{"id": "4168f5a5-c282-4f13-9c97-0b0236bd659b", "address": "fa:16:3e:a0:6a:50", "network": {"id": "88dfb29d-41e0-4df7-8f70-e52cda05620c", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1355578400-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.203", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7d4ada9ec35f42b19c6480a9101d21a4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "357d2811-e990-4985-9f9e-b158d10d3699", "external-id": "nsx-vlan-transportzone-641", "segmentation_id": 641, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4168f5a5-c2", "ovs_interfaceid": "4168f5a5-c282-4f13-9c97-0b0236bd659b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "e3c72582-41bf-48d4-8614-b516a96362a7", "address": "fa:16:3e:c0:fd:b9", "network": {"id": "88dfb29d-41e0-4df7-8f70-e52cda05620c", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1355578400-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7d4ada9ec35f42b19c6480a9101d21a4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "357d2811-e990-4985-9f9e-b158d10d3699", "external-id": "nsx-vlan-transportzone-641", "segmentation_id": 641, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape3c72582-41", "ovs_interfaceid": "e3c72582-41bf-48d4-8614-b516a96362a7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 947.615044] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-0773e564-f19d-40eb-a88a-6e28c35f232c tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: 014720fe-87b9-4e55-b9ba-295b009c7ff6] Unregistered the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 947.615044] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-0773e564-f19d-40eb-a88a-6e28c35f232c tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: 014720fe-87b9-4e55-b9ba-295b009c7ff6] Deleting contents of the VM from datastore datastore2 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 947.615044] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-0773e564-f19d-40eb-a88a-6e28c35f232c tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Deleting the datastore file [datastore2] 014720fe-87b9-4e55-b9ba-295b009c7ff6 {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 947.615044] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-888179cd-73b6-41b1-a798-6c0afabcd8f7 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.622067] env[63515]: DEBUG oslo_vmware.api [None req-0773e564-f19d-40eb-a88a-6e28c35f232c tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Waiting for the task: (returnval){ [ 947.622067] env[63515]: value = "task-1111638" [ 947.622067] env[63515]: _type = "Task" [ 947.622067] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 947.633414] env[63515]: DEBUG oslo_vmware.api [None req-0773e564-f19d-40eb-a88a-6e28c35f232c tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Task: {'id': task-1111638, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.700917] env[63515]: DEBUG oslo_concurrency.lockutils [None req-127af73c-afc8-4436-9075-a886f5e9c7d5 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.316s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 947.701493] env[63515]: DEBUG nova.compute.manager [None req-127af73c-afc8-4436-9075-a886f5e9c7d5 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] [instance: 7d572ac0-f6af-4622-96cc-e75983420222] Start building networks asynchronously for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 947.705197] env[63515]: DEBUG oslo_concurrency.lockutils [None req-857b4940-0702-491c-a977-f4c63c5dd5f5 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 8.444s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 947.706811] env[63515]: INFO nova.compute.claims [None req-857b4940-0702-491c-a977-f4c63c5dd5f5 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: 918d5b35-c46d-47c5-b59f-175831aed9d6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 947.753628] env[63515]: DEBUG oslo_concurrency.lockutils [None req-7db48089-ba7f-4a2a-9098-e95b53997aff tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Acquiring lock "refresh_cache-eab16df9-7bb5-4576-bca0-769a561c5fe9" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 947.753820] env[63515]: DEBUG oslo_concurrency.lockutils [None req-7db48089-ba7f-4a2a-9098-e95b53997aff tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Acquired lock "refresh_cache-eab16df9-7bb5-4576-bca0-769a561c5fe9" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 947.754012] env[63515]: DEBUG nova.network.neutron [None req-7db48089-ba7f-4a2a-9098-e95b53997aff tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: eab16df9-7bb5-4576-bca0-769a561c5fe9] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 947.764725] env[63515]: DEBUG oslo_concurrency.lockutils [None req-8a925afd-d6aa-4a9a-a54f-bf6262925289 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Acquiring lock "2fa098b6-ecc0-491f-8cfc-5ee335bbf943" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 947.765050] env[63515]: DEBUG oslo_concurrency.lockutils [None req-8a925afd-d6aa-4a9a-a54f-bf6262925289 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Lock "2fa098b6-ecc0-491f-8cfc-5ee335bbf943" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 947.765382] env[63515]: DEBUG oslo_concurrency.lockutils [None req-8a925afd-d6aa-4a9a-a54f-bf6262925289 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Acquiring lock "2fa098b6-ecc0-491f-8cfc-5ee335bbf943-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 947.765674] env[63515]: DEBUG oslo_concurrency.lockutils [None req-8a925afd-d6aa-4a9a-a54f-bf6262925289 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Lock "2fa098b6-ecc0-491f-8cfc-5ee335bbf943-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 947.765918] env[63515]: DEBUG oslo_concurrency.lockutils [None req-8a925afd-d6aa-4a9a-a54f-bf6262925289 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Lock "2fa098b6-ecc0-491f-8cfc-5ee335bbf943-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 947.768217] env[63515]: INFO nova.compute.manager [None req-8a925afd-d6aa-4a9a-a54f-bf6262925289 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: 2fa098b6-ecc0-491f-8cfc-5ee335bbf943] Terminating instance [ 947.770195] env[63515]: DEBUG nova.compute.manager [None req-8a925afd-d6aa-4a9a-a54f-bf6262925289 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: 2fa098b6-ecc0-491f-8cfc-5ee335bbf943] Start destroying the instance on the hypervisor. {{(pid=63515) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 947.770467] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-8a925afd-d6aa-4a9a-a54f-bf6262925289 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: 2fa098b6-ecc0-491f-8cfc-5ee335bbf943] Destroying instance {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 947.771381] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-266e2165-25e5-4263-9b10-be516b1de479 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.779666] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-8a925afd-d6aa-4a9a-a54f-bf6262925289 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: 2fa098b6-ecc0-491f-8cfc-5ee335bbf943] Powering off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 947.779968] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-96754f8d-01c2-4e13-8fdc-9c550313e83f {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.786510] env[63515]: DEBUG oslo_vmware.api [None req-8a925afd-d6aa-4a9a-a54f-bf6262925289 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Waiting for the task: (returnval){ [ 947.786510] env[63515]: value = "task-1111639" [ 947.786510] env[63515]: _type = "Task" [ 947.786510] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 947.794926] env[63515]: DEBUG oslo_vmware.api [None req-8a925afd-d6aa-4a9a-a54f-bf6262925289 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Task: {'id': task-1111639, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.919903] env[63515]: DEBUG oslo_vmware.api [None req-8fdfb203-5786-4efd-8927-4ee465e68e44 tempest-ServerMetadataTestJSON-1034558863 tempest-ServerMetadataTestJSON-1034558863-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52b844b7-5899-8a72-ec8a-9ae026a5a768, 'name': SearchDatastore_Task, 'duration_secs': 0.017822} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 947.921022] env[63515]: DEBUG oslo_concurrency.lockutils [None req-8fdfb203-5786-4efd-8927-4ee465e68e44 tempest-ServerMetadataTestJSON-1034558863 tempest-ServerMetadataTestJSON-1034558863-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 947.921022] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-8fdfb203-5786-4efd-8927-4ee465e68e44 tempest-ServerMetadataTestJSON-1034558863 tempest-ServerMetadataTestJSON-1034558863-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk to [datastore2] b25bf2ba-cf08-44ac-a524-aa87cf785575/b25bf2ba-cf08-44ac-a524-aa87cf785575.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 947.921202] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ff239fda-2088-4f67-a066-4d3cb326df5f {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.927938] env[63515]: DEBUG oslo_vmware.api [None req-8fdfb203-5786-4efd-8927-4ee465e68e44 tempest-ServerMetadataTestJSON-1034558863 tempest-ServerMetadataTestJSON-1034558863-project-member] Waiting for the task: (returnval){ [ 947.927938] env[63515]: value = "task-1111640" [ 947.927938] env[63515]: _type = "Task" [ 947.927938] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 947.937058] env[63515]: DEBUG oslo_vmware.api [None req-8fdfb203-5786-4efd-8927-4ee465e68e44 tempest-ServerMetadataTestJSON-1034558863 tempest-ServerMetadataTestJSON-1034558863-project-member] Task: {'id': task-1111640, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.056870] env[63515]: DEBUG oslo_concurrency.lockutils [req-ba527675-754d-400d-b509-eff7b0dc8952 req-1b7b37f8-bc15-4b25-a524-f350142909bf service nova] Acquiring lock "eab16df9-7bb5-4576-bca0-769a561c5fe9" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 948.056969] env[63515]: DEBUG oslo_concurrency.lockutils [req-ba527675-754d-400d-b509-eff7b0dc8952 req-1b7b37f8-bc15-4b25-a524-f350142909bf service nova] Acquired lock "eab16df9-7bb5-4576-bca0-769a561c5fe9" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 948.059012] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35c912c7-b514-443b-98c4-20311543a350 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.063129] env[63515]: DEBUG nova.compute.manager [req-09f6a403-3161-41c9-a70f-a4a80551b661 req-01cc409f-3c9a-4500-890c-702330e825ed service nova] [instance: eab16df9-7bb5-4576-bca0-769a561c5fe9] Received event network-vif-deleted-e3c72582-41bf-48d4-8614-b516a96362a7 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 948.063460] env[63515]: INFO nova.compute.manager [req-09f6a403-3161-41c9-a70f-a4a80551b661 req-01cc409f-3c9a-4500-890c-702330e825ed service nova] [instance: eab16df9-7bb5-4576-bca0-769a561c5fe9] Neutron deleted interface e3c72582-41bf-48d4-8614-b516a96362a7; detaching it from the instance and deleting it from the info cache [ 948.063815] env[63515]: DEBUG nova.network.neutron [req-09f6a403-3161-41c9-a70f-a4a80551b661 req-01cc409f-3c9a-4500-890c-702330e825ed service nova] [instance: eab16df9-7bb5-4576-bca0-769a561c5fe9] Updating instance_info_cache with network_info: [{"id": "4168f5a5-c282-4f13-9c97-0b0236bd659b", "address": "fa:16:3e:a0:6a:50", "network": {"id": "88dfb29d-41e0-4df7-8f70-e52cda05620c", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1355578400-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.203", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7d4ada9ec35f42b19c6480a9101d21a4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "357d2811-e990-4985-9f9e-b158d10d3699", "external-id": "nsx-vlan-transportzone-641", "segmentation_id": 641, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4168f5a5-c2", "ovs_interfaceid": "4168f5a5-c282-4f13-9c97-0b0236bd659b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 948.083325] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17defa09-2587-4702-9ba7-a511929726f6 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.086161] env[63515]: DEBUG oslo_concurrency.lockutils [None req-db106ad7-1693-4446-8352-ddb2ff9c25ce tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Acquiring lock "eab16df9-7bb5-4576-bca0-769a561c5fe9" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 948.112073] env[63515]: DEBUG nova.virt.vmwareapi.vmops [req-ba527675-754d-400d-b509-eff7b0dc8952 req-1b7b37f8-bc15-4b25-a524-f350142909bf service nova] [instance: eab16df9-7bb5-4576-bca0-769a561c5fe9] Reconfiguring VM to detach interface {{(pid=63515) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1968}} [ 948.112558] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-accd1d73-b68c-4f1d-86fb-17344272f140 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.137380] env[63515]: DEBUG oslo_vmware.api [None req-0773e564-f19d-40eb-a88a-6e28c35f232c tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Task: {'id': task-1111638, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.340073} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 948.139327] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-0773e564-f19d-40eb-a88a-6e28c35f232c tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Deleted the datastore file {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 948.139572] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-0773e564-f19d-40eb-a88a-6e28c35f232c tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: 014720fe-87b9-4e55-b9ba-295b009c7ff6] Deleted contents of the VM from datastore datastore2 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 948.139959] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-0773e564-f19d-40eb-a88a-6e28c35f232c tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: 014720fe-87b9-4e55-b9ba-295b009c7ff6] Instance destroyed {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 948.140420] env[63515]: INFO nova.compute.manager [None req-0773e564-f19d-40eb-a88a-6e28c35f232c tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: 014720fe-87b9-4e55-b9ba-295b009c7ff6] Took 0.61 seconds to destroy the instance on the hypervisor. [ 948.140995] env[63515]: DEBUG oslo.service.loopingcall [None req-0773e564-f19d-40eb-a88a-6e28c35f232c tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 948.141141] env[63515]: DEBUG oslo_vmware.api [req-ba527675-754d-400d-b509-eff7b0dc8952 req-1b7b37f8-bc15-4b25-a524-f350142909bf service nova] Waiting for the task: (returnval){ [ 948.141141] env[63515]: value = "task-1111641" [ 948.141141] env[63515]: _type = "Task" [ 948.141141] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 948.141324] env[63515]: DEBUG nova.compute.manager [-] [instance: 014720fe-87b9-4e55-b9ba-295b009c7ff6] Deallocating network for instance {{(pid=63515) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 948.141429] env[63515]: DEBUG nova.network.neutron [-] [instance: 014720fe-87b9-4e55-b9ba-295b009c7ff6] deallocate_for_instance() {{(pid=63515) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 948.152763] env[63515]: DEBUG oslo_vmware.api [req-ba527675-754d-400d-b509-eff7b0dc8952 req-1b7b37f8-bc15-4b25-a524-f350142909bf service nova] Task: {'id': task-1111641, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.211468] env[63515]: DEBUG nova.compute.utils [None req-127af73c-afc8-4436-9075-a886f5e9c7d5 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Using /dev/sd instead of None {{(pid=63515) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 948.215921] env[63515]: DEBUG nova.compute.manager [None req-127af73c-afc8-4436-9075-a886f5e9c7d5 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] [instance: 7d572ac0-f6af-4622-96cc-e75983420222] Allocating IP information in the background. {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 948.216180] env[63515]: DEBUG nova.network.neutron [None req-127af73c-afc8-4436-9075-a886f5e9c7d5 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] [instance: 7d572ac0-f6af-4622-96cc-e75983420222] allocate_for_instance() {{(pid=63515) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 948.298777] env[63515]: DEBUG oslo_vmware.api [None req-8a925afd-d6aa-4a9a-a54f-bf6262925289 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Task: {'id': task-1111639, 'name': PowerOffVM_Task, 'duration_secs': 0.254147} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 948.299283] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-8a925afd-d6aa-4a9a-a54f-bf6262925289 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: 2fa098b6-ecc0-491f-8cfc-5ee335bbf943] Powered off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 948.299498] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-8a925afd-d6aa-4a9a-a54f-bf6262925289 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: 2fa098b6-ecc0-491f-8cfc-5ee335bbf943] Unregistering the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 948.299779] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-7c22df7e-78a6-409b-8931-8332e4e60046 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.324700] env[63515]: DEBUG nova.policy [None req-127af73c-afc8-4436-9075-a886f5e9c7d5 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2bd575c5f972489dabe5c8057db84790', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '19ce3b05eec64475b685faad1fd801f6', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63515) authorize /opt/stack/nova/nova/policy.py:201}} [ 948.390354] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-8a925afd-d6aa-4a9a-a54f-bf6262925289 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: 2fa098b6-ecc0-491f-8cfc-5ee335bbf943] Unregistered the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 948.390636] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-8a925afd-d6aa-4a9a-a54f-bf6262925289 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: 2fa098b6-ecc0-491f-8cfc-5ee335bbf943] Deleting contents of the VM from datastore datastore2 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 948.390863] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-8a925afd-d6aa-4a9a-a54f-bf6262925289 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Deleting the datastore file [datastore2] 2fa098b6-ecc0-491f-8cfc-5ee335bbf943 {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 948.391175] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3a7c5209-2711-4b76-bfba-b2cb174d4309 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.399199] env[63515]: DEBUG oslo_vmware.api [None req-8a925afd-d6aa-4a9a-a54f-bf6262925289 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Waiting for the task: (returnval){ [ 948.399199] env[63515]: value = "task-1111643" [ 948.399199] env[63515]: _type = "Task" [ 948.399199] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 948.416493] env[63515]: DEBUG oslo_vmware.api [None req-8a925afd-d6aa-4a9a-a54f-bf6262925289 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Task: {'id': task-1111643, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.444282] env[63515]: DEBUG oslo_vmware.api [None req-8fdfb203-5786-4efd-8927-4ee465e68e44 tempest-ServerMetadataTestJSON-1034558863 tempest-ServerMetadataTestJSON-1034558863-project-member] Task: {'id': task-1111640, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.568354] env[63515]: DEBUG oslo_concurrency.lockutils [req-09f6a403-3161-41c9-a70f-a4a80551b661 req-01cc409f-3c9a-4500-890c-702330e825ed service nova] Acquiring lock "eab16df9-7bb5-4576-bca0-769a561c5fe9" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 948.657996] env[63515]: DEBUG oslo_vmware.api [req-ba527675-754d-400d-b509-eff7b0dc8952 req-1b7b37f8-bc15-4b25-a524-f350142909bf service nova] Task: {'id': task-1111641, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.722132] env[63515]: DEBUG nova.compute.manager [None req-127af73c-afc8-4436-9075-a886f5e9c7d5 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] [instance: 7d572ac0-f6af-4622-96cc-e75983420222] Start building block device mappings for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 948.788065] env[63515]: INFO nova.network.neutron [None req-7db48089-ba7f-4a2a-9098-e95b53997aff tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: eab16df9-7bb5-4576-bca0-769a561c5fe9] Port e3c72582-41bf-48d4-8614-b516a96362a7 from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 948.788065] env[63515]: DEBUG nova.network.neutron [None req-7db48089-ba7f-4a2a-9098-e95b53997aff tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: eab16df9-7bb5-4576-bca0-769a561c5fe9] Updating instance_info_cache with network_info: [{"id": "4168f5a5-c282-4f13-9c97-0b0236bd659b", "address": "fa:16:3e:a0:6a:50", "network": {"id": "88dfb29d-41e0-4df7-8f70-e52cda05620c", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1355578400-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.203", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7d4ada9ec35f42b19c6480a9101d21a4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "357d2811-e990-4985-9f9e-b158d10d3699", "external-id": "nsx-vlan-transportzone-641", "segmentation_id": 641, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4168f5a5-c2", "ovs_interfaceid": "4168f5a5-c282-4f13-9c97-0b0236bd659b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 948.909877] env[63515]: DEBUG nova.network.neutron [None req-127af73c-afc8-4436-9075-a886f5e9c7d5 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] [instance: 7d572ac0-f6af-4622-96cc-e75983420222] Successfully created port: dbfd09c2-9ec4-4821-8613-43c6d6cf4724 {{(pid=63515) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 948.920824] env[63515]: DEBUG oslo_vmware.api [None req-8a925afd-d6aa-4a9a-a54f-bf6262925289 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Task: {'id': task-1111643, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.254247} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 948.924801] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-8a925afd-d6aa-4a9a-a54f-bf6262925289 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Deleted the datastore file {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 948.925161] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-8a925afd-d6aa-4a9a-a54f-bf6262925289 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: 2fa098b6-ecc0-491f-8cfc-5ee335bbf943] Deleted contents of the VM from datastore datastore2 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 948.925474] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-8a925afd-d6aa-4a9a-a54f-bf6262925289 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: 2fa098b6-ecc0-491f-8cfc-5ee335bbf943] Instance destroyed {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 948.925773] env[63515]: INFO nova.compute.manager [None req-8a925afd-d6aa-4a9a-a54f-bf6262925289 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] [instance: 2fa098b6-ecc0-491f-8cfc-5ee335bbf943] Took 1.16 seconds to destroy the instance on the hypervisor. [ 948.926177] env[63515]: DEBUG oslo.service.loopingcall [None req-8a925afd-d6aa-4a9a-a54f-bf6262925289 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 948.927000] env[63515]: DEBUG nova.compute.manager [-] [instance: 2fa098b6-ecc0-491f-8cfc-5ee335bbf943] Deallocating network for instance {{(pid=63515) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 948.927243] env[63515]: DEBUG nova.network.neutron [-] [instance: 2fa098b6-ecc0-491f-8cfc-5ee335bbf943] deallocate_for_instance() {{(pid=63515) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 948.942441] env[63515]: DEBUG oslo_vmware.api [None req-8fdfb203-5786-4efd-8927-4ee465e68e44 tempest-ServerMetadataTestJSON-1034558863 tempest-ServerMetadataTestJSON-1034558863-project-member] Task: {'id': task-1111640, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.611202} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 948.946662] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-8fdfb203-5786-4efd-8927-4ee465e68e44 tempest-ServerMetadataTestJSON-1034558863 tempest-ServerMetadataTestJSON-1034558863-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk to [datastore2] b25bf2ba-cf08-44ac-a524-aa87cf785575/b25bf2ba-cf08-44ac-a524-aa87cf785575.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 948.946758] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-8fdfb203-5786-4efd-8927-4ee465e68e44 tempest-ServerMetadataTestJSON-1034558863 tempest-ServerMetadataTestJSON-1034558863-project-member] [instance: b25bf2ba-cf08-44ac-a524-aa87cf785575] Extending root virtual disk to 1048576 {{(pid=63515) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 948.947406] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-3b3a9de8-6bbc-419b-99f2-e21bc6e09af3 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.956379] env[63515]: DEBUG oslo_vmware.api [None req-8fdfb203-5786-4efd-8927-4ee465e68e44 tempest-ServerMetadataTestJSON-1034558863 tempest-ServerMetadataTestJSON-1034558863-project-member] Waiting for the task: (returnval){ [ 948.956379] env[63515]: value = "task-1111644" [ 948.956379] env[63515]: _type = "Task" [ 948.956379] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 948.970186] env[63515]: DEBUG oslo_vmware.api [None req-8fdfb203-5786-4efd-8927-4ee465e68e44 tempest-ServerMetadataTestJSON-1034558863 tempest-ServerMetadataTestJSON-1034558863-project-member] Task: {'id': task-1111644, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.978252] env[63515]: DEBUG nova.network.neutron [-] [instance: 014720fe-87b9-4e55-b9ba-295b009c7ff6] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 949.035044] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66617dbc-0ace-440c-aee9-2b94c5cc934a {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.044374] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e10d334-c73a-44b8-ac7c-9a77f2f74328 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.075574] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09f04205-a920-4b6a-93bc-94ad00402822 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.083188] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae15f306-23c5-402e-a0c9-3051a6dcb252 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.096918] env[63515]: DEBUG nova.compute.provider_tree [None req-857b4940-0702-491c-a977-f4c63c5dd5f5 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 949.157758] env[63515]: DEBUG oslo_vmware.api [req-ba527675-754d-400d-b509-eff7b0dc8952 req-1b7b37f8-bc15-4b25-a524-f350142909bf service nova] Task: {'id': task-1111641, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.290058] env[63515]: DEBUG oslo_concurrency.lockutils [None req-7db48089-ba7f-4a2a-9098-e95b53997aff tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Releasing lock "refresh_cache-eab16df9-7bb5-4576-bca0-769a561c5fe9" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 949.472028] env[63515]: DEBUG oslo_vmware.api [None req-8fdfb203-5786-4efd-8927-4ee465e68e44 tempest-ServerMetadataTestJSON-1034558863 tempest-ServerMetadataTestJSON-1034558863-project-member] Task: {'id': task-1111644, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.118755} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 949.472527] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-8fdfb203-5786-4efd-8927-4ee465e68e44 tempest-ServerMetadataTestJSON-1034558863 tempest-ServerMetadataTestJSON-1034558863-project-member] [instance: b25bf2ba-cf08-44ac-a524-aa87cf785575] Extended root virtual disk {{(pid=63515) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 949.473209] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-131f09d0-fc3d-4579-82cc-5ae1e71315cb {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.487478] env[63515]: INFO nova.compute.manager [-] [instance: 014720fe-87b9-4e55-b9ba-295b009c7ff6] Took 1.35 seconds to deallocate network for instance. [ 949.497180] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-8fdfb203-5786-4efd-8927-4ee465e68e44 tempest-ServerMetadataTestJSON-1034558863 tempest-ServerMetadataTestJSON-1034558863-project-member] [instance: b25bf2ba-cf08-44ac-a524-aa87cf785575] Reconfiguring VM instance instance-00000057 to attach disk [datastore2] b25bf2ba-cf08-44ac-a524-aa87cf785575/b25bf2ba-cf08-44ac-a524-aa87cf785575.vmdk or device None with type sparse {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 949.499050] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4acba295-034d-45f8-97fc-b9ea0790a9db {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.528475] env[63515]: DEBUG oslo_vmware.api [None req-8fdfb203-5786-4efd-8927-4ee465e68e44 tempest-ServerMetadataTestJSON-1034558863 tempest-ServerMetadataTestJSON-1034558863-project-member] Waiting for the task: (returnval){ [ 949.528475] env[63515]: value = "task-1111645" [ 949.528475] env[63515]: _type = "Task" [ 949.528475] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 949.539433] env[63515]: DEBUG oslo_vmware.api [None req-8fdfb203-5786-4efd-8927-4ee465e68e44 tempest-ServerMetadataTestJSON-1034558863 tempest-ServerMetadataTestJSON-1034558863-project-member] Task: {'id': task-1111645, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.600520] env[63515]: DEBUG nova.scheduler.client.report [None req-857b4940-0702-491c-a977-f4c63c5dd5f5 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 949.661026] env[63515]: DEBUG oslo_vmware.api [req-ba527675-754d-400d-b509-eff7b0dc8952 req-1b7b37f8-bc15-4b25-a524-f350142909bf service nova] Task: {'id': task-1111641, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.737325] env[63515]: DEBUG nova.compute.manager [None req-127af73c-afc8-4436-9075-a886f5e9c7d5 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] [instance: 7d572ac0-f6af-4622-96cc-e75983420222] Start spawning the instance on the hypervisor. {{(pid=63515) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 949.764157] env[63515]: DEBUG nova.compute.manager [req-593b8aab-21d2-46bb-b46d-0eb39e6f9a9d req-db177658-75c2-4ad3-8d2b-d9b400394926 service nova] [instance: 014720fe-87b9-4e55-b9ba-295b009c7ff6] Received event network-vif-deleted-406b4b98-709d-4013-9ce0-c4232abb4b1d {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 949.774540] env[63515]: DEBUG nova.virt.hardware [None req-127af73c-afc8-4436-9075-a886f5e9c7d5 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T02:52:14Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T02:51:56Z,direct_url=,disk_format='vmdk',id=8a120570-cb06-4099-b262-554ca0ad15c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b569255cc43e42e7ae3f2b4ad37c6ef4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T02:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 949.774825] env[63515]: DEBUG nova.virt.hardware [None req-127af73c-afc8-4436-9075-a886f5e9c7d5 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Flavor limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 949.774990] env[63515]: DEBUG nova.virt.hardware [None req-127af73c-afc8-4436-9075-a886f5e9c7d5 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Image limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 949.775230] env[63515]: DEBUG nova.virt.hardware [None req-127af73c-afc8-4436-9075-a886f5e9c7d5 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Flavor pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 949.775407] env[63515]: DEBUG nova.virt.hardware [None req-127af73c-afc8-4436-9075-a886f5e9c7d5 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Image pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 949.775575] env[63515]: DEBUG nova.virt.hardware [None req-127af73c-afc8-4436-9075-a886f5e9c7d5 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 949.776243] env[63515]: DEBUG nova.virt.hardware [None req-127af73c-afc8-4436-9075-a886f5e9c7d5 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 949.776243] env[63515]: DEBUG nova.virt.hardware [None req-127af73c-afc8-4436-9075-a886f5e9c7d5 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 949.776243] env[63515]: DEBUG nova.virt.hardware [None req-127af73c-afc8-4436-9075-a886f5e9c7d5 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Got 1 possible topologies {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 949.776455] env[63515]: DEBUG nova.virt.hardware [None req-127af73c-afc8-4436-9075-a886f5e9c7d5 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 949.777038] env[63515]: DEBUG nova.virt.hardware [None req-127af73c-afc8-4436-9075-a886f5e9c7d5 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 949.777786] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70c38b1c-d440-4dd2-a61a-957d49b91455 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.787341] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42839636-b446-463e-b10c-c4fdde05fea3 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.794056] env[63515]: DEBUG oslo_concurrency.lockutils [None req-7db48089-ba7f-4a2a-9098-e95b53997aff tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Lock "interface-eab16df9-7bb5-4576-bca0-769a561c5fe9-e3c72582-41bf-48d4-8614-b516a96362a7" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 9.937s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 950.025236] env[63515]: DEBUG oslo_concurrency.lockutils [None req-0773e564-f19d-40eb-a88a-6e28c35f232c tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 950.039742] env[63515]: DEBUG oslo_vmware.api [None req-8fdfb203-5786-4efd-8927-4ee465e68e44 tempest-ServerMetadataTestJSON-1034558863 tempest-ServerMetadataTestJSON-1034558863-project-member] Task: {'id': task-1111645, 'name': ReconfigVM_Task, 'duration_secs': 0.359591} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 950.040488] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-8fdfb203-5786-4efd-8927-4ee465e68e44 tempest-ServerMetadataTestJSON-1034558863 tempest-ServerMetadataTestJSON-1034558863-project-member] [instance: b25bf2ba-cf08-44ac-a524-aa87cf785575] Reconfigured VM instance instance-00000057 to attach disk [datastore2] b25bf2ba-cf08-44ac-a524-aa87cf785575/b25bf2ba-cf08-44ac-a524-aa87cf785575.vmdk or device None with type sparse {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 950.040956] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-88979e15-5c29-4b65-be16-8174c0371b34 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.048073] env[63515]: DEBUG oslo_vmware.api [None req-8fdfb203-5786-4efd-8927-4ee465e68e44 tempest-ServerMetadataTestJSON-1034558863 tempest-ServerMetadataTestJSON-1034558863-project-member] Waiting for the task: (returnval){ [ 950.048073] env[63515]: value = "task-1111646" [ 950.048073] env[63515]: _type = "Task" [ 950.048073] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 950.057090] env[63515]: DEBUG oslo_vmware.api [None req-8fdfb203-5786-4efd-8927-4ee465e68e44 tempest-ServerMetadataTestJSON-1034558863 tempest-ServerMetadataTestJSON-1034558863-project-member] Task: {'id': task-1111646, 'name': Rename_Task} progress is 5%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.088829] env[63515]: DEBUG nova.compute.manager [req-2a3f8a64-31e3-40c5-96d9-0c39e18d9103 req-e1962ad5-f899-46a5-aa45-97d012c4d3c1 service nova] [instance: 2fa098b6-ecc0-491f-8cfc-5ee335bbf943] Received event network-vif-deleted-e136c744-91c7-433e-979f-0a202f915a28 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 950.089153] env[63515]: INFO nova.compute.manager [req-2a3f8a64-31e3-40c5-96d9-0c39e18d9103 req-e1962ad5-f899-46a5-aa45-97d012c4d3c1 service nova] [instance: 2fa098b6-ecc0-491f-8cfc-5ee335bbf943] Neutron deleted interface e136c744-91c7-433e-979f-0a202f915a28; detaching it from the instance and deleting it from the info cache [ 950.089410] env[63515]: DEBUG nova.network.neutron [req-2a3f8a64-31e3-40c5-96d9-0c39e18d9103 req-e1962ad5-f899-46a5-aa45-97d012c4d3c1 service nova] [instance: 2fa098b6-ecc0-491f-8cfc-5ee335bbf943] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 950.106795] env[63515]: DEBUG oslo_concurrency.lockutils [None req-857b4940-0702-491c-a977-f4c63c5dd5f5 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.401s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 950.107386] env[63515]: DEBUG nova.compute.manager [None req-857b4940-0702-491c-a977-f4c63c5dd5f5 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: 918d5b35-c46d-47c5-b59f-175831aed9d6] Start building networks asynchronously for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 950.111075] env[63515]: DEBUG oslo_concurrency.lockutils [None req-3f7a26fd-cc84-46d9-9d28-feb378f8b5e6 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 8.651s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 950.112685] env[63515]: INFO nova.compute.claims [None req-3f7a26fd-cc84-46d9-9d28-feb378f8b5e6 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] [instance: 75f4dc79-4b73-4ddc-be03-3653a5a0797e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 950.160034] env[63515]: DEBUG oslo_vmware.api [req-ba527675-754d-400d-b509-eff7b0dc8952 req-1b7b37f8-bc15-4b25-a524-f350142909bf service nova] Task: {'id': task-1111641, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.168125] env[63515]: DEBUG nova.network.neutron [-] [instance: 2fa098b6-ecc0-491f-8cfc-5ee335bbf943] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 950.561297] env[63515]: DEBUG oslo_vmware.api [None req-8fdfb203-5786-4efd-8927-4ee465e68e44 tempest-ServerMetadataTestJSON-1034558863 tempest-ServerMetadataTestJSON-1034558863-project-member] Task: {'id': task-1111646, 'name': Rename_Task, 'duration_secs': 0.149603} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 950.561686] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-8fdfb203-5786-4efd-8927-4ee465e68e44 tempest-ServerMetadataTestJSON-1034558863 tempest-ServerMetadataTestJSON-1034558863-project-member] [instance: b25bf2ba-cf08-44ac-a524-aa87cf785575] Powering on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 950.561994] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5492e90a-1286-49cf-9205-adec5c60d185 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.569920] env[63515]: DEBUG oslo_vmware.api [None req-8fdfb203-5786-4efd-8927-4ee465e68e44 tempest-ServerMetadataTestJSON-1034558863 tempest-ServerMetadataTestJSON-1034558863-project-member] Waiting for the task: (returnval){ [ 950.569920] env[63515]: value = "task-1111647" [ 950.569920] env[63515]: _type = "Task" [ 950.569920] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 950.579220] env[63515]: DEBUG oslo_vmware.api [None req-8fdfb203-5786-4efd-8927-4ee465e68e44 tempest-ServerMetadataTestJSON-1034558863 tempest-ServerMetadataTestJSON-1034558863-project-member] Task: {'id': task-1111647, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.592155] env[63515]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-7217015f-54b1-485f-a081-270f2db05ade {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.602198] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4378c7f-0b63-4ff1-8579-6d89f5af6f05 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.618044] env[63515]: DEBUG nova.compute.utils [None req-857b4940-0702-491c-a977-f4c63c5dd5f5 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Using /dev/sd instead of None {{(pid=63515) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 950.621670] env[63515]: DEBUG nova.compute.manager [None req-857b4940-0702-491c-a977-f4c63c5dd5f5 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: 918d5b35-c46d-47c5-b59f-175831aed9d6] Allocating IP information in the background. {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 950.621831] env[63515]: DEBUG nova.network.neutron [None req-857b4940-0702-491c-a977-f4c63c5dd5f5 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: 918d5b35-c46d-47c5-b59f-175831aed9d6] allocate_for_instance() {{(pid=63515) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 950.634827] env[63515]: DEBUG nova.compute.manager [req-2a3f8a64-31e3-40c5-96d9-0c39e18d9103 req-e1962ad5-f899-46a5-aa45-97d012c4d3c1 service nova] [instance: 2fa098b6-ecc0-491f-8cfc-5ee335bbf943] Detach interface failed, port_id=e136c744-91c7-433e-979f-0a202f915a28, reason: Instance 2fa098b6-ecc0-491f-8cfc-5ee335bbf943 could not be found. {{(pid=63515) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 950.660214] env[63515]: DEBUG oslo_vmware.api [req-ba527675-754d-400d-b509-eff7b0dc8952 req-1b7b37f8-bc15-4b25-a524-f350142909bf service nova] Task: {'id': task-1111641, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.671444] env[63515]: INFO nova.compute.manager [-] [instance: 2fa098b6-ecc0-491f-8cfc-5ee335bbf943] Took 1.74 seconds to deallocate network for instance. [ 950.688706] env[63515]: DEBUG nova.policy [None req-857b4940-0702-491c-a977-f4c63c5dd5f5 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ed295ef8f0134bcaa668149ac9f4dd74', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'aef947bfaf7d4180b46b4be662e3d903', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63515) authorize /opt/stack/nova/nova/policy.py:201}} [ 951.080433] env[63515]: DEBUG oslo_vmware.api [None req-8fdfb203-5786-4efd-8927-4ee465e68e44 tempest-ServerMetadataTestJSON-1034558863 tempest-ServerMetadataTestJSON-1034558863-project-member] Task: {'id': task-1111647, 'name': PowerOnVM_Task} progress is 33%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.123308] env[63515]: DEBUG nova.compute.manager [None req-857b4940-0702-491c-a977-f4c63c5dd5f5 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: 918d5b35-c46d-47c5-b59f-175831aed9d6] Start building block device mappings for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 951.162584] env[63515]: DEBUG oslo_vmware.api [req-ba527675-754d-400d-b509-eff7b0dc8952 req-1b7b37f8-bc15-4b25-a524-f350142909bf service nova] Task: {'id': task-1111641, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.181039] env[63515]: DEBUG oslo_concurrency.lockutils [None req-8a925afd-d6aa-4a9a-a54f-bf6262925289 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 951.291440] env[63515]: DEBUG nova.network.neutron [None req-857b4940-0702-491c-a977-f4c63c5dd5f5 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: 918d5b35-c46d-47c5-b59f-175831aed9d6] Successfully created port: b4724c8b-a2c7-4f52-8a75-f46434d7b527 {{(pid=63515) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 951.389648] env[63515]: DEBUG nova.network.neutron [None req-127af73c-afc8-4436-9075-a886f5e9c7d5 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] [instance: 7d572ac0-f6af-4622-96cc-e75983420222] Successfully updated port: dbfd09c2-9ec4-4821-8613-43c6d6cf4724 {{(pid=63515) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 951.414046] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b21303e-3ca9-417c-b976-1abded2370a7 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.422922] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16a0f256-9b2c-42bb-9cb3-119096d10b52 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.456879] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-faea57a4-b19a-42c7-9aad-5ef8eb5c9245 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.464454] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d13cfd56-58c0-4dbc-a66b-088df8834946 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.477866] env[63515]: DEBUG nova.compute.provider_tree [None req-3f7a26fd-cc84-46d9-9d28-feb378f8b5e6 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Updating inventory in ProviderTree for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 951.581261] env[63515]: DEBUG oslo_vmware.api [None req-8fdfb203-5786-4efd-8927-4ee465e68e44 tempest-ServerMetadataTestJSON-1034558863 tempest-ServerMetadataTestJSON-1034558863-project-member] Task: {'id': task-1111647, 'name': PowerOnVM_Task} progress is 33%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.662525] env[63515]: DEBUG oslo_vmware.api [req-ba527675-754d-400d-b509-eff7b0dc8952 req-1b7b37f8-bc15-4b25-a524-f350142909bf service nova] Task: {'id': task-1111641, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.895479] env[63515]: DEBUG oslo_concurrency.lockutils [None req-127af73c-afc8-4436-9075-a886f5e9c7d5 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Acquiring lock "refresh_cache-7d572ac0-f6af-4622-96cc-e75983420222" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 951.895650] env[63515]: DEBUG oslo_concurrency.lockutils [None req-127af73c-afc8-4436-9075-a886f5e9c7d5 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Acquired lock "refresh_cache-7d572ac0-f6af-4622-96cc-e75983420222" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 951.896185] env[63515]: DEBUG nova.network.neutron [None req-127af73c-afc8-4436-9075-a886f5e9c7d5 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] [instance: 7d572ac0-f6af-4622-96cc-e75983420222] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 952.010418] env[63515]: DEBUG nova.scheduler.client.report [None req-3f7a26fd-cc84-46d9-9d28-feb378f8b5e6 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Updated inventory for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a with generation 128 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 952.010809] env[63515]: DEBUG nova.compute.provider_tree [None req-3f7a26fd-cc84-46d9-9d28-feb378f8b5e6 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Updating resource provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a generation from 128 to 129 during operation: update_inventory {{(pid=63515) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 952.010904] env[63515]: DEBUG nova.compute.provider_tree [None req-3f7a26fd-cc84-46d9-9d28-feb378f8b5e6 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Updating inventory in ProviderTree for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 952.083013] env[63515]: DEBUG oslo_vmware.api [None req-8fdfb203-5786-4efd-8927-4ee465e68e44 tempest-ServerMetadataTestJSON-1034558863 tempest-ServerMetadataTestJSON-1034558863-project-member] Task: {'id': task-1111647, 'name': PowerOnVM_Task} progress is 100%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.133891] env[63515]: DEBUG nova.compute.manager [None req-857b4940-0702-491c-a977-f4c63c5dd5f5 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: 918d5b35-c46d-47c5-b59f-175831aed9d6] Start spawning the instance on the hypervisor. {{(pid=63515) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 952.152449] env[63515]: DEBUG nova.compute.manager [req-6343619f-96e3-4df8-a135-d8ef272aebe2 req-1693b6bd-1e3f-493c-bd2e-27875bd61372 service nova] [instance: 7d572ac0-f6af-4622-96cc-e75983420222] Received event network-vif-plugged-dbfd09c2-9ec4-4821-8613-43c6d6cf4724 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 952.152708] env[63515]: DEBUG oslo_concurrency.lockutils [req-6343619f-96e3-4df8-a135-d8ef272aebe2 req-1693b6bd-1e3f-493c-bd2e-27875bd61372 service nova] Acquiring lock "7d572ac0-f6af-4622-96cc-e75983420222-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 952.152895] env[63515]: DEBUG oslo_concurrency.lockutils [req-6343619f-96e3-4df8-a135-d8ef272aebe2 req-1693b6bd-1e3f-493c-bd2e-27875bd61372 service nova] Lock "7d572ac0-f6af-4622-96cc-e75983420222-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 952.153120] env[63515]: DEBUG oslo_concurrency.lockutils [req-6343619f-96e3-4df8-a135-d8ef272aebe2 req-1693b6bd-1e3f-493c-bd2e-27875bd61372 service nova] Lock "7d572ac0-f6af-4622-96cc-e75983420222-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 952.153356] env[63515]: DEBUG nova.compute.manager [req-6343619f-96e3-4df8-a135-d8ef272aebe2 req-1693b6bd-1e3f-493c-bd2e-27875bd61372 service nova] [instance: 7d572ac0-f6af-4622-96cc-e75983420222] No waiting events found dispatching network-vif-plugged-dbfd09c2-9ec4-4821-8613-43c6d6cf4724 {{(pid=63515) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 952.153499] env[63515]: WARNING nova.compute.manager [req-6343619f-96e3-4df8-a135-d8ef272aebe2 req-1693b6bd-1e3f-493c-bd2e-27875bd61372 service nova] [instance: 7d572ac0-f6af-4622-96cc-e75983420222] Received unexpected event network-vif-plugged-dbfd09c2-9ec4-4821-8613-43c6d6cf4724 for instance with vm_state building and task_state spawning. [ 952.153684] env[63515]: DEBUG nova.compute.manager [req-6343619f-96e3-4df8-a135-d8ef272aebe2 req-1693b6bd-1e3f-493c-bd2e-27875bd61372 service nova] [instance: 7d572ac0-f6af-4622-96cc-e75983420222] Received event network-changed-dbfd09c2-9ec4-4821-8613-43c6d6cf4724 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 952.153894] env[63515]: DEBUG nova.compute.manager [req-6343619f-96e3-4df8-a135-d8ef272aebe2 req-1693b6bd-1e3f-493c-bd2e-27875bd61372 service nova] [instance: 7d572ac0-f6af-4622-96cc-e75983420222] Refreshing instance network info cache due to event network-changed-dbfd09c2-9ec4-4821-8613-43c6d6cf4724. {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 952.154124] env[63515]: DEBUG oslo_concurrency.lockutils [req-6343619f-96e3-4df8-a135-d8ef272aebe2 req-1693b6bd-1e3f-493c-bd2e-27875bd61372 service nova] Acquiring lock "refresh_cache-7d572ac0-f6af-4622-96cc-e75983420222" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 952.163318] env[63515]: DEBUG nova.virt.hardware [None req-857b4940-0702-491c-a977-f4c63c5dd5f5 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T02:52:14Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T02:51:56Z,direct_url=,disk_format='vmdk',id=8a120570-cb06-4099-b262-554ca0ad15c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b569255cc43e42e7ae3f2b4ad37c6ef4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T02:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 952.163659] env[63515]: DEBUG nova.virt.hardware [None req-857b4940-0702-491c-a977-f4c63c5dd5f5 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Flavor limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 952.163760] env[63515]: DEBUG nova.virt.hardware [None req-857b4940-0702-491c-a977-f4c63c5dd5f5 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Image limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 952.163896] env[63515]: DEBUG nova.virt.hardware [None req-857b4940-0702-491c-a977-f4c63c5dd5f5 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Flavor pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 952.164053] env[63515]: DEBUG nova.virt.hardware [None req-857b4940-0702-491c-a977-f4c63c5dd5f5 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Image pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 952.164237] env[63515]: DEBUG nova.virt.hardware [None req-857b4940-0702-491c-a977-f4c63c5dd5f5 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 952.164448] env[63515]: DEBUG nova.virt.hardware [None req-857b4940-0702-491c-a977-f4c63c5dd5f5 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 952.164614] env[63515]: DEBUG nova.virt.hardware [None req-857b4940-0702-491c-a977-f4c63c5dd5f5 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 952.164782] env[63515]: DEBUG nova.virt.hardware [None req-857b4940-0702-491c-a977-f4c63c5dd5f5 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Got 1 possible topologies {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 952.164946] env[63515]: DEBUG nova.virt.hardware [None req-857b4940-0702-491c-a977-f4c63c5dd5f5 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 952.165137] env[63515]: DEBUG nova.virt.hardware [None req-857b4940-0702-491c-a977-f4c63c5dd5f5 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 952.166156] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-edfd7b30-3d50-4499-a8b2-fa79e55ff47e {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.171963] env[63515]: DEBUG oslo_vmware.api [req-ba527675-754d-400d-b509-eff7b0dc8952 req-1b7b37f8-bc15-4b25-a524-f350142909bf service nova] Task: {'id': task-1111641, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.177372] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6fe58404-a696-420c-ae14-7b57ebf862fe {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.434778] env[63515]: DEBUG nova.network.neutron [None req-127af73c-afc8-4436-9075-a886f5e9c7d5 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] [instance: 7d572ac0-f6af-4622-96cc-e75983420222] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 952.516504] env[63515]: DEBUG oslo_concurrency.lockutils [None req-3f7a26fd-cc84-46d9-9d28-feb378f8b5e6 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.405s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 952.517116] env[63515]: DEBUG nova.compute.manager [None req-3f7a26fd-cc84-46d9-9d28-feb378f8b5e6 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] [instance: 75f4dc79-4b73-4ddc-be03-3653a5a0797e] Start building networks asynchronously for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 952.519722] env[63515]: DEBUG oslo_concurrency.lockutils [None req-3f7a26fd-cc84-46d9-9d28-feb378f8b5e6 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 11.055s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 952.521132] env[63515]: INFO nova.compute.claims [None req-3f7a26fd-cc84-46d9-9d28-feb378f8b5e6 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] [instance: d438dd30-b816-48a6-94c6-11e6e9bd5b71] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 952.578233] env[63515]: DEBUG nova.network.neutron [None req-127af73c-afc8-4436-9075-a886f5e9c7d5 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] [instance: 7d572ac0-f6af-4622-96cc-e75983420222] Updating instance_info_cache with network_info: [{"id": "dbfd09c2-9ec4-4821-8613-43c6d6cf4724", "address": "fa:16:3e:6d:92:3d", "network": {"id": "09fb20ca-2b17-4d46-b915-6aff85b862d8", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1815529427-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "19ce3b05eec64475b685faad1fd801f6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ded18042-834c-4792-b3e8-b1c377446432", "external-id": "nsx-vlan-transportzone-293", "segmentation_id": 293, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdbfd09c2-9e", "ovs_interfaceid": "dbfd09c2-9ec4-4821-8613-43c6d6cf4724", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 952.584937] env[63515]: DEBUG oslo_vmware.api [None req-8fdfb203-5786-4efd-8927-4ee465e68e44 tempest-ServerMetadataTestJSON-1034558863 tempest-ServerMetadataTestJSON-1034558863-project-member] Task: {'id': task-1111647, 'name': PowerOnVM_Task, 'duration_secs': 1.519387} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 952.585269] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-8fdfb203-5786-4efd-8927-4ee465e68e44 tempest-ServerMetadataTestJSON-1034558863 tempest-ServerMetadataTestJSON-1034558863-project-member] [instance: b25bf2ba-cf08-44ac-a524-aa87cf785575] Powered on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 952.585433] env[63515]: INFO nova.compute.manager [None req-8fdfb203-5786-4efd-8927-4ee465e68e44 tempest-ServerMetadataTestJSON-1034558863 tempest-ServerMetadataTestJSON-1034558863-project-member] [instance: b25bf2ba-cf08-44ac-a524-aa87cf785575] Took 8.97 seconds to spawn the instance on the hypervisor. [ 952.585620] env[63515]: DEBUG nova.compute.manager [None req-8fdfb203-5786-4efd-8927-4ee465e68e44 tempest-ServerMetadataTestJSON-1034558863 tempest-ServerMetadataTestJSON-1034558863-project-member] [instance: b25bf2ba-cf08-44ac-a524-aa87cf785575] Checking state {{(pid=63515) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 952.586455] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-edfbdcb3-b032-4c09-9a02-eac893137907 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.667636] env[63515]: DEBUG oslo_vmware.api [req-ba527675-754d-400d-b509-eff7b0dc8952 req-1b7b37f8-bc15-4b25-a524-f350142909bf service nova] Task: {'id': task-1111641, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.884326] env[63515]: DEBUG nova.network.neutron [None req-857b4940-0702-491c-a977-f4c63c5dd5f5 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: 918d5b35-c46d-47c5-b59f-175831aed9d6] Successfully updated port: b4724c8b-a2c7-4f52-8a75-f46434d7b527 {{(pid=63515) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 953.025973] env[63515]: DEBUG nova.compute.utils [None req-3f7a26fd-cc84-46d9-9d28-feb378f8b5e6 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Using /dev/sd instead of None {{(pid=63515) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 953.031229] env[63515]: DEBUG nova.compute.manager [None req-3f7a26fd-cc84-46d9-9d28-feb378f8b5e6 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] [instance: 75f4dc79-4b73-4ddc-be03-3653a5a0797e] Allocating IP information in the background. {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 953.032056] env[63515]: DEBUG nova.network.neutron [None req-3f7a26fd-cc84-46d9-9d28-feb378f8b5e6 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] [instance: 75f4dc79-4b73-4ddc-be03-3653a5a0797e] allocate_for_instance() {{(pid=63515) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 953.074849] env[63515]: DEBUG nova.policy [None req-3f7a26fd-cc84-46d9-9d28-feb378f8b5e6 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e8b57c2415b44bbd93105c6824a9e322', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3dcf6179891c4841a2901fcffe43cd31', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63515) authorize /opt/stack/nova/nova/policy.py:201}} [ 953.081377] env[63515]: DEBUG oslo_concurrency.lockutils [None req-127af73c-afc8-4436-9075-a886f5e9c7d5 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Releasing lock "refresh_cache-7d572ac0-f6af-4622-96cc-e75983420222" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 953.081716] env[63515]: DEBUG nova.compute.manager [None req-127af73c-afc8-4436-9075-a886f5e9c7d5 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] [instance: 7d572ac0-f6af-4622-96cc-e75983420222] Instance network_info: |[{"id": "dbfd09c2-9ec4-4821-8613-43c6d6cf4724", "address": "fa:16:3e:6d:92:3d", "network": {"id": "09fb20ca-2b17-4d46-b915-6aff85b862d8", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1815529427-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "19ce3b05eec64475b685faad1fd801f6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ded18042-834c-4792-b3e8-b1c377446432", "external-id": "nsx-vlan-transportzone-293", "segmentation_id": 293, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdbfd09c2-9e", "ovs_interfaceid": "dbfd09c2-9ec4-4821-8613-43c6d6cf4724", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 953.082052] env[63515]: DEBUG oslo_concurrency.lockutils [req-6343619f-96e3-4df8-a135-d8ef272aebe2 req-1693b6bd-1e3f-493c-bd2e-27875bd61372 service nova] Acquired lock "refresh_cache-7d572ac0-f6af-4622-96cc-e75983420222" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 953.082246] env[63515]: DEBUG nova.network.neutron [req-6343619f-96e3-4df8-a135-d8ef272aebe2 req-1693b6bd-1e3f-493c-bd2e-27875bd61372 service nova] [instance: 7d572ac0-f6af-4622-96cc-e75983420222] Refreshing network info cache for port dbfd09c2-9ec4-4821-8613-43c6d6cf4724 {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 953.083501] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-127af73c-afc8-4436-9075-a886f5e9c7d5 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] [instance: 7d572ac0-f6af-4622-96cc-e75983420222] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:6d:92:3d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ded18042-834c-4792-b3e8-b1c377446432', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'dbfd09c2-9ec4-4821-8613-43c6d6cf4724', 'vif_model': 'vmxnet3'}] {{(pid=63515) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 953.092345] env[63515]: DEBUG oslo.service.loopingcall [None req-127af73c-afc8-4436-9075-a886f5e9c7d5 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 953.093510] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7d572ac0-f6af-4622-96cc-e75983420222] Creating VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 953.093783] env[63515]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-0901435a-abe0-493c-8e99-f06539d45059 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.117164] env[63515]: INFO nova.compute.manager [None req-8fdfb203-5786-4efd-8927-4ee465e68e44 tempest-ServerMetadataTestJSON-1034558863 tempest-ServerMetadataTestJSON-1034558863-project-member] [instance: b25bf2ba-cf08-44ac-a524-aa87cf785575] Took 19.89 seconds to build instance. [ 953.120605] env[63515]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 953.120605] env[63515]: value = "task-1111648" [ 953.120605] env[63515]: _type = "Task" [ 953.120605] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 953.129670] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1111648, 'name': CreateVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.167101] env[63515]: DEBUG oslo_vmware.api [req-ba527675-754d-400d-b509-eff7b0dc8952 req-1b7b37f8-bc15-4b25-a524-f350142909bf service nova] Task: {'id': task-1111641, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.390030] env[63515]: DEBUG oslo_concurrency.lockutils [None req-857b4940-0702-491c-a977-f4c63c5dd5f5 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Acquiring lock "refresh_cache-918d5b35-c46d-47c5-b59f-175831aed9d6" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 953.390112] env[63515]: DEBUG oslo_concurrency.lockutils [None req-857b4940-0702-491c-a977-f4c63c5dd5f5 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Acquired lock "refresh_cache-918d5b35-c46d-47c5-b59f-175831aed9d6" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 953.390277] env[63515]: DEBUG nova.network.neutron [None req-857b4940-0702-491c-a977-f4c63c5dd5f5 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: 918d5b35-c46d-47c5-b59f-175831aed9d6] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 953.413263] env[63515]: DEBUG nova.network.neutron [None req-3f7a26fd-cc84-46d9-9d28-feb378f8b5e6 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] [instance: 75f4dc79-4b73-4ddc-be03-3653a5a0797e] Successfully created port: 00eda8ad-1a52-4c47-b065-a82c08da7980 {{(pid=63515) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 953.531134] env[63515]: DEBUG nova.compute.manager [None req-3f7a26fd-cc84-46d9-9d28-feb378f8b5e6 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] [instance: 75f4dc79-4b73-4ddc-be03-3653a5a0797e] Start building block device mappings for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 953.622493] env[63515]: DEBUG oslo_concurrency.lockutils [None req-8fdfb203-5786-4efd-8927-4ee465e68e44 tempest-ServerMetadataTestJSON-1034558863 tempest-ServerMetadataTestJSON-1034558863-project-member] Lock "b25bf2ba-cf08-44ac-a524-aa87cf785575" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 21.404s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 953.638516] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1111648, 'name': CreateVM_Task, 'duration_secs': 0.44893} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 953.638728] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7d572ac0-f6af-4622-96cc-e75983420222] Created VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 953.639526] env[63515]: DEBUG oslo_concurrency.lockutils [None req-127af73c-afc8-4436-9075-a886f5e9c7d5 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 953.639757] env[63515]: DEBUG oslo_concurrency.lockutils [None req-127af73c-afc8-4436-9075-a886f5e9c7d5 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 953.640149] env[63515]: DEBUG oslo_concurrency.lockutils [None req-127af73c-afc8-4436-9075-a886f5e9c7d5 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 953.640459] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2a69fab2-3685-422b-85fc-e940f1713163 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.646033] env[63515]: DEBUG oslo_vmware.api [None req-127af73c-afc8-4436-9075-a886f5e9c7d5 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Waiting for the task: (returnval){ [ 953.646033] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]523dcd2e-314f-df14-9b47-6a5d06e72998" [ 953.646033] env[63515]: _type = "Task" [ 953.646033] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 953.664977] env[63515]: DEBUG oslo_vmware.api [None req-127af73c-afc8-4436-9075-a886f5e9c7d5 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]523dcd2e-314f-df14-9b47-6a5d06e72998, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.677585] env[63515]: DEBUG oslo_vmware.api [req-ba527675-754d-400d-b509-eff7b0dc8952 req-1b7b37f8-bc15-4b25-a524-f350142909bf service nova] Task: {'id': task-1111641, 'name': ReconfigVM_Task} progress is 18%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.858904] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8cdea50-393a-4b93-ae54-6a6ce620b472 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.869699] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-989e01bd-bb15-4fdb-9a4a-c8e7ee03cdb1 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.906028] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-779e24ef-565e-4aa8-8c46-79f1fe08707b {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.914609] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7da686bc-def2-478c-8045-38634d782254 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.919433] env[63515]: DEBUG nova.network.neutron [req-6343619f-96e3-4df8-a135-d8ef272aebe2 req-1693b6bd-1e3f-493c-bd2e-27875bd61372 service nova] [instance: 7d572ac0-f6af-4622-96cc-e75983420222] Updated VIF entry in instance network info cache for port dbfd09c2-9ec4-4821-8613-43c6d6cf4724. {{(pid=63515) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 953.919804] env[63515]: DEBUG nova.network.neutron [req-6343619f-96e3-4df8-a135-d8ef272aebe2 req-1693b6bd-1e3f-493c-bd2e-27875bd61372 service nova] [instance: 7d572ac0-f6af-4622-96cc-e75983420222] Updating instance_info_cache with network_info: [{"id": "dbfd09c2-9ec4-4821-8613-43c6d6cf4724", "address": "fa:16:3e:6d:92:3d", "network": {"id": "09fb20ca-2b17-4d46-b915-6aff85b862d8", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1815529427-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "19ce3b05eec64475b685faad1fd801f6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ded18042-834c-4792-b3e8-b1c377446432", "external-id": "nsx-vlan-transportzone-293", "segmentation_id": 293, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdbfd09c2-9e", "ovs_interfaceid": "dbfd09c2-9ec4-4821-8613-43c6d6cf4724", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 953.931490] env[63515]: DEBUG nova.compute.provider_tree [None req-3f7a26fd-cc84-46d9-9d28-feb378f8b5e6 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 953.937800] env[63515]: DEBUG nova.network.neutron [None req-857b4940-0702-491c-a977-f4c63c5dd5f5 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: 918d5b35-c46d-47c5-b59f-175831aed9d6] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 954.085538] env[63515]: DEBUG nova.network.neutron [None req-857b4940-0702-491c-a977-f4c63c5dd5f5 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: 918d5b35-c46d-47c5-b59f-175831aed9d6] Updating instance_info_cache with network_info: [{"id": "b4724c8b-a2c7-4f52-8a75-f46434d7b527", "address": "fa:16:3e:dc:ce:b2", "network": {"id": "cec614cb-7a5b-4b16-a83d-061ff62d6677", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1501468300-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "aef947bfaf7d4180b46b4be662e3d903", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3a0ddd7d-c321-4187-bdd8-b19044ea2c4a", "external-id": "nsx-vlan-transportzone-747", "segmentation_id": 747, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb4724c8b-a2", "ovs_interfaceid": "b4724c8b-a2c7-4f52-8a75-f46434d7b527", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 954.161206] env[63515]: DEBUG oslo_vmware.api [None req-127af73c-afc8-4436-9075-a886f5e9c7d5 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]523dcd2e-314f-df14-9b47-6a5d06e72998, 'name': SearchDatastore_Task, 'duration_secs': 0.015157} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 954.161592] env[63515]: DEBUG oslo_concurrency.lockutils [None req-127af73c-afc8-4436-9075-a886f5e9c7d5 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 954.161837] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-127af73c-afc8-4436-9075-a886f5e9c7d5 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] [instance: 7d572ac0-f6af-4622-96cc-e75983420222] Processing image 8a120570-cb06-4099-b262-554ca0ad15c5 {{(pid=63515) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 954.162128] env[63515]: DEBUG oslo_concurrency.lockutils [None req-127af73c-afc8-4436-9075-a886f5e9c7d5 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 954.162304] env[63515]: DEBUG oslo_concurrency.lockutils [None req-127af73c-afc8-4436-9075-a886f5e9c7d5 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 954.162491] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-127af73c-afc8-4436-9075-a886f5e9c7d5 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 954.162805] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-4b3ec376-20de-4413-b0d7-03326fd4c1b7 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.174140] env[63515]: DEBUG oslo_vmware.api [req-ba527675-754d-400d-b509-eff7b0dc8952 req-1b7b37f8-bc15-4b25-a524-f350142909bf service nova] Task: {'id': task-1111641, 'name': ReconfigVM_Task, 'duration_secs': 5.806371} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 954.175374] env[63515]: DEBUG oslo_concurrency.lockutils [req-ba527675-754d-400d-b509-eff7b0dc8952 req-1b7b37f8-bc15-4b25-a524-f350142909bf service nova] Releasing lock "eab16df9-7bb5-4576-bca0-769a561c5fe9" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 954.175601] env[63515]: DEBUG nova.virt.vmwareapi.vmops [req-ba527675-754d-400d-b509-eff7b0dc8952 req-1b7b37f8-bc15-4b25-a524-f350142909bf service nova] [instance: eab16df9-7bb5-4576-bca0-769a561c5fe9] Reconfigured VM to detach interface {{(pid=63515) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1978}} [ 954.176154] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-127af73c-afc8-4436-9075-a886f5e9c7d5 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 954.176376] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-127af73c-afc8-4436-9075-a886f5e9c7d5 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63515) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 954.178509] env[63515]: DEBUG oslo_concurrency.lockutils [None req-db106ad7-1693-4446-8352-ddb2ff9c25ce tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Lock "eab16df9-7bb5-4576-bca0-769a561c5fe9" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 6.092s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 954.178772] env[63515]: DEBUG oslo_concurrency.lockutils [None req-db106ad7-1693-4446-8352-ddb2ff9c25ce tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Acquiring lock "eab16df9-7bb5-4576-bca0-769a561c5fe9-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 954.179119] env[63515]: DEBUG oslo_concurrency.lockutils [None req-db106ad7-1693-4446-8352-ddb2ff9c25ce tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Lock "eab16df9-7bb5-4576-bca0-769a561c5fe9-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 954.179206] env[63515]: DEBUG oslo_concurrency.lockutils [None req-db106ad7-1693-4446-8352-ddb2ff9c25ce tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Lock "eab16df9-7bb5-4576-bca0-769a561c5fe9-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 954.180810] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5d586ace-58e7-4e87-a564-9b75d4a10f1f {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.184167] env[63515]: DEBUG nova.compute.manager [req-9da6f36c-04ac-4e64-8c0e-31ea36e644b2 req-73be1d33-c709-421a-aeda-34627b7e3164 service nova] [instance: 918d5b35-c46d-47c5-b59f-175831aed9d6] Received event network-vif-plugged-b4724c8b-a2c7-4f52-8a75-f46434d7b527 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 954.184395] env[63515]: DEBUG oslo_concurrency.lockutils [req-9da6f36c-04ac-4e64-8c0e-31ea36e644b2 req-73be1d33-c709-421a-aeda-34627b7e3164 service nova] Acquiring lock "918d5b35-c46d-47c5-b59f-175831aed9d6-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 954.184632] env[63515]: DEBUG oslo_concurrency.lockutils [req-9da6f36c-04ac-4e64-8c0e-31ea36e644b2 req-73be1d33-c709-421a-aeda-34627b7e3164 service nova] Lock "918d5b35-c46d-47c5-b59f-175831aed9d6-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 954.184832] env[63515]: DEBUG oslo_concurrency.lockutils [req-9da6f36c-04ac-4e64-8c0e-31ea36e644b2 req-73be1d33-c709-421a-aeda-34627b7e3164 service nova] Lock "918d5b35-c46d-47c5-b59f-175831aed9d6-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 954.185018] env[63515]: DEBUG nova.compute.manager [req-9da6f36c-04ac-4e64-8c0e-31ea36e644b2 req-73be1d33-c709-421a-aeda-34627b7e3164 service nova] [instance: 918d5b35-c46d-47c5-b59f-175831aed9d6] No waiting events found dispatching network-vif-plugged-b4724c8b-a2c7-4f52-8a75-f46434d7b527 {{(pid=63515) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 954.185207] env[63515]: WARNING nova.compute.manager [req-9da6f36c-04ac-4e64-8c0e-31ea36e644b2 req-73be1d33-c709-421a-aeda-34627b7e3164 service nova] [instance: 918d5b35-c46d-47c5-b59f-175831aed9d6] Received unexpected event network-vif-plugged-b4724c8b-a2c7-4f52-8a75-f46434d7b527 for instance with vm_state building and task_state spawning. [ 954.185374] env[63515]: DEBUG nova.compute.manager [req-9da6f36c-04ac-4e64-8c0e-31ea36e644b2 req-73be1d33-c709-421a-aeda-34627b7e3164 service nova] [instance: 918d5b35-c46d-47c5-b59f-175831aed9d6] Received event network-changed-b4724c8b-a2c7-4f52-8a75-f46434d7b527 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 954.185533] env[63515]: DEBUG nova.compute.manager [req-9da6f36c-04ac-4e64-8c0e-31ea36e644b2 req-73be1d33-c709-421a-aeda-34627b7e3164 service nova] [instance: 918d5b35-c46d-47c5-b59f-175831aed9d6] Refreshing instance network info cache due to event network-changed-b4724c8b-a2c7-4f52-8a75-f46434d7b527. {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 954.185734] env[63515]: DEBUG oslo_concurrency.lockutils [req-9da6f36c-04ac-4e64-8c0e-31ea36e644b2 req-73be1d33-c709-421a-aeda-34627b7e3164 service nova] Acquiring lock "refresh_cache-918d5b35-c46d-47c5-b59f-175831aed9d6" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 954.186237] env[63515]: INFO nova.compute.manager [None req-db106ad7-1693-4446-8352-ddb2ff9c25ce tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: eab16df9-7bb5-4576-bca0-769a561c5fe9] Terminating instance [ 954.188797] env[63515]: DEBUG nova.compute.manager [None req-db106ad7-1693-4446-8352-ddb2ff9c25ce tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: eab16df9-7bb5-4576-bca0-769a561c5fe9] Start destroying the instance on the hypervisor. {{(pid=63515) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 954.189033] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-db106ad7-1693-4446-8352-ddb2ff9c25ce tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: eab16df9-7bb5-4576-bca0-769a561c5fe9] Destroying instance {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 954.190330] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3b7a7b8-a899-4b82-bda9-dd23c0b8174c {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.194823] env[63515]: DEBUG oslo_vmware.api [None req-127af73c-afc8-4436-9075-a886f5e9c7d5 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Waiting for the task: (returnval){ [ 954.194823] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]52eefc38-6848-df86-e1d4-7d160a498de4" [ 954.194823] env[63515]: _type = "Task" [ 954.194823] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 954.201124] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-db106ad7-1693-4446-8352-ddb2ff9c25ce tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: eab16df9-7bb5-4576-bca0-769a561c5fe9] Powering off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 954.201729] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0bc06ff6-d69d-4ddb-bf55-bc42855fb14e {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.206634] env[63515]: DEBUG oslo_vmware.api [None req-127af73c-afc8-4436-9075-a886f5e9c7d5 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52eefc38-6848-df86-e1d4-7d160a498de4, 'name': SearchDatastore_Task, 'duration_secs': 0.010049} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 954.207782] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-277a6792-9ff6-4979-9c12-309f49b51239 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.211462] env[63515]: DEBUG oslo_vmware.api [None req-db106ad7-1693-4446-8352-ddb2ff9c25ce tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Waiting for the task: (returnval){ [ 954.211462] env[63515]: value = "task-1111649" [ 954.211462] env[63515]: _type = "Task" [ 954.211462] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 954.220531] env[63515]: DEBUG oslo_vmware.api [None req-127af73c-afc8-4436-9075-a886f5e9c7d5 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Waiting for the task: (returnval){ [ 954.220531] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]52526d53-11c1-174c-ab8e-1064d644db59" [ 954.220531] env[63515]: _type = "Task" [ 954.220531] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 954.228211] env[63515]: DEBUG oslo_vmware.api [None req-db106ad7-1693-4446-8352-ddb2ff9c25ce tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Task: {'id': task-1111649, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.234905] env[63515]: DEBUG oslo_vmware.api [None req-127af73c-afc8-4436-9075-a886f5e9c7d5 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52526d53-11c1-174c-ab8e-1064d644db59, 'name': SearchDatastore_Task, 'duration_secs': 0.010141} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 954.235234] env[63515]: DEBUG oslo_concurrency.lockutils [None req-127af73c-afc8-4436-9075-a886f5e9c7d5 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 954.235557] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-127af73c-afc8-4436-9075-a886f5e9c7d5 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk to [datastore2] 7d572ac0-f6af-4622-96cc-e75983420222/7d572ac0-f6af-4622-96cc-e75983420222.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 954.235832] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7b2f755f-2fc1-4971-9296-9903549bfd4f {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.243745] env[63515]: DEBUG oslo_vmware.api [None req-127af73c-afc8-4436-9075-a886f5e9c7d5 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Waiting for the task: (returnval){ [ 954.243745] env[63515]: value = "task-1111650" [ 954.243745] env[63515]: _type = "Task" [ 954.243745] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 954.252036] env[63515]: DEBUG oslo_vmware.api [None req-127af73c-afc8-4436-9075-a886f5e9c7d5 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Task: {'id': task-1111650, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.423071] env[63515]: DEBUG oslo_concurrency.lockutils [req-6343619f-96e3-4df8-a135-d8ef272aebe2 req-1693b6bd-1e3f-493c-bd2e-27875bd61372 service nova] Releasing lock "refresh_cache-7d572ac0-f6af-4622-96cc-e75983420222" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 954.434778] env[63515]: DEBUG nova.scheduler.client.report [None req-3f7a26fd-cc84-46d9-9d28-feb378f8b5e6 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 954.546948] env[63515]: DEBUG nova.compute.manager [None req-3f7a26fd-cc84-46d9-9d28-feb378f8b5e6 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] [instance: 75f4dc79-4b73-4ddc-be03-3653a5a0797e] Start spawning the instance on the hypervisor. {{(pid=63515) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 954.574621] env[63515]: DEBUG nova.virt.hardware [None req-3f7a26fd-cc84-46d9-9d28-feb378f8b5e6 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T02:52:14Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T02:51:56Z,direct_url=,disk_format='vmdk',id=8a120570-cb06-4099-b262-554ca0ad15c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b569255cc43e42e7ae3f2b4ad37c6ef4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T02:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 954.574910] env[63515]: DEBUG nova.virt.hardware [None req-3f7a26fd-cc84-46d9-9d28-feb378f8b5e6 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Flavor limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 954.575133] env[63515]: DEBUG nova.virt.hardware [None req-3f7a26fd-cc84-46d9-9d28-feb378f8b5e6 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Image limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 954.575276] env[63515]: DEBUG nova.virt.hardware [None req-3f7a26fd-cc84-46d9-9d28-feb378f8b5e6 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Flavor pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 954.575436] env[63515]: DEBUG nova.virt.hardware [None req-3f7a26fd-cc84-46d9-9d28-feb378f8b5e6 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Image pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 954.575569] env[63515]: DEBUG nova.virt.hardware [None req-3f7a26fd-cc84-46d9-9d28-feb378f8b5e6 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 954.575788] env[63515]: DEBUG nova.virt.hardware [None req-3f7a26fd-cc84-46d9-9d28-feb378f8b5e6 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 954.575962] env[63515]: DEBUG nova.virt.hardware [None req-3f7a26fd-cc84-46d9-9d28-feb378f8b5e6 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 954.576152] env[63515]: DEBUG nova.virt.hardware [None req-3f7a26fd-cc84-46d9-9d28-feb378f8b5e6 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Got 1 possible topologies {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 954.576373] env[63515]: DEBUG nova.virt.hardware [None req-3f7a26fd-cc84-46d9-9d28-feb378f8b5e6 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 954.576599] env[63515]: DEBUG nova.virt.hardware [None req-3f7a26fd-cc84-46d9-9d28-feb378f8b5e6 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 954.577548] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-908db197-8c8e-4340-9890-28ea68d4f89e {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.586418] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a611864e-e3d6-49b5-9eb3-0c7fa8f3fe5a {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.591009] env[63515]: DEBUG oslo_concurrency.lockutils [None req-857b4940-0702-491c-a977-f4c63c5dd5f5 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Releasing lock "refresh_cache-918d5b35-c46d-47c5-b59f-175831aed9d6" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 954.591349] env[63515]: DEBUG nova.compute.manager [None req-857b4940-0702-491c-a977-f4c63c5dd5f5 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: 918d5b35-c46d-47c5-b59f-175831aed9d6] Instance network_info: |[{"id": "b4724c8b-a2c7-4f52-8a75-f46434d7b527", "address": "fa:16:3e:dc:ce:b2", "network": {"id": "cec614cb-7a5b-4b16-a83d-061ff62d6677", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1501468300-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "aef947bfaf7d4180b46b4be662e3d903", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3a0ddd7d-c321-4187-bdd8-b19044ea2c4a", "external-id": "nsx-vlan-transportzone-747", "segmentation_id": 747, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb4724c8b-a2", "ovs_interfaceid": "b4724c8b-a2c7-4f52-8a75-f46434d7b527", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 954.591672] env[63515]: DEBUG oslo_concurrency.lockutils [req-9da6f36c-04ac-4e64-8c0e-31ea36e644b2 req-73be1d33-c709-421a-aeda-34627b7e3164 service nova] Acquired lock "refresh_cache-918d5b35-c46d-47c5-b59f-175831aed9d6" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 954.591904] env[63515]: DEBUG nova.network.neutron [req-9da6f36c-04ac-4e64-8c0e-31ea36e644b2 req-73be1d33-c709-421a-aeda-34627b7e3164 service nova] [instance: 918d5b35-c46d-47c5-b59f-175831aed9d6] Refreshing network info cache for port b4724c8b-a2c7-4f52-8a75-f46434d7b527 {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 954.594019] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-857b4940-0702-491c-a977-f4c63c5dd5f5 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: 918d5b35-c46d-47c5-b59f-175831aed9d6] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:dc:ce:b2', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '3a0ddd7d-c321-4187-bdd8-b19044ea2c4a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b4724c8b-a2c7-4f52-8a75-f46434d7b527', 'vif_model': 'vmxnet3'}] {{(pid=63515) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 954.602149] env[63515]: DEBUG oslo.service.loopingcall [None req-857b4940-0702-491c-a977-f4c63c5dd5f5 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 954.603569] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 918d5b35-c46d-47c5-b59f-175831aed9d6] Creating VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 954.603909] env[63515]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-afe823cc-d091-4272-ac1f-fd67a1f6667f {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.637691] env[63515]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 954.637691] env[63515]: value = "task-1111651" [ 954.637691] env[63515]: _type = "Task" [ 954.637691] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 954.650476] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1111651, 'name': CreateVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.723921] env[63515]: DEBUG oslo_concurrency.lockutils [None req-2359bb47-ebaa-4b05-a6cb-20ad2f0f493e tempest-ServerMetadataTestJSON-1034558863 tempest-ServerMetadataTestJSON-1034558863-project-member] Acquiring lock "b25bf2ba-cf08-44ac-a524-aa87cf785575" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 954.724365] env[63515]: DEBUG oslo_concurrency.lockutils [None req-2359bb47-ebaa-4b05-a6cb-20ad2f0f493e tempest-ServerMetadataTestJSON-1034558863 tempest-ServerMetadataTestJSON-1034558863-project-member] Lock "b25bf2ba-cf08-44ac-a524-aa87cf785575" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 954.724661] env[63515]: DEBUG oslo_concurrency.lockutils [None req-2359bb47-ebaa-4b05-a6cb-20ad2f0f493e tempest-ServerMetadataTestJSON-1034558863 tempest-ServerMetadataTestJSON-1034558863-project-member] Acquiring lock "b25bf2ba-cf08-44ac-a524-aa87cf785575-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 954.724843] env[63515]: DEBUG oslo_concurrency.lockutils [None req-2359bb47-ebaa-4b05-a6cb-20ad2f0f493e tempest-ServerMetadataTestJSON-1034558863 tempest-ServerMetadataTestJSON-1034558863-project-member] Lock "b25bf2ba-cf08-44ac-a524-aa87cf785575-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 954.725043] env[63515]: DEBUG oslo_concurrency.lockutils [None req-2359bb47-ebaa-4b05-a6cb-20ad2f0f493e tempest-ServerMetadataTestJSON-1034558863 tempest-ServerMetadataTestJSON-1034558863-project-member] Lock "b25bf2ba-cf08-44ac-a524-aa87cf785575-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 954.727077] env[63515]: DEBUG oslo_vmware.api [None req-db106ad7-1693-4446-8352-ddb2ff9c25ce tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Task: {'id': task-1111649, 'name': PowerOffVM_Task, 'duration_secs': 0.210876} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 954.727606] env[63515]: INFO nova.compute.manager [None req-2359bb47-ebaa-4b05-a6cb-20ad2f0f493e tempest-ServerMetadataTestJSON-1034558863 tempest-ServerMetadataTestJSON-1034558863-project-member] [instance: b25bf2ba-cf08-44ac-a524-aa87cf785575] Terminating instance [ 954.729144] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-db106ad7-1693-4446-8352-ddb2ff9c25ce tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: eab16df9-7bb5-4576-bca0-769a561c5fe9] Powered off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 954.729328] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-db106ad7-1693-4446-8352-ddb2ff9c25ce tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: eab16df9-7bb5-4576-bca0-769a561c5fe9] Unregistering the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 954.729936] env[63515]: DEBUG nova.compute.manager [None req-2359bb47-ebaa-4b05-a6cb-20ad2f0f493e tempest-ServerMetadataTestJSON-1034558863 tempest-ServerMetadataTestJSON-1034558863-project-member] [instance: b25bf2ba-cf08-44ac-a524-aa87cf785575] Start destroying the instance on the hypervisor. {{(pid=63515) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 954.730169] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-2359bb47-ebaa-4b05-a6cb-20ad2f0f493e tempest-ServerMetadataTestJSON-1034558863 tempest-ServerMetadataTestJSON-1034558863-project-member] [instance: b25bf2ba-cf08-44ac-a524-aa87cf785575] Destroying instance {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 954.730487] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-4807a186-b2c1-43bd-93f1-22cd1b831ae4 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.732775] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8110991-86e5-4f4a-8e0f-58619cc05bcd {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.741295] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-2359bb47-ebaa-4b05-a6cb-20ad2f0f493e tempest-ServerMetadataTestJSON-1034558863 tempest-ServerMetadataTestJSON-1034558863-project-member] [instance: b25bf2ba-cf08-44ac-a524-aa87cf785575] Powering off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 954.741511] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b6826524-beb2-42cf-84bd-f3393036f86a {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.749578] env[63515]: DEBUG oslo_vmware.api [None req-2359bb47-ebaa-4b05-a6cb-20ad2f0f493e tempest-ServerMetadataTestJSON-1034558863 tempest-ServerMetadataTestJSON-1034558863-project-member] Waiting for the task: (returnval){ [ 954.749578] env[63515]: value = "task-1111653" [ 954.749578] env[63515]: _type = "Task" [ 954.749578] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 954.756476] env[63515]: DEBUG oslo_vmware.api [None req-127af73c-afc8-4436-9075-a886f5e9c7d5 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Task: {'id': task-1111650, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.474709} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 954.757174] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-127af73c-afc8-4436-9075-a886f5e9c7d5 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk to [datastore2] 7d572ac0-f6af-4622-96cc-e75983420222/7d572ac0-f6af-4622-96cc-e75983420222.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 954.757429] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-127af73c-afc8-4436-9075-a886f5e9c7d5 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] [instance: 7d572ac0-f6af-4622-96cc-e75983420222] Extending root virtual disk to 1048576 {{(pid=63515) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 954.757674] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-957cd50d-6b36-466e-b555-d0bc3882335a {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.762670] env[63515]: DEBUG oslo_vmware.api [None req-2359bb47-ebaa-4b05-a6cb-20ad2f0f493e tempest-ServerMetadataTestJSON-1034558863 tempest-ServerMetadataTestJSON-1034558863-project-member] Task: {'id': task-1111653, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.767736] env[63515]: DEBUG oslo_vmware.api [None req-127af73c-afc8-4436-9075-a886f5e9c7d5 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Waiting for the task: (returnval){ [ 954.767736] env[63515]: value = "task-1111654" [ 954.767736] env[63515]: _type = "Task" [ 954.767736] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 954.775540] env[63515]: DEBUG oslo_vmware.api [None req-127af73c-afc8-4436-9075-a886f5e9c7d5 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Task: {'id': task-1111654, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.809140] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-db106ad7-1693-4446-8352-ddb2ff9c25ce tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: eab16df9-7bb5-4576-bca0-769a561c5fe9] Unregistered the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 954.809478] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-db106ad7-1693-4446-8352-ddb2ff9c25ce tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: eab16df9-7bb5-4576-bca0-769a561c5fe9] Deleting contents of the VM from datastore datastore1 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 954.809715] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-db106ad7-1693-4446-8352-ddb2ff9c25ce tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Deleting the datastore file [datastore1] eab16df9-7bb5-4576-bca0-769a561c5fe9 {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 954.810029] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-8c7d9cb5-4d33-4121-a4f7-69faaa80c488 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.816813] env[63515]: DEBUG oslo_vmware.api [None req-db106ad7-1693-4446-8352-ddb2ff9c25ce tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Waiting for the task: (returnval){ [ 954.816813] env[63515]: value = "task-1111655" [ 954.816813] env[63515]: _type = "Task" [ 954.816813] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 954.825412] env[63515]: DEBUG oslo_vmware.api [None req-db106ad7-1693-4446-8352-ddb2ff9c25ce tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Task: {'id': task-1111655, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.922946] env[63515]: DEBUG oslo_concurrency.lockutils [None req-2099b375-1441-4c83-88b0-3c0c09203cee tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Acquiring lock "86b1ddcf-585c-41d9-84fe-17b794f1abbc" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 954.923352] env[63515]: DEBUG oslo_concurrency.lockutils [None req-2099b375-1441-4c83-88b0-3c0c09203cee tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Lock "86b1ddcf-585c-41d9-84fe-17b794f1abbc" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 954.942076] env[63515]: DEBUG oslo_concurrency.lockutils [None req-3f7a26fd-cc84-46d9-9d28-feb378f8b5e6 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.422s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 954.942623] env[63515]: DEBUG nova.compute.manager [None req-3f7a26fd-cc84-46d9-9d28-feb378f8b5e6 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] [instance: d438dd30-b816-48a6-94c6-11e6e9bd5b71] Start building networks asynchronously for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 954.946830] env[63515]: DEBUG oslo_concurrency.lockutils [None req-b94389d0-616d-4e7e-a415-7b0c3a16a683 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 10.633s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 955.063301] env[63515]: DEBUG nova.network.neutron [None req-3f7a26fd-cc84-46d9-9d28-feb378f8b5e6 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] [instance: 75f4dc79-4b73-4ddc-be03-3653a5a0797e] Successfully updated port: 00eda8ad-1a52-4c47-b065-a82c08da7980 {{(pid=63515) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 955.147976] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1111651, 'name': CreateVM_Task, 'duration_secs': 0.438076} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 955.148159] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 918d5b35-c46d-47c5-b59f-175831aed9d6] Created VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 955.148839] env[63515]: DEBUG oslo_concurrency.lockutils [None req-857b4940-0702-491c-a977-f4c63c5dd5f5 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 955.149024] env[63515]: DEBUG oslo_concurrency.lockutils [None req-857b4940-0702-491c-a977-f4c63c5dd5f5 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 955.149352] env[63515]: DEBUG oslo_concurrency.lockutils [None req-857b4940-0702-491c-a977-f4c63c5dd5f5 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 955.149614] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5d032736-0aac-4dc0-ba9b-cb6872c26d5c {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.154317] env[63515]: DEBUG oslo_vmware.api [None req-857b4940-0702-491c-a977-f4c63c5dd5f5 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Waiting for the task: (returnval){ [ 955.154317] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]52430bd0-f37c-dc0a-120b-1be4204b9a6e" [ 955.154317] env[63515]: _type = "Task" [ 955.154317] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 955.162827] env[63515]: DEBUG oslo_vmware.api [None req-857b4940-0702-491c-a977-f4c63c5dd5f5 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52430bd0-f37c-dc0a-120b-1be4204b9a6e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.259838] env[63515]: DEBUG oslo_vmware.api [None req-2359bb47-ebaa-4b05-a6cb-20ad2f0f493e tempest-ServerMetadataTestJSON-1034558863 tempest-ServerMetadataTestJSON-1034558863-project-member] Task: {'id': task-1111653, 'name': PowerOffVM_Task, 'duration_secs': 0.213359} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 955.260111] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-2359bb47-ebaa-4b05-a6cb-20ad2f0f493e tempest-ServerMetadataTestJSON-1034558863 tempest-ServerMetadataTestJSON-1034558863-project-member] [instance: b25bf2ba-cf08-44ac-a524-aa87cf785575] Powered off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 955.260301] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-2359bb47-ebaa-4b05-a6cb-20ad2f0f493e tempest-ServerMetadataTestJSON-1034558863 tempest-ServerMetadataTestJSON-1034558863-project-member] [instance: b25bf2ba-cf08-44ac-a524-aa87cf785575] Unregistering the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 955.260602] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-fc1dd98f-b9ca-4cb2-9cd6-4dcdb8c91480 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.278038] env[63515]: DEBUG oslo_vmware.api [None req-127af73c-afc8-4436-9075-a886f5e9c7d5 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Task: {'id': task-1111654, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.101988} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 955.278857] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-127af73c-afc8-4436-9075-a886f5e9c7d5 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] [instance: 7d572ac0-f6af-4622-96cc-e75983420222] Extended root virtual disk {{(pid=63515) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 955.279927] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1584350c-a304-4210-834b-5c0eb0941778 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.305336] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-127af73c-afc8-4436-9075-a886f5e9c7d5 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] [instance: 7d572ac0-f6af-4622-96cc-e75983420222] Reconfiguring VM instance instance-00000058 to attach disk [datastore2] 7d572ac0-f6af-4622-96cc-e75983420222/7d572ac0-f6af-4622-96cc-e75983420222.vmdk or device None with type sparse {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 955.308084] env[63515]: DEBUG nova.network.neutron [req-9da6f36c-04ac-4e64-8c0e-31ea36e644b2 req-73be1d33-c709-421a-aeda-34627b7e3164 service nova] [instance: 918d5b35-c46d-47c5-b59f-175831aed9d6] Updated VIF entry in instance network info cache for port b4724c8b-a2c7-4f52-8a75-f46434d7b527. {{(pid=63515) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 955.308433] env[63515]: DEBUG nova.network.neutron [req-9da6f36c-04ac-4e64-8c0e-31ea36e644b2 req-73be1d33-c709-421a-aeda-34627b7e3164 service nova] [instance: 918d5b35-c46d-47c5-b59f-175831aed9d6] Updating instance_info_cache with network_info: [{"id": "b4724c8b-a2c7-4f52-8a75-f46434d7b527", "address": "fa:16:3e:dc:ce:b2", "network": {"id": "cec614cb-7a5b-4b16-a83d-061ff62d6677", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1501468300-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "aef947bfaf7d4180b46b4be662e3d903", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3a0ddd7d-c321-4187-bdd8-b19044ea2c4a", "external-id": "nsx-vlan-transportzone-747", "segmentation_id": 747, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb4724c8b-a2", "ovs_interfaceid": "b4724c8b-a2c7-4f52-8a75-f46434d7b527", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 955.309598] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b5d9030e-f9e4-4232-b75f-e50b3e1fc91e {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.332326] env[63515]: DEBUG oslo_vmware.rw_handles [None req-8e6475b3-95eb-4b44-a2f3-5ddd5db39f52 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52c2c74c-f165-626c-5ca1-1950d805bd0e/disk-0.vmdk. {{(pid=63515) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 955.333534] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0f688ee-7e93-485c-84d8-cc98befa2c53 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.341164] env[63515]: DEBUG oslo_vmware.api [None req-db106ad7-1693-4446-8352-ddb2ff9c25ce tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Task: {'id': task-1111655, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.207146} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 955.341374] env[63515]: DEBUG oslo_vmware.api [None req-127af73c-afc8-4436-9075-a886f5e9c7d5 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Waiting for the task: (returnval){ [ 955.341374] env[63515]: value = "task-1111657" [ 955.341374] env[63515]: _type = "Task" [ 955.341374] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 955.342026] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-db106ad7-1693-4446-8352-ddb2ff9c25ce tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Deleted the datastore file {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 955.342229] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-db106ad7-1693-4446-8352-ddb2ff9c25ce tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: eab16df9-7bb5-4576-bca0-769a561c5fe9] Deleted contents of the VM from datastore datastore1 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 955.342409] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-db106ad7-1693-4446-8352-ddb2ff9c25ce tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: eab16df9-7bb5-4576-bca0-769a561c5fe9] Instance destroyed {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 955.342605] env[63515]: INFO nova.compute.manager [None req-db106ad7-1693-4446-8352-ddb2ff9c25ce tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: eab16df9-7bb5-4576-bca0-769a561c5fe9] Took 1.15 seconds to destroy the instance on the hypervisor. [ 955.342821] env[63515]: DEBUG oslo.service.loopingcall [None req-db106ad7-1693-4446-8352-ddb2ff9c25ce tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 955.343514] env[63515]: DEBUG nova.compute.manager [-] [instance: eab16df9-7bb5-4576-bca0-769a561c5fe9] Deallocating network for instance {{(pid=63515) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 955.343620] env[63515]: DEBUG nova.network.neutron [-] [instance: eab16df9-7bb5-4576-bca0-769a561c5fe9] deallocate_for_instance() {{(pid=63515) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 955.350427] env[63515]: DEBUG oslo_vmware.rw_handles [None req-8e6475b3-95eb-4b44-a2f3-5ddd5db39f52 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52c2c74c-f165-626c-5ca1-1950d805bd0e/disk-0.vmdk is in state: ready. {{(pid=63515) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 955.350690] env[63515]: ERROR oslo_vmware.rw_handles [None req-8e6475b3-95eb-4b44-a2f3-5ddd5db39f52 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52c2c74c-f165-626c-5ca1-1950d805bd0e/disk-0.vmdk due to incomplete transfer. [ 955.351055] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-2359bb47-ebaa-4b05-a6cb-20ad2f0f493e tempest-ServerMetadataTestJSON-1034558863 tempest-ServerMetadataTestJSON-1034558863-project-member] [instance: b25bf2ba-cf08-44ac-a524-aa87cf785575] Unregistered the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 955.351055] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-2359bb47-ebaa-4b05-a6cb-20ad2f0f493e tempest-ServerMetadataTestJSON-1034558863 tempest-ServerMetadataTestJSON-1034558863-project-member] [instance: b25bf2ba-cf08-44ac-a524-aa87cf785575] Deleting contents of the VM from datastore datastore2 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 955.351264] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-2359bb47-ebaa-4b05-a6cb-20ad2f0f493e tempest-ServerMetadataTestJSON-1034558863 tempest-ServerMetadataTestJSON-1034558863-project-member] Deleting the datastore file [datastore2] b25bf2ba-cf08-44ac-a524-aa87cf785575 {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 955.351955] env[63515]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-326543a9-5e76-4a7d-8820-7be99903e8bd {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.353364] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f3d12830-93a8-4270-a148-09126fee5187 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.358103] env[63515]: DEBUG oslo_vmware.api [None req-127af73c-afc8-4436-9075-a886f5e9c7d5 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Task: {'id': task-1111657, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.364030] env[63515]: DEBUG oslo_vmware.rw_handles [None req-8e6475b3-95eb-4b44-a2f3-5ddd5db39f52 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52c2c74c-f165-626c-5ca1-1950d805bd0e/disk-0.vmdk. {{(pid=63515) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 955.364030] env[63515]: DEBUG nova.virt.vmwareapi.images [None req-8e6475b3-95eb-4b44-a2f3-5ddd5db39f52 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 68766cd4-84be-475b-8494-d7ab43a9e969] Uploaded image 4978423f-d581-473b-85c6-34574f05d1d5 to the Glance image server {{(pid=63515) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 955.365847] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-8e6475b3-95eb-4b44-a2f3-5ddd5db39f52 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 68766cd4-84be-475b-8494-d7ab43a9e969] Destroying the VM {{(pid=63515) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 955.367137] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-f551190e-cbf3-4919-9301-e10b8d1e02fe {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.369107] env[63515]: DEBUG oslo_vmware.api [None req-2359bb47-ebaa-4b05-a6cb-20ad2f0f493e tempest-ServerMetadataTestJSON-1034558863 tempest-ServerMetadataTestJSON-1034558863-project-member] Waiting for the task: (returnval){ [ 955.369107] env[63515]: value = "task-1111658" [ 955.369107] env[63515]: _type = "Task" [ 955.369107] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 955.373578] env[63515]: DEBUG oslo_vmware.api [None req-8e6475b3-95eb-4b44-a2f3-5ddd5db39f52 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Waiting for the task: (returnval){ [ 955.373578] env[63515]: value = "task-1111659" [ 955.373578] env[63515]: _type = "Task" [ 955.373578] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 955.380168] env[63515]: DEBUG oslo_vmware.api [None req-2359bb47-ebaa-4b05-a6cb-20ad2f0f493e tempest-ServerMetadataTestJSON-1034558863 tempest-ServerMetadataTestJSON-1034558863-project-member] Task: {'id': task-1111658, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.385640] env[63515]: DEBUG oslo_vmware.api [None req-8e6475b3-95eb-4b44-a2f3-5ddd5db39f52 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Task: {'id': task-1111659, 'name': Destroy_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.426829] env[63515]: DEBUG nova.compute.utils [None req-2099b375-1441-4c83-88b0-3c0c09203cee tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Using /dev/sd instead of None {{(pid=63515) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 955.434600] env[63515]: DEBUG neutronclient.v2_0.client [-] Error message: {"NeutronError": {"type": "PortNotFound", "message": "Port ff4043f6-7576-4093-80ed-d36240b37f2f could not be found.", "detail": ""}} {{(pid=63515) _handle_fault_response /opt/stack/data/venv/lib/python3.10/site-packages/neutronclient/v2_0/client.py:262}} [ 955.434880] env[63515]: DEBUG nova.network.neutron [-] Unable to show port ff4043f6-7576-4093-80ed-d36240b37f2f as it no longer exists. {{(pid=63515) _unbind_ports /opt/stack/nova/nova/network/neutron.py:666}} [ 955.447289] env[63515]: DEBUG nova.compute.utils [None req-3f7a26fd-cc84-46d9-9d28-feb378f8b5e6 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Using /dev/sd instead of None {{(pid=63515) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 955.448651] env[63515]: DEBUG nova.compute.manager [None req-3f7a26fd-cc84-46d9-9d28-feb378f8b5e6 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] [instance: d438dd30-b816-48a6-94c6-11e6e9bd5b71] Allocating IP information in the background. {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 955.448827] env[63515]: DEBUG nova.network.neutron [None req-3f7a26fd-cc84-46d9-9d28-feb378f8b5e6 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] [instance: d438dd30-b816-48a6-94c6-11e6e9bd5b71] allocate_for_instance() {{(pid=63515) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 955.519795] env[63515]: DEBUG nova.policy [None req-3f7a26fd-cc84-46d9-9d28-feb378f8b5e6 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e8b57c2415b44bbd93105c6824a9e322', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3dcf6179891c4841a2901fcffe43cd31', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63515) authorize /opt/stack/nova/nova/policy.py:201}} [ 955.566081] env[63515]: DEBUG oslo_concurrency.lockutils [None req-3f7a26fd-cc84-46d9-9d28-feb378f8b5e6 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Acquiring lock "refresh_cache-75f4dc79-4b73-4ddc-be03-3653a5a0797e" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 955.566242] env[63515]: DEBUG oslo_concurrency.lockutils [None req-3f7a26fd-cc84-46d9-9d28-feb378f8b5e6 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Acquired lock "refresh_cache-75f4dc79-4b73-4ddc-be03-3653a5a0797e" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 955.566447] env[63515]: DEBUG nova.network.neutron [None req-3f7a26fd-cc84-46d9-9d28-feb378f8b5e6 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] [instance: 75f4dc79-4b73-4ddc-be03-3653a5a0797e] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 955.671728] env[63515]: DEBUG oslo_vmware.api [None req-857b4940-0702-491c-a977-f4c63c5dd5f5 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52430bd0-f37c-dc0a-120b-1be4204b9a6e, 'name': SearchDatastore_Task, 'duration_secs': 0.011241} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 955.672404] env[63515]: DEBUG oslo_concurrency.lockutils [None req-857b4940-0702-491c-a977-f4c63c5dd5f5 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 955.672791] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-857b4940-0702-491c-a977-f4c63c5dd5f5 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: 918d5b35-c46d-47c5-b59f-175831aed9d6] Processing image 8a120570-cb06-4099-b262-554ca0ad15c5 {{(pid=63515) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 955.673245] env[63515]: DEBUG oslo_concurrency.lockutils [None req-857b4940-0702-491c-a977-f4c63c5dd5f5 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 955.673477] env[63515]: DEBUG oslo_concurrency.lockutils [None req-857b4940-0702-491c-a977-f4c63c5dd5f5 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 955.674018] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-857b4940-0702-491c-a977-f4c63c5dd5f5 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 955.674449] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2fabd1ea-ec67-49c4-9e34-6829efeee6ef {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.689324] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-857b4940-0702-491c-a977-f4c63c5dd5f5 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 955.689934] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-857b4940-0702-491c-a977-f4c63c5dd5f5 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63515) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 955.690361] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4abcc836-390a-4ec3-929d-c648b479b7dc {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.696458] env[63515]: DEBUG oslo_vmware.api [None req-857b4940-0702-491c-a977-f4c63c5dd5f5 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Waiting for the task: (returnval){ [ 955.696458] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]5233a445-1cd1-1f3b-913d-fafdbbc52e55" [ 955.696458] env[63515]: _type = "Task" [ 955.696458] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 955.708240] env[63515]: DEBUG oslo_vmware.api [None req-857b4940-0702-491c-a977-f4c63c5dd5f5 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]5233a445-1cd1-1f3b-913d-fafdbbc52e55, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.742448] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10fe806a-7c89-465a-9ba4-997df31ca2ca {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.750026] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8132c767-e841-453f-b646-3cb9f1670828 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.782755] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42e62666-2cb3-453f-9aab-a6976d44e153 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.793650] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69d2053e-80d6-4a63-8f5b-81d8a182e0df {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.809014] env[63515]: DEBUG nova.compute.provider_tree [None req-b94389d0-616d-4e7e-a415-7b0c3a16a683 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 955.826301] env[63515]: DEBUG oslo_concurrency.lockutils [req-9da6f36c-04ac-4e64-8c0e-31ea36e644b2 req-73be1d33-c709-421a-aeda-34627b7e3164 service nova] Releasing lock "refresh_cache-918d5b35-c46d-47c5-b59f-175831aed9d6" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 955.852998] env[63515]: DEBUG oslo_vmware.api [None req-127af73c-afc8-4436-9075-a886f5e9c7d5 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Task: {'id': task-1111657, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.884211] env[63515]: DEBUG oslo_vmware.api [None req-2359bb47-ebaa-4b05-a6cb-20ad2f0f493e tempest-ServerMetadataTestJSON-1034558863 tempest-ServerMetadataTestJSON-1034558863-project-member] Task: {'id': task-1111658, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.185414} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 955.886241] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-2359bb47-ebaa-4b05-a6cb-20ad2f0f493e tempest-ServerMetadataTestJSON-1034558863 tempest-ServerMetadataTestJSON-1034558863-project-member] Deleted the datastore file {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 955.886241] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-2359bb47-ebaa-4b05-a6cb-20ad2f0f493e tempest-ServerMetadataTestJSON-1034558863 tempest-ServerMetadataTestJSON-1034558863-project-member] [instance: b25bf2ba-cf08-44ac-a524-aa87cf785575] Deleted contents of the VM from datastore datastore2 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 955.886241] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-2359bb47-ebaa-4b05-a6cb-20ad2f0f493e tempest-ServerMetadataTestJSON-1034558863 tempest-ServerMetadataTestJSON-1034558863-project-member] [instance: b25bf2ba-cf08-44ac-a524-aa87cf785575] Instance destroyed {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 955.886241] env[63515]: INFO nova.compute.manager [None req-2359bb47-ebaa-4b05-a6cb-20ad2f0f493e tempest-ServerMetadataTestJSON-1034558863 tempest-ServerMetadataTestJSON-1034558863-project-member] [instance: b25bf2ba-cf08-44ac-a524-aa87cf785575] Took 1.16 seconds to destroy the instance on the hypervisor. [ 955.889769] env[63515]: DEBUG oslo.service.loopingcall [None req-2359bb47-ebaa-4b05-a6cb-20ad2f0f493e tempest-ServerMetadataTestJSON-1034558863 tempest-ServerMetadataTestJSON-1034558863-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 955.889769] env[63515]: DEBUG nova.compute.manager [-] [instance: b25bf2ba-cf08-44ac-a524-aa87cf785575] Deallocating network for instance {{(pid=63515) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 955.889769] env[63515]: DEBUG nova.network.neutron [-] [instance: b25bf2ba-cf08-44ac-a524-aa87cf785575] deallocate_for_instance() {{(pid=63515) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 955.892774] env[63515]: DEBUG oslo_vmware.api [None req-8e6475b3-95eb-4b44-a2f3-5ddd5db39f52 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Task: {'id': task-1111659, 'name': Destroy_Task, 'duration_secs': 0.416092} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 955.893338] env[63515]: INFO nova.virt.vmwareapi.vm_util [None req-8e6475b3-95eb-4b44-a2f3-5ddd5db39f52 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 68766cd4-84be-475b-8494-d7ab43a9e969] Destroyed the VM [ 955.893578] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-8e6475b3-95eb-4b44-a2f3-5ddd5db39f52 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 68766cd4-84be-475b-8494-d7ab43a9e969] Deleting Snapshot of the VM instance {{(pid=63515) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 955.893834] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-811eccd7-ca61-49c8-a23a-cf6265d3a5e3 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.901695] env[63515]: DEBUG oslo_vmware.api [None req-8e6475b3-95eb-4b44-a2f3-5ddd5db39f52 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Waiting for the task: (returnval){ [ 955.901695] env[63515]: value = "task-1111660" [ 955.901695] env[63515]: _type = "Task" [ 955.901695] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 955.910444] env[63515]: DEBUG oslo_vmware.api [None req-8e6475b3-95eb-4b44-a2f3-5ddd5db39f52 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Task: {'id': task-1111660, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.929380] env[63515]: DEBUG oslo_concurrency.lockutils [None req-2099b375-1441-4c83-88b0-3c0c09203cee tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Lock "86b1ddcf-585c-41d9-84fe-17b794f1abbc" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.006s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 955.947360] env[63515]: DEBUG nova.network.neutron [None req-3f7a26fd-cc84-46d9-9d28-feb378f8b5e6 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] [instance: d438dd30-b816-48a6-94c6-11e6e9bd5b71] Successfully created port: a673b83a-c9d4-425b-b6e3-aad0d17243ba {{(pid=63515) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 955.952503] env[63515]: DEBUG nova.compute.manager [None req-3f7a26fd-cc84-46d9-9d28-feb378f8b5e6 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] [instance: d438dd30-b816-48a6-94c6-11e6e9bd5b71] Start building block device mappings for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 956.118530] env[63515]: DEBUG nova.network.neutron [None req-3f7a26fd-cc84-46d9-9d28-feb378f8b5e6 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] [instance: 75f4dc79-4b73-4ddc-be03-3653a5a0797e] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 956.208572] env[63515]: DEBUG oslo_vmware.api [None req-857b4940-0702-491c-a977-f4c63c5dd5f5 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]5233a445-1cd1-1f3b-913d-fafdbbc52e55, 'name': SearchDatastore_Task, 'duration_secs': 0.011913} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 956.209578] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-08ac104e-6bc8-4517-bd94-a55eba5b54dc {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.218715] env[63515]: DEBUG oslo_vmware.api [None req-857b4940-0702-491c-a977-f4c63c5dd5f5 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Waiting for the task: (returnval){ [ 956.218715] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]521860d3-6992-b044-e43b-1378368b65f1" [ 956.218715] env[63515]: _type = "Task" [ 956.218715] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 956.229968] env[63515]: DEBUG oslo_vmware.api [None req-857b4940-0702-491c-a977-f4c63c5dd5f5 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]521860d3-6992-b044-e43b-1378368b65f1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.292850] env[63515]: DEBUG nova.network.neutron [None req-3f7a26fd-cc84-46d9-9d28-feb378f8b5e6 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] [instance: 75f4dc79-4b73-4ddc-be03-3653a5a0797e] Updating instance_info_cache with network_info: [{"id": "00eda8ad-1a52-4c47-b065-a82c08da7980", "address": "fa:16:3e:02:56:38", "network": {"id": "549c4fb2-0ed6-44ee-bfd1-9e59bb7aed27", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-949461070-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3dcf6179891c4841a2901fcffe43cd31", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8c3e2368-4a35-4aa5-9135-23daedbbf9ef", "external-id": "nsx-vlan-transportzone-125", "segmentation_id": 125, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap00eda8ad-1a", "ovs_interfaceid": "00eda8ad-1a52-4c47-b065-a82c08da7980", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 956.314017] env[63515]: DEBUG nova.scheduler.client.report [None req-b94389d0-616d-4e7e-a415-7b0c3a16a683 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 956.354804] env[63515]: DEBUG oslo_vmware.api [None req-127af73c-afc8-4436-9075-a886f5e9c7d5 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Task: {'id': task-1111657, 'name': ReconfigVM_Task, 'duration_secs': 0.9613} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 956.354804] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-127af73c-afc8-4436-9075-a886f5e9c7d5 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] [instance: 7d572ac0-f6af-4622-96cc-e75983420222] Reconfigured VM instance instance-00000058 to attach disk [datastore2] 7d572ac0-f6af-4622-96cc-e75983420222/7d572ac0-f6af-4622-96cc-e75983420222.vmdk or device None with type sparse {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 956.355418] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b1bcf5bb-52b7-4ad8-abcf-b507ffb79734 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.363035] env[63515]: DEBUG oslo_vmware.api [None req-127af73c-afc8-4436-9075-a886f5e9c7d5 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Waiting for the task: (returnval){ [ 956.363035] env[63515]: value = "task-1111661" [ 956.363035] env[63515]: _type = "Task" [ 956.363035] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 956.372056] env[63515]: DEBUG oslo_vmware.api [None req-127af73c-afc8-4436-9075-a886f5e9c7d5 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Task: {'id': task-1111661, 'name': Rename_Task} progress is 5%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.384564] env[63515]: DEBUG nova.network.neutron [-] [instance: eab16df9-7bb5-4576-bca0-769a561c5fe9] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 956.414898] env[63515]: DEBUG oslo_vmware.api [None req-8e6475b3-95eb-4b44-a2f3-5ddd5db39f52 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Task: {'id': task-1111660, 'name': RemoveSnapshot_Task} progress is 12%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.528247] env[63515]: DEBUG nova.compute.manager [req-93d4a1f0-6b3c-4bf4-b86d-1539b55c5a3c req-bde21bed-9194-4f8c-b217-f21a000b1f75 service nova] [instance: 75f4dc79-4b73-4ddc-be03-3653a5a0797e] Received event network-vif-plugged-00eda8ad-1a52-4c47-b065-a82c08da7980 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 956.529075] env[63515]: DEBUG oslo_concurrency.lockutils [req-93d4a1f0-6b3c-4bf4-b86d-1539b55c5a3c req-bde21bed-9194-4f8c-b217-f21a000b1f75 service nova] Acquiring lock "75f4dc79-4b73-4ddc-be03-3653a5a0797e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 956.529075] env[63515]: DEBUG oslo_concurrency.lockutils [req-93d4a1f0-6b3c-4bf4-b86d-1539b55c5a3c req-bde21bed-9194-4f8c-b217-f21a000b1f75 service nova] Lock "75f4dc79-4b73-4ddc-be03-3653a5a0797e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 956.529075] env[63515]: DEBUG oslo_concurrency.lockutils [req-93d4a1f0-6b3c-4bf4-b86d-1539b55c5a3c req-bde21bed-9194-4f8c-b217-f21a000b1f75 service nova] Lock "75f4dc79-4b73-4ddc-be03-3653a5a0797e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 956.529224] env[63515]: DEBUG nova.compute.manager [req-93d4a1f0-6b3c-4bf4-b86d-1539b55c5a3c req-bde21bed-9194-4f8c-b217-f21a000b1f75 service nova] [instance: 75f4dc79-4b73-4ddc-be03-3653a5a0797e] No waiting events found dispatching network-vif-plugged-00eda8ad-1a52-4c47-b065-a82c08da7980 {{(pid=63515) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 956.529890] env[63515]: WARNING nova.compute.manager [req-93d4a1f0-6b3c-4bf4-b86d-1539b55c5a3c req-bde21bed-9194-4f8c-b217-f21a000b1f75 service nova] [instance: 75f4dc79-4b73-4ddc-be03-3653a5a0797e] Received unexpected event network-vif-plugged-00eda8ad-1a52-4c47-b065-a82c08da7980 for instance with vm_state building and task_state spawning. [ 956.529890] env[63515]: DEBUG nova.compute.manager [req-93d4a1f0-6b3c-4bf4-b86d-1539b55c5a3c req-bde21bed-9194-4f8c-b217-f21a000b1f75 service nova] [instance: 75f4dc79-4b73-4ddc-be03-3653a5a0797e] Received event network-changed-00eda8ad-1a52-4c47-b065-a82c08da7980 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 956.529890] env[63515]: DEBUG nova.compute.manager [req-93d4a1f0-6b3c-4bf4-b86d-1539b55c5a3c req-bde21bed-9194-4f8c-b217-f21a000b1f75 service nova] [instance: 75f4dc79-4b73-4ddc-be03-3653a5a0797e] Refreshing instance network info cache due to event network-changed-00eda8ad-1a52-4c47-b065-a82c08da7980. {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 956.530066] env[63515]: DEBUG oslo_concurrency.lockutils [req-93d4a1f0-6b3c-4bf4-b86d-1539b55c5a3c req-bde21bed-9194-4f8c-b217-f21a000b1f75 service nova] Acquiring lock "refresh_cache-75f4dc79-4b73-4ddc-be03-3653a5a0797e" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 956.696833] env[63515]: DEBUG nova.network.neutron [-] [instance: b25bf2ba-cf08-44ac-a524-aa87cf785575] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 956.729251] env[63515]: DEBUG oslo_vmware.api [None req-857b4940-0702-491c-a977-f4c63c5dd5f5 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]521860d3-6992-b044-e43b-1378368b65f1, 'name': SearchDatastore_Task, 'duration_secs': 0.023767} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 956.729543] env[63515]: DEBUG oslo_concurrency.lockutils [None req-857b4940-0702-491c-a977-f4c63c5dd5f5 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 956.729804] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-857b4940-0702-491c-a977-f4c63c5dd5f5 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk to [datastore2] 918d5b35-c46d-47c5-b59f-175831aed9d6/918d5b35-c46d-47c5-b59f-175831aed9d6.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 956.730079] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-07c846e9-2ce6-48ba-bffd-afa9754d5d53 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.736042] env[63515]: DEBUG oslo_vmware.api [None req-857b4940-0702-491c-a977-f4c63c5dd5f5 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Waiting for the task: (returnval){ [ 956.736042] env[63515]: value = "task-1111662" [ 956.736042] env[63515]: _type = "Task" [ 956.736042] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 956.743388] env[63515]: DEBUG oslo_vmware.api [None req-857b4940-0702-491c-a977-f4c63c5dd5f5 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Task: {'id': task-1111662, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.795478] env[63515]: DEBUG oslo_concurrency.lockutils [None req-3f7a26fd-cc84-46d9-9d28-feb378f8b5e6 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Releasing lock "refresh_cache-75f4dc79-4b73-4ddc-be03-3653a5a0797e" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 956.795811] env[63515]: DEBUG nova.compute.manager [None req-3f7a26fd-cc84-46d9-9d28-feb378f8b5e6 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] [instance: 75f4dc79-4b73-4ddc-be03-3653a5a0797e] Instance network_info: |[{"id": "00eda8ad-1a52-4c47-b065-a82c08da7980", "address": "fa:16:3e:02:56:38", "network": {"id": "549c4fb2-0ed6-44ee-bfd1-9e59bb7aed27", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-949461070-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3dcf6179891c4841a2901fcffe43cd31", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8c3e2368-4a35-4aa5-9135-23daedbbf9ef", "external-id": "nsx-vlan-transportzone-125", "segmentation_id": 125, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap00eda8ad-1a", "ovs_interfaceid": "00eda8ad-1a52-4c47-b065-a82c08da7980", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 956.796131] env[63515]: DEBUG oslo_concurrency.lockutils [req-93d4a1f0-6b3c-4bf4-b86d-1539b55c5a3c req-bde21bed-9194-4f8c-b217-f21a000b1f75 service nova] Acquired lock "refresh_cache-75f4dc79-4b73-4ddc-be03-3653a5a0797e" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 956.796402] env[63515]: DEBUG nova.network.neutron [req-93d4a1f0-6b3c-4bf4-b86d-1539b55c5a3c req-bde21bed-9194-4f8c-b217-f21a000b1f75 service nova] [instance: 75f4dc79-4b73-4ddc-be03-3653a5a0797e] Refreshing network info cache for port 00eda8ad-1a52-4c47-b065-a82c08da7980 {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 956.797637] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-3f7a26fd-cc84-46d9-9d28-feb378f8b5e6 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] [instance: 75f4dc79-4b73-4ddc-be03-3653a5a0797e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:02:56:38', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '8c3e2368-4a35-4aa5-9135-23daedbbf9ef', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '00eda8ad-1a52-4c47-b065-a82c08da7980', 'vif_model': 'vmxnet3'}] {{(pid=63515) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 956.805143] env[63515]: DEBUG oslo.service.loopingcall [None req-3f7a26fd-cc84-46d9-9d28-feb378f8b5e6 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 956.807885] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 75f4dc79-4b73-4ddc-be03-3653a5a0797e] Creating VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 956.808368] env[63515]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-1064fd38-a4f3-4749-8ce5-ca46915d3296 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.831316] env[63515]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 956.831316] env[63515]: value = "task-1111663" [ 956.831316] env[63515]: _type = "Task" [ 956.831316] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 956.839540] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1111663, 'name': CreateVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.872603] env[63515]: DEBUG oslo_vmware.api [None req-127af73c-afc8-4436-9075-a886f5e9c7d5 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Task: {'id': task-1111661, 'name': Rename_Task, 'duration_secs': 0.147315} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 956.872889] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-127af73c-afc8-4436-9075-a886f5e9c7d5 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] [instance: 7d572ac0-f6af-4622-96cc-e75983420222] Powering on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 956.873186] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e44c7539-1e9d-431b-b58c-bb0cd4893d4e {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.880809] env[63515]: DEBUG oslo_vmware.api [None req-127af73c-afc8-4436-9075-a886f5e9c7d5 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Waiting for the task: (returnval){ [ 956.880809] env[63515]: value = "task-1111664" [ 956.880809] env[63515]: _type = "Task" [ 956.880809] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 956.887717] env[63515]: INFO nova.compute.manager [-] [instance: eab16df9-7bb5-4576-bca0-769a561c5fe9] Took 1.54 seconds to deallocate network for instance. [ 956.895871] env[63515]: DEBUG oslo_vmware.api [None req-127af73c-afc8-4436-9075-a886f5e9c7d5 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Task: {'id': task-1111664, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.912794] env[63515]: DEBUG oslo_vmware.api [None req-8e6475b3-95eb-4b44-a2f3-5ddd5db39f52 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Task: {'id': task-1111660, 'name': RemoveSnapshot_Task, 'duration_secs': 0.549148} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 956.913118] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-8e6475b3-95eb-4b44-a2f3-5ddd5db39f52 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 68766cd4-84be-475b-8494-d7ab43a9e969] Deleted Snapshot of the VM instance {{(pid=63515) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 956.913362] env[63515]: INFO nova.compute.manager [None req-8e6475b3-95eb-4b44-a2f3-5ddd5db39f52 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 68766cd4-84be-475b-8494-d7ab43a9e969] Took 14.80 seconds to snapshot the instance on the hypervisor. [ 956.964812] env[63515]: DEBUG nova.compute.manager [None req-3f7a26fd-cc84-46d9-9d28-feb378f8b5e6 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] [instance: d438dd30-b816-48a6-94c6-11e6e9bd5b71] Start spawning the instance on the hypervisor. {{(pid=63515) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 956.997983] env[63515]: DEBUG nova.virt.hardware [None req-3f7a26fd-cc84-46d9-9d28-feb378f8b5e6 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T02:52:14Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T02:51:56Z,direct_url=,disk_format='vmdk',id=8a120570-cb06-4099-b262-554ca0ad15c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b569255cc43e42e7ae3f2b4ad37c6ef4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T02:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 956.999023] env[63515]: DEBUG nova.virt.hardware [None req-3f7a26fd-cc84-46d9-9d28-feb378f8b5e6 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Flavor limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 956.999362] env[63515]: DEBUG nova.virt.hardware [None req-3f7a26fd-cc84-46d9-9d28-feb378f8b5e6 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Image limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 956.999531] env[63515]: DEBUG nova.virt.hardware [None req-3f7a26fd-cc84-46d9-9d28-feb378f8b5e6 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Flavor pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 956.999657] env[63515]: DEBUG nova.virt.hardware [None req-3f7a26fd-cc84-46d9-9d28-feb378f8b5e6 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Image pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 956.999812] env[63515]: DEBUG nova.virt.hardware [None req-3f7a26fd-cc84-46d9-9d28-feb378f8b5e6 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 957.000133] env[63515]: DEBUG nova.virt.hardware [None req-3f7a26fd-cc84-46d9-9d28-feb378f8b5e6 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 957.000318] env[63515]: DEBUG nova.virt.hardware [None req-3f7a26fd-cc84-46d9-9d28-feb378f8b5e6 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 957.000710] env[63515]: DEBUG nova.virt.hardware [None req-3f7a26fd-cc84-46d9-9d28-feb378f8b5e6 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Got 1 possible topologies {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 957.000979] env[63515]: DEBUG nova.virt.hardware [None req-3f7a26fd-cc84-46d9-9d28-feb378f8b5e6 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 957.001210] env[63515]: DEBUG nova.virt.hardware [None req-3f7a26fd-cc84-46d9-9d28-feb378f8b5e6 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 957.002196] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47d734eb-e292-46ff-9b58-bdc2d76e31c9 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.006435] env[63515]: DEBUG oslo_concurrency.lockutils [None req-2099b375-1441-4c83-88b0-3c0c09203cee tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Acquiring lock "86b1ddcf-585c-41d9-84fe-17b794f1abbc" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 957.006724] env[63515]: DEBUG oslo_concurrency.lockutils [None req-2099b375-1441-4c83-88b0-3c0c09203cee tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Lock "86b1ddcf-585c-41d9-84fe-17b794f1abbc" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 957.006959] env[63515]: INFO nova.compute.manager [None req-2099b375-1441-4c83-88b0-3c0c09203cee tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] [instance: 86b1ddcf-585c-41d9-84fe-17b794f1abbc] Attaching volume aa5c9d63-11af-4c5d-ae0b-9aa149332c19 to /dev/sdb [ 957.021639] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47120d61-c521-434c-820f-8f7f05c5f8b0 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.060383] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b26ed56-e215-4822-873b-1e9314e3c321 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.068785] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7b25714-210e-43cd-8499-6df4f1d35121 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.086702] env[63515]: DEBUG nova.virt.block_device [None req-2099b375-1441-4c83-88b0-3c0c09203cee tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] [instance: 86b1ddcf-585c-41d9-84fe-17b794f1abbc] Updating existing volume attachment record: ee08a52f-84be-46c2-87ae-3b39a2239ed8 {{(pid=63515) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 957.152706] env[63515]: DEBUG nova.network.neutron [req-93d4a1f0-6b3c-4bf4-b86d-1539b55c5a3c req-bde21bed-9194-4f8c-b217-f21a000b1f75 service nova] [instance: 75f4dc79-4b73-4ddc-be03-3653a5a0797e] Updated VIF entry in instance network info cache for port 00eda8ad-1a52-4c47-b065-a82c08da7980. {{(pid=63515) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 957.153229] env[63515]: DEBUG nova.network.neutron [req-93d4a1f0-6b3c-4bf4-b86d-1539b55c5a3c req-bde21bed-9194-4f8c-b217-f21a000b1f75 service nova] [instance: 75f4dc79-4b73-4ddc-be03-3653a5a0797e] Updating instance_info_cache with network_info: [{"id": "00eda8ad-1a52-4c47-b065-a82c08da7980", "address": "fa:16:3e:02:56:38", "network": {"id": "549c4fb2-0ed6-44ee-bfd1-9e59bb7aed27", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-949461070-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3dcf6179891c4841a2901fcffe43cd31", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8c3e2368-4a35-4aa5-9135-23daedbbf9ef", "external-id": "nsx-vlan-transportzone-125", "segmentation_id": 125, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap00eda8ad-1a", "ovs_interfaceid": "00eda8ad-1a52-4c47-b065-a82c08da7980", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 957.201298] env[63515]: INFO nova.compute.manager [-] [instance: b25bf2ba-cf08-44ac-a524-aa87cf785575] Took 1.31 seconds to deallocate network for instance. [ 957.246517] env[63515]: DEBUG oslo_vmware.api [None req-857b4940-0702-491c-a977-f4c63c5dd5f5 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Task: {'id': task-1111662, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.495946} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 957.246812] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-857b4940-0702-491c-a977-f4c63c5dd5f5 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk to [datastore2] 918d5b35-c46d-47c5-b59f-175831aed9d6/918d5b35-c46d-47c5-b59f-175831aed9d6.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 957.247046] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-857b4940-0702-491c-a977-f4c63c5dd5f5 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: 918d5b35-c46d-47c5-b59f-175831aed9d6] Extending root virtual disk to 1048576 {{(pid=63515) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 957.247299] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-e5140808-043b-4417-88f8-f689b4c2ebc9 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.252734] env[63515]: DEBUG oslo_vmware.api [None req-857b4940-0702-491c-a977-f4c63c5dd5f5 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Waiting for the task: (returnval){ [ 957.252734] env[63515]: value = "task-1111665" [ 957.252734] env[63515]: _type = "Task" [ 957.252734] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 957.264698] env[63515]: DEBUG oslo_vmware.api [None req-857b4940-0702-491c-a977-f4c63c5dd5f5 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Task: {'id': task-1111665, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.328709] env[63515]: DEBUG oslo_concurrency.lockutils [None req-b94389d0-616d-4e7e-a415-7b0c3a16a683 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.382s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 957.332586] env[63515]: DEBUG oslo_concurrency.lockutils [None req-165bb947-b1a3-4f80-91f4-6990bed3b114 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 9.792s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 957.334807] env[63515]: INFO nova.compute.claims [None req-165bb947-b1a3-4f80-91f4-6990bed3b114 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: d024b6b2-dd10-4112-89e8-ced57efe8208] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 957.348851] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1111663, 'name': CreateVM_Task} progress is 99%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.394154] env[63515]: DEBUG oslo_vmware.api [None req-127af73c-afc8-4436-9075-a886f5e9c7d5 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Task: {'id': task-1111664, 'name': PowerOnVM_Task} progress is 89%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.397406] env[63515]: DEBUG oslo_concurrency.lockutils [None req-db106ad7-1693-4446-8352-ddb2ff9c25ce tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 957.469433] env[63515]: DEBUG nova.compute.manager [None req-8e6475b3-95eb-4b44-a2f3-5ddd5db39f52 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 68766cd4-84be-475b-8494-d7ab43a9e969] Found 2 images (rotation: 2) {{(pid=63515) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4554}} [ 957.651537] env[63515]: DEBUG nova.network.neutron [None req-3f7a26fd-cc84-46d9-9d28-feb378f8b5e6 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] [instance: d438dd30-b816-48a6-94c6-11e6e9bd5b71] Successfully updated port: a673b83a-c9d4-425b-b6e3-aad0d17243ba {{(pid=63515) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 957.656036] env[63515]: DEBUG oslo_concurrency.lockutils [req-93d4a1f0-6b3c-4bf4-b86d-1539b55c5a3c req-bde21bed-9194-4f8c-b217-f21a000b1f75 service nova] Releasing lock "refresh_cache-75f4dc79-4b73-4ddc-be03-3653a5a0797e" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 957.656298] env[63515]: DEBUG nova.compute.manager [req-93d4a1f0-6b3c-4bf4-b86d-1539b55c5a3c req-bde21bed-9194-4f8c-b217-f21a000b1f75 service nova] [instance: eab16df9-7bb5-4576-bca0-769a561c5fe9] Received event network-vif-deleted-4168f5a5-c282-4f13-9c97-0b0236bd659b {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 957.694190] env[63515]: DEBUG nova.compute.manager [req-1f480e08-5069-46c2-98a2-20b07c96ec15 req-30c2baf9-5e9d-4974-bd79-316cd0840b79 service nova] [instance: d438dd30-b816-48a6-94c6-11e6e9bd5b71] Received event network-vif-plugged-a673b83a-c9d4-425b-b6e3-aad0d17243ba {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 957.694516] env[63515]: DEBUG oslo_concurrency.lockutils [req-1f480e08-5069-46c2-98a2-20b07c96ec15 req-30c2baf9-5e9d-4974-bd79-316cd0840b79 service nova] Acquiring lock "d438dd30-b816-48a6-94c6-11e6e9bd5b71-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 957.694734] env[63515]: DEBUG oslo_concurrency.lockutils [req-1f480e08-5069-46c2-98a2-20b07c96ec15 req-30c2baf9-5e9d-4974-bd79-316cd0840b79 service nova] Lock "d438dd30-b816-48a6-94c6-11e6e9bd5b71-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 957.694896] env[63515]: DEBUG oslo_concurrency.lockutils [req-1f480e08-5069-46c2-98a2-20b07c96ec15 req-30c2baf9-5e9d-4974-bd79-316cd0840b79 service nova] Lock "d438dd30-b816-48a6-94c6-11e6e9bd5b71-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 957.695095] env[63515]: DEBUG nova.compute.manager [req-1f480e08-5069-46c2-98a2-20b07c96ec15 req-30c2baf9-5e9d-4974-bd79-316cd0840b79 service nova] [instance: d438dd30-b816-48a6-94c6-11e6e9bd5b71] No waiting events found dispatching network-vif-plugged-a673b83a-c9d4-425b-b6e3-aad0d17243ba {{(pid=63515) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 957.695274] env[63515]: WARNING nova.compute.manager [req-1f480e08-5069-46c2-98a2-20b07c96ec15 req-30c2baf9-5e9d-4974-bd79-316cd0840b79 service nova] [instance: d438dd30-b816-48a6-94c6-11e6e9bd5b71] Received unexpected event network-vif-plugged-a673b83a-c9d4-425b-b6e3-aad0d17243ba for instance with vm_state building and task_state spawning. [ 957.708557] env[63515]: DEBUG oslo_concurrency.lockutils [None req-2359bb47-ebaa-4b05-a6cb-20ad2f0f493e tempest-ServerMetadataTestJSON-1034558863 tempest-ServerMetadataTestJSON-1034558863-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 957.762631] env[63515]: DEBUG oslo_vmware.api [None req-857b4940-0702-491c-a977-f4c63c5dd5f5 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Task: {'id': task-1111665, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.067684} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 957.762925] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-857b4940-0702-491c-a977-f4c63c5dd5f5 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: 918d5b35-c46d-47c5-b59f-175831aed9d6] Extended root virtual disk {{(pid=63515) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 957.763746] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d5b155e-98ce-416c-9dfc-4875ca7870bb {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.785445] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-857b4940-0702-491c-a977-f4c63c5dd5f5 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: 918d5b35-c46d-47c5-b59f-175831aed9d6] Reconfiguring VM instance instance-00000059 to attach disk [datastore2] 918d5b35-c46d-47c5-b59f-175831aed9d6/918d5b35-c46d-47c5-b59f-175831aed9d6.vmdk or device None with type sparse {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 957.785730] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b384505a-796f-4957-9491-f2c19dfd3383 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.805094] env[63515]: DEBUG oslo_vmware.api [None req-857b4940-0702-491c-a977-f4c63c5dd5f5 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Waiting for the task: (returnval){ [ 957.805094] env[63515]: value = "task-1111667" [ 957.805094] env[63515]: _type = "Task" [ 957.805094] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 957.813342] env[63515]: DEBUG oslo_vmware.api [None req-857b4940-0702-491c-a977-f4c63c5dd5f5 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Task: {'id': task-1111667, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.856245] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1111663, 'name': CreateVM_Task, 'duration_secs': 0.595736} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 957.858296] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 75f4dc79-4b73-4ddc-be03-3653a5a0797e] Created VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 957.861209] env[63515]: DEBUG oslo_concurrency.lockutils [None req-3f7a26fd-cc84-46d9-9d28-feb378f8b5e6 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 957.861356] env[63515]: DEBUG oslo_concurrency.lockutils [None req-3f7a26fd-cc84-46d9-9d28-feb378f8b5e6 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 957.861693] env[63515]: DEBUG oslo_concurrency.lockutils [None req-3f7a26fd-cc84-46d9-9d28-feb378f8b5e6 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 957.862527] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0fa23719-18d9-4efe-a85a-9676239351d3 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.869627] env[63515]: DEBUG oslo_vmware.api [None req-3f7a26fd-cc84-46d9-9d28-feb378f8b5e6 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Waiting for the task: (returnval){ [ 957.869627] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]52ecb461-ca9b-be7f-0004-fca94dbc7f2f" [ 957.869627] env[63515]: _type = "Task" [ 957.869627] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 957.881888] env[63515]: DEBUG oslo_vmware.api [None req-3f7a26fd-cc84-46d9-9d28-feb378f8b5e6 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52ecb461-ca9b-be7f-0004-fca94dbc7f2f, 'name': SearchDatastore_Task, 'duration_secs': 0.009267} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 957.882212] env[63515]: DEBUG oslo_concurrency.lockutils [None req-3f7a26fd-cc84-46d9-9d28-feb378f8b5e6 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 957.882429] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-3f7a26fd-cc84-46d9-9d28-feb378f8b5e6 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] [instance: 75f4dc79-4b73-4ddc-be03-3653a5a0797e] Processing image 8a120570-cb06-4099-b262-554ca0ad15c5 {{(pid=63515) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 957.882665] env[63515]: DEBUG oslo_concurrency.lockutils [None req-3f7a26fd-cc84-46d9-9d28-feb378f8b5e6 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 957.882812] env[63515]: DEBUG oslo_concurrency.lockutils [None req-3f7a26fd-cc84-46d9-9d28-feb378f8b5e6 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 957.882989] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-3f7a26fd-cc84-46d9-9d28-feb378f8b5e6 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 957.883310] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-8f6d034d-e15f-4377-ad1a-02d93cab9e73 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.893637] env[63515]: DEBUG oslo_vmware.api [None req-127af73c-afc8-4436-9075-a886f5e9c7d5 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Task: {'id': task-1111664, 'name': PowerOnVM_Task, 'duration_secs': 0.618126} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 957.894386] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-127af73c-afc8-4436-9075-a886f5e9c7d5 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] [instance: 7d572ac0-f6af-4622-96cc-e75983420222] Powered on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 957.894732] env[63515]: INFO nova.compute.manager [None req-127af73c-afc8-4436-9075-a886f5e9c7d5 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] [instance: 7d572ac0-f6af-4622-96cc-e75983420222] Took 8.16 seconds to spawn the instance on the hypervisor. [ 957.894817] env[63515]: DEBUG nova.compute.manager [None req-127af73c-afc8-4436-9075-a886f5e9c7d5 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] [instance: 7d572ac0-f6af-4622-96cc-e75983420222] Checking state {{(pid=63515) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 957.895346] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-3f7a26fd-cc84-46d9-9d28-feb378f8b5e6 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 957.895511] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-3f7a26fd-cc84-46d9-9d28-feb378f8b5e6 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63515) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 957.896660] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32cc652a-0582-44b5-bef8-a5bc75cd54ff {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.899158] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b9c692a6-8d5a-4e66-b7a2-5eab3934a186 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.912017] env[63515]: INFO nova.scheduler.client.report [None req-b94389d0-616d-4e7e-a415-7b0c3a16a683 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Deleted allocation for migration c4203df0-81f5-4cda-85fc-996941b25046 [ 957.912017] env[63515]: DEBUG oslo_vmware.api [None req-3f7a26fd-cc84-46d9-9d28-feb378f8b5e6 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Waiting for the task: (returnval){ [ 957.912017] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]528b06c0-c25f-edec-72af-d5a954910c10" [ 957.912017] env[63515]: _type = "Task" [ 957.912017] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 957.923949] env[63515]: DEBUG oslo_vmware.api [None req-3f7a26fd-cc84-46d9-9d28-feb378f8b5e6 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]528b06c0-c25f-edec-72af-d5a954910c10, 'name': SearchDatastore_Task, 'duration_secs': 0.010324} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 957.925062] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c0a17374-fe64-415a-9f48-74798c7828ee {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.930548] env[63515]: DEBUG oslo_vmware.api [None req-3f7a26fd-cc84-46d9-9d28-feb378f8b5e6 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Waiting for the task: (returnval){ [ 957.930548] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]52f7f922-35e9-9b27-3bf2-7c79a82a6bc1" [ 957.930548] env[63515]: _type = "Task" [ 957.930548] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 957.938102] env[63515]: DEBUG oslo_vmware.api [None req-3f7a26fd-cc84-46d9-9d28-feb378f8b5e6 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52f7f922-35e9-9b27-3bf2-7c79a82a6bc1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.154207] env[63515]: DEBUG oslo_concurrency.lockutils [None req-3f7a26fd-cc84-46d9-9d28-feb378f8b5e6 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Acquiring lock "refresh_cache-d438dd30-b816-48a6-94c6-11e6e9bd5b71" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 958.154411] env[63515]: DEBUG oslo_concurrency.lockutils [None req-3f7a26fd-cc84-46d9-9d28-feb378f8b5e6 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Acquired lock "refresh_cache-d438dd30-b816-48a6-94c6-11e6e9bd5b71" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 958.155672] env[63515]: DEBUG nova.network.neutron [None req-3f7a26fd-cc84-46d9-9d28-feb378f8b5e6 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] [instance: d438dd30-b816-48a6-94c6-11e6e9bd5b71] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 958.315362] env[63515]: DEBUG oslo_vmware.api [None req-857b4940-0702-491c-a977-f4c63c5dd5f5 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Task: {'id': task-1111667, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.363416] env[63515]: DEBUG nova.compute.manager [None req-57194bc7-9fdc-4bda-bd6d-4b604b59fffa tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 68766cd4-84be-475b-8494-d7ab43a9e969] Checking state {{(pid=63515) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 958.364058] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07a0317f-f146-46e2-ac8c-3c9553ebf8c4 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.419933] env[63515]: DEBUG oslo_concurrency.lockutils [None req-b94389d0-616d-4e7e-a415-7b0c3a16a683 tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Lock "829e72d7-a56a-451c-b98a-94253e5900d5" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 17.422s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 958.420142] env[63515]: INFO nova.compute.manager [None req-127af73c-afc8-4436-9075-a886f5e9c7d5 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] [instance: 7d572ac0-f6af-4622-96cc-e75983420222] Took 22.48 seconds to build instance. [ 958.441865] env[63515]: DEBUG oslo_vmware.api [None req-3f7a26fd-cc84-46d9-9d28-feb378f8b5e6 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52f7f922-35e9-9b27-3bf2-7c79a82a6bc1, 'name': SearchDatastore_Task, 'duration_secs': 0.008655} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 958.444601] env[63515]: DEBUG oslo_concurrency.lockutils [None req-3f7a26fd-cc84-46d9-9d28-feb378f8b5e6 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 958.444883] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-3f7a26fd-cc84-46d9-9d28-feb378f8b5e6 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk to [datastore2] 75f4dc79-4b73-4ddc-be03-3653a5a0797e/75f4dc79-4b73-4ddc-be03-3653a5a0797e.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 958.445362] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a5169891-13e4-4ce8-a889-fe461c4f8f3a {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.454211] env[63515]: DEBUG oslo_vmware.api [None req-3f7a26fd-cc84-46d9-9d28-feb378f8b5e6 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Waiting for the task: (returnval){ [ 958.454211] env[63515]: value = "task-1111668" [ 958.454211] env[63515]: _type = "Task" [ 958.454211] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 958.466269] env[63515]: DEBUG oslo_vmware.api [None req-3f7a26fd-cc84-46d9-9d28-feb378f8b5e6 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Task: {'id': task-1111668, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.553781] env[63515]: DEBUG nova.compute.manager [req-8993ac7b-b5b4-446f-8bc3-78e8bf343a90 req-174e828c-f3ea-4e7a-b90f-2967bb5f27ff service nova] [instance: b25bf2ba-cf08-44ac-a524-aa87cf785575] Received event network-vif-deleted-5d351a64-ccc8-4d4c-9057-7ef8d2c3edbd {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 958.636835] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5115b19-0cca-4dcb-8547-d11b82539d81 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.651658] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90a31873-6841-42d2-8cf1-bc97da10f7ef {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.697991] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba7a25dc-0d39-4772-be0e-184a7fbc9638 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.706463] env[63515]: DEBUG nova.network.neutron [None req-3f7a26fd-cc84-46d9-9d28-feb378f8b5e6 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] [instance: d438dd30-b816-48a6-94c6-11e6e9bd5b71] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 958.709507] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a34816a-55d2-4743-9b71-05328e762e74 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.725929] env[63515]: DEBUG nova.compute.provider_tree [None req-165bb947-b1a3-4f80-91f4-6990bed3b114 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 958.820249] env[63515]: DEBUG oslo_vmware.api [None req-857b4940-0702-491c-a977-f4c63c5dd5f5 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Task: {'id': task-1111667, 'name': ReconfigVM_Task, 'duration_secs': 0.823578} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 958.820990] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-857b4940-0702-491c-a977-f4c63c5dd5f5 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: 918d5b35-c46d-47c5-b59f-175831aed9d6] Reconfigured VM instance instance-00000059 to attach disk [datastore2] 918d5b35-c46d-47c5-b59f-175831aed9d6/918d5b35-c46d-47c5-b59f-175831aed9d6.vmdk or device None with type sparse {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 958.821684] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e80771b2-fac5-4e8d-b43e-f4f9b26a1039 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.832016] env[63515]: DEBUG oslo_vmware.api [None req-857b4940-0702-491c-a977-f4c63c5dd5f5 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Waiting for the task: (returnval){ [ 958.832016] env[63515]: value = "task-1111669" [ 958.832016] env[63515]: _type = "Task" [ 958.832016] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 958.842022] env[63515]: DEBUG oslo_vmware.api [None req-857b4940-0702-491c-a977-f4c63c5dd5f5 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Task: {'id': task-1111669, 'name': Rename_Task} progress is 5%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.876817] env[63515]: INFO nova.compute.manager [None req-57194bc7-9fdc-4bda-bd6d-4b604b59fffa tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 68766cd4-84be-475b-8494-d7ab43a9e969] instance snapshotting [ 958.877648] env[63515]: DEBUG nova.objects.instance [None req-57194bc7-9fdc-4bda-bd6d-4b604b59fffa tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Lazy-loading 'flavor' on Instance uuid 68766cd4-84be-475b-8494-d7ab43a9e969 {{(pid=63515) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 958.922183] env[63515]: DEBUG oslo_concurrency.lockutils [None req-127af73c-afc8-4436-9075-a886f5e9c7d5 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Lock "7d572ac0-f6af-4622-96cc-e75983420222" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 23.988s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 958.933848] env[63515]: DEBUG nova.network.neutron [None req-3f7a26fd-cc84-46d9-9d28-feb378f8b5e6 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] [instance: d438dd30-b816-48a6-94c6-11e6e9bd5b71] Updating instance_info_cache with network_info: [{"id": "a673b83a-c9d4-425b-b6e3-aad0d17243ba", "address": "fa:16:3e:bb:9c:97", "network": {"id": "549c4fb2-0ed6-44ee-bfd1-9e59bb7aed27", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-949461070-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3dcf6179891c4841a2901fcffe43cd31", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8c3e2368-4a35-4aa5-9135-23daedbbf9ef", "external-id": "nsx-vlan-transportzone-125", "segmentation_id": 125, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa673b83a-c9", "ovs_interfaceid": "a673b83a-c9d4-425b-b6e3-aad0d17243ba", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 958.969413] env[63515]: DEBUG oslo_vmware.api [None req-3f7a26fd-cc84-46d9-9d28-feb378f8b5e6 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Task: {'id': task-1111668, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.232105] env[63515]: DEBUG nova.scheduler.client.report [None req-165bb947-b1a3-4f80-91f4-6990bed3b114 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 959.245135] env[63515]: DEBUG oslo_concurrency.lockutils [None req-8ce4ce82-9ad9-4201-af2c-019ea542780e tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Acquiring lock "829e72d7-a56a-451c-b98a-94253e5900d5" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 959.245135] env[63515]: DEBUG oslo_concurrency.lockutils [None req-8ce4ce82-9ad9-4201-af2c-019ea542780e tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Lock "829e72d7-a56a-451c-b98a-94253e5900d5" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 959.245135] env[63515]: DEBUG oslo_concurrency.lockutils [None req-8ce4ce82-9ad9-4201-af2c-019ea542780e tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Acquiring lock "829e72d7-a56a-451c-b98a-94253e5900d5-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 959.245135] env[63515]: DEBUG oslo_concurrency.lockutils [None req-8ce4ce82-9ad9-4201-af2c-019ea542780e tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Lock "829e72d7-a56a-451c-b98a-94253e5900d5-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 959.245135] env[63515]: DEBUG oslo_concurrency.lockutils [None req-8ce4ce82-9ad9-4201-af2c-019ea542780e tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Lock "829e72d7-a56a-451c-b98a-94253e5900d5-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 959.247735] env[63515]: INFO nova.compute.manager [None req-8ce4ce82-9ad9-4201-af2c-019ea542780e tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: 829e72d7-a56a-451c-b98a-94253e5900d5] Terminating instance [ 959.249797] env[63515]: DEBUG nova.compute.manager [None req-8ce4ce82-9ad9-4201-af2c-019ea542780e tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: 829e72d7-a56a-451c-b98a-94253e5900d5] Start destroying the instance on the hypervisor. {{(pid=63515) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 959.250008] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-8ce4ce82-9ad9-4201-af2c-019ea542780e tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: 829e72d7-a56a-451c-b98a-94253e5900d5] Destroying instance {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 959.256310] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d76fb0d1-6f9c-41d4-939b-d7431a621700 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.265984] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-8ce4ce82-9ad9-4201-af2c-019ea542780e tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: 829e72d7-a56a-451c-b98a-94253e5900d5] Powering off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 959.266274] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-47b7f60f-f1ed-46ab-84ea-b99496fd6eea {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.272497] env[63515]: DEBUG oslo_vmware.api [None req-8ce4ce82-9ad9-4201-af2c-019ea542780e tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Waiting for the task: (returnval){ [ 959.272497] env[63515]: value = "task-1111670" [ 959.272497] env[63515]: _type = "Task" [ 959.272497] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 959.283785] env[63515]: DEBUG oslo_vmware.api [None req-8ce4ce82-9ad9-4201-af2c-019ea542780e tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Task: {'id': task-1111670, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.342505] env[63515]: DEBUG oslo_vmware.api [None req-857b4940-0702-491c-a977-f4c63c5dd5f5 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Task: {'id': task-1111669, 'name': Rename_Task, 'duration_secs': 0.267262} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 959.342838] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-857b4940-0702-491c-a977-f4c63c5dd5f5 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: 918d5b35-c46d-47c5-b59f-175831aed9d6] Powering on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 959.343117] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e25140df-9792-4e53-ac15-59bd4d05eaf4 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.349170] env[63515]: DEBUG oslo_vmware.api [None req-857b4940-0702-491c-a977-f4c63c5dd5f5 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Waiting for the task: (returnval){ [ 959.349170] env[63515]: value = "task-1111672" [ 959.349170] env[63515]: _type = "Task" [ 959.349170] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 959.358325] env[63515]: DEBUG oslo_vmware.api [None req-857b4940-0702-491c-a977-f4c63c5dd5f5 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Task: {'id': task-1111672, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.383984] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bde57ccb-7134-4d0f-b1d8-7d717affc33e {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.404940] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b227428-c878-4dd8-ac3c-e0e965b36b80 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.437291] env[63515]: DEBUG oslo_concurrency.lockutils [None req-3f7a26fd-cc84-46d9-9d28-feb378f8b5e6 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Releasing lock "refresh_cache-d438dd30-b816-48a6-94c6-11e6e9bd5b71" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 959.437770] env[63515]: DEBUG nova.compute.manager [None req-3f7a26fd-cc84-46d9-9d28-feb378f8b5e6 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] [instance: d438dd30-b816-48a6-94c6-11e6e9bd5b71] Instance network_info: |[{"id": "a673b83a-c9d4-425b-b6e3-aad0d17243ba", "address": "fa:16:3e:bb:9c:97", "network": {"id": "549c4fb2-0ed6-44ee-bfd1-9e59bb7aed27", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-949461070-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3dcf6179891c4841a2901fcffe43cd31", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8c3e2368-4a35-4aa5-9135-23daedbbf9ef", "external-id": "nsx-vlan-transportzone-125", "segmentation_id": 125, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa673b83a-c9", "ovs_interfaceid": "a673b83a-c9d4-425b-b6e3-aad0d17243ba", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 959.438393] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-3f7a26fd-cc84-46d9-9d28-feb378f8b5e6 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] [instance: d438dd30-b816-48a6-94c6-11e6e9bd5b71] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:bb:9c:97', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '8c3e2368-4a35-4aa5-9135-23daedbbf9ef', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a673b83a-c9d4-425b-b6e3-aad0d17243ba', 'vif_model': 'vmxnet3'}] {{(pid=63515) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 959.448239] env[63515]: DEBUG oslo.service.loopingcall [None req-3f7a26fd-cc84-46d9-9d28-feb378f8b5e6 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 959.449462] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d438dd30-b816-48a6-94c6-11e6e9bd5b71] Creating VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 959.449655] env[63515]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ea585cac-1f8a-42af-8afc-8f1f440a2deb {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.478609] env[63515]: DEBUG oslo_vmware.api [None req-3f7a26fd-cc84-46d9-9d28-feb378f8b5e6 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Task: {'id': task-1111668, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.699561} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 959.479961] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-3f7a26fd-cc84-46d9-9d28-feb378f8b5e6 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk to [datastore2] 75f4dc79-4b73-4ddc-be03-3653a5a0797e/75f4dc79-4b73-4ddc-be03-3653a5a0797e.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 959.480230] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-3f7a26fd-cc84-46d9-9d28-feb378f8b5e6 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] [instance: 75f4dc79-4b73-4ddc-be03-3653a5a0797e] Extending root virtual disk to 1048576 {{(pid=63515) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 959.480498] env[63515]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 959.480498] env[63515]: value = "task-1111673" [ 959.480498] env[63515]: _type = "Task" [ 959.480498] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 959.480726] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-7908cd8d-9ad3-45aa-b875-dfd675ad7d88 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.492519] env[63515]: DEBUG oslo_vmware.api [None req-3f7a26fd-cc84-46d9-9d28-feb378f8b5e6 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Waiting for the task: (returnval){ [ 959.492519] env[63515]: value = "task-1111674" [ 959.492519] env[63515]: _type = "Task" [ 959.492519] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 959.499746] env[63515]: DEBUG oslo_vmware.api [None req-3f7a26fd-cc84-46d9-9d28-feb378f8b5e6 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Task: {'id': task-1111674, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.725479] env[63515]: DEBUG nova.compute.manager [req-ce5a100f-49e3-456b-91fa-51270b84a114 req-ab374f20-802c-4ead-b87b-4c73afa6ce62 service nova] [instance: d438dd30-b816-48a6-94c6-11e6e9bd5b71] Received event network-changed-a673b83a-c9d4-425b-b6e3-aad0d17243ba {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 959.725730] env[63515]: DEBUG nova.compute.manager [req-ce5a100f-49e3-456b-91fa-51270b84a114 req-ab374f20-802c-4ead-b87b-4c73afa6ce62 service nova] [instance: d438dd30-b816-48a6-94c6-11e6e9bd5b71] Refreshing instance network info cache due to event network-changed-a673b83a-c9d4-425b-b6e3-aad0d17243ba. {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 959.725967] env[63515]: DEBUG oslo_concurrency.lockutils [req-ce5a100f-49e3-456b-91fa-51270b84a114 req-ab374f20-802c-4ead-b87b-4c73afa6ce62 service nova] Acquiring lock "refresh_cache-d438dd30-b816-48a6-94c6-11e6e9bd5b71" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 959.726136] env[63515]: DEBUG oslo_concurrency.lockutils [req-ce5a100f-49e3-456b-91fa-51270b84a114 req-ab374f20-802c-4ead-b87b-4c73afa6ce62 service nova] Acquired lock "refresh_cache-d438dd30-b816-48a6-94c6-11e6e9bd5b71" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 959.726309] env[63515]: DEBUG nova.network.neutron [req-ce5a100f-49e3-456b-91fa-51270b84a114 req-ab374f20-802c-4ead-b87b-4c73afa6ce62 service nova] [instance: d438dd30-b816-48a6-94c6-11e6e9bd5b71] Refreshing network info cache for port a673b83a-c9d4-425b-b6e3-aad0d17243ba {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 959.733781] env[63515]: DEBUG oslo_concurrency.lockutils [None req-165bb947-b1a3-4f80-91f4-6990bed3b114 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.402s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 959.734310] env[63515]: DEBUG nova.compute.manager [None req-165bb947-b1a3-4f80-91f4-6990bed3b114 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: d024b6b2-dd10-4112-89e8-ced57efe8208] Start building networks asynchronously for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 959.736839] env[63515]: DEBUG oslo_concurrency.lockutils [None req-0773e564-f19d-40eb-a88a-6e28c35f232c tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 9.712s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 959.737073] env[63515]: DEBUG nova.objects.instance [None req-0773e564-f19d-40eb-a88a-6e28c35f232c tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Lazy-loading 'resources' on Instance uuid 014720fe-87b9-4e55-b9ba-295b009c7ff6 {{(pid=63515) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 959.783478] env[63515]: DEBUG oslo_vmware.api [None req-8ce4ce82-9ad9-4201-af2c-019ea542780e tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Task: {'id': task-1111670, 'name': PowerOffVM_Task, 'duration_secs': 0.182665} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 959.783767] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-8ce4ce82-9ad9-4201-af2c-019ea542780e tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: 829e72d7-a56a-451c-b98a-94253e5900d5] Powered off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 959.783940] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-8ce4ce82-9ad9-4201-af2c-019ea542780e tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: 829e72d7-a56a-451c-b98a-94253e5900d5] Unregistering the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 959.784214] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-8a4755ef-4fba-47a3-8fe3-71e495135213 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.834473] env[63515]: DEBUG oslo_concurrency.lockutils [None req-a5676663-6933-430d-aa5e-aeffa94052ee tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Acquiring lock "6a66dd9b-7b24-4b70-b58f-0830cf39376c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 959.834734] env[63515]: DEBUG oslo_concurrency.lockutils [None req-a5676663-6933-430d-aa5e-aeffa94052ee tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Lock "6a66dd9b-7b24-4b70-b58f-0830cf39376c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 959.861136] env[63515]: DEBUG oslo_vmware.api [None req-857b4940-0702-491c-a977-f4c63c5dd5f5 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Task: {'id': task-1111672, 'name': PowerOnVM_Task} progress is 78%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.862815] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-8ce4ce82-9ad9-4201-af2c-019ea542780e tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: 829e72d7-a56a-451c-b98a-94253e5900d5] Unregistered the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 959.863078] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-8ce4ce82-9ad9-4201-af2c-019ea542780e tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: 829e72d7-a56a-451c-b98a-94253e5900d5] Deleting contents of the VM from datastore datastore1 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 959.863295] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-8ce4ce82-9ad9-4201-af2c-019ea542780e tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Deleting the datastore file [datastore1] 829e72d7-a56a-451c-b98a-94253e5900d5 {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 959.863608] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-8d1bfdaf-d96b-48eb-b345-1cb1e0cfee3c {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.869428] env[63515]: DEBUG oslo_vmware.api [None req-8ce4ce82-9ad9-4201-af2c-019ea542780e tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Waiting for the task: (returnval){ [ 959.869428] env[63515]: value = "task-1111676" [ 959.869428] env[63515]: _type = "Task" [ 959.869428] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 959.877828] env[63515]: DEBUG oslo_vmware.api [None req-8ce4ce82-9ad9-4201-af2c-019ea542780e tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Task: {'id': task-1111676, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.916567] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-57194bc7-9fdc-4bda-bd6d-4b604b59fffa tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 68766cd4-84be-475b-8494-d7ab43a9e969] Creating Snapshot of the VM instance {{(pid=63515) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 959.916927] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-2d263fe0-05d2-43a3-a670-7d7a93bc1254 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.924400] env[63515]: DEBUG oslo_vmware.api [None req-57194bc7-9fdc-4bda-bd6d-4b604b59fffa tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Waiting for the task: (returnval){ [ 959.924400] env[63515]: value = "task-1111677" [ 959.924400] env[63515]: _type = "Task" [ 959.924400] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 959.933852] env[63515]: DEBUG oslo_vmware.api [None req-57194bc7-9fdc-4bda-bd6d-4b604b59fffa tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Task: {'id': task-1111677, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.993740] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1111673, 'name': CreateVM_Task} progress is 99%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.001661] env[63515]: DEBUG oslo_vmware.api [None req-3f7a26fd-cc84-46d9-9d28-feb378f8b5e6 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Task: {'id': task-1111674, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.138048} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 960.001919] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-3f7a26fd-cc84-46d9-9d28-feb378f8b5e6 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] [instance: 75f4dc79-4b73-4ddc-be03-3653a5a0797e] Extended root virtual disk {{(pid=63515) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 960.002868] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-244796a1-76a5-44e6-aad2-15abeea05eb1 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.027169] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-3f7a26fd-cc84-46d9-9d28-feb378f8b5e6 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] [instance: 75f4dc79-4b73-4ddc-be03-3653a5a0797e] Reconfiguring VM instance instance-0000005a to attach disk [datastore2] 75f4dc79-4b73-4ddc-be03-3653a5a0797e/75f4dc79-4b73-4ddc-be03-3653a5a0797e.vmdk or device None with type sparse {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 960.027513] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-387d52ed-0503-4609-8ab8-c8df1b24e45f {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.046431] env[63515]: DEBUG oslo_vmware.api [None req-3f7a26fd-cc84-46d9-9d28-feb378f8b5e6 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Waiting for the task: (returnval){ [ 960.046431] env[63515]: value = "task-1111678" [ 960.046431] env[63515]: _type = "Task" [ 960.046431] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 960.054798] env[63515]: DEBUG oslo_vmware.api [None req-3f7a26fd-cc84-46d9-9d28-feb378f8b5e6 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Task: {'id': task-1111678, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.247564] env[63515]: DEBUG nova.compute.utils [None req-165bb947-b1a3-4f80-91f4-6990bed3b114 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Using /dev/sd instead of None {{(pid=63515) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 960.250684] env[63515]: DEBUG nova.compute.manager [None req-165bb947-b1a3-4f80-91f4-6990bed3b114 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: d024b6b2-dd10-4112-89e8-ced57efe8208] Allocating IP information in the background. {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 960.250919] env[63515]: DEBUG nova.network.neutron [None req-165bb947-b1a3-4f80-91f4-6990bed3b114 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: d024b6b2-dd10-4112-89e8-ced57efe8208] allocate_for_instance() {{(pid=63515) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 960.338626] env[63515]: DEBUG nova.compute.manager [None req-a5676663-6933-430d-aa5e-aeffa94052ee tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: 6a66dd9b-7b24-4b70-b58f-0830cf39376c] Starting instance... {{(pid=63515) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 960.344721] env[63515]: DEBUG nova.policy [None req-165bb947-b1a3-4f80-91f4-6990bed3b114 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9b19bb2d32d84c019541c3b2e711a202', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '299fdeff647f486390366d5bbf911518', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63515) authorize /opt/stack/nova/nova/policy.py:201}} [ 960.364089] env[63515]: DEBUG oslo_vmware.api [None req-857b4940-0702-491c-a977-f4c63c5dd5f5 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Task: {'id': task-1111672, 'name': PowerOnVM_Task, 'duration_secs': 0.882986} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 960.366982] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-857b4940-0702-491c-a977-f4c63c5dd5f5 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: 918d5b35-c46d-47c5-b59f-175831aed9d6] Powered on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 960.367238] env[63515]: INFO nova.compute.manager [None req-857b4940-0702-491c-a977-f4c63c5dd5f5 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: 918d5b35-c46d-47c5-b59f-175831aed9d6] Took 8.23 seconds to spawn the instance on the hypervisor. [ 960.367429] env[63515]: DEBUG nova.compute.manager [None req-857b4940-0702-491c-a977-f4c63c5dd5f5 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: 918d5b35-c46d-47c5-b59f-175831aed9d6] Checking state {{(pid=63515) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 960.368770] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b90770c1-7408-4a5f-8adb-34757f4f9d01 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.383840] env[63515]: DEBUG oslo_vmware.api [None req-8ce4ce82-9ad9-4201-af2c-019ea542780e tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Task: {'id': task-1111676, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.155623} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 960.384367] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-8ce4ce82-9ad9-4201-af2c-019ea542780e tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Deleted the datastore file {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 960.384585] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-8ce4ce82-9ad9-4201-af2c-019ea542780e tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: 829e72d7-a56a-451c-b98a-94253e5900d5] Deleted contents of the VM from datastore datastore1 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 960.384763] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-8ce4ce82-9ad9-4201-af2c-019ea542780e tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: 829e72d7-a56a-451c-b98a-94253e5900d5] Instance destroyed {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 960.385926] env[63515]: INFO nova.compute.manager [None req-8ce4ce82-9ad9-4201-af2c-019ea542780e tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: 829e72d7-a56a-451c-b98a-94253e5900d5] Took 1.13 seconds to destroy the instance on the hypervisor. [ 960.385926] env[63515]: DEBUG oslo.service.loopingcall [None req-8ce4ce82-9ad9-4201-af2c-019ea542780e tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 960.385926] env[63515]: DEBUG nova.compute.manager [-] [instance: 829e72d7-a56a-451c-b98a-94253e5900d5] Deallocating network for instance {{(pid=63515) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 960.385926] env[63515]: DEBUG nova.network.neutron [-] [instance: 829e72d7-a56a-451c-b98a-94253e5900d5] deallocate_for_instance() {{(pid=63515) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 960.434129] env[63515]: DEBUG oslo_vmware.api [None req-57194bc7-9fdc-4bda-bd6d-4b604b59fffa tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Task: {'id': task-1111677, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.506243] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1111673, 'name': CreateVM_Task, 'duration_secs': 0.574667} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 960.511636] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d438dd30-b816-48a6-94c6-11e6e9bd5b71] Created VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 960.514292] env[63515]: DEBUG oslo_concurrency.lockutils [None req-3f7a26fd-cc84-46d9-9d28-feb378f8b5e6 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 960.514292] env[63515]: DEBUG oslo_concurrency.lockutils [None req-3f7a26fd-cc84-46d9-9d28-feb378f8b5e6 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 960.514292] env[63515]: DEBUG oslo_concurrency.lockutils [None req-3f7a26fd-cc84-46d9-9d28-feb378f8b5e6 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 960.514292] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-32ca8ecd-00bf-4bb4-bdd7-0b3ec4ba2da1 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.523073] env[63515]: DEBUG oslo_vmware.api [None req-3f7a26fd-cc84-46d9-9d28-feb378f8b5e6 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Waiting for the task: (returnval){ [ 960.523073] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]529aac59-c4fb-5ab9-df1a-79a04bc1d264" [ 960.523073] env[63515]: _type = "Task" [ 960.523073] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 960.534887] env[63515]: DEBUG oslo_vmware.api [None req-3f7a26fd-cc84-46d9-9d28-feb378f8b5e6 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]529aac59-c4fb-5ab9-df1a-79a04bc1d264, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.535939] env[63515]: DEBUG nova.network.neutron [req-ce5a100f-49e3-456b-91fa-51270b84a114 req-ab374f20-802c-4ead-b87b-4c73afa6ce62 service nova] [instance: d438dd30-b816-48a6-94c6-11e6e9bd5b71] Updated VIF entry in instance network info cache for port a673b83a-c9d4-425b-b6e3-aad0d17243ba. {{(pid=63515) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 960.536078] env[63515]: DEBUG nova.network.neutron [req-ce5a100f-49e3-456b-91fa-51270b84a114 req-ab374f20-802c-4ead-b87b-4c73afa6ce62 service nova] [instance: d438dd30-b816-48a6-94c6-11e6e9bd5b71] Updating instance_info_cache with network_info: [{"id": "a673b83a-c9d4-425b-b6e3-aad0d17243ba", "address": "fa:16:3e:bb:9c:97", "network": {"id": "549c4fb2-0ed6-44ee-bfd1-9e59bb7aed27", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-949461070-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3dcf6179891c4841a2901fcffe43cd31", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8c3e2368-4a35-4aa5-9135-23daedbbf9ef", "external-id": "nsx-vlan-transportzone-125", "segmentation_id": 125, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa673b83a-c9", "ovs_interfaceid": "a673b83a-c9d4-425b-b6e3-aad0d17243ba", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 960.556781] env[63515]: DEBUG oslo_vmware.api [None req-3f7a26fd-cc84-46d9-9d28-feb378f8b5e6 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Task: {'id': task-1111678, 'name': ReconfigVM_Task, 'duration_secs': 0.320798} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 960.556916] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-3f7a26fd-cc84-46d9-9d28-feb378f8b5e6 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] [instance: 75f4dc79-4b73-4ddc-be03-3653a5a0797e] Reconfigured VM instance instance-0000005a to attach disk [datastore2] 75f4dc79-4b73-4ddc-be03-3653a5a0797e/75f4dc79-4b73-4ddc-be03-3653a5a0797e.vmdk or device None with type sparse {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 960.559995] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b1ebb81c-d928-4d04-9192-4500b543e906 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.566983] env[63515]: DEBUG oslo_vmware.api [None req-3f7a26fd-cc84-46d9-9d28-feb378f8b5e6 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Waiting for the task: (returnval){ [ 960.566983] env[63515]: value = "task-1111679" [ 960.566983] env[63515]: _type = "Task" [ 960.566983] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 960.576084] env[63515]: DEBUG oslo_vmware.api [None req-3f7a26fd-cc84-46d9-9d28-feb378f8b5e6 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Task: {'id': task-1111679, 'name': Rename_Task} progress is 5%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.588348] env[63515]: DEBUG nova.compute.manager [req-54df9ac6-0a5f-4573-8321-1b0b14258020 req-4047b8bd-6af2-49d4-aff9-ebfb6d11dad0 service nova] [instance: 7d572ac0-f6af-4622-96cc-e75983420222] Received event network-changed-dbfd09c2-9ec4-4821-8613-43c6d6cf4724 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 960.588557] env[63515]: DEBUG nova.compute.manager [req-54df9ac6-0a5f-4573-8321-1b0b14258020 req-4047b8bd-6af2-49d4-aff9-ebfb6d11dad0 service nova] [instance: 7d572ac0-f6af-4622-96cc-e75983420222] Refreshing instance network info cache due to event network-changed-dbfd09c2-9ec4-4821-8613-43c6d6cf4724. {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 960.588855] env[63515]: DEBUG oslo_concurrency.lockutils [req-54df9ac6-0a5f-4573-8321-1b0b14258020 req-4047b8bd-6af2-49d4-aff9-ebfb6d11dad0 service nova] Acquiring lock "refresh_cache-7d572ac0-f6af-4622-96cc-e75983420222" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 960.589041] env[63515]: DEBUG oslo_concurrency.lockutils [req-54df9ac6-0a5f-4573-8321-1b0b14258020 req-4047b8bd-6af2-49d4-aff9-ebfb6d11dad0 service nova] Acquired lock "refresh_cache-7d572ac0-f6af-4622-96cc-e75983420222" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 960.589238] env[63515]: DEBUG nova.network.neutron [req-54df9ac6-0a5f-4573-8321-1b0b14258020 req-4047b8bd-6af2-49d4-aff9-ebfb6d11dad0 service nova] [instance: 7d572ac0-f6af-4622-96cc-e75983420222] Refreshing network info cache for port dbfd09c2-9ec4-4821-8613-43c6d6cf4724 {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 960.599268] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cbd83767-984c-4225-9c82-01d893b4737c {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.609792] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0eb3d628-9c0c-474c-87c4-ce8218f2b89e {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.639557] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66f59853-c1fe-4585-a1fd-ace807690f49 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.647335] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a3ae02a-10d3-463d-a45e-0017d2c67f60 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.661133] env[63515]: DEBUG nova.compute.provider_tree [None req-0773e564-f19d-40eb-a88a-6e28c35f232c tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 960.756654] env[63515]: DEBUG nova.compute.manager [None req-165bb947-b1a3-4f80-91f4-6990bed3b114 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: d024b6b2-dd10-4112-89e8-ced57efe8208] Start building block device mappings for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 960.834769] env[63515]: DEBUG nova.network.neutron [None req-165bb947-b1a3-4f80-91f4-6990bed3b114 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: d024b6b2-dd10-4112-89e8-ced57efe8208] Successfully created port: 06e24fbd-1278-4395-8108-f966a61c7b6c {{(pid=63515) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 960.866280] env[63515]: DEBUG oslo_concurrency.lockutils [None req-a5676663-6933-430d-aa5e-aeffa94052ee tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 960.894670] env[63515]: INFO nova.compute.manager [None req-857b4940-0702-491c-a977-f4c63c5dd5f5 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: 918d5b35-c46d-47c5-b59f-175831aed9d6] Took 21.65 seconds to build instance. [ 960.935803] env[63515]: DEBUG oslo_vmware.api [None req-57194bc7-9fdc-4bda-bd6d-4b604b59fffa tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Task: {'id': task-1111677, 'name': CreateSnapshot_Task, 'duration_secs': 0.632549} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 960.936166] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-57194bc7-9fdc-4bda-bd6d-4b604b59fffa tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 68766cd4-84be-475b-8494-d7ab43a9e969] Created Snapshot of the VM instance {{(pid=63515) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 960.936943] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-411b9a51-aabb-4b92-8d55-ae2024d3d574 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.037855] env[63515]: DEBUG oslo_vmware.api [None req-3f7a26fd-cc84-46d9-9d28-feb378f8b5e6 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]529aac59-c4fb-5ab9-df1a-79a04bc1d264, 'name': SearchDatastore_Task, 'duration_secs': 0.011177} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 961.038631] env[63515]: DEBUG oslo_concurrency.lockutils [req-ce5a100f-49e3-456b-91fa-51270b84a114 req-ab374f20-802c-4ead-b87b-4c73afa6ce62 service nova] Releasing lock "refresh_cache-d438dd30-b816-48a6-94c6-11e6e9bd5b71" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 961.038878] env[63515]: DEBUG oslo_concurrency.lockutils [None req-3f7a26fd-cc84-46d9-9d28-feb378f8b5e6 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 961.039165] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-3f7a26fd-cc84-46d9-9d28-feb378f8b5e6 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] [instance: d438dd30-b816-48a6-94c6-11e6e9bd5b71] Processing image 8a120570-cb06-4099-b262-554ca0ad15c5 {{(pid=63515) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 961.039419] env[63515]: DEBUG oslo_concurrency.lockutils [None req-3f7a26fd-cc84-46d9-9d28-feb378f8b5e6 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 961.039574] env[63515]: DEBUG oslo_concurrency.lockutils [None req-3f7a26fd-cc84-46d9-9d28-feb378f8b5e6 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 961.039789] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-3f7a26fd-cc84-46d9-9d28-feb378f8b5e6 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 961.040337] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-31280e9b-be90-43f5-b688-75fb8728af63 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.049040] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-3f7a26fd-cc84-46d9-9d28-feb378f8b5e6 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 961.049157] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-3f7a26fd-cc84-46d9-9d28-feb378f8b5e6 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63515) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 961.049819] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-35ed53b3-01b5-437e-9993-21d5628c71fc {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.055082] env[63515]: DEBUG oslo_vmware.api [None req-3f7a26fd-cc84-46d9-9d28-feb378f8b5e6 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Waiting for the task: (returnval){ [ 961.055082] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]523c9586-06e1-346a-1294-ccdfeb680140" [ 961.055082] env[63515]: _type = "Task" [ 961.055082] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 961.062795] env[63515]: DEBUG oslo_vmware.api [None req-3f7a26fd-cc84-46d9-9d28-feb378f8b5e6 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]523c9586-06e1-346a-1294-ccdfeb680140, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.074146] env[63515]: DEBUG oslo_vmware.api [None req-3f7a26fd-cc84-46d9-9d28-feb378f8b5e6 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Task: {'id': task-1111679, 'name': Rename_Task, 'duration_secs': 0.148512} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 961.074480] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-3f7a26fd-cc84-46d9-9d28-feb378f8b5e6 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] [instance: 75f4dc79-4b73-4ddc-be03-3653a5a0797e] Powering on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 961.074745] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2c33eea7-ee27-4287-af3a-6076935312f8 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.080335] env[63515]: DEBUG oslo_vmware.api [None req-3f7a26fd-cc84-46d9-9d28-feb378f8b5e6 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Waiting for the task: (returnval){ [ 961.080335] env[63515]: value = "task-1111680" [ 961.080335] env[63515]: _type = "Task" [ 961.080335] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 961.164546] env[63515]: DEBUG nova.scheduler.client.report [None req-0773e564-f19d-40eb-a88a-6e28c35f232c tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 961.305997] env[63515]: DEBUG nova.network.neutron [-] [instance: 829e72d7-a56a-451c-b98a-94253e5900d5] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 961.329649] env[63515]: DEBUG nova.network.neutron [req-54df9ac6-0a5f-4573-8321-1b0b14258020 req-4047b8bd-6af2-49d4-aff9-ebfb6d11dad0 service nova] [instance: 7d572ac0-f6af-4622-96cc-e75983420222] Updated VIF entry in instance network info cache for port dbfd09c2-9ec4-4821-8613-43c6d6cf4724. {{(pid=63515) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 961.330064] env[63515]: DEBUG nova.network.neutron [req-54df9ac6-0a5f-4573-8321-1b0b14258020 req-4047b8bd-6af2-49d4-aff9-ebfb6d11dad0 service nova] [instance: 7d572ac0-f6af-4622-96cc-e75983420222] Updating instance_info_cache with network_info: [{"id": "dbfd09c2-9ec4-4821-8613-43c6d6cf4724", "address": "fa:16:3e:6d:92:3d", "network": {"id": "09fb20ca-2b17-4d46-b915-6aff85b862d8", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1815529427-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.244", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "19ce3b05eec64475b685faad1fd801f6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ded18042-834c-4792-b3e8-b1c377446432", "external-id": "nsx-vlan-transportzone-293", "segmentation_id": 293, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdbfd09c2-9e", "ovs_interfaceid": "dbfd09c2-9ec4-4821-8613-43c6d6cf4724", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 961.399417] env[63515]: DEBUG oslo_concurrency.lockutils [None req-857b4940-0702-491c-a977-f4c63c5dd5f5 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Lock "918d5b35-c46d-47c5-b59f-175831aed9d6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 23.166s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 961.455035] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-57194bc7-9fdc-4bda-bd6d-4b604b59fffa tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 68766cd4-84be-475b-8494-d7ab43a9e969] Creating linked-clone VM from snapshot {{(pid=63515) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 961.455210] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-fd300b2a-989d-4831-b417-4d7c516c46be {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.464714] env[63515]: DEBUG oslo_vmware.api [None req-57194bc7-9fdc-4bda-bd6d-4b604b59fffa tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Waiting for the task: (returnval){ [ 961.464714] env[63515]: value = "task-1111681" [ 961.464714] env[63515]: _type = "Task" [ 961.464714] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 961.472538] env[63515]: DEBUG oslo_vmware.api [None req-57194bc7-9fdc-4bda-bd6d-4b604b59fffa tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Task: {'id': task-1111681, 'name': CloneVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.565472] env[63515]: DEBUG oslo_vmware.api [None req-3f7a26fd-cc84-46d9-9d28-feb378f8b5e6 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]523c9586-06e1-346a-1294-ccdfeb680140, 'name': SearchDatastore_Task, 'duration_secs': 0.009195} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 961.566457] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-40872292-e7cd-4527-bc7e-abf5467b65bf {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.572443] env[63515]: DEBUG oslo_vmware.api [None req-3f7a26fd-cc84-46d9-9d28-feb378f8b5e6 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Waiting for the task: (returnval){ [ 961.572443] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]523902e1-abf8-26dd-f24c-463b864cd0fe" [ 961.572443] env[63515]: _type = "Task" [ 961.572443] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 961.582182] env[63515]: DEBUG oslo_vmware.api [None req-3f7a26fd-cc84-46d9-9d28-feb378f8b5e6 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]523902e1-abf8-26dd-f24c-463b864cd0fe, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.593426] env[63515]: DEBUG oslo_vmware.api [None req-3f7a26fd-cc84-46d9-9d28-feb378f8b5e6 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Task: {'id': task-1111680, 'name': PowerOnVM_Task} progress is 100%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.644206] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-2099b375-1441-4c83-88b0-3c0c09203cee tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] [instance: 86b1ddcf-585c-41d9-84fe-17b794f1abbc] Volume attach. Driver type: vmdk {{(pid=63515) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 961.644571] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-2099b375-1441-4c83-88b0-3c0c09203cee tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] [instance: 86b1ddcf-585c-41d9-84fe-17b794f1abbc] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-243544', 'volume_id': 'aa5c9d63-11af-4c5d-ae0b-9aa149332c19', 'name': 'volume-aa5c9d63-11af-4c5d-ae0b-9aa149332c19', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '86b1ddcf-585c-41d9-84fe-17b794f1abbc', 'attached_at': '', 'detached_at': '', 'volume_id': 'aa5c9d63-11af-4c5d-ae0b-9aa149332c19', 'serial': 'aa5c9d63-11af-4c5d-ae0b-9aa149332c19'} {{(pid=63515) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 961.645481] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eeb2d04b-3218-47c4-9ef6-810c9dae7bdc {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.661886] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6253aacc-6a5e-4ba1-8c2d-267ca7fe00f9 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.680023] env[63515]: DEBUG oslo_concurrency.lockutils [None req-0773e564-f19d-40eb-a88a-6e28c35f232c tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.943s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 961.689213] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-2099b375-1441-4c83-88b0-3c0c09203cee tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] [instance: 86b1ddcf-585c-41d9-84fe-17b794f1abbc] Reconfiguring VM instance instance-00000051 to attach disk [datastore1] volume-aa5c9d63-11af-4c5d-ae0b-9aa149332c19/volume-aa5c9d63-11af-4c5d-ae0b-9aa149332c19.vmdk or device None with type thin {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 961.689891] env[63515]: DEBUG oslo_concurrency.lockutils [None req-8a925afd-d6aa-4a9a-a54f-bf6262925289 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 10.509s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 961.690154] env[63515]: DEBUG nova.objects.instance [None req-8a925afd-d6aa-4a9a-a54f-bf6262925289 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Lazy-loading 'resources' on Instance uuid 2fa098b6-ecc0-491f-8cfc-5ee335bbf943 {{(pid=63515) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 961.691759] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-49f481f5-65c8-4dee-9691-797290c86710 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.717641] env[63515]: DEBUG oslo_vmware.api [None req-2099b375-1441-4c83-88b0-3c0c09203cee tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Waiting for the task: (returnval){ [ 961.717641] env[63515]: value = "task-1111682" [ 961.717641] env[63515]: _type = "Task" [ 961.717641] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 961.719103] env[63515]: INFO nova.scheduler.client.report [None req-0773e564-f19d-40eb-a88a-6e28c35f232c tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Deleted allocations for instance 014720fe-87b9-4e55-b9ba-295b009c7ff6 [ 961.733590] env[63515]: DEBUG oslo_vmware.api [None req-2099b375-1441-4c83-88b0-3c0c09203cee tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Task: {'id': task-1111682, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.766784] env[63515]: DEBUG nova.compute.manager [None req-165bb947-b1a3-4f80-91f4-6990bed3b114 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: d024b6b2-dd10-4112-89e8-ced57efe8208] Start spawning the instance on the hypervisor. {{(pid=63515) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 961.798394] env[63515]: DEBUG nova.virt.hardware [None req-165bb947-b1a3-4f80-91f4-6990bed3b114 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T02:52:14Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T02:51:56Z,direct_url=,disk_format='vmdk',id=8a120570-cb06-4099-b262-554ca0ad15c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b569255cc43e42e7ae3f2b4ad37c6ef4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T02:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 961.799054] env[63515]: DEBUG nova.virt.hardware [None req-165bb947-b1a3-4f80-91f4-6990bed3b114 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Flavor limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 961.799054] env[63515]: DEBUG nova.virt.hardware [None req-165bb947-b1a3-4f80-91f4-6990bed3b114 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Image limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 961.799054] env[63515]: DEBUG nova.virt.hardware [None req-165bb947-b1a3-4f80-91f4-6990bed3b114 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Flavor pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 961.799237] env[63515]: DEBUG nova.virt.hardware [None req-165bb947-b1a3-4f80-91f4-6990bed3b114 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Image pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 961.799330] env[63515]: DEBUG nova.virt.hardware [None req-165bb947-b1a3-4f80-91f4-6990bed3b114 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 961.799564] env[63515]: DEBUG nova.virt.hardware [None req-165bb947-b1a3-4f80-91f4-6990bed3b114 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 961.799745] env[63515]: DEBUG nova.virt.hardware [None req-165bb947-b1a3-4f80-91f4-6990bed3b114 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 961.799923] env[63515]: DEBUG nova.virt.hardware [None req-165bb947-b1a3-4f80-91f4-6990bed3b114 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Got 1 possible topologies {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 961.800126] env[63515]: DEBUG nova.virt.hardware [None req-165bb947-b1a3-4f80-91f4-6990bed3b114 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 961.800329] env[63515]: DEBUG nova.virt.hardware [None req-165bb947-b1a3-4f80-91f4-6990bed3b114 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 961.801222] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b5888c1-4716-4b06-8b8d-7492d09d6a64 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.811658] env[63515]: INFO nova.compute.manager [-] [instance: 829e72d7-a56a-451c-b98a-94253e5900d5] Took 1.42 seconds to deallocate network for instance. [ 961.813011] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-957e9461-826b-483b-8d65-8794d62a52b8 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.833436] env[63515]: DEBUG oslo_concurrency.lockutils [req-54df9ac6-0a5f-4573-8321-1b0b14258020 req-4047b8bd-6af2-49d4-aff9-ebfb6d11dad0 service nova] Releasing lock "refresh_cache-7d572ac0-f6af-4622-96cc-e75983420222" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 961.959258] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f037b2fc-b05a-48da-9a64-84483cad6839 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.970457] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3ee7c75-3735-4eb6-868f-58d42979faef {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.978709] env[63515]: DEBUG oslo_vmware.api [None req-57194bc7-9fdc-4bda-bd6d-4b604b59fffa tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Task: {'id': task-1111681, 'name': CloneVM_Task} progress is 94%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.005361] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d3c44f0-fecd-43de-a5c5-1467f926aa61 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.013436] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad784ba4-41e5-456b-a835-13b2d3d0b2a0 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.026443] env[63515]: DEBUG nova.compute.provider_tree [None req-8a925afd-d6aa-4a9a-a54f-bf6262925289 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 962.082577] env[63515]: DEBUG oslo_vmware.api [None req-3f7a26fd-cc84-46d9-9d28-feb378f8b5e6 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]523902e1-abf8-26dd-f24c-463b864cd0fe, 'name': SearchDatastore_Task, 'duration_secs': 0.017686} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 962.085782] env[63515]: DEBUG oslo_concurrency.lockutils [None req-3f7a26fd-cc84-46d9-9d28-feb378f8b5e6 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 962.086110] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-3f7a26fd-cc84-46d9-9d28-feb378f8b5e6 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk to [datastore1] d438dd30-b816-48a6-94c6-11e6e9bd5b71/d438dd30-b816-48a6-94c6-11e6e9bd5b71.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 962.086393] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-22c1f2fa-0f7f-48ce-93b6-a39715774479 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.093429] env[63515]: DEBUG oslo_vmware.api [None req-3f7a26fd-cc84-46d9-9d28-feb378f8b5e6 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Task: {'id': task-1111680, 'name': PowerOnVM_Task, 'duration_secs': 0.523898} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 962.094553] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-3f7a26fd-cc84-46d9-9d28-feb378f8b5e6 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] [instance: 75f4dc79-4b73-4ddc-be03-3653a5a0797e] Powered on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 962.094794] env[63515]: INFO nova.compute.manager [None req-3f7a26fd-cc84-46d9-9d28-feb378f8b5e6 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] [instance: 75f4dc79-4b73-4ddc-be03-3653a5a0797e] Took 7.55 seconds to spawn the instance on the hypervisor. [ 962.094985] env[63515]: DEBUG nova.compute.manager [None req-3f7a26fd-cc84-46d9-9d28-feb378f8b5e6 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] [instance: 75f4dc79-4b73-4ddc-be03-3653a5a0797e] Checking state {{(pid=63515) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 962.095312] env[63515]: DEBUG oslo_vmware.api [None req-3f7a26fd-cc84-46d9-9d28-feb378f8b5e6 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Waiting for the task: (returnval){ [ 962.095312] env[63515]: value = "task-1111683" [ 962.095312] env[63515]: _type = "Task" [ 962.095312] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 962.096021] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb647421-d4a5-4d8e-ab44-f39b48e37cc9 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.112388] env[63515]: DEBUG oslo_vmware.api [None req-3f7a26fd-cc84-46d9-9d28-feb378f8b5e6 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Task: {'id': task-1111683, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.232472] env[63515]: DEBUG oslo_vmware.api [None req-2099b375-1441-4c83-88b0-3c0c09203cee tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Task: {'id': task-1111682, 'name': ReconfigVM_Task, 'duration_secs': 0.406792} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 962.232472] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-2099b375-1441-4c83-88b0-3c0c09203cee tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] [instance: 86b1ddcf-585c-41d9-84fe-17b794f1abbc] Reconfigured VM instance instance-00000051 to attach disk [datastore1] volume-aa5c9d63-11af-4c5d-ae0b-9aa149332c19/volume-aa5c9d63-11af-4c5d-ae0b-9aa149332c19.vmdk or device None with type thin {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 962.238114] env[63515]: DEBUG oslo_concurrency.lockutils [None req-0773e564-f19d-40eb-a88a-6e28c35f232c tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Lock "014720fe-87b9-4e55-b9ba-295b009c7ff6" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 14.718s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 962.239224] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-44cca68f-a6ac-4968-b0b9-88bc1d9434e8 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.251066] env[63515]: DEBUG oslo_concurrency.lockutils [None req-81bc1898-e097-4e8c-b6d2-4f02be1d9d2a tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Acquiring lock "918d5b35-c46d-47c5-b59f-175831aed9d6" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 962.251380] env[63515]: DEBUG oslo_concurrency.lockutils [None req-81bc1898-e097-4e8c-b6d2-4f02be1d9d2a tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Lock "918d5b35-c46d-47c5-b59f-175831aed9d6" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 962.251594] env[63515]: DEBUG nova.compute.manager [None req-81bc1898-e097-4e8c-b6d2-4f02be1d9d2a tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: 918d5b35-c46d-47c5-b59f-175831aed9d6] Checking state {{(pid=63515) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 962.252743] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-600fe7e0-5c07-4a34-9614-eaf532118d45 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.260095] env[63515]: DEBUG nova.compute.manager [None req-81bc1898-e097-4e8c-b6d2-4f02be1d9d2a tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: 918d5b35-c46d-47c5-b59f-175831aed9d6] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=63515) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 962.260723] env[63515]: DEBUG nova.objects.instance [None req-81bc1898-e097-4e8c-b6d2-4f02be1d9d2a tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Lazy-loading 'flavor' on Instance uuid 918d5b35-c46d-47c5-b59f-175831aed9d6 {{(pid=63515) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 962.264846] env[63515]: DEBUG oslo_vmware.api [None req-2099b375-1441-4c83-88b0-3c0c09203cee tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Waiting for the task: (returnval){ [ 962.264846] env[63515]: value = "task-1111684" [ 962.264846] env[63515]: _type = "Task" [ 962.264846] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 962.278099] env[63515]: DEBUG oslo_vmware.api [None req-2099b375-1441-4c83-88b0-3c0c09203cee tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Task: {'id': task-1111684, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.323652] env[63515]: DEBUG oslo_concurrency.lockutils [None req-8ce4ce82-9ad9-4201-af2c-019ea542780e tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 962.433321] env[63515]: DEBUG nova.network.neutron [None req-165bb947-b1a3-4f80-91f4-6990bed3b114 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: d024b6b2-dd10-4112-89e8-ced57efe8208] Successfully updated port: 06e24fbd-1278-4395-8108-f966a61c7b6c {{(pid=63515) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 962.477900] env[63515]: DEBUG oslo_vmware.api [None req-57194bc7-9fdc-4bda-bd6d-4b604b59fffa tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Task: {'id': task-1111681, 'name': CloneVM_Task} progress is 95%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.529939] env[63515]: DEBUG nova.scheduler.client.report [None req-8a925afd-d6aa-4a9a-a54f-bf6262925289 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 962.611900] env[63515]: DEBUG oslo_vmware.api [None req-3f7a26fd-cc84-46d9-9d28-feb378f8b5e6 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Task: {'id': task-1111683, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.49223} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 962.612599] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-3f7a26fd-cc84-46d9-9d28-feb378f8b5e6 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk to [datastore1] d438dd30-b816-48a6-94c6-11e6e9bd5b71/d438dd30-b816-48a6-94c6-11e6e9bd5b71.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 962.612868] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-3f7a26fd-cc84-46d9-9d28-feb378f8b5e6 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] [instance: d438dd30-b816-48a6-94c6-11e6e9bd5b71] Extending root virtual disk to 1048576 {{(pid=63515) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 962.617128] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a2014d1e-9342-4f9b-949b-afb73147783e {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.622260] env[63515]: INFO nova.compute.manager [None req-3f7a26fd-cc84-46d9-9d28-feb378f8b5e6 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] [instance: 75f4dc79-4b73-4ddc-be03-3653a5a0797e] Took 21.18 seconds to build instance. [ 962.625739] env[63515]: DEBUG oslo_vmware.api [None req-3f7a26fd-cc84-46d9-9d28-feb378f8b5e6 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Waiting for the task: (returnval){ [ 962.625739] env[63515]: value = "task-1111685" [ 962.625739] env[63515]: _type = "Task" [ 962.625739] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 962.637163] env[63515]: DEBUG oslo_vmware.api [None req-3f7a26fd-cc84-46d9-9d28-feb378f8b5e6 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Task: {'id': task-1111685, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.757338] env[63515]: DEBUG nova.compute.manager [req-9c250657-cb66-4ddf-90d0-610336e43aa9 req-fc28d6a1-c709-4c24-acfc-30e78a6096ee service nova] [instance: 829e72d7-a56a-451c-b98a-94253e5900d5] Received event network-vif-deleted-856de677-b1c8-479b-91d9-34d31a703a63 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 962.757338] env[63515]: DEBUG nova.compute.manager [req-9c250657-cb66-4ddf-90d0-610336e43aa9 req-fc28d6a1-c709-4c24-acfc-30e78a6096ee service nova] [instance: d024b6b2-dd10-4112-89e8-ced57efe8208] Received event network-vif-plugged-06e24fbd-1278-4395-8108-f966a61c7b6c {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 962.757794] env[63515]: DEBUG oslo_concurrency.lockutils [req-9c250657-cb66-4ddf-90d0-610336e43aa9 req-fc28d6a1-c709-4c24-acfc-30e78a6096ee service nova] Acquiring lock "d024b6b2-dd10-4112-89e8-ced57efe8208-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 962.757852] env[63515]: DEBUG oslo_concurrency.lockutils [req-9c250657-cb66-4ddf-90d0-610336e43aa9 req-fc28d6a1-c709-4c24-acfc-30e78a6096ee service nova] Lock "d024b6b2-dd10-4112-89e8-ced57efe8208-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 962.758158] env[63515]: DEBUG oslo_concurrency.lockutils [req-9c250657-cb66-4ddf-90d0-610336e43aa9 req-fc28d6a1-c709-4c24-acfc-30e78a6096ee service nova] Lock "d024b6b2-dd10-4112-89e8-ced57efe8208-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 962.758378] env[63515]: DEBUG nova.compute.manager [req-9c250657-cb66-4ddf-90d0-610336e43aa9 req-fc28d6a1-c709-4c24-acfc-30e78a6096ee service nova] [instance: d024b6b2-dd10-4112-89e8-ced57efe8208] No waiting events found dispatching network-vif-plugged-06e24fbd-1278-4395-8108-f966a61c7b6c {{(pid=63515) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 962.758614] env[63515]: WARNING nova.compute.manager [req-9c250657-cb66-4ddf-90d0-610336e43aa9 req-fc28d6a1-c709-4c24-acfc-30e78a6096ee service nova] [instance: d024b6b2-dd10-4112-89e8-ced57efe8208] Received unexpected event network-vif-plugged-06e24fbd-1278-4395-8108-f966a61c7b6c for instance with vm_state building and task_state spawning. [ 962.759000] env[63515]: DEBUG nova.compute.manager [req-9c250657-cb66-4ddf-90d0-610336e43aa9 req-fc28d6a1-c709-4c24-acfc-30e78a6096ee service nova] [instance: d024b6b2-dd10-4112-89e8-ced57efe8208] Received event network-changed-06e24fbd-1278-4395-8108-f966a61c7b6c {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 962.759329] env[63515]: DEBUG nova.compute.manager [req-9c250657-cb66-4ddf-90d0-610336e43aa9 req-fc28d6a1-c709-4c24-acfc-30e78a6096ee service nova] [instance: d024b6b2-dd10-4112-89e8-ced57efe8208] Refreshing instance network info cache due to event network-changed-06e24fbd-1278-4395-8108-f966a61c7b6c. {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 962.759601] env[63515]: DEBUG oslo_concurrency.lockutils [req-9c250657-cb66-4ddf-90d0-610336e43aa9 req-fc28d6a1-c709-4c24-acfc-30e78a6096ee service nova] Acquiring lock "refresh_cache-d024b6b2-dd10-4112-89e8-ced57efe8208" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 962.759824] env[63515]: DEBUG oslo_concurrency.lockutils [req-9c250657-cb66-4ddf-90d0-610336e43aa9 req-fc28d6a1-c709-4c24-acfc-30e78a6096ee service nova] Acquired lock "refresh_cache-d024b6b2-dd10-4112-89e8-ced57efe8208" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 962.760058] env[63515]: DEBUG nova.network.neutron [req-9c250657-cb66-4ddf-90d0-610336e43aa9 req-fc28d6a1-c709-4c24-acfc-30e78a6096ee service nova] [instance: d024b6b2-dd10-4112-89e8-ced57efe8208] Refreshing network info cache for port 06e24fbd-1278-4395-8108-f966a61c7b6c {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 962.771218] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-81bc1898-e097-4e8c-b6d2-4f02be1d9d2a tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: 918d5b35-c46d-47c5-b59f-175831aed9d6] Powering off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 962.771698] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9dca6211-e1d1-4449-ad73-163d0f6dee15 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.778899] env[63515]: DEBUG oslo_vmware.api [None req-2099b375-1441-4c83-88b0-3c0c09203cee tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Task: {'id': task-1111684, 'name': ReconfigVM_Task, 'duration_secs': 0.383526} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 962.780156] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-2099b375-1441-4c83-88b0-3c0c09203cee tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] [instance: 86b1ddcf-585c-41d9-84fe-17b794f1abbc] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-243544', 'volume_id': 'aa5c9d63-11af-4c5d-ae0b-9aa149332c19', 'name': 'volume-aa5c9d63-11af-4c5d-ae0b-9aa149332c19', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '86b1ddcf-585c-41d9-84fe-17b794f1abbc', 'attached_at': '', 'detached_at': '', 'volume_id': 'aa5c9d63-11af-4c5d-ae0b-9aa149332c19', 'serial': 'aa5c9d63-11af-4c5d-ae0b-9aa149332c19'} {{(pid=63515) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 962.781668] env[63515]: DEBUG oslo_vmware.api [None req-81bc1898-e097-4e8c-b6d2-4f02be1d9d2a tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Waiting for the task: (returnval){ [ 962.781668] env[63515]: value = "task-1111686" [ 962.781668] env[63515]: _type = "Task" [ 962.781668] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 962.791483] env[63515]: DEBUG oslo_vmware.api [None req-81bc1898-e097-4e8c-b6d2-4f02be1d9d2a tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Task: {'id': task-1111686, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.936727] env[63515]: DEBUG oslo_concurrency.lockutils [None req-165bb947-b1a3-4f80-91f4-6990bed3b114 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Acquiring lock "refresh_cache-d024b6b2-dd10-4112-89e8-ced57efe8208" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 962.978027] env[63515]: DEBUG oslo_vmware.api [None req-57194bc7-9fdc-4bda-bd6d-4b604b59fffa tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Task: {'id': task-1111681, 'name': CloneVM_Task, 'duration_secs': 1.383218} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 962.978353] env[63515]: INFO nova.virt.vmwareapi.vmops [None req-57194bc7-9fdc-4bda-bd6d-4b604b59fffa tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 68766cd4-84be-475b-8494-d7ab43a9e969] Created linked-clone VM from snapshot [ 962.979224] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ea94219-98ad-4c01-9d0b-204d234b991b {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.987064] env[63515]: DEBUG nova.virt.vmwareapi.images [None req-57194bc7-9fdc-4bda-bd6d-4b604b59fffa tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 68766cd4-84be-475b-8494-d7ab43a9e969] Uploading image fc46b835-451b-4001-9747-41c9d5d93184 {{(pid=63515) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 963.013721] env[63515]: DEBUG oslo_vmware.rw_handles [None req-57194bc7-9fdc-4bda-bd6d-4b604b59fffa tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 963.013721] env[63515]: value = "vm-243547" [ 963.013721] env[63515]: _type = "VirtualMachine" [ 963.013721] env[63515]: }. {{(pid=63515) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 963.014134] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-4047433f-1db4-4dd3-ba6e-08cd15e22c93 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.020569] env[63515]: DEBUG oslo_vmware.rw_handles [None req-57194bc7-9fdc-4bda-bd6d-4b604b59fffa tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Lease: (returnval){ [ 963.020569] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]52dae6f3-838f-b062-f414-c53d2ea24ca8" [ 963.020569] env[63515]: _type = "HttpNfcLease" [ 963.020569] env[63515]: } obtained for exporting VM: (result){ [ 963.020569] env[63515]: value = "vm-243547" [ 963.020569] env[63515]: _type = "VirtualMachine" [ 963.020569] env[63515]: }. {{(pid=63515) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 963.020851] env[63515]: DEBUG oslo_vmware.api [None req-57194bc7-9fdc-4bda-bd6d-4b604b59fffa tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Waiting for the lease: (returnval){ [ 963.020851] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]52dae6f3-838f-b062-f414-c53d2ea24ca8" [ 963.020851] env[63515]: _type = "HttpNfcLease" [ 963.020851] env[63515]: } to be ready. {{(pid=63515) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 963.027849] env[63515]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 963.027849] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]52dae6f3-838f-b062-f414-c53d2ea24ca8" [ 963.027849] env[63515]: _type = "HttpNfcLease" [ 963.027849] env[63515]: } is initializing. {{(pid=63515) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 963.034736] env[63515]: DEBUG oslo_concurrency.lockutils [None req-8a925afd-d6aa-4a9a-a54f-bf6262925289 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.345s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 963.037059] env[63515]: DEBUG oslo_concurrency.lockutils [None req-db106ad7-1693-4446-8352-ddb2ff9c25ce tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 5.640s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 963.037335] env[63515]: DEBUG nova.objects.instance [None req-db106ad7-1693-4446-8352-ddb2ff9c25ce tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Lazy-loading 'resources' on Instance uuid eab16df9-7bb5-4576-bca0-769a561c5fe9 {{(pid=63515) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 963.051192] env[63515]: INFO nova.scheduler.client.report [None req-8a925afd-d6aa-4a9a-a54f-bf6262925289 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Deleted allocations for instance 2fa098b6-ecc0-491f-8cfc-5ee335bbf943 [ 963.124320] env[63515]: DEBUG oslo_concurrency.lockutils [None req-3f7a26fd-cc84-46d9-9d28-feb378f8b5e6 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Lock "75f4dc79-4b73-4ddc-be03-3653a5a0797e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 22.690s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 963.135868] env[63515]: DEBUG oslo_vmware.api [None req-3f7a26fd-cc84-46d9-9d28-feb378f8b5e6 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Task: {'id': task-1111685, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.136927} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 963.136204] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-3f7a26fd-cc84-46d9-9d28-feb378f8b5e6 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] [instance: d438dd30-b816-48a6-94c6-11e6e9bd5b71] Extended root virtual disk {{(pid=63515) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 963.137077] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82131edd-14a0-46ea-813f-1933afa4304a {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.160629] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-3f7a26fd-cc84-46d9-9d28-feb378f8b5e6 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] [instance: d438dd30-b816-48a6-94c6-11e6e9bd5b71] Reconfiguring VM instance instance-0000005b to attach disk [datastore1] d438dd30-b816-48a6-94c6-11e6e9bd5b71/d438dd30-b816-48a6-94c6-11e6e9bd5b71.vmdk or device None with type sparse {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 963.160922] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4093a55d-a95e-4928-8b27-6d3d77f988af {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.179806] env[63515]: DEBUG oslo_vmware.api [None req-3f7a26fd-cc84-46d9-9d28-feb378f8b5e6 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Waiting for the task: (returnval){ [ 963.179806] env[63515]: value = "task-1111688" [ 963.179806] env[63515]: _type = "Task" [ 963.179806] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 963.188322] env[63515]: DEBUG oslo_vmware.api [None req-3f7a26fd-cc84-46d9-9d28-feb378f8b5e6 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Task: {'id': task-1111688, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.291527] env[63515]: DEBUG nova.network.neutron [req-9c250657-cb66-4ddf-90d0-610336e43aa9 req-fc28d6a1-c709-4c24-acfc-30e78a6096ee service nova] [instance: d024b6b2-dd10-4112-89e8-ced57efe8208] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 963.298586] env[63515]: DEBUG oslo_vmware.api [None req-81bc1898-e097-4e8c-b6d2-4f02be1d9d2a tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Task: {'id': task-1111686, 'name': PowerOffVM_Task, 'duration_secs': 0.384052} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 963.299230] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-81bc1898-e097-4e8c-b6d2-4f02be1d9d2a tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: 918d5b35-c46d-47c5-b59f-175831aed9d6] Powered off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 963.299230] env[63515]: DEBUG nova.compute.manager [None req-81bc1898-e097-4e8c-b6d2-4f02be1d9d2a tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: 918d5b35-c46d-47c5-b59f-175831aed9d6] Checking state {{(pid=63515) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 963.299839] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e0a817e-c6b0-4ce9-bc97-b1203d5b27e3 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.373639] env[63515]: DEBUG nova.network.neutron [req-9c250657-cb66-4ddf-90d0-610336e43aa9 req-fc28d6a1-c709-4c24-acfc-30e78a6096ee service nova] [instance: d024b6b2-dd10-4112-89e8-ced57efe8208] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 963.451496] env[63515]: DEBUG oslo_concurrency.lockutils [None req-113af8fa-af69-464c-ac2e-e5cf96a3de06 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Acquiring lock "61cf2f9c-5d35-46c4-a477-d2cef9fcd5ae" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 963.451669] env[63515]: DEBUG oslo_concurrency.lockutils [None req-113af8fa-af69-464c-ac2e-e5cf96a3de06 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Lock "61cf2f9c-5d35-46c4-a477-d2cef9fcd5ae" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 963.528604] env[63515]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 963.528604] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]52dae6f3-838f-b062-f414-c53d2ea24ca8" [ 963.528604] env[63515]: _type = "HttpNfcLease" [ 963.528604] env[63515]: } is ready. {{(pid=63515) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 963.528996] env[63515]: DEBUG oslo_vmware.rw_handles [None req-57194bc7-9fdc-4bda-bd6d-4b604b59fffa tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 963.528996] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]52dae6f3-838f-b062-f414-c53d2ea24ca8" [ 963.528996] env[63515]: _type = "HttpNfcLease" [ 963.528996] env[63515]: }. {{(pid=63515) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 963.529747] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c097e31-d552-4a44-983c-5f1e99dcf706 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.537271] env[63515]: DEBUG oslo_vmware.rw_handles [None req-57194bc7-9fdc-4bda-bd6d-4b604b59fffa tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52cbe900-84a3-8fda-b5af-75b7f0e7d6a0/disk-0.vmdk from lease info. {{(pid=63515) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 963.537455] env[63515]: DEBUG oslo_vmware.rw_handles [None req-57194bc7-9fdc-4bda-bd6d-4b604b59fffa tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52cbe900-84a3-8fda-b5af-75b7f0e7d6a0/disk-0.vmdk for reading. {{(pid=63515) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 963.609351] env[63515]: DEBUG oslo_concurrency.lockutils [None req-8a925afd-d6aa-4a9a-a54f-bf6262925289 tempest-ImagesTestJSON-1492052665 tempest-ImagesTestJSON-1492052665-project-member] Lock "2fa098b6-ecc0-491f-8cfc-5ee335bbf943" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 15.844s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 963.639515] env[63515]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-8a717cca-4510-4a2e-9aa0-e797c5129269 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.692457] env[63515]: DEBUG oslo_vmware.api [None req-3f7a26fd-cc84-46d9-9d28-feb378f8b5e6 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Task: {'id': task-1111688, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.813385] env[63515]: DEBUG oslo_concurrency.lockutils [None req-81bc1898-e097-4e8c-b6d2-4f02be1d9d2a tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Lock "918d5b35-c46d-47c5-b59f-175831aed9d6" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.562s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 963.817576] env[63515]: DEBUG nova.objects.instance [None req-2099b375-1441-4c83-88b0-3c0c09203cee tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Lazy-loading 'flavor' on Instance uuid 86b1ddcf-585c-41d9-84fe-17b794f1abbc {{(pid=63515) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 963.877852] env[63515]: DEBUG oslo_concurrency.lockutils [req-9c250657-cb66-4ddf-90d0-610336e43aa9 req-fc28d6a1-c709-4c24-acfc-30e78a6096ee service nova] Releasing lock "refresh_cache-d024b6b2-dd10-4112-89e8-ced57efe8208" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 963.878704] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3afaf0b0-be84-465c-b483-42b2cf55291f {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.881499] env[63515]: DEBUG oslo_concurrency.lockutils [None req-165bb947-b1a3-4f80-91f4-6990bed3b114 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Acquired lock "refresh_cache-d024b6b2-dd10-4112-89e8-ced57efe8208" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 963.882507] env[63515]: DEBUG nova.network.neutron [None req-165bb947-b1a3-4f80-91f4-6990bed3b114 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: d024b6b2-dd10-4112-89e8-ced57efe8208] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 963.888770] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8683e930-a884-4ae4-8753-505d510d20f6 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.931554] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0c73d0b-6ae8-415b-9bf1-634cd5655214 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.940623] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2136ea6a-ec0f-463d-9b41-59f9f59fc4d2 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.957234] env[63515]: DEBUG nova.compute.manager [None req-113af8fa-af69-464c-ac2e-e5cf96a3de06 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: 61cf2f9c-5d35-46c4-a477-d2cef9fcd5ae] Starting instance... {{(pid=63515) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 963.960552] env[63515]: DEBUG nova.compute.provider_tree [None req-db106ad7-1693-4446-8352-ddb2ff9c25ce tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 964.195634] env[63515]: DEBUG oslo_vmware.api [None req-3f7a26fd-cc84-46d9-9d28-feb378f8b5e6 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Task: {'id': task-1111688, 'name': ReconfigVM_Task, 'duration_secs': 0.762297} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 964.195946] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-3f7a26fd-cc84-46d9-9d28-feb378f8b5e6 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] [instance: d438dd30-b816-48a6-94c6-11e6e9bd5b71] Reconfigured VM instance instance-0000005b to attach disk [datastore1] d438dd30-b816-48a6-94c6-11e6e9bd5b71/d438dd30-b816-48a6-94c6-11e6e9bd5b71.vmdk or device None with type sparse {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 964.196887] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-2d1f3b95-7ade-40f9-9d9b-37432bf89895 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.209040] env[63515]: DEBUG oslo_vmware.api [None req-3f7a26fd-cc84-46d9-9d28-feb378f8b5e6 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Waiting for the task: (returnval){ [ 964.209040] env[63515]: value = "task-1111689" [ 964.209040] env[63515]: _type = "Task" [ 964.209040] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 964.215827] env[63515]: DEBUG oslo_vmware.api [None req-3f7a26fd-cc84-46d9-9d28-feb378f8b5e6 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Task: {'id': task-1111689, 'name': Rename_Task} progress is 5%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.322558] env[63515]: DEBUG oslo_concurrency.lockutils [None req-2099b375-1441-4c83-88b0-3c0c09203cee tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Lock "86b1ddcf-585c-41d9-84fe-17b794f1abbc" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.316s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 964.429950] env[63515]: DEBUG nova.network.neutron [None req-165bb947-b1a3-4f80-91f4-6990bed3b114 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: d024b6b2-dd10-4112-89e8-ced57efe8208] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 964.464148] env[63515]: DEBUG nova.scheduler.client.report [None req-db106ad7-1693-4446-8352-ddb2ff9c25ce tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 964.488040] env[63515]: DEBUG oslo_concurrency.lockutils [None req-113af8fa-af69-464c-ac2e-e5cf96a3de06 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 964.643625] env[63515]: DEBUG nova.network.neutron [None req-165bb947-b1a3-4f80-91f4-6990bed3b114 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: d024b6b2-dd10-4112-89e8-ced57efe8208] Updating instance_info_cache with network_info: [{"id": "06e24fbd-1278-4395-8108-f966a61c7b6c", "address": "fa:16:3e:bd:f5:8c", "network": {"id": "581d448e-be1b-4d5b-8a6e-c27326aaffc5", "bridge": "br-int", "label": "tempest-ServersTestJSON-1769610073-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "299fdeff647f486390366d5bbf911518", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fb143ef7-8271-4a8a-a4aa-8eba9a89f6a1", "external-id": "nsx-vlan-transportzone-504", "segmentation_id": 504, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap06e24fbd-12", "ovs_interfaceid": "06e24fbd-1278-4395-8108-f966a61c7b6c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 964.717056] env[63515]: DEBUG oslo_vmware.api [None req-3f7a26fd-cc84-46d9-9d28-feb378f8b5e6 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Task: {'id': task-1111689, 'name': Rename_Task} progress is 14%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.869147] env[63515]: INFO nova.compute.manager [None req-16f0ea70-a973-4a3c-bc2e-292ffb4571ac tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: 918d5b35-c46d-47c5-b59f-175831aed9d6] Rebuilding instance [ 964.924738] env[63515]: DEBUG nova.compute.manager [None req-16f0ea70-a973-4a3c-bc2e-292ffb4571ac tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: 918d5b35-c46d-47c5-b59f-175831aed9d6] Checking state {{(pid=63515) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 964.925644] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6271acb5-2e49-4af0-9d99-cb7bcad35f78 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.971594] env[63515]: DEBUG oslo_concurrency.lockutils [None req-db106ad7-1693-4446-8352-ddb2ff9c25ce tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.934s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 964.975046] env[63515]: DEBUG oslo_concurrency.lockutils [None req-2359bb47-ebaa-4b05-a6cb-20ad2f0f493e tempest-ServerMetadataTestJSON-1034558863 tempest-ServerMetadataTestJSON-1034558863-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 7.266s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 964.979881] env[63515]: DEBUG nova.objects.instance [None req-2359bb47-ebaa-4b05-a6cb-20ad2f0f493e tempest-ServerMetadataTestJSON-1034558863 tempest-ServerMetadataTestJSON-1034558863-project-member] Lazy-loading 'resources' on Instance uuid b25bf2ba-cf08-44ac-a524-aa87cf785575 {{(pid=63515) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 965.002184] env[63515]: INFO nova.scheduler.client.report [None req-db106ad7-1693-4446-8352-ddb2ff9c25ce tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Deleted allocations for instance eab16df9-7bb5-4576-bca0-769a561c5fe9 [ 965.148590] env[63515]: DEBUG oslo_concurrency.lockutils [None req-165bb947-b1a3-4f80-91f4-6990bed3b114 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Releasing lock "refresh_cache-d024b6b2-dd10-4112-89e8-ced57efe8208" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 965.148941] env[63515]: DEBUG nova.compute.manager [None req-165bb947-b1a3-4f80-91f4-6990bed3b114 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: d024b6b2-dd10-4112-89e8-ced57efe8208] Instance network_info: |[{"id": "06e24fbd-1278-4395-8108-f966a61c7b6c", "address": "fa:16:3e:bd:f5:8c", "network": {"id": "581d448e-be1b-4d5b-8a6e-c27326aaffc5", "bridge": "br-int", "label": "tempest-ServersTestJSON-1769610073-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "299fdeff647f486390366d5bbf911518", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fb143ef7-8271-4a8a-a4aa-8eba9a89f6a1", "external-id": "nsx-vlan-transportzone-504", "segmentation_id": 504, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap06e24fbd-12", "ovs_interfaceid": "06e24fbd-1278-4395-8108-f966a61c7b6c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 965.149407] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-165bb947-b1a3-4f80-91f4-6990bed3b114 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: d024b6b2-dd10-4112-89e8-ced57efe8208] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:bd:f5:8c', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'fb143ef7-8271-4a8a-a4aa-8eba9a89f6a1', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '06e24fbd-1278-4395-8108-f966a61c7b6c', 'vif_model': 'vmxnet3'}] {{(pid=63515) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 965.157849] env[63515]: DEBUG oslo.service.loopingcall [None req-165bb947-b1a3-4f80-91f4-6990bed3b114 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 965.158198] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d024b6b2-dd10-4112-89e8-ced57efe8208] Creating VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 965.158520] env[63515]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e7515c61-d143-41fc-8ebe-7d74da852d16 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.179321] env[63515]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 965.179321] env[63515]: value = "task-1111690" [ 965.179321] env[63515]: _type = "Task" [ 965.179321] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 965.188522] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1111690, 'name': CreateVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.217027] env[63515]: DEBUG oslo_vmware.api [None req-3f7a26fd-cc84-46d9-9d28-feb378f8b5e6 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Task: {'id': task-1111689, 'name': Rename_Task} progress is 14%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.331420] env[63515]: DEBUG oslo_concurrency.lockutils [None req-60a0611f-4742-4524-a115-5b0171ec85a1 tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Acquiring lock "86b1ddcf-585c-41d9-84fe-17b794f1abbc" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 965.331986] env[63515]: DEBUG oslo_concurrency.lockutils [None req-60a0611f-4742-4524-a115-5b0171ec85a1 tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Lock "86b1ddcf-585c-41d9-84fe-17b794f1abbc" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.001s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 965.441772] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-16f0ea70-a973-4a3c-bc2e-292ffb4571ac tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: 918d5b35-c46d-47c5-b59f-175831aed9d6] Powering off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 965.442141] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5187991b-2fd4-49d3-a518-7d27ae801b52 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.449165] env[63515]: DEBUG oslo_vmware.api [None req-16f0ea70-a973-4a3c-bc2e-292ffb4571ac tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Waiting for the task: (returnval){ [ 965.449165] env[63515]: value = "task-1111691" [ 965.449165] env[63515]: _type = "Task" [ 965.449165] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 965.459592] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-16f0ea70-a973-4a3c-bc2e-292ffb4571ac tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: 918d5b35-c46d-47c5-b59f-175831aed9d6] VM already powered off {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 965.459953] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-16f0ea70-a973-4a3c-bc2e-292ffb4571ac tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: 918d5b35-c46d-47c5-b59f-175831aed9d6] Destroying instance {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 965.460866] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07e079aa-fd65-44eb-a2b9-ce2297c7b6e5 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.468919] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-16f0ea70-a973-4a3c-bc2e-292ffb4571ac tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: 918d5b35-c46d-47c5-b59f-175831aed9d6] Unregistering the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 965.469317] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-bbe444f6-fe60-4966-b346-cef35772ea51 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.516251] env[63515]: DEBUG oslo_concurrency.lockutils [None req-db106ad7-1693-4446-8352-ddb2ff9c25ce tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Lock "eab16df9-7bb5-4576-bca0-769a561c5fe9" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 11.338s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 965.518951] env[63515]: DEBUG oslo_concurrency.lockutils [req-09f6a403-3161-41c9-a70f-a4a80551b661 req-01cc409f-3c9a-4500-890c-702330e825ed service nova] Acquired lock "eab16df9-7bb5-4576-bca0-769a561c5fe9" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 965.519558] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e782b29f-e288-4f5c-a2b2-305c4d58045b {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.531779] env[63515]: WARNING suds.client [-] Web service reported a SOAP processing fault using an unexpected HTTP status code 200. Reporting as an internal server error. [ 965.532012] env[63515]: DEBUG oslo_vmware.api [-] Fault list: [ManagedObjectNotFound] {{(pid=63515) _invoke_api /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:337}} [ 965.536668] env[63515]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-7f751b61-0752-406b-bae7-c88fae5d8e30 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.541973] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-16f0ea70-a973-4a3c-bc2e-292ffb4571ac tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: 918d5b35-c46d-47c5-b59f-175831aed9d6] Unregistered the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 965.542245] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-16f0ea70-a973-4a3c-bc2e-292ffb4571ac tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: 918d5b35-c46d-47c5-b59f-175831aed9d6] Deleting contents of the VM from datastore datastore2 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 965.545961] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-16f0ea70-a973-4a3c-bc2e-292ffb4571ac tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Deleting the datastore file [datastore2] 918d5b35-c46d-47c5-b59f-175831aed9d6 {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 965.547349] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-883d2c4c-fad6-470a-b1e7-27fca4a2e095 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.554650] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4fcf12c0-f6b3-43e4-9f78-1703f4ca7ea2 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.572283] env[63515]: DEBUG oslo_vmware.api [None req-16f0ea70-a973-4a3c-bc2e-292ffb4571ac tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Waiting for the task: (returnval){ [ 965.572283] env[63515]: value = "task-1111693" [ 965.572283] env[63515]: _type = "Task" [ 965.572283] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 965.582834] env[63515]: DEBUG oslo_vmware.api [None req-16f0ea70-a973-4a3c-bc2e-292ffb4571ac tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Task: {'id': task-1111693, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.598747] env[63515]: ERROR root [req-09f6a403-3161-41c9-a70f-a4a80551b661 req-01cc409f-3c9a-4500-890c-702330e825ed service nova] Original exception being dropped: ['Traceback (most recent call last):\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py", line 377, in request_handler\n response = request(managed_object, **kwargs)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/suds/client.py", line 586, in __call__\n return client.invoke(args, kwargs)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/suds/client.py", line 728, in invoke\n result = self.send(soapenv, timeout=timeout)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/suds/client.py", line 777, in send\n return self.process_reply(reply.message, None, None)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/suds/client.py", line 840, in process_reply\n raise WebFault(fault, replyroot)\n', "suds.WebFault: Server raised fault: 'The object 'vim.VirtualMachine:vm-243502' has already been deleted or has not been completely created'\n", '\nDuring handling of the above exception, another exception occurred:\n\n', 'Traceback (most recent call last):\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py", line 301, in _invoke_api\n return api_method(*args, **kwargs)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/vim_util.py", line 480, in get_object_property\n props = get_object_properties(vim, moref, [property_name],\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/vim_util.py", line 360, in get_object_properties\n retrieve_result = vim.RetrievePropertiesEx(\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py", line 413, in request_handler\n raise exceptions.VimFaultException(fault_list, fault_string,\n', "oslo_vmware.exceptions.VimFaultException: The object 'vim.VirtualMachine:vm-243502' has already been deleted or has not been completely created\nCause: Server raised fault: 'The object 'vim.VirtualMachine:vm-243502' has already been deleted or has not been completely created'\nFaults: [ManagedObjectNotFound]\nDetails: {'obj': 'vm-243502'}\n", '\nDuring handling of the above exception, another exception occurred:\n\n', 'Traceback (most recent call last):\n', ' File "/opt/stack/nova/nova/virt/vmwareapi/session.py", line 123, in _call_method\n return self.invoke_api(module, method, self.vim, *args,\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py", line 358, in invoke_api\n return _invoke_api(module, method, *args, **kwargs)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py", line 122, in func\n return evt.wait()\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait\n result = hub.switch()\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch\n return self.greenlet.switch()\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/common/loopingcall.py", line 122, in _inner\n idle = self.f(*self.args, **self.kw)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py", line 96, in _func\n result = f(*args, **kwargs)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py", line 341, in _invoke_api\n raise clazz(str(excep),\n', "oslo_vmware.exceptions.ManagedObjectNotFoundException: The object 'vim.VirtualMachine:vm-243502' has already been deleted or has not been completely created\nCause: Server raised fault: 'The object 'vim.VirtualMachine:vm-243502' has already been deleted or has not been completely created'\nFaults: [ManagedObjectNotFound]\nDetails: {'obj': 'vm-243502'}\n"]: nova.exception.InstanceNotFound: Instance eab16df9-7bb5-4576-bca0-769a561c5fe9 could not be found. [ 965.599116] env[63515]: DEBUG oslo_concurrency.lockutils [req-09f6a403-3161-41c9-a70f-a4a80551b661 req-01cc409f-3c9a-4500-890c-702330e825ed service nova] Releasing lock "eab16df9-7bb5-4576-bca0-769a561c5fe9" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 965.599226] env[63515]: DEBUG nova.compute.manager [req-09f6a403-3161-41c9-a70f-a4a80551b661 req-01cc409f-3c9a-4500-890c-702330e825ed service nova] [instance: eab16df9-7bb5-4576-bca0-769a561c5fe9] Detach interface failed, port_id=e3c72582-41bf-48d4-8614-b516a96362a7, reason: Instance eab16df9-7bb5-4576-bca0-769a561c5fe9 could not be found. {{(pid=63515) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 965.695520] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1111690, 'name': CreateVM_Task, 'duration_secs': 0.399838} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 965.696298] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d024b6b2-dd10-4112-89e8-ced57efe8208] Created VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 965.697833] env[63515]: DEBUG oslo_concurrency.lockutils [None req-165bb947-b1a3-4f80-91f4-6990bed3b114 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 965.698142] env[63515]: DEBUG oslo_concurrency.lockutils [None req-165bb947-b1a3-4f80-91f4-6990bed3b114 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 965.699209] env[63515]: DEBUG oslo_concurrency.lockutils [None req-165bb947-b1a3-4f80-91f4-6990bed3b114 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 965.699891] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-59644cf8-da9b-40ed-8b5f-8e5e6623b205 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.707965] env[63515]: DEBUG oslo_vmware.api [None req-165bb947-b1a3-4f80-91f4-6990bed3b114 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Waiting for the task: (returnval){ [ 965.707965] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]5292bd37-6c77-2352-3633-9a5728c80575" [ 965.707965] env[63515]: _type = "Task" [ 965.707965] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 965.722731] env[63515]: DEBUG oslo_vmware.api [None req-165bb947-b1a3-4f80-91f4-6990bed3b114 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]5292bd37-6c77-2352-3633-9a5728c80575, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.726245] env[63515]: DEBUG oslo_vmware.api [None req-3f7a26fd-cc84-46d9-9d28-feb378f8b5e6 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Task: {'id': task-1111689, 'name': Rename_Task, 'duration_secs': 1.237617} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 965.728635] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-3f7a26fd-cc84-46d9-9d28-feb378f8b5e6 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] [instance: d438dd30-b816-48a6-94c6-11e6e9bd5b71] Powering on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 965.729086] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5c69f4c9-a7ad-4f65-bd42-f2d4a7733e58 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.742032] env[63515]: DEBUG oslo_vmware.api [None req-3f7a26fd-cc84-46d9-9d28-feb378f8b5e6 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Waiting for the task: (returnval){ [ 965.742032] env[63515]: value = "task-1111694" [ 965.742032] env[63515]: _type = "Task" [ 965.742032] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 965.750463] env[63515]: DEBUG oslo_vmware.api [None req-3f7a26fd-cc84-46d9-9d28-feb378f8b5e6 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Task: {'id': task-1111694, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.764515] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03c01e07-b0e8-4440-9c26-9f60235ee4c2 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.772167] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef70cee4-e71e-42f5-acdf-249ecbcde929 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.805327] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2597751c-9520-4780-b86d-4da352eba700 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.813075] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8945eef4-5b25-4aac-b968-e496e2c66071 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.826355] env[63515]: DEBUG nova.compute.provider_tree [None req-2359bb47-ebaa-4b05-a6cb-20ad2f0f493e tempest-ServerMetadataTestJSON-1034558863 tempest-ServerMetadataTestJSON-1034558863-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 965.837066] env[63515]: DEBUG nova.compute.utils [None req-60a0611f-4742-4524-a115-5b0171ec85a1 tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Using /dev/sd instead of None {{(pid=63515) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 966.085388] env[63515]: DEBUG oslo_vmware.api [None req-16f0ea70-a973-4a3c-bc2e-292ffb4571ac tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Task: {'id': task-1111693, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.224349} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 966.085797] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-16f0ea70-a973-4a3c-bc2e-292ffb4571ac tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Deleted the datastore file {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 966.085846] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-16f0ea70-a973-4a3c-bc2e-292ffb4571ac tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: 918d5b35-c46d-47c5-b59f-175831aed9d6] Deleted contents of the VM from datastore datastore2 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 966.086048] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-16f0ea70-a973-4a3c-bc2e-292ffb4571ac tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: 918d5b35-c46d-47c5-b59f-175831aed9d6] Instance destroyed {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 966.221068] env[63515]: DEBUG oslo_vmware.api [None req-165bb947-b1a3-4f80-91f4-6990bed3b114 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]5292bd37-6c77-2352-3633-9a5728c80575, 'name': SearchDatastore_Task, 'duration_secs': 0.017046} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 966.221068] env[63515]: DEBUG oslo_concurrency.lockutils [None req-165bb947-b1a3-4f80-91f4-6990bed3b114 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 966.221068] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-165bb947-b1a3-4f80-91f4-6990bed3b114 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: d024b6b2-dd10-4112-89e8-ced57efe8208] Processing image 8a120570-cb06-4099-b262-554ca0ad15c5 {{(pid=63515) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 966.221068] env[63515]: DEBUG oslo_concurrency.lockutils [None req-165bb947-b1a3-4f80-91f4-6990bed3b114 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 966.221068] env[63515]: DEBUG oslo_concurrency.lockutils [None req-165bb947-b1a3-4f80-91f4-6990bed3b114 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 966.221068] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-165bb947-b1a3-4f80-91f4-6990bed3b114 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 966.221068] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-cab00aaa-48ed-494c-ac0e-6c6a72be1177 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.233027] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-165bb947-b1a3-4f80-91f4-6990bed3b114 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 966.233027] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-165bb947-b1a3-4f80-91f4-6990bed3b114 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63515) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 966.233027] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a0c4da9c-b9e4-4ccb-87e3-fb9e1268bcb9 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.237646] env[63515]: DEBUG oslo_vmware.api [None req-165bb947-b1a3-4f80-91f4-6990bed3b114 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Waiting for the task: (returnval){ [ 966.237646] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]521936c9-5cab-750e-0f29-f0d7af891e96" [ 966.237646] env[63515]: _type = "Task" [ 966.237646] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 966.248521] env[63515]: DEBUG oslo_vmware.api [None req-165bb947-b1a3-4f80-91f4-6990bed3b114 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]521936c9-5cab-750e-0f29-f0d7af891e96, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.253894] env[63515]: DEBUG oslo_vmware.api [None req-3f7a26fd-cc84-46d9-9d28-feb378f8b5e6 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Task: {'id': task-1111694, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.331022] env[63515]: DEBUG nova.scheduler.client.report [None req-2359bb47-ebaa-4b05-a6cb-20ad2f0f493e tempest-ServerMetadataTestJSON-1034558863 tempest-ServerMetadataTestJSON-1034558863-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 966.338981] env[63515]: DEBUG oslo_concurrency.lockutils [None req-60a0611f-4742-4524-a115-5b0171ec85a1 tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Lock "86b1ddcf-585c-41d9-84fe-17b794f1abbc" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.007s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 966.758285] env[63515]: DEBUG oslo_vmware.api [None req-165bb947-b1a3-4f80-91f4-6990bed3b114 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]521936c9-5cab-750e-0f29-f0d7af891e96, 'name': SearchDatastore_Task, 'duration_secs': 0.012996} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 966.765414] env[63515]: DEBUG oslo_vmware.api [None req-3f7a26fd-cc84-46d9-9d28-feb378f8b5e6 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Task: {'id': task-1111694, 'name': PowerOnVM_Task, 'duration_secs': 0.571406} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 966.765808] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-734562aa-71fc-4daa-b39d-5e3b0cfaed07 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.769825] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-3f7a26fd-cc84-46d9-9d28-feb378f8b5e6 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] [instance: d438dd30-b816-48a6-94c6-11e6e9bd5b71] Powered on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 966.770267] env[63515]: INFO nova.compute.manager [None req-3f7a26fd-cc84-46d9-9d28-feb378f8b5e6 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] [instance: d438dd30-b816-48a6-94c6-11e6e9bd5b71] Took 9.81 seconds to spawn the instance on the hypervisor. [ 966.770646] env[63515]: DEBUG nova.compute.manager [None req-3f7a26fd-cc84-46d9-9d28-feb378f8b5e6 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] [instance: d438dd30-b816-48a6-94c6-11e6e9bd5b71] Checking state {{(pid=63515) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 966.771990] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74a9f37a-2494-43c0-b027-483b3d514004 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.780100] env[63515]: DEBUG oslo_vmware.api [None req-165bb947-b1a3-4f80-91f4-6990bed3b114 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Waiting for the task: (returnval){ [ 966.780100] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]525f02c8-b68d-b6db-361b-d431b9883a5b" [ 966.780100] env[63515]: _type = "Task" [ 966.780100] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 966.798411] env[63515]: DEBUG oslo_vmware.api [None req-165bb947-b1a3-4f80-91f4-6990bed3b114 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]525f02c8-b68d-b6db-361b-d431b9883a5b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.836014] env[63515]: DEBUG oslo_concurrency.lockutils [None req-2359bb47-ebaa-4b05-a6cb-20ad2f0f493e tempest-ServerMetadataTestJSON-1034558863 tempest-ServerMetadataTestJSON-1034558863-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.860s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 966.838217] env[63515]: DEBUG oslo_concurrency.lockutils [None req-a5676663-6933-430d-aa5e-aeffa94052ee tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.972s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 966.840051] env[63515]: INFO nova.compute.claims [None req-a5676663-6933-430d-aa5e-aeffa94052ee tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: 6a66dd9b-7b24-4b70-b58f-0830cf39376c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 966.860459] env[63515]: INFO nova.scheduler.client.report [None req-2359bb47-ebaa-4b05-a6cb-20ad2f0f493e tempest-ServerMetadataTestJSON-1034558863 tempest-ServerMetadataTestJSON-1034558863-project-member] Deleted allocations for instance b25bf2ba-cf08-44ac-a524-aa87cf785575 [ 967.136643] env[63515]: DEBUG nova.virt.hardware [None req-16f0ea70-a973-4a3c-bc2e-292ffb4571ac tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T02:52:14Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T02:51:56Z,direct_url=,disk_format='vmdk',id=8a120570-cb06-4099-b262-554ca0ad15c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b569255cc43e42e7ae3f2b4ad37c6ef4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T02:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 967.137103] env[63515]: DEBUG nova.virt.hardware [None req-16f0ea70-a973-4a3c-bc2e-292ffb4571ac tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Flavor limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 967.137315] env[63515]: DEBUG nova.virt.hardware [None req-16f0ea70-a973-4a3c-bc2e-292ffb4571ac tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Image limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 967.137522] env[63515]: DEBUG nova.virt.hardware [None req-16f0ea70-a973-4a3c-bc2e-292ffb4571ac tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Flavor pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 967.137674] env[63515]: DEBUG nova.virt.hardware [None req-16f0ea70-a973-4a3c-bc2e-292ffb4571ac tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Image pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 967.137936] env[63515]: DEBUG nova.virt.hardware [None req-16f0ea70-a973-4a3c-bc2e-292ffb4571ac tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 967.138230] env[63515]: DEBUG nova.virt.hardware [None req-16f0ea70-a973-4a3c-bc2e-292ffb4571ac tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 967.138405] env[63515]: DEBUG nova.virt.hardware [None req-16f0ea70-a973-4a3c-bc2e-292ffb4571ac tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 967.138770] env[63515]: DEBUG nova.virt.hardware [None req-16f0ea70-a973-4a3c-bc2e-292ffb4571ac tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Got 1 possible topologies {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 967.138980] env[63515]: DEBUG nova.virt.hardware [None req-16f0ea70-a973-4a3c-bc2e-292ffb4571ac tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 967.139190] env[63515]: DEBUG nova.virt.hardware [None req-16f0ea70-a973-4a3c-bc2e-292ffb4571ac tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 967.140693] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9c88049-3008-4b42-b1c2-8014badb23ba {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.152106] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b48ad69-f8c8-46f8-ba68-4da80b1c20bc {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.167815] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-16f0ea70-a973-4a3c-bc2e-292ffb4571ac tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: 918d5b35-c46d-47c5-b59f-175831aed9d6] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:dc:ce:b2', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '3a0ddd7d-c321-4187-bdd8-b19044ea2c4a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b4724c8b-a2c7-4f52-8a75-f46434d7b527', 'vif_model': 'vmxnet3'}] {{(pid=63515) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 967.175329] env[63515]: DEBUG oslo.service.loopingcall [None req-16f0ea70-a973-4a3c-bc2e-292ffb4571ac tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 967.175620] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 918d5b35-c46d-47c5-b59f-175831aed9d6] Creating VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 967.175844] env[63515]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-43cbf2be-e410-404f-834d-f28737b61703 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.195400] env[63515]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 967.195400] env[63515]: value = "task-1111695" [ 967.195400] env[63515]: _type = "Task" [ 967.195400] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 967.203260] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1111695, 'name': CreateVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.296105] env[63515]: DEBUG oslo_vmware.api [None req-165bb947-b1a3-4f80-91f4-6990bed3b114 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]525f02c8-b68d-b6db-361b-d431b9883a5b, 'name': SearchDatastore_Task, 'duration_secs': 0.018249} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 967.297991] env[63515]: DEBUG oslo_concurrency.lockutils [None req-165bb947-b1a3-4f80-91f4-6990bed3b114 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 967.298208] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-165bb947-b1a3-4f80-91f4-6990bed3b114 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk to [datastore1] d024b6b2-dd10-4112-89e8-ced57efe8208/d024b6b2-dd10-4112-89e8-ced57efe8208.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 967.298675] env[63515]: INFO nova.compute.manager [None req-3f7a26fd-cc84-46d9-9d28-feb378f8b5e6 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] [instance: d438dd30-b816-48a6-94c6-11e6e9bd5b71] Took 25.85 seconds to build instance. [ 967.300061] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-67e33504-53d6-4c48-8ef0-02fa463e6cec {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.306244] env[63515]: DEBUG oslo_vmware.api [None req-165bb947-b1a3-4f80-91f4-6990bed3b114 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Waiting for the task: (returnval){ [ 967.306244] env[63515]: value = "task-1111696" [ 967.306244] env[63515]: _type = "Task" [ 967.306244] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 967.315604] env[63515]: DEBUG oslo_vmware.api [None req-165bb947-b1a3-4f80-91f4-6990bed3b114 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Task: {'id': task-1111696, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.368260] env[63515]: DEBUG oslo_concurrency.lockutils [None req-2359bb47-ebaa-4b05-a6cb-20ad2f0f493e tempest-ServerMetadataTestJSON-1034558863 tempest-ServerMetadataTestJSON-1034558863-project-member] Lock "b25bf2ba-cf08-44ac-a524-aa87cf785575" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 12.644s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 967.421696] env[63515]: DEBUG oslo_concurrency.lockutils [None req-60a0611f-4742-4524-a115-5b0171ec85a1 tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Acquiring lock "86b1ddcf-585c-41d9-84fe-17b794f1abbc" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 967.421978] env[63515]: DEBUG oslo_concurrency.lockutils [None req-60a0611f-4742-4524-a115-5b0171ec85a1 tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Lock "86b1ddcf-585c-41d9-84fe-17b794f1abbc" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 967.422255] env[63515]: INFO nova.compute.manager [None req-60a0611f-4742-4524-a115-5b0171ec85a1 tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] [instance: 86b1ddcf-585c-41d9-84fe-17b794f1abbc] Attaching volume faf85999-e3c6-4036-a18e-7d126676fad9 to /dev/sdc [ 967.467023] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4453162f-bc98-4e95-a3da-30e93cb7fe70 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.472590] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-789ba9b3-d39b-47b1-a3cd-85e7184e1dae {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.488157] env[63515]: DEBUG nova.virt.block_device [None req-60a0611f-4742-4524-a115-5b0171ec85a1 tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] [instance: 86b1ddcf-585c-41d9-84fe-17b794f1abbc] Updating existing volume attachment record: 3fc33c7f-6d5d-49bc-b919-a954211ab63a {{(pid=63515) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 967.630654] env[63515]: DEBUG oslo_concurrency.lockutils [None req-c547db4d-5471-41ef-8a57-e1e58211e41e tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Acquiring lock "2a29f37d-e786-48d3-b126-4854d6e7c67c" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 967.630953] env[63515]: DEBUG oslo_concurrency.lockutils [None req-c547db4d-5471-41ef-8a57-e1e58211e41e tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Lock "2a29f37d-e786-48d3-b126-4854d6e7c67c" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.001s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 967.706634] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1111695, 'name': CreateVM_Task} progress is 99%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.802558] env[63515]: DEBUG oslo_concurrency.lockutils [None req-3f7a26fd-cc84-46d9-9d28-feb378f8b5e6 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Lock "d438dd30-b816-48a6-94c6-11e6e9bd5b71" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 27.362s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 967.820863] env[63515]: DEBUG oslo_vmware.api [None req-165bb947-b1a3-4f80-91f4-6990bed3b114 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Task: {'id': task-1111696, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.103508] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9501802d-5226-404d-b2d3-309a3cf6fb9d {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.112318] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2e35293-15fc-41be-9638-e2f1db41672f {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.145095] env[63515]: INFO nova.compute.manager [None req-c547db4d-5471-41ef-8a57-e1e58211e41e tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] [instance: 2a29f37d-e786-48d3-b126-4854d6e7c67c] Detaching volume dea169a3-b753-4313-853a-cbf3a4c05ef8 [ 968.149552] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76cb3642-0578-4e15-b25a-2574d942e9a5 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.159873] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fddca070-7ec9-4dc9-8b1b-42052c845b57 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.175566] env[63515]: DEBUG nova.compute.provider_tree [None req-a5676663-6933-430d-aa5e-aeffa94052ee tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Updating inventory in ProviderTree for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 968.191121] env[63515]: INFO nova.virt.block_device [None req-c547db4d-5471-41ef-8a57-e1e58211e41e tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] [instance: 2a29f37d-e786-48d3-b126-4854d6e7c67c] Attempting to driver detach volume dea169a3-b753-4313-853a-cbf3a4c05ef8 from mountpoint /dev/sdb [ 968.191402] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-c547db4d-5471-41ef-8a57-e1e58211e41e tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] [instance: 2a29f37d-e786-48d3-b126-4854d6e7c67c] Volume detach. Driver type: vmdk {{(pid=63515) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 968.191607] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-c547db4d-5471-41ef-8a57-e1e58211e41e tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] [instance: 2a29f37d-e786-48d3-b126-4854d6e7c67c] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-243524', 'volume_id': 'dea169a3-b753-4313-853a-cbf3a4c05ef8', 'name': 'volume-dea169a3-b753-4313-853a-cbf3a4c05ef8', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '2a29f37d-e786-48d3-b126-4854d6e7c67c', 'attached_at': '', 'detached_at': '', 'volume_id': 'dea169a3-b753-4313-853a-cbf3a4c05ef8', 'serial': 'dea169a3-b753-4313-853a-cbf3a4c05ef8'} {{(pid=63515) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 968.192755] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7266289-f914-460e-8a19-1184795b4a29 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.217906] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb98e1fd-b235-44d4-95ab-5dc16b295025 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.223855] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1111695, 'name': CreateVM_Task, 'duration_secs': 0.513583} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 968.224812] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 918d5b35-c46d-47c5-b59f-175831aed9d6] Created VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 968.225258] env[63515]: DEBUG oslo_concurrency.lockutils [None req-16f0ea70-a973-4a3c-bc2e-292ffb4571ac tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 968.225430] env[63515]: DEBUG oslo_concurrency.lockutils [None req-16f0ea70-a973-4a3c-bc2e-292ffb4571ac tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 968.225800] env[63515]: DEBUG oslo_concurrency.lockutils [None req-16f0ea70-a973-4a3c-bc2e-292ffb4571ac tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 968.228170] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-82349f01-c9b1-4b4e-9bd7-2f96b38372fd {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.230219] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b69f2281-cfc6-4746-8677-b800a3971c4d {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.235298] env[63515]: DEBUG oslo_vmware.api [None req-16f0ea70-a973-4a3c-bc2e-292ffb4571ac tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Waiting for the task: (returnval){ [ 968.235298] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]5292ef76-ce19-8595-9ec0-216b93954753" [ 968.235298] env[63515]: _type = "Task" [ 968.235298] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 968.254497] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4d66877-18e8-4e9b-af30-5787fb58bd41 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.263996] env[63515]: DEBUG oslo_vmware.api [None req-16f0ea70-a973-4a3c-bc2e-292ffb4571ac tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]5292ef76-ce19-8595-9ec0-216b93954753, 'name': SearchDatastore_Task, 'duration_secs': 0.012156} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 968.274972] env[63515]: DEBUG oslo_concurrency.lockutils [None req-16f0ea70-a973-4a3c-bc2e-292ffb4571ac tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 968.275326] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-16f0ea70-a973-4a3c-bc2e-292ffb4571ac tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: 918d5b35-c46d-47c5-b59f-175831aed9d6] Processing image 8a120570-cb06-4099-b262-554ca0ad15c5 {{(pid=63515) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 968.275584] env[63515]: DEBUG oslo_concurrency.lockutils [None req-16f0ea70-a973-4a3c-bc2e-292ffb4571ac tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 968.275891] env[63515]: DEBUG oslo_concurrency.lockutils [None req-16f0ea70-a973-4a3c-bc2e-292ffb4571ac tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 968.275997] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-16f0ea70-a973-4a3c-bc2e-292ffb4571ac tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 968.276656] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-c547db4d-5471-41ef-8a57-e1e58211e41e tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] The volume has not been displaced from its original location: [datastore2] volume-dea169a3-b753-4313-853a-cbf3a4c05ef8/volume-dea169a3-b753-4313-853a-cbf3a4c05ef8.vmdk. No consolidation needed. {{(pid=63515) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 968.282061] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-c547db4d-5471-41ef-8a57-e1e58211e41e tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] [instance: 2a29f37d-e786-48d3-b126-4854d6e7c67c] Reconfiguring VM instance instance-00000045 to detach disk 2001 {{(pid=63515) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 968.284500] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e1d23b92-d688-4921-bfb8-644c87834318 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.285387] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e4b6b363-4914-42ef-89e1-d02ed2013695 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.306644] env[63515]: DEBUG oslo_vmware.api [None req-c547db4d-5471-41ef-8a57-e1e58211e41e tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Waiting for the task: (returnval){ [ 968.306644] env[63515]: value = "task-1111698" [ 968.306644] env[63515]: _type = "Task" [ 968.306644] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 968.311661] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-16f0ea70-a973-4a3c-bc2e-292ffb4571ac tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 968.311661] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-16f0ea70-a973-4a3c-bc2e-292ffb4571ac tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63515) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 968.315099] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-09dad698-fb63-48b2-9735-e2a793bdec74 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.324177] env[63515]: DEBUG oslo_vmware.api [None req-c547db4d-5471-41ef-8a57-e1e58211e41e tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Task: {'id': task-1111698, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.327791] env[63515]: DEBUG oslo_vmware.api [None req-16f0ea70-a973-4a3c-bc2e-292ffb4571ac tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Waiting for the task: (returnval){ [ 968.327791] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]5211a921-2b53-8dc9-eacb-79d72afc2a41" [ 968.327791] env[63515]: _type = "Task" [ 968.327791] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 968.328049] env[63515]: DEBUG oslo_vmware.api [None req-165bb947-b1a3-4f80-91f4-6990bed3b114 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Task: {'id': task-1111696, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.691585} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 968.328382] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-165bb947-b1a3-4f80-91f4-6990bed3b114 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk to [datastore1] d024b6b2-dd10-4112-89e8-ced57efe8208/d024b6b2-dd10-4112-89e8-ced57efe8208.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 968.328779] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-165bb947-b1a3-4f80-91f4-6990bed3b114 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: d024b6b2-dd10-4112-89e8-ced57efe8208] Extending root virtual disk to 1048576 {{(pid=63515) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 968.332111] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-2b691aba-b946-4d21-a360-7624920f8c4b {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.340938] env[63515]: DEBUG oslo_vmware.api [None req-16f0ea70-a973-4a3c-bc2e-292ffb4571ac tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]5211a921-2b53-8dc9-eacb-79d72afc2a41, 'name': SearchDatastore_Task, 'duration_secs': 0.014885} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 968.342862] env[63515]: DEBUG oslo_vmware.api [None req-165bb947-b1a3-4f80-91f4-6990bed3b114 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Waiting for the task: (returnval){ [ 968.342862] env[63515]: value = "task-1111699" [ 968.342862] env[63515]: _type = "Task" [ 968.342862] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 968.343041] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9f8f5139-8d0e-439a-8bfb-f11083759d85 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.352295] env[63515]: DEBUG oslo_vmware.api [None req-16f0ea70-a973-4a3c-bc2e-292ffb4571ac tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Waiting for the task: (returnval){ [ 968.352295] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]5254bcff-067f-ba8e-ca9d-1200b83c2543" [ 968.352295] env[63515]: _type = "Task" [ 968.352295] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 968.355703] env[63515]: DEBUG oslo_vmware.api [None req-165bb947-b1a3-4f80-91f4-6990bed3b114 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Task: {'id': task-1111699, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.366445] env[63515]: DEBUG oslo_concurrency.lockutils [None req-adce163d-60b0-4e35-be4e-020840038d00 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Acquiring lock "724b05bc-ce4b-4f99-ad14-f3abea5567f2" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 968.366695] env[63515]: DEBUG oslo_concurrency.lockutils [None req-adce163d-60b0-4e35-be4e-020840038d00 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Lock "724b05bc-ce4b-4f99-ad14-f3abea5567f2" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 968.371710] env[63515]: DEBUG oslo_vmware.api [None req-16f0ea70-a973-4a3c-bc2e-292ffb4571ac tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]5254bcff-067f-ba8e-ca9d-1200b83c2543, 'name': SearchDatastore_Task, 'duration_secs': 0.010535} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 968.372456] env[63515]: DEBUG oslo_concurrency.lockutils [None req-16f0ea70-a973-4a3c-bc2e-292ffb4571ac tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 968.372723] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-16f0ea70-a973-4a3c-bc2e-292ffb4571ac tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk to [datastore1] 918d5b35-c46d-47c5-b59f-175831aed9d6/918d5b35-c46d-47c5-b59f-175831aed9d6.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 968.374294] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-24e8cc40-1012-4772-908e-cabaa36f0d94 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.380054] env[63515]: DEBUG oslo_vmware.api [None req-16f0ea70-a973-4a3c-bc2e-292ffb4571ac tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Waiting for the task: (returnval){ [ 968.380054] env[63515]: value = "task-1111700" [ 968.380054] env[63515]: _type = "Task" [ 968.380054] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 968.388990] env[63515]: DEBUG oslo_vmware.api [None req-16f0ea70-a973-4a3c-bc2e-292ffb4571ac tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Task: {'id': task-1111700, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.696810] env[63515]: ERROR nova.scheduler.client.report [None req-a5676663-6933-430d-aa5e-aeffa94052ee tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [req-9c10025c-788c-4e41-ac75-95ffaedbbe5f] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 2bd86232-2b6f-44d5-9057-1a3a6b27185a. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-9c10025c-788c-4e41-ac75-95ffaedbbe5f"}]} [ 968.715120] env[63515]: DEBUG nova.scheduler.client.report [None req-a5676663-6933-430d-aa5e-aeffa94052ee tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Refreshing inventories for resource provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 968.730801] env[63515]: DEBUG nova.scheduler.client.report [None req-a5676663-6933-430d-aa5e-aeffa94052ee tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Updating ProviderTree inventory for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 968.731096] env[63515]: DEBUG nova.compute.provider_tree [None req-a5676663-6933-430d-aa5e-aeffa94052ee tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Updating inventory in ProviderTree for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 968.742655] env[63515]: DEBUG nova.scheduler.client.report [None req-a5676663-6933-430d-aa5e-aeffa94052ee tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Refreshing aggregate associations for resource provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a, aggregates: None {{(pid=63515) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 968.761097] env[63515]: DEBUG nova.scheduler.client.report [None req-a5676663-6933-430d-aa5e-aeffa94052ee tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Refreshing trait associations for resource provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a, traits: COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=63515) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 968.804437] env[63515]: DEBUG oslo_concurrency.lockutils [None req-c7884167-3670-45b6-b67d-b6f36de44d1a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Acquiring lock "75f4dc79-4b73-4ddc-be03-3653a5a0797e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 968.804437] env[63515]: DEBUG oslo_concurrency.lockutils [None req-c7884167-3670-45b6-b67d-b6f36de44d1a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Lock "75f4dc79-4b73-4ddc-be03-3653a5a0797e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 968.804437] env[63515]: DEBUG oslo_concurrency.lockutils [None req-c7884167-3670-45b6-b67d-b6f36de44d1a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Acquiring lock "75f4dc79-4b73-4ddc-be03-3653a5a0797e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 968.804593] env[63515]: DEBUG oslo_concurrency.lockutils [None req-c7884167-3670-45b6-b67d-b6f36de44d1a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Lock "75f4dc79-4b73-4ddc-be03-3653a5a0797e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 968.804809] env[63515]: DEBUG oslo_concurrency.lockutils [None req-c7884167-3670-45b6-b67d-b6f36de44d1a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Lock "75f4dc79-4b73-4ddc-be03-3653a5a0797e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 968.808516] env[63515]: INFO nova.compute.manager [None req-c7884167-3670-45b6-b67d-b6f36de44d1a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] [instance: 75f4dc79-4b73-4ddc-be03-3653a5a0797e] Terminating instance [ 968.818927] env[63515]: DEBUG nova.compute.manager [None req-c7884167-3670-45b6-b67d-b6f36de44d1a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] [instance: 75f4dc79-4b73-4ddc-be03-3653a5a0797e] Start destroying the instance on the hypervisor. {{(pid=63515) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 968.819078] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-c7884167-3670-45b6-b67d-b6f36de44d1a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] [instance: 75f4dc79-4b73-4ddc-be03-3653a5a0797e] Destroying instance {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 968.820794] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b16a3e57-2e4d-48b9-a754-9ea8a1fe2247 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.827697] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-c7884167-3670-45b6-b67d-b6f36de44d1a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] [instance: 75f4dc79-4b73-4ddc-be03-3653a5a0797e] Powering off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 968.830351] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f1ca5473-eb05-4dfb-99d9-886ba92bf28d {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.832143] env[63515]: DEBUG oslo_vmware.api [None req-c547db4d-5471-41ef-8a57-e1e58211e41e tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Task: {'id': task-1111698, 'name': ReconfigVM_Task, 'duration_secs': 0.314403} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 968.833110] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-c547db4d-5471-41ef-8a57-e1e58211e41e tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] [instance: 2a29f37d-e786-48d3-b126-4854d6e7c67c] Reconfigured VM instance instance-00000045 to detach disk 2001 {{(pid=63515) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 968.840274] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-99aa9604-6619-412b-80cf-5d7c7ff51961 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.853348] env[63515]: DEBUG oslo_vmware.api [None req-c7884167-3670-45b6-b67d-b6f36de44d1a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Waiting for the task: (returnval){ [ 968.853348] env[63515]: value = "task-1111701" [ 968.853348] env[63515]: _type = "Task" [ 968.853348] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 968.862902] env[63515]: DEBUG oslo_vmware.api [None req-c547db4d-5471-41ef-8a57-e1e58211e41e tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Waiting for the task: (returnval){ [ 968.862902] env[63515]: value = "task-1111702" [ 968.862902] env[63515]: _type = "Task" [ 968.862902] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 968.872542] env[63515]: DEBUG oslo_concurrency.lockutils [None req-59dbe5ed-eaca-4998-8a9a-682a5345ed51 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Acquiring lock "d438dd30-b816-48a6-94c6-11e6e9bd5b71" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 968.872542] env[63515]: DEBUG oslo_concurrency.lockutils [None req-59dbe5ed-eaca-4998-8a9a-682a5345ed51 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Lock "d438dd30-b816-48a6-94c6-11e6e9bd5b71" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 968.872542] env[63515]: DEBUG oslo_concurrency.lockutils [None req-59dbe5ed-eaca-4998-8a9a-682a5345ed51 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Acquiring lock "d438dd30-b816-48a6-94c6-11e6e9bd5b71-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 968.872542] env[63515]: DEBUG oslo_concurrency.lockutils [None req-59dbe5ed-eaca-4998-8a9a-682a5345ed51 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Lock "d438dd30-b816-48a6-94c6-11e6e9bd5b71-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 968.872542] env[63515]: DEBUG oslo_concurrency.lockutils [None req-59dbe5ed-eaca-4998-8a9a-682a5345ed51 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Lock "d438dd30-b816-48a6-94c6-11e6e9bd5b71-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 968.873073] env[63515]: INFO nova.compute.manager [None req-59dbe5ed-eaca-4998-8a9a-682a5345ed51 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] [instance: d438dd30-b816-48a6-94c6-11e6e9bd5b71] Terminating instance [ 968.877906] env[63515]: DEBUG oslo_vmware.api [None req-165bb947-b1a3-4f80-91f4-6990bed3b114 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Task: {'id': task-1111699, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.082784} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 968.881919] env[63515]: DEBUG nova.compute.manager [None req-59dbe5ed-eaca-4998-8a9a-682a5345ed51 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] [instance: d438dd30-b816-48a6-94c6-11e6e9bd5b71] Start destroying the instance on the hypervisor. {{(pid=63515) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 968.882163] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-59dbe5ed-eaca-4998-8a9a-682a5345ed51 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] [instance: d438dd30-b816-48a6-94c6-11e6e9bd5b71] Destroying instance {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 968.882547] env[63515]: DEBUG nova.compute.manager [None req-adce163d-60b0-4e35-be4e-020840038d00 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: 724b05bc-ce4b-4f99-ad14-f3abea5567f2] Starting instance... {{(pid=63515) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 968.885779] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-165bb947-b1a3-4f80-91f4-6990bed3b114 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: d024b6b2-dd10-4112-89e8-ced57efe8208] Extended root virtual disk {{(pid=63515) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 968.886281] env[63515]: DEBUG oslo_vmware.api [None req-c7884167-3670-45b6-b67d-b6f36de44d1a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Task: {'id': task-1111701, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.887042] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d7c24cd-ad23-4d4c-a32d-3bb1b56b214b {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.893375] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bbaf1394-5a96-4d1c-a52b-a6446a4b98fa {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.905413] env[63515]: DEBUG oslo_vmware.api [None req-c547db4d-5471-41ef-8a57-e1e58211e41e tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Task: {'id': task-1111702, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.909640] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-59dbe5ed-eaca-4998-8a9a-682a5345ed51 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] [instance: d438dd30-b816-48a6-94c6-11e6e9bd5b71] Powering off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 968.923244] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c5c0c106-f38e-4df8-a588-c48e3f104cc6 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.925250] env[63515]: DEBUG oslo_vmware.api [None req-16f0ea70-a973-4a3c-bc2e-292ffb4571ac tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Task: {'id': task-1111700, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.935877] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-165bb947-b1a3-4f80-91f4-6990bed3b114 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: d024b6b2-dd10-4112-89e8-ced57efe8208] Reconfiguring VM instance instance-0000005c to attach disk [datastore1] d024b6b2-dd10-4112-89e8-ced57efe8208/d024b6b2-dd10-4112-89e8-ced57efe8208.vmdk or device None with type sparse {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 968.940096] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-657cc009-76c9-4099-93b7-d35e079030f5 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.962436] env[63515]: DEBUG oslo_vmware.api [None req-59dbe5ed-eaca-4998-8a9a-682a5345ed51 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Waiting for the task: (returnval){ [ 968.962436] env[63515]: value = "task-1111703" [ 968.962436] env[63515]: _type = "Task" [ 968.962436] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 968.963024] env[63515]: DEBUG oslo_vmware.api [None req-165bb947-b1a3-4f80-91f4-6990bed3b114 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Waiting for the task: (returnval){ [ 968.963024] env[63515]: value = "task-1111704" [ 968.963024] env[63515]: _type = "Task" [ 968.963024] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 968.982243] env[63515]: DEBUG oslo_vmware.api [None req-165bb947-b1a3-4f80-91f4-6990bed3b114 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Task: {'id': task-1111704, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.982528] env[63515]: DEBUG oslo_vmware.api [None req-59dbe5ed-eaca-4998-8a9a-682a5345ed51 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Task: {'id': task-1111703, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.113529] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04e02037-f655-40b4-8001-515be514270c {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.124285] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6252b50-7885-4307-9d4f-ee4dc1793e89 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.165795] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aaf4d944-1976-45c4-80cc-8f8ccdb8ed36 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.175072] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0615d256-e76f-4dae-9184-5e719093c65c {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.194103] env[63515]: DEBUG nova.compute.provider_tree [None req-a5676663-6933-430d-aa5e-aeffa94052ee tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Updating inventory in ProviderTree for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 969.369053] env[63515]: DEBUG oslo_vmware.api [None req-c7884167-3670-45b6-b67d-b6f36de44d1a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Task: {'id': task-1111701, 'name': PowerOffVM_Task, 'duration_secs': 0.286783} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 969.372972] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-c7884167-3670-45b6-b67d-b6f36de44d1a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] [instance: 75f4dc79-4b73-4ddc-be03-3653a5a0797e] Powered off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 969.373204] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-c7884167-3670-45b6-b67d-b6f36de44d1a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] [instance: 75f4dc79-4b73-4ddc-be03-3653a5a0797e] Unregistering the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 969.373499] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f021d7b5-671b-40da-9fe0-569a0ae6299a {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.380921] env[63515]: DEBUG oslo_vmware.api [None req-c547db4d-5471-41ef-8a57-e1e58211e41e tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Task: {'id': task-1111702, 'name': ReconfigVM_Task, 'duration_secs': 0.201908} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 969.381370] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-c547db4d-5471-41ef-8a57-e1e58211e41e tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] [instance: 2a29f37d-e786-48d3-b126-4854d6e7c67c] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-243524', 'volume_id': 'dea169a3-b753-4313-853a-cbf3a4c05ef8', 'name': 'volume-dea169a3-b753-4313-853a-cbf3a4c05ef8', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '2a29f37d-e786-48d3-b126-4854d6e7c67c', 'attached_at': '', 'detached_at': '', 'volume_id': 'dea169a3-b753-4313-853a-cbf3a4c05ef8', 'serial': 'dea169a3-b753-4313-853a-cbf3a4c05ef8'} {{(pid=63515) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 969.401182] env[63515]: DEBUG oslo_vmware.api [None req-16f0ea70-a973-4a3c-bc2e-292ffb4571ac tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Task: {'id': task-1111700, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.854988} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 969.401313] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-16f0ea70-a973-4a3c-bc2e-292ffb4571ac tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk to [datastore1] 918d5b35-c46d-47c5-b59f-175831aed9d6/918d5b35-c46d-47c5-b59f-175831aed9d6.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 969.401615] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-16f0ea70-a973-4a3c-bc2e-292ffb4571ac tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: 918d5b35-c46d-47c5-b59f-175831aed9d6] Extending root virtual disk to 1048576 {{(pid=63515) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 969.401915] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-40497a87-d722-451a-9a72-19cecbe36373 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.404945] env[63515]: DEBUG oslo_concurrency.lockutils [None req-adce163d-60b0-4e35-be4e-020840038d00 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 969.409845] env[63515]: DEBUG oslo_vmware.api [None req-16f0ea70-a973-4a3c-bc2e-292ffb4571ac tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Waiting for the task: (returnval){ [ 969.409845] env[63515]: value = "task-1111706" [ 969.409845] env[63515]: _type = "Task" [ 969.409845] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 969.418748] env[63515]: DEBUG oslo_vmware.api [None req-16f0ea70-a973-4a3c-bc2e-292ffb4571ac tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Task: {'id': task-1111706, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.458868] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-c7884167-3670-45b6-b67d-b6f36de44d1a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] [instance: 75f4dc79-4b73-4ddc-be03-3653a5a0797e] Unregistered the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 969.459150] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-c7884167-3670-45b6-b67d-b6f36de44d1a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] [instance: 75f4dc79-4b73-4ddc-be03-3653a5a0797e] Deleting contents of the VM from datastore datastore2 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 969.459375] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-c7884167-3670-45b6-b67d-b6f36de44d1a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Deleting the datastore file [datastore2] 75f4dc79-4b73-4ddc-be03-3653a5a0797e {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 969.459694] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4fc7c8da-cb33-4f9c-8578-f3013724a517 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.468418] env[63515]: DEBUG oslo_vmware.api [None req-c7884167-3670-45b6-b67d-b6f36de44d1a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Waiting for the task: (returnval){ [ 969.468418] env[63515]: value = "task-1111707" [ 969.468418] env[63515]: _type = "Task" [ 969.468418] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 969.478744] env[63515]: DEBUG oslo_vmware.api [None req-59dbe5ed-eaca-4998-8a9a-682a5345ed51 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Task: {'id': task-1111703, 'name': PowerOffVM_Task, 'duration_secs': 0.313668} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 969.479515] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-59dbe5ed-eaca-4998-8a9a-682a5345ed51 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] [instance: d438dd30-b816-48a6-94c6-11e6e9bd5b71] Powered off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 969.479803] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-59dbe5ed-eaca-4998-8a9a-682a5345ed51 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] [instance: d438dd30-b816-48a6-94c6-11e6e9bd5b71] Unregistering the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 969.483068] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-08c67788-581c-44d2-acff-df4feecbd9d8 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.484752] env[63515]: DEBUG oslo_vmware.api [None req-165bb947-b1a3-4f80-91f4-6990bed3b114 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Task: {'id': task-1111704, 'name': ReconfigVM_Task, 'duration_secs': 0.508624} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 969.487905] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-165bb947-b1a3-4f80-91f4-6990bed3b114 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: d024b6b2-dd10-4112-89e8-ced57efe8208] Reconfigured VM instance instance-0000005c to attach disk [datastore1] d024b6b2-dd10-4112-89e8-ced57efe8208/d024b6b2-dd10-4112-89e8-ced57efe8208.vmdk or device None with type sparse {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 969.488722] env[63515]: DEBUG oslo_vmware.api [None req-c7884167-3670-45b6-b67d-b6f36de44d1a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Task: {'id': task-1111707, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.489356] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-bacb363d-5044-4f82-be04-c4765688fd52 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.496239] env[63515]: DEBUG oslo_vmware.api [None req-165bb947-b1a3-4f80-91f4-6990bed3b114 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Waiting for the task: (returnval){ [ 969.496239] env[63515]: value = "task-1111709" [ 969.496239] env[63515]: _type = "Task" [ 969.496239] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 969.505019] env[63515]: DEBUG oslo_vmware.api [None req-165bb947-b1a3-4f80-91f4-6990bed3b114 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Task: {'id': task-1111709, 'name': Rename_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.730282] env[63515]: DEBUG nova.scheduler.client.report [None req-a5676663-6933-430d-aa5e-aeffa94052ee tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Updated inventory for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a with generation 133 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 969.730470] env[63515]: DEBUG nova.compute.provider_tree [None req-a5676663-6933-430d-aa5e-aeffa94052ee tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Updating resource provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a generation from 133 to 134 during operation: update_inventory {{(pid=63515) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 969.730660] env[63515]: DEBUG nova.compute.provider_tree [None req-a5676663-6933-430d-aa5e-aeffa94052ee tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Updating inventory in ProviderTree for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 969.921554] env[63515]: DEBUG oslo_vmware.api [None req-16f0ea70-a973-4a3c-bc2e-292ffb4571ac tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Task: {'id': task-1111706, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.084232} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 969.921887] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-16f0ea70-a973-4a3c-bc2e-292ffb4571ac tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: 918d5b35-c46d-47c5-b59f-175831aed9d6] Extended root virtual disk {{(pid=63515) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 969.922831] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6baa7a66-e08e-455d-9d77-3af8d7d17dae {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.946886] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-16f0ea70-a973-4a3c-bc2e-292ffb4571ac tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: 918d5b35-c46d-47c5-b59f-175831aed9d6] Reconfiguring VM instance instance-00000059 to attach disk [datastore1] 918d5b35-c46d-47c5-b59f-175831aed9d6/918d5b35-c46d-47c5-b59f-175831aed9d6.vmdk or device None with type sparse {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 969.948489] env[63515]: DEBUG nova.objects.instance [None req-c547db4d-5471-41ef-8a57-e1e58211e41e tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Lazy-loading 'flavor' on Instance uuid 2a29f37d-e786-48d3-b126-4854d6e7c67c {{(pid=63515) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 969.949850] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bbea72fb-b905-46cf-add9-716460dc69b6 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.973750] env[63515]: DEBUG oslo_vmware.api [None req-16f0ea70-a973-4a3c-bc2e-292ffb4571ac tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Waiting for the task: (returnval){ [ 969.973750] env[63515]: value = "task-1111710" [ 969.973750] env[63515]: _type = "Task" [ 969.973750] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 969.985948] env[63515]: DEBUG oslo_vmware.api [None req-c7884167-3670-45b6-b67d-b6f36de44d1a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Task: {'id': task-1111707, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.170403} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 969.989162] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-c7884167-3670-45b6-b67d-b6f36de44d1a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Deleted the datastore file {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 969.989397] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-c7884167-3670-45b6-b67d-b6f36de44d1a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] [instance: 75f4dc79-4b73-4ddc-be03-3653a5a0797e] Deleted contents of the VM from datastore datastore2 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 969.989591] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-c7884167-3670-45b6-b67d-b6f36de44d1a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] [instance: 75f4dc79-4b73-4ddc-be03-3653a5a0797e] Instance destroyed {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 969.989775] env[63515]: INFO nova.compute.manager [None req-c7884167-3670-45b6-b67d-b6f36de44d1a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] [instance: 75f4dc79-4b73-4ddc-be03-3653a5a0797e] Took 1.17 seconds to destroy the instance on the hypervisor. [ 969.990048] env[63515]: DEBUG oslo.service.loopingcall [None req-c7884167-3670-45b6-b67d-b6f36de44d1a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 969.990540] env[63515]: DEBUG oslo_vmware.api [None req-16f0ea70-a973-4a3c-bc2e-292ffb4571ac tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Task: {'id': task-1111710, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.990775] env[63515]: DEBUG nova.compute.manager [-] [instance: 75f4dc79-4b73-4ddc-be03-3653a5a0797e] Deallocating network for instance {{(pid=63515) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 969.990875] env[63515]: DEBUG nova.network.neutron [-] [instance: 75f4dc79-4b73-4ddc-be03-3653a5a0797e] deallocate_for_instance() {{(pid=63515) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 970.006184] env[63515]: DEBUG oslo_vmware.api [None req-165bb947-b1a3-4f80-91f4-6990bed3b114 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Task: {'id': task-1111709, 'name': Rename_Task, 'duration_secs': 0.168427} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 970.007057] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-165bb947-b1a3-4f80-91f4-6990bed3b114 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: d024b6b2-dd10-4112-89e8-ced57efe8208] Powering on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 970.007057] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-20bee7be-9f7e-4a85-9e99-534775a5e237 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.012493] env[63515]: DEBUG oslo_vmware.api [None req-165bb947-b1a3-4f80-91f4-6990bed3b114 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Waiting for the task: (returnval){ [ 970.012493] env[63515]: value = "task-1111711" [ 970.012493] env[63515]: _type = "Task" [ 970.012493] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 970.020435] env[63515]: DEBUG oslo_vmware.api [None req-165bb947-b1a3-4f80-91f4-6990bed3b114 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Task: {'id': task-1111711, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.235414] env[63515]: DEBUG oslo_concurrency.lockutils [None req-a5676663-6933-430d-aa5e-aeffa94052ee tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.397s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 970.235936] env[63515]: DEBUG nova.compute.manager [None req-a5676663-6933-430d-aa5e-aeffa94052ee tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: 6a66dd9b-7b24-4b70-b58f-0830cf39376c] Start building networks asynchronously for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 970.238598] env[63515]: DEBUG oslo_concurrency.lockutils [None req-8ce4ce82-9ad9-4201-af2c-019ea542780e tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 7.915s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 970.238818] env[63515]: DEBUG oslo_concurrency.lockutils [None req-8ce4ce82-9ad9-4201-af2c-019ea542780e tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 970.241470] env[63515]: DEBUG oslo_concurrency.lockutils [None req-113af8fa-af69-464c-ac2e-e5cf96a3de06 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.753s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 970.242634] env[63515]: INFO nova.compute.claims [None req-113af8fa-af69-464c-ac2e-e5cf96a3de06 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: 61cf2f9c-5d35-46c4-a477-d2cef9fcd5ae] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 970.261878] env[63515]: DEBUG nova.compute.manager [req-8e3580e0-be33-4033-996d-1325d352465a req-61977bb9-906d-4298-a7cc-b9fa636c8115 service nova] [instance: 75f4dc79-4b73-4ddc-be03-3653a5a0797e] Received event network-vif-deleted-00eda8ad-1a52-4c47-b065-a82c08da7980 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 970.262706] env[63515]: INFO nova.compute.manager [req-8e3580e0-be33-4033-996d-1325d352465a req-61977bb9-906d-4298-a7cc-b9fa636c8115 service nova] [instance: 75f4dc79-4b73-4ddc-be03-3653a5a0797e] Neutron deleted interface 00eda8ad-1a52-4c47-b065-a82c08da7980; detaching it from the instance and deleting it from the info cache [ 970.262706] env[63515]: DEBUG nova.network.neutron [req-8e3580e0-be33-4033-996d-1325d352465a req-61977bb9-906d-4298-a7cc-b9fa636c8115 service nova] [instance: 75f4dc79-4b73-4ddc-be03-3653a5a0797e] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 970.267028] env[63515]: INFO nova.scheduler.client.report [None req-8ce4ce82-9ad9-4201-af2c-019ea542780e tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Deleted allocations for instance 829e72d7-a56a-451c-b98a-94253e5900d5 [ 970.292127] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-59dbe5ed-eaca-4998-8a9a-682a5345ed51 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] [instance: d438dd30-b816-48a6-94c6-11e6e9bd5b71] Unregistered the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 970.292439] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-59dbe5ed-eaca-4998-8a9a-682a5345ed51 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] [instance: d438dd30-b816-48a6-94c6-11e6e9bd5b71] Deleting contents of the VM from datastore datastore1 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 970.292581] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-59dbe5ed-eaca-4998-8a9a-682a5345ed51 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Deleting the datastore file [datastore1] d438dd30-b816-48a6-94c6-11e6e9bd5b71 {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 970.292934] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-da215963-dc9e-4e74-ab5c-2232360fae9f {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.300630] env[63515]: DEBUG oslo_vmware.api [None req-59dbe5ed-eaca-4998-8a9a-682a5345ed51 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Waiting for the task: (returnval){ [ 970.300630] env[63515]: value = "task-1111712" [ 970.300630] env[63515]: _type = "Task" [ 970.300630] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 970.312593] env[63515]: DEBUG oslo_vmware.api [None req-59dbe5ed-eaca-4998-8a9a-682a5345ed51 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Task: {'id': task-1111712, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.469816] env[63515]: DEBUG oslo_concurrency.lockutils [None req-c547db4d-5471-41ef-8a57-e1e58211e41e tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Lock "2a29f37d-e786-48d3-b126-4854d6e7c67c" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 2.839s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 970.489996] env[63515]: DEBUG oslo_vmware.api [None req-16f0ea70-a973-4a3c-bc2e-292ffb4571ac tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Task: {'id': task-1111710, 'name': ReconfigVM_Task, 'duration_secs': 0.329811} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 970.490325] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-16f0ea70-a973-4a3c-bc2e-292ffb4571ac tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: 918d5b35-c46d-47c5-b59f-175831aed9d6] Reconfigured VM instance instance-00000059 to attach disk [datastore1] 918d5b35-c46d-47c5-b59f-175831aed9d6/918d5b35-c46d-47c5-b59f-175831aed9d6.vmdk or device None with type sparse {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 970.490942] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-0a15d928-d992-43d8-841f-b1f59607539f {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.497903] env[63515]: DEBUG oslo_vmware.api [None req-16f0ea70-a973-4a3c-bc2e-292ffb4571ac tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Waiting for the task: (returnval){ [ 970.497903] env[63515]: value = "task-1111713" [ 970.497903] env[63515]: _type = "Task" [ 970.497903] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 970.513811] env[63515]: DEBUG oslo_vmware.api [None req-16f0ea70-a973-4a3c-bc2e-292ffb4571ac tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Task: {'id': task-1111713, 'name': Rename_Task} progress is 6%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.522794] env[63515]: DEBUG oslo_vmware.api [None req-165bb947-b1a3-4f80-91f4-6990bed3b114 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Task: {'id': task-1111711, 'name': PowerOnVM_Task} progress is 100%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.716963] env[63515]: DEBUG nova.network.neutron [-] [instance: 75f4dc79-4b73-4ddc-be03-3653a5a0797e] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 970.746826] env[63515]: DEBUG nova.compute.utils [None req-a5676663-6933-430d-aa5e-aeffa94052ee tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Using /dev/sd instead of None {{(pid=63515) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 970.750202] env[63515]: DEBUG nova.compute.manager [None req-a5676663-6933-430d-aa5e-aeffa94052ee tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: 6a66dd9b-7b24-4b70-b58f-0830cf39376c] Allocating IP information in the background. {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 970.750394] env[63515]: DEBUG nova.network.neutron [None req-a5676663-6933-430d-aa5e-aeffa94052ee tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: 6a66dd9b-7b24-4b70-b58f-0830cf39376c] allocate_for_instance() {{(pid=63515) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 970.766107] env[63515]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-8837f787-e52f-405c-bbdb-b0d8ae5c8408 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.777777] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3efc636e-ef1b-4aa0-bbcf-d7854b59ebc1 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.789237] env[63515]: DEBUG oslo_concurrency.lockutils [None req-8ce4ce82-9ad9-4201-af2c-019ea542780e tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Lock "829e72d7-a56a-451c-b98a-94253e5900d5" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 11.546s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 970.802311] env[63515]: DEBUG nova.policy [None req-a5676663-6933-430d-aa5e-aeffa94052ee tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b5de3e36ae88482eb795894592e76c7c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f9c98e9f6020475490aaa7e76d907ab7', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63515) authorize /opt/stack/nova/nova/policy.py:201}} [ 970.817842] env[63515]: DEBUG nova.compute.manager [req-8e3580e0-be33-4033-996d-1325d352465a req-61977bb9-906d-4298-a7cc-b9fa636c8115 service nova] [instance: 75f4dc79-4b73-4ddc-be03-3653a5a0797e] Detach interface failed, port_id=00eda8ad-1a52-4c47-b065-a82c08da7980, reason: Instance 75f4dc79-4b73-4ddc-be03-3653a5a0797e could not be found. {{(pid=63515) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 970.822891] env[63515]: DEBUG oslo_vmware.api [None req-59dbe5ed-eaca-4998-8a9a-682a5345ed51 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Task: {'id': task-1111712, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.275254} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 970.823172] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-59dbe5ed-eaca-4998-8a9a-682a5345ed51 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Deleted the datastore file {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 970.823913] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-59dbe5ed-eaca-4998-8a9a-682a5345ed51 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] [instance: d438dd30-b816-48a6-94c6-11e6e9bd5b71] Deleted contents of the VM from datastore datastore1 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 970.824153] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-59dbe5ed-eaca-4998-8a9a-682a5345ed51 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] [instance: d438dd30-b816-48a6-94c6-11e6e9bd5b71] Instance destroyed {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 970.824361] env[63515]: INFO nova.compute.manager [None req-59dbe5ed-eaca-4998-8a9a-682a5345ed51 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] [instance: d438dd30-b816-48a6-94c6-11e6e9bd5b71] Took 1.94 seconds to destroy the instance on the hypervisor. [ 970.824633] env[63515]: DEBUG oslo.service.loopingcall [None req-59dbe5ed-eaca-4998-8a9a-682a5345ed51 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 970.825100] env[63515]: DEBUG nova.compute.manager [-] [instance: d438dd30-b816-48a6-94c6-11e6e9bd5b71] Deallocating network for instance {{(pid=63515) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 970.825208] env[63515]: DEBUG nova.network.neutron [-] [instance: d438dd30-b816-48a6-94c6-11e6e9bd5b71] deallocate_for_instance() {{(pid=63515) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 970.962771] env[63515]: DEBUG oslo_concurrency.lockutils [None req-e595b7dd-52b8-43f9-9fcf-6cb888cef16d tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Acquiring lock "2a29f37d-e786-48d3-b126-4854d6e7c67c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 970.968973] env[63515]: DEBUG oslo_concurrency.lockutils [None req-e595b7dd-52b8-43f9-9fcf-6cb888cef16d tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Lock "2a29f37d-e786-48d3-b126-4854d6e7c67c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.004s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 970.968973] env[63515]: DEBUG oslo_concurrency.lockutils [None req-e595b7dd-52b8-43f9-9fcf-6cb888cef16d tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Acquiring lock "2a29f37d-e786-48d3-b126-4854d6e7c67c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 970.968973] env[63515]: DEBUG oslo_concurrency.lockutils [None req-e595b7dd-52b8-43f9-9fcf-6cb888cef16d tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Lock "2a29f37d-e786-48d3-b126-4854d6e7c67c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 970.968973] env[63515]: DEBUG oslo_concurrency.lockutils [None req-e595b7dd-52b8-43f9-9fcf-6cb888cef16d tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Lock "2a29f37d-e786-48d3-b126-4854d6e7c67c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 970.970362] env[63515]: INFO nova.compute.manager [None req-e595b7dd-52b8-43f9-9fcf-6cb888cef16d tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] [instance: 2a29f37d-e786-48d3-b126-4854d6e7c67c] Terminating instance [ 970.973638] env[63515]: DEBUG nova.compute.manager [None req-e595b7dd-52b8-43f9-9fcf-6cb888cef16d tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] [instance: 2a29f37d-e786-48d3-b126-4854d6e7c67c] Start destroying the instance on the hypervisor. {{(pid=63515) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 970.974021] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-e595b7dd-52b8-43f9-9fcf-6cb888cef16d tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] [instance: 2a29f37d-e786-48d3-b126-4854d6e7c67c] Destroying instance {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 970.975318] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d744ce4b-1abd-4cc3-9526-887e3efd95a1 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.986441] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-e595b7dd-52b8-43f9-9fcf-6cb888cef16d tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] [instance: 2a29f37d-e786-48d3-b126-4854d6e7c67c] Powering off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 970.986441] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5e79c386-f5b4-4018-8043-4745cf10cf2b {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.995538] env[63515]: DEBUG oslo_vmware.api [None req-e595b7dd-52b8-43f9-9fcf-6cb888cef16d tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Waiting for the task: (returnval){ [ 970.995538] env[63515]: value = "task-1111714" [ 970.995538] env[63515]: _type = "Task" [ 970.995538] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 971.013314] env[63515]: DEBUG oslo_vmware.api [None req-e595b7dd-52b8-43f9-9fcf-6cb888cef16d tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Task: {'id': task-1111714, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.017778] env[63515]: DEBUG oslo_vmware.api [None req-16f0ea70-a973-4a3c-bc2e-292ffb4571ac tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Task: {'id': task-1111713, 'name': Rename_Task, 'duration_secs': 0.185658} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 971.022077] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-16f0ea70-a973-4a3c-bc2e-292ffb4571ac tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: 918d5b35-c46d-47c5-b59f-175831aed9d6] Powering on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 971.024700] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2c9dd71e-087c-4498-8f54-f0ec1e996cb2 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.036812] env[63515]: DEBUG oslo_vmware.api [None req-165bb947-b1a3-4f80-91f4-6990bed3b114 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Task: {'id': task-1111711, 'name': PowerOnVM_Task, 'duration_secs': 0.517232} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 971.037171] env[63515]: DEBUG oslo_vmware.api [None req-16f0ea70-a973-4a3c-bc2e-292ffb4571ac tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Waiting for the task: (returnval){ [ 971.037171] env[63515]: value = "task-1111715" [ 971.037171] env[63515]: _type = "Task" [ 971.037171] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 971.037432] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-165bb947-b1a3-4f80-91f4-6990bed3b114 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: d024b6b2-dd10-4112-89e8-ced57efe8208] Powered on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 971.037636] env[63515]: INFO nova.compute.manager [None req-165bb947-b1a3-4f80-91f4-6990bed3b114 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: d024b6b2-dd10-4112-89e8-ced57efe8208] Took 9.27 seconds to spawn the instance on the hypervisor. [ 971.037846] env[63515]: DEBUG nova.compute.manager [None req-165bb947-b1a3-4f80-91f4-6990bed3b114 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: d024b6b2-dd10-4112-89e8-ced57efe8208] Checking state {{(pid=63515) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 971.038867] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08ae568e-3879-461e-9c88-49203581bf61 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.052461] env[63515]: DEBUG oslo_vmware.api [None req-16f0ea70-a973-4a3c-bc2e-292ffb4571ac tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Task: {'id': task-1111715, 'name': PowerOnVM_Task} progress is 33%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.220037] env[63515]: INFO nova.compute.manager [-] [instance: 75f4dc79-4b73-4ddc-be03-3653a5a0797e] Took 1.23 seconds to deallocate network for instance. [ 971.261982] env[63515]: DEBUG nova.compute.manager [None req-a5676663-6933-430d-aa5e-aeffa94052ee tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: 6a66dd9b-7b24-4b70-b58f-0830cf39376c] Start building block device mappings for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 971.274953] env[63515]: DEBUG nova.network.neutron [None req-a5676663-6933-430d-aa5e-aeffa94052ee tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: 6a66dd9b-7b24-4b70-b58f-0830cf39376c] Successfully created port: b190896f-9020-45f3-bb87-f1fe6401b017 {{(pid=63515) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 971.477019] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66934075-7209-46f7-b82f-3d3b9faf53fc {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.483897] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7cdca5df-a6a4-4f8c-b04b-f2de59502ffb {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.529427] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd6677dc-6276-45d8-8cc0-b5341c755986 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.541693] env[63515]: DEBUG oslo_vmware.api [None req-e595b7dd-52b8-43f9-9fcf-6cb888cef16d tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Task: {'id': task-1111714, 'name': PowerOffVM_Task, 'duration_secs': 0.217168} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 971.549178] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-e595b7dd-52b8-43f9-9fcf-6cb888cef16d tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] [instance: 2a29f37d-e786-48d3-b126-4854d6e7c67c] Powered off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 971.549430] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-e595b7dd-52b8-43f9-9fcf-6cb888cef16d tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] [instance: 2a29f37d-e786-48d3-b126-4854d6e7c67c] Unregistering the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 971.549758] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3a839b47-3359-4b08-9862-3b8bb58d19d8 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.552522] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dbee383a-b104-4424-8fb9-e72b15199487 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.568966] env[63515]: DEBUG oslo_vmware.api [None req-16f0ea70-a973-4a3c-bc2e-292ffb4571ac tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Task: {'id': task-1111715, 'name': PowerOnVM_Task, 'duration_secs': 0.511148} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 971.571102] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-16f0ea70-a973-4a3c-bc2e-292ffb4571ac tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: 918d5b35-c46d-47c5-b59f-175831aed9d6] Powered on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 971.571475] env[63515]: DEBUG nova.compute.manager [None req-16f0ea70-a973-4a3c-bc2e-292ffb4571ac tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: 918d5b35-c46d-47c5-b59f-175831aed9d6] Checking state {{(pid=63515) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 971.572501] env[63515]: INFO nova.compute.manager [None req-165bb947-b1a3-4f80-91f4-6990bed3b114 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: d024b6b2-dd10-4112-89e8-ced57efe8208] Took 24.06 seconds to build instance. [ 971.582525] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d18461ba-9ed5-4c86-a7b7-fdb1a46365e8 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.587648] env[63515]: DEBUG nova.compute.provider_tree [None req-113af8fa-af69-464c-ac2e-e5cf96a3de06 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Updating inventory in ProviderTree for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 971.589741] env[63515]: DEBUG oslo_concurrency.lockutils [None req-165bb947-b1a3-4f80-91f4-6990bed3b114 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Lock "d024b6b2-dd10-4112-89e8-ced57efe8208" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 25.091s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 971.713984] env[63515]: DEBUG nova.network.neutron [-] [instance: d438dd30-b816-48a6-94c6-11e6e9bd5b71] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 971.727562] env[63515]: DEBUG oslo_concurrency.lockutils [None req-c7884167-3670-45b6-b67d-b6f36de44d1a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 971.815928] env[63515]: DEBUG oslo_vmware.rw_handles [None req-57194bc7-9fdc-4bda-bd6d-4b604b59fffa tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52cbe900-84a3-8fda-b5af-75b7f0e7d6a0/disk-0.vmdk. {{(pid=63515) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 971.817139] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e5a1bab-6e6f-4ae4-8ede-9d3821fb73ec {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.823181] env[63515]: DEBUG oslo_vmware.rw_handles [None req-57194bc7-9fdc-4bda-bd6d-4b604b59fffa tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52cbe900-84a3-8fda-b5af-75b7f0e7d6a0/disk-0.vmdk is in state: ready. {{(pid=63515) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 971.823363] env[63515]: ERROR oslo_vmware.rw_handles [None req-57194bc7-9fdc-4bda-bd6d-4b604b59fffa tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52cbe900-84a3-8fda-b5af-75b7f0e7d6a0/disk-0.vmdk due to incomplete transfer. [ 971.823587] env[63515]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-5e20c277-0a2b-4d83-acd0-38c3f922dab6 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.831517] env[63515]: DEBUG oslo_vmware.rw_handles [None req-57194bc7-9fdc-4bda-bd6d-4b604b59fffa tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52cbe900-84a3-8fda-b5af-75b7f0e7d6a0/disk-0.vmdk. {{(pid=63515) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 971.832377] env[63515]: DEBUG nova.virt.vmwareapi.images [None req-57194bc7-9fdc-4bda-bd6d-4b604b59fffa tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 68766cd4-84be-475b-8494-d7ab43a9e969] Uploaded image fc46b835-451b-4001-9747-41c9d5d93184 to the Glance image server {{(pid=63515) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 971.833977] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-57194bc7-9fdc-4bda-bd6d-4b604b59fffa tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 68766cd4-84be-475b-8494-d7ab43a9e969] Destroying the VM {{(pid=63515) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 971.834219] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-35bce39f-a0b7-402b-9dcb-9e5ba7bd7764 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.839148] env[63515]: DEBUG oslo_vmware.api [None req-57194bc7-9fdc-4bda-bd6d-4b604b59fffa tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Waiting for the task: (returnval){ [ 971.839148] env[63515]: value = "task-1111718" [ 971.839148] env[63515]: _type = "Task" [ 971.839148] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 971.847327] env[63515]: DEBUG oslo_vmware.api [None req-57194bc7-9fdc-4bda-bd6d-4b604b59fffa tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Task: {'id': task-1111718, 'name': Destroy_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.097867] env[63515]: INFO nova.compute.manager [None req-16f0ea70-a973-4a3c-bc2e-292ffb4571ac tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: 918d5b35-c46d-47c5-b59f-175831aed9d6] bringing vm to original state: 'stopped' [ 972.126289] env[63515]: DEBUG nova.scheduler.client.report [None req-113af8fa-af69-464c-ac2e-e5cf96a3de06 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Updated inventory for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a with generation 134 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 972.126707] env[63515]: DEBUG nova.compute.provider_tree [None req-113af8fa-af69-464c-ac2e-e5cf96a3de06 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Updating resource provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a generation from 134 to 135 during operation: update_inventory {{(pid=63515) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 972.127141] env[63515]: DEBUG nova.compute.provider_tree [None req-113af8fa-af69-464c-ac2e-e5cf96a3de06 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Updating inventory in ProviderTree for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 972.153567] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-e595b7dd-52b8-43f9-9fcf-6cb888cef16d tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] [instance: 2a29f37d-e786-48d3-b126-4854d6e7c67c] Unregistered the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 972.154199] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-e595b7dd-52b8-43f9-9fcf-6cb888cef16d tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] [instance: 2a29f37d-e786-48d3-b126-4854d6e7c67c] Deleting contents of the VM from datastore datastore1 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 972.154199] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-e595b7dd-52b8-43f9-9fcf-6cb888cef16d tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Deleting the datastore file [datastore1] 2a29f37d-e786-48d3-b126-4854d6e7c67c {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 972.154570] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ba8afaf0-9342-443e-b39b-53c47b6b71f3 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.162051] env[63515]: DEBUG oslo_vmware.api [None req-e595b7dd-52b8-43f9-9fcf-6cb888cef16d tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Waiting for the task: (returnval){ [ 972.162051] env[63515]: value = "task-1111719" [ 972.162051] env[63515]: _type = "Task" [ 972.162051] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 972.170354] env[63515]: DEBUG oslo_vmware.api [None req-e595b7dd-52b8-43f9-9fcf-6cb888cef16d tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Task: {'id': task-1111719, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.216978] env[63515]: INFO nova.compute.manager [-] [instance: d438dd30-b816-48a6-94c6-11e6e9bd5b71] Took 1.39 seconds to deallocate network for instance. [ 972.275675] env[63515]: DEBUG nova.compute.manager [None req-a5676663-6933-430d-aa5e-aeffa94052ee tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: 6a66dd9b-7b24-4b70-b58f-0830cf39376c] Start spawning the instance on the hypervisor. {{(pid=63515) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 972.289828] env[63515]: DEBUG nova.compute.manager [req-7fd26f2f-f8c8-4530-82ff-cd758d5a494c req-37bb3212-06d8-4a97-a287-6e7dc9882d85 service nova] [instance: d438dd30-b816-48a6-94c6-11e6e9bd5b71] Received event network-vif-deleted-a673b83a-c9d4-425b-b6e3-aad0d17243ba {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 972.300788] env[63515]: DEBUG nova.virt.hardware [None req-a5676663-6933-430d-aa5e-aeffa94052ee tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T02:52:14Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T02:51:56Z,direct_url=,disk_format='vmdk',id=8a120570-cb06-4099-b262-554ca0ad15c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b569255cc43e42e7ae3f2b4ad37c6ef4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T02:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 972.301120] env[63515]: DEBUG nova.virt.hardware [None req-a5676663-6933-430d-aa5e-aeffa94052ee tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Flavor limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 972.301307] env[63515]: DEBUG nova.virt.hardware [None req-a5676663-6933-430d-aa5e-aeffa94052ee tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Image limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 972.301501] env[63515]: DEBUG nova.virt.hardware [None req-a5676663-6933-430d-aa5e-aeffa94052ee tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Flavor pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 972.301657] env[63515]: DEBUG nova.virt.hardware [None req-a5676663-6933-430d-aa5e-aeffa94052ee tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Image pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 972.301936] env[63515]: DEBUG nova.virt.hardware [None req-a5676663-6933-430d-aa5e-aeffa94052ee tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 972.302035] env[63515]: DEBUG nova.virt.hardware [None req-a5676663-6933-430d-aa5e-aeffa94052ee tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 972.302195] env[63515]: DEBUG nova.virt.hardware [None req-a5676663-6933-430d-aa5e-aeffa94052ee tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 972.302367] env[63515]: DEBUG nova.virt.hardware [None req-a5676663-6933-430d-aa5e-aeffa94052ee tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Got 1 possible topologies {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 972.302535] env[63515]: DEBUG nova.virt.hardware [None req-a5676663-6933-430d-aa5e-aeffa94052ee tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 972.302744] env[63515]: DEBUG nova.virt.hardware [None req-a5676663-6933-430d-aa5e-aeffa94052ee tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 972.303657] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b92d2bb-48bc-4b84-bcbd-e47c9c372139 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.313747] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4bffc7dc-2897-4d8d-aab3-94122497f42f {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.348665] env[63515]: DEBUG oslo_vmware.api [None req-57194bc7-9fdc-4bda-bd6d-4b604b59fffa tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Task: {'id': task-1111718, 'name': Destroy_Task, 'duration_secs': 0.344567} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 972.348964] env[63515]: INFO nova.virt.vmwareapi.vm_util [None req-57194bc7-9fdc-4bda-bd6d-4b604b59fffa tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 68766cd4-84be-475b-8494-d7ab43a9e969] Destroyed the VM [ 972.349089] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-57194bc7-9fdc-4bda-bd6d-4b604b59fffa tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 68766cd4-84be-475b-8494-d7ab43a9e969] Deleting Snapshot of the VM instance {{(pid=63515) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 972.349333] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-61668097-db37-4d18-ab35-fc1d933d4129 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.355742] env[63515]: DEBUG oslo_vmware.api [None req-57194bc7-9fdc-4bda-bd6d-4b604b59fffa tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Waiting for the task: (returnval){ [ 972.355742] env[63515]: value = "task-1111720" [ 972.355742] env[63515]: _type = "Task" [ 972.355742] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 972.363722] env[63515]: DEBUG oslo_vmware.api [None req-57194bc7-9fdc-4bda-bd6d-4b604b59fffa tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Task: {'id': task-1111720, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.634302] env[63515]: DEBUG oslo_concurrency.lockutils [None req-113af8fa-af69-464c-ac2e-e5cf96a3de06 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.393s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 972.634869] env[63515]: DEBUG nova.compute.manager [None req-113af8fa-af69-464c-ac2e-e5cf96a3de06 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: 61cf2f9c-5d35-46c4-a477-d2cef9fcd5ae] Start building networks asynchronously for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 972.638138] env[63515]: DEBUG oslo_concurrency.lockutils [None req-adce163d-60b0-4e35-be4e-020840038d00 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.233s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 972.640295] env[63515]: INFO nova.compute.claims [None req-adce163d-60b0-4e35-be4e-020840038d00 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: 724b05bc-ce4b-4f99-ad14-f3abea5567f2] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 972.673914] env[63515]: DEBUG oslo_vmware.api [None req-e595b7dd-52b8-43f9-9fcf-6cb888cef16d tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Task: {'id': task-1111719, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.420186} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 972.674134] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-e595b7dd-52b8-43f9-9fcf-6cb888cef16d tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Deleted the datastore file {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 972.674256] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-e595b7dd-52b8-43f9-9fcf-6cb888cef16d tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] [instance: 2a29f37d-e786-48d3-b126-4854d6e7c67c] Deleted contents of the VM from datastore datastore1 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 972.674513] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-e595b7dd-52b8-43f9-9fcf-6cb888cef16d tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] [instance: 2a29f37d-e786-48d3-b126-4854d6e7c67c] Instance destroyed {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 972.674750] env[63515]: INFO nova.compute.manager [None req-e595b7dd-52b8-43f9-9fcf-6cb888cef16d tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] [instance: 2a29f37d-e786-48d3-b126-4854d6e7c67c] Took 1.70 seconds to destroy the instance on the hypervisor. [ 972.675117] env[63515]: DEBUG oslo.service.loopingcall [None req-e595b7dd-52b8-43f9-9fcf-6cb888cef16d tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 972.675373] env[63515]: DEBUG nova.compute.manager [-] [instance: 2a29f37d-e786-48d3-b126-4854d6e7c67c] Deallocating network for instance {{(pid=63515) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 972.675476] env[63515]: DEBUG nova.network.neutron [-] [instance: 2a29f37d-e786-48d3-b126-4854d6e7c67c] deallocate_for_instance() {{(pid=63515) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 972.723073] env[63515]: DEBUG oslo_concurrency.lockutils [None req-59dbe5ed-eaca-4998-8a9a-682a5345ed51 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 972.865491] env[63515]: DEBUG oslo_vmware.api [None req-57194bc7-9fdc-4bda-bd6d-4b604b59fffa tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Task: {'id': task-1111720, 'name': RemoveSnapshot_Task, 'duration_secs': 0.495562} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 972.865770] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-57194bc7-9fdc-4bda-bd6d-4b604b59fffa tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 68766cd4-84be-475b-8494-d7ab43a9e969] Deleted Snapshot of the VM instance {{(pid=63515) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 972.866009] env[63515]: INFO nova.compute.manager [None req-57194bc7-9fdc-4bda-bd6d-4b604b59fffa tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 68766cd4-84be-475b-8494-d7ab43a9e969] Took 13.48 seconds to snapshot the instance on the hypervisor. [ 972.872944] env[63515]: DEBUG oslo_concurrency.lockutils [None req-978a15da-7dcc-413b-825d-e9c53aaf3163 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Acquiring lock "461a0c4c-dfdb-4669-ac72-8913c0277298" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 972.873186] env[63515]: DEBUG oslo_concurrency.lockutils [None req-978a15da-7dcc-413b-825d-e9c53aaf3163 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Lock "461a0c4c-dfdb-4669-ac72-8913c0277298" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 973.106080] env[63515]: DEBUG oslo_concurrency.lockutils [None req-16f0ea70-a973-4a3c-bc2e-292ffb4571ac tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Acquiring lock "918d5b35-c46d-47c5-b59f-175831aed9d6" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 973.106400] env[63515]: DEBUG oslo_concurrency.lockutils [None req-16f0ea70-a973-4a3c-bc2e-292ffb4571ac tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Lock "918d5b35-c46d-47c5-b59f-175831aed9d6" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 973.106868] env[63515]: DEBUG nova.compute.manager [None req-16f0ea70-a973-4a3c-bc2e-292ffb4571ac tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: 918d5b35-c46d-47c5-b59f-175831aed9d6] Checking state {{(pid=63515) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 973.108077] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a61743b8-b073-46a3-b9fa-81b1f0a4b369 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.115063] env[63515]: DEBUG nova.compute.manager [None req-16f0ea70-a973-4a3c-bc2e-292ffb4571ac tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: 918d5b35-c46d-47c5-b59f-175831aed9d6] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=63515) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 973.121016] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-16f0ea70-a973-4a3c-bc2e-292ffb4571ac tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: 918d5b35-c46d-47c5-b59f-175831aed9d6] Powering off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 973.121016] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1c2151fd-6d6c-4559-bb3f-c50e5ace7cd4 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.124553] env[63515]: DEBUG oslo_vmware.api [None req-16f0ea70-a973-4a3c-bc2e-292ffb4571ac tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Waiting for the task: (returnval){ [ 973.124553] env[63515]: value = "task-1111721" [ 973.124553] env[63515]: _type = "Task" [ 973.124553] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 973.133544] env[63515]: DEBUG oslo_vmware.api [None req-16f0ea70-a973-4a3c-bc2e-292ffb4571ac tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Task: {'id': task-1111721, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.144152] env[63515]: DEBUG nova.compute.utils [None req-113af8fa-af69-464c-ac2e-e5cf96a3de06 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Using /dev/sd instead of None {{(pid=63515) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 973.148268] env[63515]: DEBUG nova.network.neutron [None req-a5676663-6933-430d-aa5e-aeffa94052ee tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: 6a66dd9b-7b24-4b70-b58f-0830cf39376c] Successfully updated port: b190896f-9020-45f3-bb87-f1fe6401b017 {{(pid=63515) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 973.149374] env[63515]: DEBUG nova.compute.manager [None req-113af8fa-af69-464c-ac2e-e5cf96a3de06 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: 61cf2f9c-5d35-46c4-a477-d2cef9fcd5ae] Allocating IP information in the background. {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 973.149550] env[63515]: DEBUG nova.network.neutron [None req-113af8fa-af69-464c-ac2e-e5cf96a3de06 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: 61cf2f9c-5d35-46c4-a477-d2cef9fcd5ae] allocate_for_instance() {{(pid=63515) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 973.228357] env[63515]: DEBUG nova.policy [None req-113af8fa-af69-464c-ac2e-e5cf96a3de06 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f58ac812970845dc825202b842feb5d0', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ea099ce07bfb4a8da014d9303cf552eb', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63515) authorize /opt/stack/nova/nova/policy.py:201}} [ 973.375349] env[63515]: DEBUG nova.compute.manager [None req-978a15da-7dcc-413b-825d-e9c53aaf3163 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: 461a0c4c-dfdb-4669-ac72-8913c0277298] Starting instance... {{(pid=63515) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 973.434848] env[63515]: DEBUG nova.compute.manager [None req-57194bc7-9fdc-4bda-bd6d-4b604b59fffa tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 68766cd4-84be-475b-8494-d7ab43a9e969] Found 3 images (rotation: 2) {{(pid=63515) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4554}} [ 973.435074] env[63515]: DEBUG nova.compute.manager [None req-57194bc7-9fdc-4bda-bd6d-4b604b59fffa tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 68766cd4-84be-475b-8494-d7ab43a9e969] Rotating out 1 backups {{(pid=63515) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4562}} [ 973.435264] env[63515]: DEBUG nova.compute.manager [None req-57194bc7-9fdc-4bda-bd6d-4b604b59fffa tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 68766cd4-84be-475b-8494-d7ab43a9e969] Deleting image ca8a2595-00a8-4a1d-8bdc-e37e8a7f93e5 {{(pid=63515) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4567}} [ 973.630083] env[63515]: DEBUG nova.network.neutron [None req-113af8fa-af69-464c-ac2e-e5cf96a3de06 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: 61cf2f9c-5d35-46c4-a477-d2cef9fcd5ae] Successfully created port: 99e47951-e384-4aa7-a394-3d0a6bade8d7 {{(pid=63515) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 973.637213] env[63515]: DEBUG oslo_vmware.api [None req-16f0ea70-a973-4a3c-bc2e-292ffb4571ac tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Task: {'id': task-1111721, 'name': PowerOffVM_Task, 'duration_secs': 0.21772} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 973.637487] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-16f0ea70-a973-4a3c-bc2e-292ffb4571ac tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: 918d5b35-c46d-47c5-b59f-175831aed9d6] Powered off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 973.637721] env[63515]: DEBUG nova.compute.manager [None req-16f0ea70-a973-4a3c-bc2e-292ffb4571ac tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: 918d5b35-c46d-47c5-b59f-175831aed9d6] Checking state {{(pid=63515) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 973.638534] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8c7006c-0255-4e4f-b7e9-e892396e3e7a {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.648649] env[63515]: DEBUG nova.compute.manager [None req-113af8fa-af69-464c-ac2e-e5cf96a3de06 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: 61cf2f9c-5d35-46c4-a477-d2cef9fcd5ae] Start building block device mappings for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 973.655079] env[63515]: DEBUG oslo_concurrency.lockutils [None req-a5676663-6933-430d-aa5e-aeffa94052ee tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Acquiring lock "refresh_cache-6a66dd9b-7b24-4b70-b58f-0830cf39376c" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 973.655203] env[63515]: DEBUG oslo_concurrency.lockutils [None req-a5676663-6933-430d-aa5e-aeffa94052ee tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Acquired lock "refresh_cache-6a66dd9b-7b24-4b70-b58f-0830cf39376c" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 973.655368] env[63515]: DEBUG nova.network.neutron [None req-a5676663-6933-430d-aa5e-aeffa94052ee tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: 6a66dd9b-7b24-4b70-b58f-0830cf39376c] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 973.898762] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71691787-8e7a-4d0f-9dca-f9241a72af1f {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.903111] env[63515]: DEBUG oslo_concurrency.lockutils [None req-978a15da-7dcc-413b-825d-e9c53aaf3163 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 973.908982] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c5843f3-806d-4bcc-b5c6-76706ffc2fd1 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.938498] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c67f0f29-bf63-4a6a-afa9-d12f96cb9df4 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.946158] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f04b06a7-53d1-43d5-8952-e3637caf8473 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.960381] env[63515]: DEBUG nova.network.neutron [-] [instance: 2a29f37d-e786-48d3-b126-4854d6e7c67c] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 973.961966] env[63515]: DEBUG nova.compute.provider_tree [None req-adce163d-60b0-4e35-be4e-020840038d00 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 974.054226] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-60a0611f-4742-4524-a115-5b0171ec85a1 tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] [instance: 86b1ddcf-585c-41d9-84fe-17b794f1abbc] Volume attach. Driver type: vmdk {{(pid=63515) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 974.054492] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-60a0611f-4742-4524-a115-5b0171ec85a1 tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] [instance: 86b1ddcf-585c-41d9-84fe-17b794f1abbc] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-243550', 'volume_id': 'faf85999-e3c6-4036-a18e-7d126676fad9', 'name': 'volume-faf85999-e3c6-4036-a18e-7d126676fad9', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '86b1ddcf-585c-41d9-84fe-17b794f1abbc', 'attached_at': '', 'detached_at': '', 'volume_id': 'faf85999-e3c6-4036-a18e-7d126676fad9', 'serial': 'faf85999-e3c6-4036-a18e-7d126676fad9'} {{(pid=63515) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 974.055398] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2a1f279-6796-445a-acb3-536b0d982631 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.072701] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57f6c223-b359-4221-9d86-1972c3019f68 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.101766] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-60a0611f-4742-4524-a115-5b0171ec85a1 tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] [instance: 86b1ddcf-585c-41d9-84fe-17b794f1abbc] Reconfiguring VM instance instance-00000051 to attach disk [datastore1] volume-faf85999-e3c6-4036-a18e-7d126676fad9/volume-faf85999-e3c6-4036-a18e-7d126676fad9.vmdk or device None with type thin {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 974.102468] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b4a4d91c-400d-4a69-aa47-24f2599083d7 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.120429] env[63515]: DEBUG oslo_vmware.api [None req-60a0611f-4742-4524-a115-5b0171ec85a1 tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Waiting for the task: (returnval){ [ 974.120429] env[63515]: value = "task-1111722" [ 974.120429] env[63515]: _type = "Task" [ 974.120429] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 974.129466] env[63515]: DEBUG oslo_vmware.api [None req-60a0611f-4742-4524-a115-5b0171ec85a1 tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Task: {'id': task-1111722, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.155883] env[63515]: DEBUG oslo_concurrency.lockutils [None req-16f0ea70-a973-4a3c-bc2e-292ffb4571ac tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Lock "918d5b35-c46d-47c5-b59f-175831aed9d6" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.049s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 974.196133] env[63515]: DEBUG nova.network.neutron [None req-a5676663-6933-430d-aa5e-aeffa94052ee tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: 6a66dd9b-7b24-4b70-b58f-0830cf39376c] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 974.318753] env[63515]: DEBUG nova.compute.manager [req-840a4814-3a9c-47ac-9eda-0d0cd91b4285 req-c9b1b4f5-b36f-4c5e-818c-c35387addd5e service nova] [instance: 6a66dd9b-7b24-4b70-b58f-0830cf39376c] Received event network-vif-plugged-b190896f-9020-45f3-bb87-f1fe6401b017 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 974.319025] env[63515]: DEBUG oslo_concurrency.lockutils [req-840a4814-3a9c-47ac-9eda-0d0cd91b4285 req-c9b1b4f5-b36f-4c5e-818c-c35387addd5e service nova] Acquiring lock "6a66dd9b-7b24-4b70-b58f-0830cf39376c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 974.319290] env[63515]: DEBUG oslo_concurrency.lockutils [req-840a4814-3a9c-47ac-9eda-0d0cd91b4285 req-c9b1b4f5-b36f-4c5e-818c-c35387addd5e service nova] Lock "6a66dd9b-7b24-4b70-b58f-0830cf39376c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 974.319485] env[63515]: DEBUG oslo_concurrency.lockutils [req-840a4814-3a9c-47ac-9eda-0d0cd91b4285 req-c9b1b4f5-b36f-4c5e-818c-c35387addd5e service nova] Lock "6a66dd9b-7b24-4b70-b58f-0830cf39376c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 974.319664] env[63515]: DEBUG nova.compute.manager [req-840a4814-3a9c-47ac-9eda-0d0cd91b4285 req-c9b1b4f5-b36f-4c5e-818c-c35387addd5e service nova] [instance: 6a66dd9b-7b24-4b70-b58f-0830cf39376c] No waiting events found dispatching network-vif-plugged-b190896f-9020-45f3-bb87-f1fe6401b017 {{(pid=63515) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 974.319833] env[63515]: WARNING nova.compute.manager [req-840a4814-3a9c-47ac-9eda-0d0cd91b4285 req-c9b1b4f5-b36f-4c5e-818c-c35387addd5e service nova] [instance: 6a66dd9b-7b24-4b70-b58f-0830cf39376c] Received unexpected event network-vif-plugged-b190896f-9020-45f3-bb87-f1fe6401b017 for instance with vm_state building and task_state spawning. [ 974.319999] env[63515]: DEBUG nova.compute.manager [req-840a4814-3a9c-47ac-9eda-0d0cd91b4285 req-c9b1b4f5-b36f-4c5e-818c-c35387addd5e service nova] [instance: 6a66dd9b-7b24-4b70-b58f-0830cf39376c] Received event network-changed-b190896f-9020-45f3-bb87-f1fe6401b017 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 974.320188] env[63515]: DEBUG nova.compute.manager [req-840a4814-3a9c-47ac-9eda-0d0cd91b4285 req-c9b1b4f5-b36f-4c5e-818c-c35387addd5e service nova] [instance: 6a66dd9b-7b24-4b70-b58f-0830cf39376c] Refreshing instance network info cache due to event network-changed-b190896f-9020-45f3-bb87-f1fe6401b017. {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 974.320390] env[63515]: DEBUG oslo_concurrency.lockutils [req-840a4814-3a9c-47ac-9eda-0d0cd91b4285 req-c9b1b4f5-b36f-4c5e-818c-c35387addd5e service nova] Acquiring lock "refresh_cache-6a66dd9b-7b24-4b70-b58f-0830cf39376c" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 974.328248] env[63515]: DEBUG nova.network.neutron [None req-a5676663-6933-430d-aa5e-aeffa94052ee tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: 6a66dd9b-7b24-4b70-b58f-0830cf39376c] Updating instance_info_cache with network_info: [{"id": "b190896f-9020-45f3-bb87-f1fe6401b017", "address": "fa:16:3e:f4:14:9f", "network": {"id": "15fdaabe-6793-400f-9737-2cdff8f07238", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1175204866-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f9c98e9f6020475490aaa7e76d907ab7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7d377d75-3add-4a15-8691-74b2eb010924", "external-id": "nsx-vlan-transportzone-71", "segmentation_id": 71, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb190896f-90", "ovs_interfaceid": "b190896f-9020-45f3-bb87-f1fe6401b017", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 974.464812] env[63515]: DEBUG nova.scheduler.client.report [None req-adce163d-60b0-4e35-be4e-020840038d00 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 974.468182] env[63515]: INFO nova.compute.manager [-] [instance: 2a29f37d-e786-48d3-b126-4854d6e7c67c] Took 1.79 seconds to deallocate network for instance. [ 974.630938] env[63515]: DEBUG oslo_vmware.api [None req-60a0611f-4742-4524-a115-5b0171ec85a1 tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Task: {'id': task-1111722, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.658447] env[63515]: DEBUG nova.compute.manager [None req-113af8fa-af69-464c-ac2e-e5cf96a3de06 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: 61cf2f9c-5d35-46c4-a477-d2cef9fcd5ae] Start spawning the instance on the hypervisor. {{(pid=63515) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 974.669684] env[63515]: DEBUG oslo_concurrency.lockutils [None req-16f0ea70-a973-4a3c-bc2e-292ffb4571ac tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 974.681413] env[63515]: DEBUG nova.virt.hardware [None req-113af8fa-af69-464c-ac2e-e5cf96a3de06 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T02:52:14Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T02:51:56Z,direct_url=,disk_format='vmdk',id=8a120570-cb06-4099-b262-554ca0ad15c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b569255cc43e42e7ae3f2b4ad37c6ef4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T02:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 974.681703] env[63515]: DEBUG nova.virt.hardware [None req-113af8fa-af69-464c-ac2e-e5cf96a3de06 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Flavor limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 974.681878] env[63515]: DEBUG nova.virt.hardware [None req-113af8fa-af69-464c-ac2e-e5cf96a3de06 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Image limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 974.682097] env[63515]: DEBUG nova.virt.hardware [None req-113af8fa-af69-464c-ac2e-e5cf96a3de06 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Flavor pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 974.682275] env[63515]: DEBUG nova.virt.hardware [None req-113af8fa-af69-464c-ac2e-e5cf96a3de06 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Image pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 974.682441] env[63515]: DEBUG nova.virt.hardware [None req-113af8fa-af69-464c-ac2e-e5cf96a3de06 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 974.682673] env[63515]: DEBUG nova.virt.hardware [None req-113af8fa-af69-464c-ac2e-e5cf96a3de06 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 974.682848] env[63515]: DEBUG nova.virt.hardware [None req-113af8fa-af69-464c-ac2e-e5cf96a3de06 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 974.683359] env[63515]: DEBUG nova.virt.hardware [None req-113af8fa-af69-464c-ac2e-e5cf96a3de06 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Got 1 possible topologies {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 974.683676] env[63515]: DEBUG nova.virt.hardware [None req-113af8fa-af69-464c-ac2e-e5cf96a3de06 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 974.683912] env[63515]: DEBUG nova.virt.hardware [None req-113af8fa-af69-464c-ac2e-e5cf96a3de06 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 974.684872] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-929dca30-d5bc-46df-b670-2ae718a55355 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.693424] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f848cf2-0813-4c11-9990-c721b6a7566b {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.831178] env[63515]: DEBUG oslo_concurrency.lockutils [None req-a5676663-6933-430d-aa5e-aeffa94052ee tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Releasing lock "refresh_cache-6a66dd9b-7b24-4b70-b58f-0830cf39376c" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 974.831916] env[63515]: DEBUG nova.compute.manager [None req-a5676663-6933-430d-aa5e-aeffa94052ee tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: 6a66dd9b-7b24-4b70-b58f-0830cf39376c] Instance network_info: |[{"id": "b190896f-9020-45f3-bb87-f1fe6401b017", "address": "fa:16:3e:f4:14:9f", "network": {"id": "15fdaabe-6793-400f-9737-2cdff8f07238", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1175204866-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f9c98e9f6020475490aaa7e76d907ab7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7d377d75-3add-4a15-8691-74b2eb010924", "external-id": "nsx-vlan-transportzone-71", "segmentation_id": 71, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb190896f-90", "ovs_interfaceid": "b190896f-9020-45f3-bb87-f1fe6401b017", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 974.831916] env[63515]: DEBUG oslo_concurrency.lockutils [req-840a4814-3a9c-47ac-9eda-0d0cd91b4285 req-c9b1b4f5-b36f-4c5e-818c-c35387addd5e service nova] Acquired lock "refresh_cache-6a66dd9b-7b24-4b70-b58f-0830cf39376c" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 974.832181] env[63515]: DEBUG nova.network.neutron [req-840a4814-3a9c-47ac-9eda-0d0cd91b4285 req-c9b1b4f5-b36f-4c5e-818c-c35387addd5e service nova] [instance: 6a66dd9b-7b24-4b70-b58f-0830cf39376c] Refreshing network info cache for port b190896f-9020-45f3-bb87-f1fe6401b017 {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 974.833784] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-a5676663-6933-430d-aa5e-aeffa94052ee tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: 6a66dd9b-7b24-4b70-b58f-0830cf39376c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:f4:14:9f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '7d377d75-3add-4a15-8691-74b2eb010924', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b190896f-9020-45f3-bb87-f1fe6401b017', 'vif_model': 'vmxnet3'}] {{(pid=63515) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 974.842874] env[63515]: DEBUG oslo.service.loopingcall [None req-a5676663-6933-430d-aa5e-aeffa94052ee tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 974.845519] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6a66dd9b-7b24-4b70-b58f-0830cf39376c] Creating VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 974.845519] env[63515]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-2816812b-5253-4d37-b592-c0f9381662db {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.865135] env[63515]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 974.865135] env[63515]: value = "task-1111723" [ 974.865135] env[63515]: _type = "Task" [ 974.865135] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 974.873735] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1111723, 'name': CreateVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.971178] env[63515]: DEBUG oslo_concurrency.lockutils [None req-adce163d-60b0-4e35-be4e-020840038d00 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.332s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 974.971178] env[63515]: DEBUG nova.compute.manager [None req-adce163d-60b0-4e35-be4e-020840038d00 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: 724b05bc-ce4b-4f99-ad14-f3abea5567f2] Start building networks asynchronously for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 974.973826] env[63515]: DEBUG oslo_concurrency.lockutils [None req-c7884167-3670-45b6-b67d-b6f36de44d1a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.246s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 974.974073] env[63515]: DEBUG nova.objects.instance [None req-c7884167-3670-45b6-b67d-b6f36de44d1a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Lazy-loading 'resources' on Instance uuid 75f4dc79-4b73-4ddc-be03-3653a5a0797e {{(pid=63515) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 974.975950] env[63515]: DEBUG oslo_concurrency.lockutils [None req-e595b7dd-52b8-43f9-9fcf-6cb888cef16d tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 975.139015] env[63515]: DEBUG oslo_vmware.api [None req-60a0611f-4742-4524-a115-5b0171ec85a1 tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Task: {'id': task-1111722, 'name': ReconfigVM_Task, 'duration_secs': 0.647201} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 975.139015] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-60a0611f-4742-4524-a115-5b0171ec85a1 tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] [instance: 86b1ddcf-585c-41d9-84fe-17b794f1abbc] Reconfigured VM instance instance-00000051 to attach disk [datastore1] volume-faf85999-e3c6-4036-a18e-7d126676fad9/volume-faf85999-e3c6-4036-a18e-7d126676fad9.vmdk or device None with type thin {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 975.141958] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-42f74798-23ec-4868-97b6-a0afad83636b {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.158768] env[63515]: DEBUG oslo_vmware.api [None req-60a0611f-4742-4524-a115-5b0171ec85a1 tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Waiting for the task: (returnval){ [ 975.158768] env[63515]: value = "task-1111724" [ 975.158768] env[63515]: _type = "Task" [ 975.158768] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 975.167625] env[63515]: DEBUG oslo_vmware.api [None req-60a0611f-4742-4524-a115-5b0171ec85a1 tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Task: {'id': task-1111724, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.333185] env[63515]: DEBUG nova.network.neutron [None req-113af8fa-af69-464c-ac2e-e5cf96a3de06 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: 61cf2f9c-5d35-46c4-a477-d2cef9fcd5ae] Successfully updated port: 99e47951-e384-4aa7-a394-3d0a6bade8d7 {{(pid=63515) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 975.375664] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1111723, 'name': CreateVM_Task, 'duration_secs': 0.300458} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 975.375835] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6a66dd9b-7b24-4b70-b58f-0830cf39376c] Created VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 975.376585] env[63515]: DEBUG oslo_concurrency.lockutils [None req-a5676663-6933-430d-aa5e-aeffa94052ee tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 975.376769] env[63515]: DEBUG oslo_concurrency.lockutils [None req-a5676663-6933-430d-aa5e-aeffa94052ee tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 975.377172] env[63515]: DEBUG oslo_concurrency.lockutils [None req-a5676663-6933-430d-aa5e-aeffa94052ee tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 975.377347] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0a667454-b603-49ec-a51c-14c571b27e35 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.384341] env[63515]: DEBUG oslo_vmware.api [None req-a5676663-6933-430d-aa5e-aeffa94052ee tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Waiting for the task: (returnval){ [ 975.384341] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]526f6231-e556-7223-f6f8-b796999e136a" [ 975.384341] env[63515]: _type = "Task" [ 975.384341] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 975.393064] env[63515]: DEBUG oslo_vmware.api [None req-a5676663-6933-430d-aa5e-aeffa94052ee tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]526f6231-e556-7223-f6f8-b796999e136a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.480610] env[63515]: DEBUG nova.compute.utils [None req-adce163d-60b0-4e35-be4e-020840038d00 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Using /dev/sd instead of None {{(pid=63515) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 975.482319] env[63515]: DEBUG nova.compute.manager [None req-adce163d-60b0-4e35-be4e-020840038d00 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: 724b05bc-ce4b-4f99-ad14-f3abea5567f2] Allocating IP information in the background. {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 975.482435] env[63515]: DEBUG nova.network.neutron [None req-adce163d-60b0-4e35-be4e-020840038d00 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: 724b05bc-ce4b-4f99-ad14-f3abea5567f2] allocate_for_instance() {{(pid=63515) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 975.641622] env[63515]: DEBUG nova.policy [None req-adce163d-60b0-4e35-be4e-020840038d00 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '16880abf1c5b4341800c94ada2c756c7', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7d4ada9ec35f42b19c6480a9101d21a4', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63515) authorize /opt/stack/nova/nova/policy.py:201}} [ 975.669394] env[63515]: DEBUG oslo_vmware.api [None req-60a0611f-4742-4524-a115-5b0171ec85a1 tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Task: {'id': task-1111724, 'name': ReconfigVM_Task, 'duration_secs': 0.139856} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 975.672124] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-60a0611f-4742-4524-a115-5b0171ec85a1 tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] [instance: 86b1ddcf-585c-41d9-84fe-17b794f1abbc] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-243550', 'volume_id': 'faf85999-e3c6-4036-a18e-7d126676fad9', 'name': 'volume-faf85999-e3c6-4036-a18e-7d126676fad9', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '86b1ddcf-585c-41d9-84fe-17b794f1abbc', 'attached_at': '', 'detached_at': '', 'volume_id': 'faf85999-e3c6-4036-a18e-7d126676fad9', 'serial': 'faf85999-e3c6-4036-a18e-7d126676fad9'} {{(pid=63515) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 975.725837] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-968428cd-be27-4e32-b14f-ad9ef7a5cbe4 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.734126] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d691abcc-c600-4a18-bb23-4b664e34b993 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.769016] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82d18de3-41b8-47b6-b48a-e4bc64284c84 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.777487] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cdab1cc1-0411-453e-9313-38b95aac72b2 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.792824] env[63515]: DEBUG nova.compute.provider_tree [None req-c7884167-3670-45b6-b67d-b6f36de44d1a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 975.836347] env[63515]: DEBUG oslo_concurrency.lockutils [None req-113af8fa-af69-464c-ac2e-e5cf96a3de06 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Acquiring lock "refresh_cache-61cf2f9c-5d35-46c4-a477-d2cef9fcd5ae" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 975.836568] env[63515]: DEBUG oslo_concurrency.lockutils [None req-113af8fa-af69-464c-ac2e-e5cf96a3de06 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Acquired lock "refresh_cache-61cf2f9c-5d35-46c4-a477-d2cef9fcd5ae" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 975.836735] env[63515]: DEBUG nova.network.neutron [None req-113af8fa-af69-464c-ac2e-e5cf96a3de06 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: 61cf2f9c-5d35-46c4-a477-d2cef9fcd5ae] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 975.853853] env[63515]: DEBUG oslo_concurrency.lockutils [None req-ee146559-ead8-41ee-8bac-7272cc90bf4b tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Acquiring lock "918d5b35-c46d-47c5-b59f-175831aed9d6" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 975.854028] env[63515]: DEBUG oslo_concurrency.lockutils [None req-ee146559-ead8-41ee-8bac-7272cc90bf4b tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Lock "918d5b35-c46d-47c5-b59f-175831aed9d6" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 975.854391] env[63515]: DEBUG oslo_concurrency.lockutils [None req-ee146559-ead8-41ee-8bac-7272cc90bf4b tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Acquiring lock "918d5b35-c46d-47c5-b59f-175831aed9d6-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 975.854513] env[63515]: DEBUG oslo_concurrency.lockutils [None req-ee146559-ead8-41ee-8bac-7272cc90bf4b tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Lock "918d5b35-c46d-47c5-b59f-175831aed9d6-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 975.855343] env[63515]: DEBUG oslo_concurrency.lockutils [None req-ee146559-ead8-41ee-8bac-7272cc90bf4b tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Lock "918d5b35-c46d-47c5-b59f-175831aed9d6-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 975.857533] env[63515]: INFO nova.compute.manager [None req-ee146559-ead8-41ee-8bac-7272cc90bf4b tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: 918d5b35-c46d-47c5-b59f-175831aed9d6] Terminating instance [ 975.860196] env[63515]: DEBUG nova.compute.manager [None req-ee146559-ead8-41ee-8bac-7272cc90bf4b tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: 918d5b35-c46d-47c5-b59f-175831aed9d6] Start destroying the instance on the hypervisor. {{(pid=63515) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 975.860378] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-ee146559-ead8-41ee-8bac-7272cc90bf4b tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: 918d5b35-c46d-47c5-b59f-175831aed9d6] Destroying instance {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 975.861893] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dbc1a66d-4e20-4a4b-a31f-0cb31d22881b {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.870919] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-ee146559-ead8-41ee-8bac-7272cc90bf4b tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: 918d5b35-c46d-47c5-b59f-175831aed9d6] Unregistering the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 975.871530] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b16f463c-60e0-4f8c-9978-d88dc745aa46 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.889362] env[63515]: DEBUG nova.network.neutron [req-840a4814-3a9c-47ac-9eda-0d0cd91b4285 req-c9b1b4f5-b36f-4c5e-818c-c35387addd5e service nova] [instance: 6a66dd9b-7b24-4b70-b58f-0830cf39376c] Updated VIF entry in instance network info cache for port b190896f-9020-45f3-bb87-f1fe6401b017. {{(pid=63515) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 975.889737] env[63515]: DEBUG nova.network.neutron [req-840a4814-3a9c-47ac-9eda-0d0cd91b4285 req-c9b1b4f5-b36f-4c5e-818c-c35387addd5e service nova] [instance: 6a66dd9b-7b24-4b70-b58f-0830cf39376c] Updating instance_info_cache with network_info: [{"id": "b190896f-9020-45f3-bb87-f1fe6401b017", "address": "fa:16:3e:f4:14:9f", "network": {"id": "15fdaabe-6793-400f-9737-2cdff8f07238", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1175204866-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f9c98e9f6020475490aaa7e76d907ab7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7d377d75-3add-4a15-8691-74b2eb010924", "external-id": "nsx-vlan-transportzone-71", "segmentation_id": 71, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb190896f-90", "ovs_interfaceid": "b190896f-9020-45f3-bb87-f1fe6401b017", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 975.898695] env[63515]: DEBUG oslo_vmware.api [None req-a5676663-6933-430d-aa5e-aeffa94052ee tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]526f6231-e556-7223-f6f8-b796999e136a, 'name': SearchDatastore_Task, 'duration_secs': 0.010273} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 975.898856] env[63515]: DEBUG oslo_concurrency.lockutils [None req-a5676663-6933-430d-aa5e-aeffa94052ee tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 975.899054] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-a5676663-6933-430d-aa5e-aeffa94052ee tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: 6a66dd9b-7b24-4b70-b58f-0830cf39376c] Processing image 8a120570-cb06-4099-b262-554ca0ad15c5 {{(pid=63515) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 975.899336] env[63515]: DEBUG oslo_concurrency.lockutils [None req-a5676663-6933-430d-aa5e-aeffa94052ee tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 975.899482] env[63515]: DEBUG oslo_concurrency.lockutils [None req-a5676663-6933-430d-aa5e-aeffa94052ee tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 975.899676] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-a5676663-6933-430d-aa5e-aeffa94052ee tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 975.899958] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e2fcffbd-5cc4-46da-8724-314b91b07d64 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.911157] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-a5676663-6933-430d-aa5e-aeffa94052ee tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 975.911533] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-a5676663-6933-430d-aa5e-aeffa94052ee tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63515) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 975.913290] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-afa5540f-a488-4798-9f18-477ffa5728a7 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.921446] env[63515]: DEBUG oslo_vmware.api [None req-a5676663-6933-430d-aa5e-aeffa94052ee tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Waiting for the task: (returnval){ [ 975.921446] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]520327ed-e0a0-7159-cd74-c344f0891af0" [ 975.921446] env[63515]: _type = "Task" [ 975.921446] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 975.935452] env[63515]: DEBUG oslo_vmware.api [None req-a5676663-6933-430d-aa5e-aeffa94052ee tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]520327ed-e0a0-7159-cd74-c344f0891af0, 'name': SearchDatastore_Task, 'duration_secs': 0.009775} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 975.936507] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cc84e79e-8011-4796-a1e8-7f32e35b995f {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.940635] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-ee146559-ead8-41ee-8bac-7272cc90bf4b tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: 918d5b35-c46d-47c5-b59f-175831aed9d6] Unregistered the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 975.940849] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-ee146559-ead8-41ee-8bac-7272cc90bf4b tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: 918d5b35-c46d-47c5-b59f-175831aed9d6] Deleting contents of the VM from datastore datastore1 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 975.941044] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-ee146559-ead8-41ee-8bac-7272cc90bf4b tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Deleting the datastore file [datastore1] 918d5b35-c46d-47c5-b59f-175831aed9d6 {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 975.941729] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-098c62e7-226e-4158-a74b-e1d4042a7ff3 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.944738] env[63515]: DEBUG oslo_vmware.api [None req-a5676663-6933-430d-aa5e-aeffa94052ee tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Waiting for the task: (returnval){ [ 975.944738] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]52ecacb7-f25f-c43e-1aae-821ea5ee7980" [ 975.944738] env[63515]: _type = "Task" [ 975.944738] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 975.950052] env[63515]: DEBUG oslo_vmware.api [None req-ee146559-ead8-41ee-8bac-7272cc90bf4b tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Waiting for the task: (returnval){ [ 975.950052] env[63515]: value = "task-1111726" [ 975.950052] env[63515]: _type = "Task" [ 975.950052] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 975.956648] env[63515]: DEBUG oslo_vmware.api [None req-a5676663-6933-430d-aa5e-aeffa94052ee tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52ecacb7-f25f-c43e-1aae-821ea5ee7980, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.961438] env[63515]: DEBUG oslo_vmware.api [None req-ee146559-ead8-41ee-8bac-7272cc90bf4b tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Task: {'id': task-1111726, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.988550] env[63515]: DEBUG nova.compute.manager [None req-adce163d-60b0-4e35-be4e-020840038d00 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: 724b05bc-ce4b-4f99-ad14-f3abea5567f2] Start building block device mappings for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 976.215253] env[63515]: DEBUG nova.network.neutron [None req-adce163d-60b0-4e35-be4e-020840038d00 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: 724b05bc-ce4b-4f99-ad14-f3abea5567f2] Successfully created port: 3dee43b6-2650-42e1-aa5a-4994a3ec7f05 {{(pid=63515) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 976.299528] env[63515]: DEBUG nova.scheduler.client.report [None req-c7884167-3670-45b6-b67d-b6f36de44d1a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 976.353610] env[63515]: DEBUG nova.compute.manager [req-b5e42f37-adb8-499e-96ff-802c95558f0d req-0065585d-1402-4b5e-b588-cce8280b9840 service nova] [instance: 61cf2f9c-5d35-46c4-a477-d2cef9fcd5ae] Received event network-vif-plugged-99e47951-e384-4aa7-a394-3d0a6bade8d7 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 976.353890] env[63515]: DEBUG oslo_concurrency.lockutils [req-b5e42f37-adb8-499e-96ff-802c95558f0d req-0065585d-1402-4b5e-b588-cce8280b9840 service nova] Acquiring lock "61cf2f9c-5d35-46c4-a477-d2cef9fcd5ae-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 976.354143] env[63515]: DEBUG oslo_concurrency.lockutils [req-b5e42f37-adb8-499e-96ff-802c95558f0d req-0065585d-1402-4b5e-b588-cce8280b9840 service nova] Lock "61cf2f9c-5d35-46c4-a477-d2cef9fcd5ae-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 976.354320] env[63515]: DEBUG oslo_concurrency.lockutils [req-b5e42f37-adb8-499e-96ff-802c95558f0d req-0065585d-1402-4b5e-b588-cce8280b9840 service nova] Lock "61cf2f9c-5d35-46c4-a477-d2cef9fcd5ae-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 976.354512] env[63515]: DEBUG nova.compute.manager [req-b5e42f37-adb8-499e-96ff-802c95558f0d req-0065585d-1402-4b5e-b588-cce8280b9840 service nova] [instance: 61cf2f9c-5d35-46c4-a477-d2cef9fcd5ae] No waiting events found dispatching network-vif-plugged-99e47951-e384-4aa7-a394-3d0a6bade8d7 {{(pid=63515) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 976.354679] env[63515]: WARNING nova.compute.manager [req-b5e42f37-adb8-499e-96ff-802c95558f0d req-0065585d-1402-4b5e-b588-cce8280b9840 service nova] [instance: 61cf2f9c-5d35-46c4-a477-d2cef9fcd5ae] Received unexpected event network-vif-plugged-99e47951-e384-4aa7-a394-3d0a6bade8d7 for instance with vm_state building and task_state spawning. [ 976.354843] env[63515]: DEBUG nova.compute.manager [req-b5e42f37-adb8-499e-96ff-802c95558f0d req-0065585d-1402-4b5e-b588-cce8280b9840 service nova] [instance: 61cf2f9c-5d35-46c4-a477-d2cef9fcd5ae] Received event network-changed-99e47951-e384-4aa7-a394-3d0a6bade8d7 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 976.355011] env[63515]: DEBUG nova.compute.manager [req-b5e42f37-adb8-499e-96ff-802c95558f0d req-0065585d-1402-4b5e-b588-cce8280b9840 service nova] [instance: 61cf2f9c-5d35-46c4-a477-d2cef9fcd5ae] Refreshing instance network info cache due to event network-changed-99e47951-e384-4aa7-a394-3d0a6bade8d7. {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 976.355555] env[63515]: DEBUG oslo_concurrency.lockutils [req-b5e42f37-adb8-499e-96ff-802c95558f0d req-0065585d-1402-4b5e-b588-cce8280b9840 service nova] Acquiring lock "refresh_cache-61cf2f9c-5d35-46c4-a477-d2cef9fcd5ae" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 976.386987] env[63515]: DEBUG nova.network.neutron [None req-113af8fa-af69-464c-ac2e-e5cf96a3de06 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: 61cf2f9c-5d35-46c4-a477-d2cef9fcd5ae] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 976.392554] env[63515]: DEBUG oslo_concurrency.lockutils [req-840a4814-3a9c-47ac-9eda-0d0cd91b4285 req-c9b1b4f5-b36f-4c5e-818c-c35387addd5e service nova] Releasing lock "refresh_cache-6a66dd9b-7b24-4b70-b58f-0830cf39376c" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 976.392843] env[63515]: DEBUG nova.compute.manager [req-840a4814-3a9c-47ac-9eda-0d0cd91b4285 req-c9b1b4f5-b36f-4c5e-818c-c35387addd5e service nova] [instance: 2a29f37d-e786-48d3-b126-4854d6e7c67c] Received event network-vif-deleted-956cd361-0d42-4914-85d3-5f75a72a68a1 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 976.460079] env[63515]: DEBUG oslo_vmware.api [None req-a5676663-6933-430d-aa5e-aeffa94052ee tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52ecacb7-f25f-c43e-1aae-821ea5ee7980, 'name': SearchDatastore_Task, 'duration_secs': 0.011051} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 976.460806] env[63515]: DEBUG oslo_concurrency.lockutils [None req-a5676663-6933-430d-aa5e-aeffa94052ee tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 976.461119] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-a5676663-6933-430d-aa5e-aeffa94052ee tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk to [datastore2] 6a66dd9b-7b24-4b70-b58f-0830cf39376c/6a66dd9b-7b24-4b70-b58f-0830cf39376c.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 976.461403] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2cfc2fa6-127a-431b-9a8d-565962ca8fd0 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.466986] env[63515]: DEBUG oslo_concurrency.lockutils [None req-24c0a5e0-a9b9-485a-b28e-8920d1542ede tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Acquiring lock "68766cd4-84be-475b-8494-d7ab43a9e969" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 976.467260] env[63515]: DEBUG oslo_concurrency.lockutils [None req-24c0a5e0-a9b9-485a-b28e-8920d1542ede tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Lock "68766cd4-84be-475b-8494-d7ab43a9e969" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 976.467434] env[63515]: DEBUG nova.compute.manager [None req-24c0a5e0-a9b9-485a-b28e-8920d1542ede tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 68766cd4-84be-475b-8494-d7ab43a9e969] Checking state {{(pid=63515) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 976.467887] env[63515]: DEBUG oslo_vmware.api [None req-ee146559-ead8-41ee-8bac-7272cc90bf4b tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Task: {'id': task-1111726, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.209799} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 976.468776] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c3266ad-b9e0-4d8a-8a4d-57d17efa9c6f {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.471299] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-ee146559-ead8-41ee-8bac-7272cc90bf4b tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Deleted the datastore file {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 976.471495] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-ee146559-ead8-41ee-8bac-7272cc90bf4b tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: 918d5b35-c46d-47c5-b59f-175831aed9d6] Deleted contents of the VM from datastore datastore1 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 976.471677] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-ee146559-ead8-41ee-8bac-7272cc90bf4b tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: 918d5b35-c46d-47c5-b59f-175831aed9d6] Instance destroyed {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 976.471854] env[63515]: INFO nova.compute.manager [None req-ee146559-ead8-41ee-8bac-7272cc90bf4b tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: 918d5b35-c46d-47c5-b59f-175831aed9d6] Took 0.61 seconds to destroy the instance on the hypervisor. [ 976.472127] env[63515]: DEBUG oslo.service.loopingcall [None req-ee146559-ead8-41ee-8bac-7272cc90bf4b tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 976.473418] env[63515]: DEBUG nova.compute.manager [-] [instance: 918d5b35-c46d-47c5-b59f-175831aed9d6] Deallocating network for instance {{(pid=63515) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 976.473600] env[63515]: DEBUG nova.network.neutron [-] [instance: 918d5b35-c46d-47c5-b59f-175831aed9d6] deallocate_for_instance() {{(pid=63515) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 976.475538] env[63515]: DEBUG oslo_vmware.api [None req-a5676663-6933-430d-aa5e-aeffa94052ee tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Waiting for the task: (returnval){ [ 976.475538] env[63515]: value = "task-1111727" [ 976.475538] env[63515]: _type = "Task" [ 976.475538] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 976.481434] env[63515]: DEBUG nova.compute.manager [None req-24c0a5e0-a9b9-485a-b28e-8920d1542ede tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 68766cd4-84be-475b-8494-d7ab43a9e969] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=63515) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 976.482620] env[63515]: DEBUG nova.objects.instance [None req-24c0a5e0-a9b9-485a-b28e-8920d1542ede tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Lazy-loading 'flavor' on Instance uuid 68766cd4-84be-475b-8494-d7ab43a9e969 {{(pid=63515) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 976.491329] env[63515]: DEBUG oslo_vmware.api [None req-a5676663-6933-430d-aa5e-aeffa94052ee tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Task: {'id': task-1111727, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 976.640891] env[63515]: DEBUG nova.network.neutron [None req-113af8fa-af69-464c-ac2e-e5cf96a3de06 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: 61cf2f9c-5d35-46c4-a477-d2cef9fcd5ae] Updating instance_info_cache with network_info: [{"id": "99e47951-e384-4aa7-a394-3d0a6bade8d7", "address": "fa:16:3e:34:06:3c", "network": {"id": "3e5914c8-0cd4-49ee-81b4-1f3e280440bb", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-641632277-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ea099ce07bfb4a8da014d9303cf552eb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "225b6979-9329-403b-91fa-138bd41f6e83", "external-id": "nsx-vlan-transportzone-38", "segmentation_id": 38, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap99e47951-e3", "ovs_interfaceid": "99e47951-e384-4aa7-a394-3d0a6bade8d7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 976.744088] env[63515]: DEBUG nova.objects.instance [None req-60a0611f-4742-4524-a115-5b0171ec85a1 tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Lazy-loading 'flavor' on Instance uuid 86b1ddcf-585c-41d9-84fe-17b794f1abbc {{(pid=63515) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 976.805847] env[63515]: DEBUG oslo_concurrency.lockutils [None req-c7884167-3670-45b6-b67d-b6f36de44d1a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.832s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 976.809017] env[63515]: DEBUG oslo_concurrency.lockutils [None req-59dbe5ed-eaca-4998-8a9a-682a5345ed51 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 4.086s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 976.809317] env[63515]: DEBUG nova.objects.instance [None req-59dbe5ed-eaca-4998-8a9a-682a5345ed51 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Lazy-loading 'resources' on Instance uuid d438dd30-b816-48a6-94c6-11e6e9bd5b71 {{(pid=63515) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 976.838101] env[63515]: INFO nova.scheduler.client.report [None req-c7884167-3670-45b6-b67d-b6f36de44d1a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Deleted allocations for instance 75f4dc79-4b73-4ddc-be03-3653a5a0797e [ 976.879246] env[63515]: DEBUG nova.compute.manager [req-0be601c9-6251-4ffd-810f-c9b60d8b96d4 req-2670a339-d1c1-4dc8-814d-82b433516d2b service nova] [instance: 918d5b35-c46d-47c5-b59f-175831aed9d6] Received event network-vif-deleted-b4724c8b-a2c7-4f52-8a75-f46434d7b527 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 976.879442] env[63515]: INFO nova.compute.manager [req-0be601c9-6251-4ffd-810f-c9b60d8b96d4 req-2670a339-d1c1-4dc8-814d-82b433516d2b service nova] [instance: 918d5b35-c46d-47c5-b59f-175831aed9d6] Neutron deleted interface b4724c8b-a2c7-4f52-8a75-f46434d7b527; detaching it from the instance and deleting it from the info cache [ 976.879530] env[63515]: DEBUG nova.network.neutron [req-0be601c9-6251-4ffd-810f-c9b60d8b96d4 req-2670a339-d1c1-4dc8-814d-82b433516d2b service nova] [instance: 918d5b35-c46d-47c5-b59f-175831aed9d6] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 976.988017] env[63515]: DEBUG oslo_vmware.api [None req-a5676663-6933-430d-aa5e-aeffa94052ee tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Task: {'id': task-1111727, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.483391} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 976.988214] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-a5676663-6933-430d-aa5e-aeffa94052ee tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk to [datastore2] 6a66dd9b-7b24-4b70-b58f-0830cf39376c/6a66dd9b-7b24-4b70-b58f-0830cf39376c.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 976.988439] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-a5676663-6933-430d-aa5e-aeffa94052ee tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: 6a66dd9b-7b24-4b70-b58f-0830cf39376c] Extending root virtual disk to 1048576 {{(pid=63515) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 976.988769] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-88fadee8-f220-4b9f-bb6a-0ba44b0c9d4e {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.991591] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-24c0a5e0-a9b9-485a-b28e-8920d1542ede tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 68766cd4-84be-475b-8494-d7ab43a9e969] Powering off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 976.991800] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-807d0e87-b3a5-4bd4-b49f-5130c36e279b {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.000110] env[63515]: DEBUG nova.compute.manager [None req-adce163d-60b0-4e35-be4e-020840038d00 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: 724b05bc-ce4b-4f99-ad14-f3abea5567f2] Start spawning the instance on the hypervisor. {{(pid=63515) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 977.002319] env[63515]: DEBUG oslo_vmware.api [None req-a5676663-6933-430d-aa5e-aeffa94052ee tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Waiting for the task: (returnval){ [ 977.002319] env[63515]: value = "task-1111728" [ 977.002319] env[63515]: _type = "Task" [ 977.002319] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 977.004359] env[63515]: DEBUG oslo_vmware.api [None req-24c0a5e0-a9b9-485a-b28e-8920d1542ede tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Waiting for the task: (returnval){ [ 977.004359] env[63515]: value = "task-1111729" [ 977.004359] env[63515]: _type = "Task" [ 977.004359] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 977.016304] env[63515]: DEBUG oslo_vmware.api [None req-a5676663-6933-430d-aa5e-aeffa94052ee tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Task: {'id': task-1111728, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.019704] env[63515]: DEBUG oslo_vmware.api [None req-24c0a5e0-a9b9-485a-b28e-8920d1542ede tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Task: {'id': task-1111729, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.033057] env[63515]: DEBUG nova.virt.hardware [None req-adce163d-60b0-4e35-be4e-020840038d00 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T02:52:14Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T02:51:56Z,direct_url=,disk_format='vmdk',id=8a120570-cb06-4099-b262-554ca0ad15c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b569255cc43e42e7ae3f2b4ad37c6ef4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T02:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 977.033335] env[63515]: DEBUG nova.virt.hardware [None req-adce163d-60b0-4e35-be4e-020840038d00 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Flavor limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 977.033500] env[63515]: DEBUG nova.virt.hardware [None req-adce163d-60b0-4e35-be4e-020840038d00 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Image limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 977.033689] env[63515]: DEBUG nova.virt.hardware [None req-adce163d-60b0-4e35-be4e-020840038d00 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Flavor pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 977.033839] env[63515]: DEBUG nova.virt.hardware [None req-adce163d-60b0-4e35-be4e-020840038d00 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Image pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 977.033992] env[63515]: DEBUG nova.virt.hardware [None req-adce163d-60b0-4e35-be4e-020840038d00 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 977.034243] env[63515]: DEBUG nova.virt.hardware [None req-adce163d-60b0-4e35-be4e-020840038d00 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 977.034447] env[63515]: DEBUG nova.virt.hardware [None req-adce163d-60b0-4e35-be4e-020840038d00 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 977.034664] env[63515]: DEBUG nova.virt.hardware [None req-adce163d-60b0-4e35-be4e-020840038d00 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Got 1 possible topologies {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 977.034848] env[63515]: DEBUG nova.virt.hardware [None req-adce163d-60b0-4e35-be4e-020840038d00 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 977.035071] env[63515]: DEBUG nova.virt.hardware [None req-adce163d-60b0-4e35-be4e-020840038d00 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 977.035989] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-900f6bad-221b-48e9-ad0f-2b4b17a1b798 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.045393] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42d3d6c9-3783-4b48-9a38-25b3ad5e71f3 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.144289] env[63515]: DEBUG oslo_concurrency.lockutils [None req-113af8fa-af69-464c-ac2e-e5cf96a3de06 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Releasing lock "refresh_cache-61cf2f9c-5d35-46c4-a477-d2cef9fcd5ae" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 977.144666] env[63515]: DEBUG nova.compute.manager [None req-113af8fa-af69-464c-ac2e-e5cf96a3de06 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: 61cf2f9c-5d35-46c4-a477-d2cef9fcd5ae] Instance network_info: |[{"id": "99e47951-e384-4aa7-a394-3d0a6bade8d7", "address": "fa:16:3e:34:06:3c", "network": {"id": "3e5914c8-0cd4-49ee-81b4-1f3e280440bb", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-641632277-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ea099ce07bfb4a8da014d9303cf552eb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "225b6979-9329-403b-91fa-138bd41f6e83", "external-id": "nsx-vlan-transportzone-38", "segmentation_id": 38, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap99e47951-e3", "ovs_interfaceid": "99e47951-e384-4aa7-a394-3d0a6bade8d7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 977.145417] env[63515]: DEBUG oslo_concurrency.lockutils [req-b5e42f37-adb8-499e-96ff-802c95558f0d req-0065585d-1402-4b5e-b588-cce8280b9840 service nova] Acquired lock "refresh_cache-61cf2f9c-5d35-46c4-a477-d2cef9fcd5ae" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 977.145612] env[63515]: DEBUG nova.network.neutron [req-b5e42f37-adb8-499e-96ff-802c95558f0d req-0065585d-1402-4b5e-b588-cce8280b9840 service nova] [instance: 61cf2f9c-5d35-46c4-a477-d2cef9fcd5ae] Refreshing network info cache for port 99e47951-e384-4aa7-a394-3d0a6bade8d7 {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 977.146809] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-113af8fa-af69-464c-ac2e-e5cf96a3de06 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: 61cf2f9c-5d35-46c4-a477-d2cef9fcd5ae] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:34:06:3c', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '225b6979-9329-403b-91fa-138bd41f6e83', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '99e47951-e384-4aa7-a394-3d0a6bade8d7', 'vif_model': 'vmxnet3'}] {{(pid=63515) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 977.154929] env[63515]: DEBUG oslo.service.loopingcall [None req-113af8fa-af69-464c-ac2e-e5cf96a3de06 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 977.155494] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 61cf2f9c-5d35-46c4-a477-d2cef9fcd5ae] Creating VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 977.156265] env[63515]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f5f2cac1-2477-4309-b011-897e304cc180 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.175907] env[63515]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 977.175907] env[63515]: value = "task-1111730" [ 977.175907] env[63515]: _type = "Task" [ 977.175907] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 977.184313] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1111730, 'name': CreateVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.251503] env[63515]: DEBUG oslo_concurrency.lockutils [None req-60a0611f-4742-4524-a115-5b0171ec85a1 tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Lock "86b1ddcf-585c-41d9-84fe-17b794f1abbc" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 9.829s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 977.265531] env[63515]: DEBUG oslo_concurrency.lockutils [None req-7d3c29b2-1cec-41c4-90f7-528df42289ec tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Acquiring lock "86b1ddcf-585c-41d9-84fe-17b794f1abbc" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 977.265797] env[63515]: DEBUG oslo_concurrency.lockutils [None req-7d3c29b2-1cec-41c4-90f7-528df42289ec tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Lock "86b1ddcf-585c-41d9-84fe-17b794f1abbc" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 977.349579] env[63515]: DEBUG oslo_concurrency.lockutils [None req-c7884167-3670-45b6-b67d-b6f36de44d1a tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Lock "75f4dc79-4b73-4ddc-be03-3653a5a0797e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.545s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 977.352105] env[63515]: DEBUG nova.network.neutron [-] [instance: 918d5b35-c46d-47c5-b59f-175831aed9d6] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 977.383877] env[63515]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-8493be61-27c3-43fe-bd24-8f09c195fcb1 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.393688] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d1c49c3-918f-402d-964c-344dc34556ed {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.429227] env[63515]: DEBUG nova.compute.manager [req-0be601c9-6251-4ffd-810f-c9b60d8b96d4 req-2670a339-d1c1-4dc8-814d-82b433516d2b service nova] [instance: 918d5b35-c46d-47c5-b59f-175831aed9d6] Detach interface failed, port_id=b4724c8b-a2c7-4f52-8a75-f46434d7b527, reason: Instance 918d5b35-c46d-47c5-b59f-175831aed9d6 could not be found. {{(pid=63515) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 977.519552] env[63515]: DEBUG oslo_vmware.api [None req-a5676663-6933-430d-aa5e-aeffa94052ee tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Task: {'id': task-1111728, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.087728} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 977.519815] env[63515]: DEBUG oslo_vmware.api [None req-24c0a5e0-a9b9-485a-b28e-8920d1542ede tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Task: {'id': task-1111729, 'name': PowerOffVM_Task, 'duration_secs': 0.190046} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 977.520226] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-a5676663-6933-430d-aa5e-aeffa94052ee tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: 6a66dd9b-7b24-4b70-b58f-0830cf39376c] Extended root virtual disk {{(pid=63515) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 977.520338] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-24c0a5e0-a9b9-485a-b28e-8920d1542ede tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 68766cd4-84be-475b-8494-d7ab43a9e969] Powered off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 977.520514] env[63515]: DEBUG nova.compute.manager [None req-24c0a5e0-a9b9-485a-b28e-8920d1542ede tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 68766cd4-84be-475b-8494-d7ab43a9e969] Checking state {{(pid=63515) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 977.521259] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99f7929c-095d-491f-8b33-0e4feb9d073f {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.524084] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d1d3063-385c-4347-ab28-e565b49853d2 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.527876] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f27a3338-c5da-4b2c-90b5-85a2b4fb718d {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.553943] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-a5676663-6933-430d-aa5e-aeffa94052ee tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: 6a66dd9b-7b24-4b70-b58f-0830cf39376c] Reconfiguring VM instance instance-0000005d to attach disk [datastore2] 6a66dd9b-7b24-4b70-b58f-0830cf39376c/6a66dd9b-7b24-4b70-b58f-0830cf39376c.vmdk or device None with type sparse {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 977.555036] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40979f9b-d407-4b6b-bca1-d926b6f42629 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.558561] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6c4e607e-d820-48dc-b695-6540083508b5 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.606345] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a38ae6b7-2e4f-460b-bad3-9172765cfe3f {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.609594] env[63515]: DEBUG oslo_vmware.api [None req-a5676663-6933-430d-aa5e-aeffa94052ee tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Waiting for the task: (returnval){ [ 977.609594] env[63515]: value = "task-1111731" [ 977.609594] env[63515]: _type = "Task" [ 977.609594] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 977.616598] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76e18e69-7409-4424-87d8-976d7b6f88e9 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.623351] env[63515]: DEBUG oslo_vmware.api [None req-a5676663-6933-430d-aa5e-aeffa94052ee tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Task: {'id': task-1111731, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.633850] env[63515]: DEBUG nova.compute.provider_tree [None req-59dbe5ed-eaca-4998-8a9a-682a5345ed51 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 977.685863] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1111730, 'name': CreateVM_Task} progress is 25%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.768744] env[63515]: INFO nova.compute.manager [None req-7d3c29b2-1cec-41c4-90f7-528df42289ec tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] [instance: 86b1ddcf-585c-41d9-84fe-17b794f1abbc] Detaching volume aa5c9d63-11af-4c5d-ae0b-9aa149332c19 [ 977.816627] env[63515]: INFO nova.virt.block_device [None req-7d3c29b2-1cec-41c4-90f7-528df42289ec tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] [instance: 86b1ddcf-585c-41d9-84fe-17b794f1abbc] Attempting to driver detach volume aa5c9d63-11af-4c5d-ae0b-9aa149332c19 from mountpoint /dev/sdb [ 977.816906] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-7d3c29b2-1cec-41c4-90f7-528df42289ec tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] [instance: 86b1ddcf-585c-41d9-84fe-17b794f1abbc] Volume detach. Driver type: vmdk {{(pid=63515) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 977.817112] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-7d3c29b2-1cec-41c4-90f7-528df42289ec tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] [instance: 86b1ddcf-585c-41d9-84fe-17b794f1abbc] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-243544', 'volume_id': 'aa5c9d63-11af-4c5d-ae0b-9aa149332c19', 'name': 'volume-aa5c9d63-11af-4c5d-ae0b-9aa149332c19', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '86b1ddcf-585c-41d9-84fe-17b794f1abbc', 'attached_at': '', 'detached_at': '', 'volume_id': 'aa5c9d63-11af-4c5d-ae0b-9aa149332c19', 'serial': 'aa5c9d63-11af-4c5d-ae0b-9aa149332c19'} {{(pid=63515) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 977.818071] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f332a36-7863-400c-940e-02b0bc748900 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.844732] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77b7d60c-2136-42d0-a4ab-11ffb4920d05 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.852196] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dffb1b9e-d238-457f-82ca-57241fcfc19f {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.854663] env[63515]: INFO nova.compute.manager [-] [instance: 918d5b35-c46d-47c5-b59f-175831aed9d6] Took 1.38 seconds to deallocate network for instance. [ 977.879286] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-320a6dd7-70d0-4a1e-a8c5-aeb84cb0f3c3 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.895077] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-7d3c29b2-1cec-41c4-90f7-528df42289ec tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] The volume has not been displaced from its original location: [datastore1] volume-aa5c9d63-11af-4c5d-ae0b-9aa149332c19/volume-aa5c9d63-11af-4c5d-ae0b-9aa149332c19.vmdk. No consolidation needed. {{(pid=63515) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 977.900636] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-7d3c29b2-1cec-41c4-90f7-528df42289ec tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] [instance: 86b1ddcf-585c-41d9-84fe-17b794f1abbc] Reconfiguring VM instance instance-00000051 to detach disk 2001 {{(pid=63515) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 977.901824] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8790a678-5ef9-4637-b92f-29e898813522 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.918227] env[63515]: DEBUG nova.network.neutron [req-b5e42f37-adb8-499e-96ff-802c95558f0d req-0065585d-1402-4b5e-b588-cce8280b9840 service nova] [instance: 61cf2f9c-5d35-46c4-a477-d2cef9fcd5ae] Updated VIF entry in instance network info cache for port 99e47951-e384-4aa7-a394-3d0a6bade8d7. {{(pid=63515) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 977.918590] env[63515]: DEBUG nova.network.neutron [req-b5e42f37-adb8-499e-96ff-802c95558f0d req-0065585d-1402-4b5e-b588-cce8280b9840 service nova] [instance: 61cf2f9c-5d35-46c4-a477-d2cef9fcd5ae] Updating instance_info_cache with network_info: [{"id": "99e47951-e384-4aa7-a394-3d0a6bade8d7", "address": "fa:16:3e:34:06:3c", "network": {"id": "3e5914c8-0cd4-49ee-81b4-1f3e280440bb", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-641632277-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ea099ce07bfb4a8da014d9303cf552eb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "225b6979-9329-403b-91fa-138bd41f6e83", "external-id": "nsx-vlan-transportzone-38", "segmentation_id": 38, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap99e47951-e3", "ovs_interfaceid": "99e47951-e384-4aa7-a394-3d0a6bade8d7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 977.921256] env[63515]: DEBUG oslo_vmware.api [None req-7d3c29b2-1cec-41c4-90f7-528df42289ec tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Waiting for the task: (returnval){ [ 977.921256] env[63515]: value = "task-1111732" [ 977.921256] env[63515]: _type = "Task" [ 977.921256] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 977.931542] env[63515]: DEBUG oslo_vmware.api [None req-7d3c29b2-1cec-41c4-90f7-528df42289ec tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Task: {'id': task-1111732, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.975887] env[63515]: DEBUG nova.network.neutron [None req-adce163d-60b0-4e35-be4e-020840038d00 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: 724b05bc-ce4b-4f99-ad14-f3abea5567f2] Successfully updated port: 3dee43b6-2650-42e1-aa5a-4994a3ec7f05 {{(pid=63515) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 978.057398] env[63515]: DEBUG oslo_concurrency.lockutils [None req-24c0a5e0-a9b9-485a-b28e-8920d1542ede tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Lock "68766cd4-84be-475b-8494-d7ab43a9e969" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.590s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 978.120256] env[63515]: DEBUG oslo_vmware.api [None req-a5676663-6933-430d-aa5e-aeffa94052ee tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Task: {'id': task-1111731, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.136561] env[63515]: DEBUG nova.scheduler.client.report [None req-59dbe5ed-eaca-4998-8a9a-682a5345ed51 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 978.186728] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1111730, 'name': CreateVM_Task} progress is 25%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.379056] env[63515]: DEBUG oslo_concurrency.lockutils [None req-ee146559-ead8-41ee-8bac-7272cc90bf4b tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 978.421816] env[63515]: DEBUG oslo_concurrency.lockutils [req-b5e42f37-adb8-499e-96ff-802c95558f0d req-0065585d-1402-4b5e-b588-cce8280b9840 service nova] Releasing lock "refresh_cache-61cf2f9c-5d35-46c4-a477-d2cef9fcd5ae" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 978.430597] env[63515]: DEBUG oslo_vmware.api [None req-7d3c29b2-1cec-41c4-90f7-528df42289ec tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Task: {'id': task-1111732, 'name': ReconfigVM_Task, 'duration_secs': 0.40204} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 978.430905] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-7d3c29b2-1cec-41c4-90f7-528df42289ec tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] [instance: 86b1ddcf-585c-41d9-84fe-17b794f1abbc] Reconfigured VM instance instance-00000051 to detach disk 2001 {{(pid=63515) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 978.436594] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a303201d-4967-4645-a5f3-7032a0fc4dbc {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.447596] env[63515]: DEBUG nova.compute.manager [req-7e541203-6727-4400-92e2-e174b4fe0e8a req-5dfabda1-173a-45df-954a-f44ff1b4a410 service nova] [instance: 724b05bc-ce4b-4f99-ad14-f3abea5567f2] Received event network-vif-plugged-3dee43b6-2650-42e1-aa5a-4994a3ec7f05 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 978.447866] env[63515]: DEBUG oslo_concurrency.lockutils [req-7e541203-6727-4400-92e2-e174b4fe0e8a req-5dfabda1-173a-45df-954a-f44ff1b4a410 service nova] Acquiring lock "724b05bc-ce4b-4f99-ad14-f3abea5567f2-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 978.448149] env[63515]: DEBUG oslo_concurrency.lockutils [req-7e541203-6727-4400-92e2-e174b4fe0e8a req-5dfabda1-173a-45df-954a-f44ff1b4a410 service nova] Lock "724b05bc-ce4b-4f99-ad14-f3abea5567f2-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 978.448342] env[63515]: DEBUG oslo_concurrency.lockutils [req-7e541203-6727-4400-92e2-e174b4fe0e8a req-5dfabda1-173a-45df-954a-f44ff1b4a410 service nova] Lock "724b05bc-ce4b-4f99-ad14-f3abea5567f2-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 978.448543] env[63515]: DEBUG nova.compute.manager [req-7e541203-6727-4400-92e2-e174b4fe0e8a req-5dfabda1-173a-45df-954a-f44ff1b4a410 service nova] [instance: 724b05bc-ce4b-4f99-ad14-f3abea5567f2] No waiting events found dispatching network-vif-plugged-3dee43b6-2650-42e1-aa5a-4994a3ec7f05 {{(pid=63515) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 978.448777] env[63515]: WARNING nova.compute.manager [req-7e541203-6727-4400-92e2-e174b4fe0e8a req-5dfabda1-173a-45df-954a-f44ff1b4a410 service nova] [instance: 724b05bc-ce4b-4f99-ad14-f3abea5567f2] Received unexpected event network-vif-plugged-3dee43b6-2650-42e1-aa5a-4994a3ec7f05 for instance with vm_state building and task_state spawning. [ 978.448993] env[63515]: DEBUG nova.compute.manager [req-7e541203-6727-4400-92e2-e174b4fe0e8a req-5dfabda1-173a-45df-954a-f44ff1b4a410 service nova] [instance: 724b05bc-ce4b-4f99-ad14-f3abea5567f2] Received event network-changed-3dee43b6-2650-42e1-aa5a-4994a3ec7f05 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 978.449223] env[63515]: DEBUG nova.compute.manager [req-7e541203-6727-4400-92e2-e174b4fe0e8a req-5dfabda1-173a-45df-954a-f44ff1b4a410 service nova] [instance: 724b05bc-ce4b-4f99-ad14-f3abea5567f2] Refreshing instance network info cache due to event network-changed-3dee43b6-2650-42e1-aa5a-4994a3ec7f05. {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 978.449466] env[63515]: DEBUG oslo_concurrency.lockutils [req-7e541203-6727-4400-92e2-e174b4fe0e8a req-5dfabda1-173a-45df-954a-f44ff1b4a410 service nova] Acquiring lock "refresh_cache-724b05bc-ce4b-4f99-ad14-f3abea5567f2" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 978.449652] env[63515]: DEBUG oslo_concurrency.lockutils [req-7e541203-6727-4400-92e2-e174b4fe0e8a req-5dfabda1-173a-45df-954a-f44ff1b4a410 service nova] Acquired lock "refresh_cache-724b05bc-ce4b-4f99-ad14-f3abea5567f2" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 978.449847] env[63515]: DEBUG nova.network.neutron [req-7e541203-6727-4400-92e2-e174b4fe0e8a req-5dfabda1-173a-45df-954a-f44ff1b4a410 service nova] [instance: 724b05bc-ce4b-4f99-ad14-f3abea5567f2] Refreshing network info cache for port 3dee43b6-2650-42e1-aa5a-4994a3ec7f05 {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 978.458718] env[63515]: DEBUG oslo_vmware.api [None req-7d3c29b2-1cec-41c4-90f7-528df42289ec tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Waiting for the task: (returnval){ [ 978.458718] env[63515]: value = "task-1111733" [ 978.458718] env[63515]: _type = "Task" [ 978.458718] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 978.469430] env[63515]: DEBUG oslo_vmware.api [None req-7d3c29b2-1cec-41c4-90f7-528df42289ec tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Task: {'id': task-1111733, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.477768] env[63515]: DEBUG oslo_concurrency.lockutils [None req-adce163d-60b0-4e35-be4e-020840038d00 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Acquiring lock "refresh_cache-724b05bc-ce4b-4f99-ad14-f3abea5567f2" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 978.621912] env[63515]: DEBUG oslo_vmware.api [None req-a5676663-6933-430d-aa5e-aeffa94052ee tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Task: {'id': task-1111731, 'name': ReconfigVM_Task, 'duration_secs': 0.91934} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 978.622362] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-a5676663-6933-430d-aa5e-aeffa94052ee tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: 6a66dd9b-7b24-4b70-b58f-0830cf39376c] Reconfigured VM instance instance-0000005d to attach disk [datastore2] 6a66dd9b-7b24-4b70-b58f-0830cf39376c/6a66dd9b-7b24-4b70-b58f-0830cf39376c.vmdk or device None with type sparse {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 978.623247] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ce53b59f-275d-4b6f-a620-c163f817ac08 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.630307] env[63515]: DEBUG oslo_vmware.api [None req-a5676663-6933-430d-aa5e-aeffa94052ee tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Waiting for the task: (returnval){ [ 978.630307] env[63515]: value = "task-1111734" [ 978.630307] env[63515]: _type = "Task" [ 978.630307] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 978.638446] env[63515]: DEBUG oslo_vmware.api [None req-a5676663-6933-430d-aa5e-aeffa94052ee tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Task: {'id': task-1111734, 'name': Rename_Task} progress is 5%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.641313] env[63515]: DEBUG oslo_concurrency.lockutils [None req-59dbe5ed-eaca-4998-8a9a-682a5345ed51 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.832s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 978.643856] env[63515]: DEBUG oslo_concurrency.lockutils [None req-978a15da-7dcc-413b-825d-e9c53aaf3163 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.741s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 978.645364] env[63515]: INFO nova.compute.claims [None req-978a15da-7dcc-413b-825d-e9c53aaf3163 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: 461a0c4c-dfdb-4669-ac72-8913c0277298] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 978.669884] env[63515]: INFO nova.scheduler.client.report [None req-59dbe5ed-eaca-4998-8a9a-682a5345ed51 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Deleted allocations for instance d438dd30-b816-48a6-94c6-11e6e9bd5b71 [ 978.687151] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1111730, 'name': CreateVM_Task, 'duration_secs': 1.283466} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 978.687324] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 61cf2f9c-5d35-46c4-a477-d2cef9fcd5ae] Created VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 978.688226] env[63515]: DEBUG oslo_concurrency.lockutils [None req-113af8fa-af69-464c-ac2e-e5cf96a3de06 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 978.688396] env[63515]: DEBUG oslo_concurrency.lockutils [None req-113af8fa-af69-464c-ac2e-e5cf96a3de06 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 978.688699] env[63515]: DEBUG oslo_concurrency.lockutils [None req-113af8fa-af69-464c-ac2e-e5cf96a3de06 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 978.688964] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f8c992a0-a28f-4cbe-90e6-c072e08bae80 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.694454] env[63515]: DEBUG oslo_vmware.api [None req-113af8fa-af69-464c-ac2e-e5cf96a3de06 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Waiting for the task: (returnval){ [ 978.694454] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]52a22f0f-fe71-d9ac-f0c3-88ae05ef1343" [ 978.694454] env[63515]: _type = "Task" [ 978.694454] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 978.703589] env[63515]: DEBUG oslo_vmware.api [None req-113af8fa-af69-464c-ac2e-e5cf96a3de06 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52a22f0f-fe71-d9ac-f0c3-88ae05ef1343, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.969183] env[63515]: DEBUG oslo_vmware.api [None req-7d3c29b2-1cec-41c4-90f7-528df42289ec tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Task: {'id': task-1111733, 'name': ReconfigVM_Task, 'duration_secs': 0.200222} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 978.969500] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-7d3c29b2-1cec-41c4-90f7-528df42289ec tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] [instance: 86b1ddcf-585c-41d9-84fe-17b794f1abbc] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-243544', 'volume_id': 'aa5c9d63-11af-4c5d-ae0b-9aa149332c19', 'name': 'volume-aa5c9d63-11af-4c5d-ae0b-9aa149332c19', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '86b1ddcf-585c-41d9-84fe-17b794f1abbc', 'attached_at': '', 'detached_at': '', 'volume_id': 'aa5c9d63-11af-4c5d-ae0b-9aa149332c19', 'serial': 'aa5c9d63-11af-4c5d-ae0b-9aa149332c19'} {{(pid=63515) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 978.981489] env[63515]: DEBUG nova.network.neutron [req-7e541203-6727-4400-92e2-e174b4fe0e8a req-5dfabda1-173a-45df-954a-f44ff1b4a410 service nova] [instance: 724b05bc-ce4b-4f99-ad14-f3abea5567f2] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 979.053620] env[63515]: DEBUG nova.network.neutron [req-7e541203-6727-4400-92e2-e174b4fe0e8a req-5dfabda1-173a-45df-954a-f44ff1b4a410 service nova] [instance: 724b05bc-ce4b-4f99-ad14-f3abea5567f2] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 979.141477] env[63515]: DEBUG oslo_vmware.api [None req-a5676663-6933-430d-aa5e-aeffa94052ee tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Task: {'id': task-1111734, 'name': Rename_Task, 'duration_secs': 0.23402} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 979.141744] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-a5676663-6933-430d-aa5e-aeffa94052ee tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: 6a66dd9b-7b24-4b70-b58f-0830cf39376c] Powering on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 979.141984] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8e3b2ef0-0ea3-4113-83a0-195036814dab {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.148506] env[63515]: DEBUG oslo_vmware.api [None req-a5676663-6933-430d-aa5e-aeffa94052ee tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Waiting for the task: (returnval){ [ 979.148506] env[63515]: value = "task-1111735" [ 979.148506] env[63515]: _type = "Task" [ 979.148506] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 979.157984] env[63515]: DEBUG oslo_vmware.api [None req-a5676663-6933-430d-aa5e-aeffa94052ee tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Task: {'id': task-1111735, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 979.177769] env[63515]: DEBUG oslo_concurrency.lockutils [None req-59dbe5ed-eaca-4998-8a9a-682a5345ed51 tempest-MultipleCreateTestJSON-1619805082 tempest-MultipleCreateTestJSON-1619805082-project-member] Lock "d438dd30-b816-48a6-94c6-11e6e9bd5b71" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 10.307s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 979.205103] env[63515]: DEBUG oslo_vmware.api [None req-113af8fa-af69-464c-ac2e-e5cf96a3de06 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52a22f0f-fe71-d9ac-f0c3-88ae05ef1343, 'name': SearchDatastore_Task, 'duration_secs': 0.009305} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 979.205413] env[63515]: DEBUG oslo_concurrency.lockutils [None req-113af8fa-af69-464c-ac2e-e5cf96a3de06 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 979.205650] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-113af8fa-af69-464c-ac2e-e5cf96a3de06 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: 61cf2f9c-5d35-46c4-a477-d2cef9fcd5ae] Processing image 8a120570-cb06-4099-b262-554ca0ad15c5 {{(pid=63515) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 979.205892] env[63515]: DEBUG oslo_concurrency.lockutils [None req-113af8fa-af69-464c-ac2e-e5cf96a3de06 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 979.206057] env[63515]: DEBUG oslo_concurrency.lockutils [None req-113af8fa-af69-464c-ac2e-e5cf96a3de06 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 979.206247] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-113af8fa-af69-464c-ac2e-e5cf96a3de06 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 979.206522] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e62b9957-8271-437b-aa5e-264a6e953fc9 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.213966] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-113af8fa-af69-464c-ac2e-e5cf96a3de06 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 979.214129] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-113af8fa-af69-464c-ac2e-e5cf96a3de06 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63515) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 979.214834] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-61ec6ed3-e415-4138-b54c-f03c6c45a078 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.219818] env[63515]: DEBUG oslo_vmware.api [None req-113af8fa-af69-464c-ac2e-e5cf96a3de06 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Waiting for the task: (returnval){ [ 979.219818] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]5295caf4-22ef-3315-5433-22e50e5a0c52" [ 979.219818] env[63515]: _type = "Task" [ 979.219818] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 979.226760] env[63515]: DEBUG oslo_vmware.api [None req-113af8fa-af69-464c-ac2e-e5cf96a3de06 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]5295caf4-22ef-3315-5433-22e50e5a0c52, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 979.510971] env[63515]: DEBUG nova.objects.instance [None req-7d3c29b2-1cec-41c4-90f7-528df42289ec tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Lazy-loading 'flavor' on Instance uuid 86b1ddcf-585c-41d9-84fe-17b794f1abbc {{(pid=63515) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 979.556672] env[63515]: DEBUG oslo_concurrency.lockutils [req-7e541203-6727-4400-92e2-e174b4fe0e8a req-5dfabda1-173a-45df-954a-f44ff1b4a410 service nova] Releasing lock "refresh_cache-724b05bc-ce4b-4f99-ad14-f3abea5567f2" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 979.557033] env[63515]: DEBUG oslo_concurrency.lockutils [None req-adce163d-60b0-4e35-be4e-020840038d00 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Acquired lock "refresh_cache-724b05bc-ce4b-4f99-ad14-f3abea5567f2" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 979.557202] env[63515]: DEBUG nova.network.neutron [None req-adce163d-60b0-4e35-be4e-020840038d00 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: 724b05bc-ce4b-4f99-ad14-f3abea5567f2] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 979.660608] env[63515]: DEBUG oslo_vmware.api [None req-a5676663-6933-430d-aa5e-aeffa94052ee tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Task: {'id': task-1111735, 'name': PowerOnVM_Task} progress is 89%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 979.730954] env[63515]: DEBUG oslo_vmware.api [None req-113af8fa-af69-464c-ac2e-e5cf96a3de06 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]5295caf4-22ef-3315-5433-22e50e5a0c52, 'name': SearchDatastore_Task, 'duration_secs': 0.007908} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 979.731781] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e05261a1-6c2f-438b-a12e-2193254ee81b {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.737887] env[63515]: DEBUG oslo_vmware.api [None req-113af8fa-af69-464c-ac2e-e5cf96a3de06 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Waiting for the task: (returnval){ [ 979.737887] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]52169573-1b93-81f0-e1a7-1d9bbde5dc73" [ 979.737887] env[63515]: _type = "Task" [ 979.737887] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 979.749067] env[63515]: DEBUG oslo_vmware.api [None req-113af8fa-af69-464c-ac2e-e5cf96a3de06 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52169573-1b93-81f0-e1a7-1d9bbde5dc73, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 979.763786] env[63515]: DEBUG nova.compute.manager [None req-7c387ab3-2055-4517-9874-052fe02db033 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 68766cd4-84be-475b-8494-d7ab43a9e969] Stashing vm_state: stopped {{(pid=63515) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 979.870645] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b42cc246-1e43-4a09-950d-d91812714fd9 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.881879] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-117fdb2a-e017-45fb-bb93-368da9791d43 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.921157] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-096bfdd5-327a-4a71-a465-1f2ef2c087cf {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.929159] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f01d275d-bfb8-4043-bac0-22930526ab34 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.944940] env[63515]: DEBUG nova.compute.provider_tree [None req-978a15da-7dcc-413b-825d-e9c53aaf3163 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 980.107820] env[63515]: DEBUG nova.network.neutron [None req-adce163d-60b0-4e35-be4e-020840038d00 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: 724b05bc-ce4b-4f99-ad14-f3abea5567f2] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 980.159188] env[63515]: DEBUG oslo_vmware.api [None req-a5676663-6933-430d-aa5e-aeffa94052ee tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Task: {'id': task-1111735, 'name': PowerOnVM_Task, 'duration_secs': 0.692084} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 980.159481] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-a5676663-6933-430d-aa5e-aeffa94052ee tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: 6a66dd9b-7b24-4b70-b58f-0830cf39376c] Powered on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 980.159692] env[63515]: INFO nova.compute.manager [None req-a5676663-6933-430d-aa5e-aeffa94052ee tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: 6a66dd9b-7b24-4b70-b58f-0830cf39376c] Took 7.88 seconds to spawn the instance on the hypervisor. [ 980.159881] env[63515]: DEBUG nova.compute.manager [None req-a5676663-6933-430d-aa5e-aeffa94052ee tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: 6a66dd9b-7b24-4b70-b58f-0830cf39376c] Checking state {{(pid=63515) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 980.162625] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb12d80f-da30-4271-903b-b8faf05a3407 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.252065] env[63515]: DEBUG oslo_vmware.api [None req-113af8fa-af69-464c-ac2e-e5cf96a3de06 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52169573-1b93-81f0-e1a7-1d9bbde5dc73, 'name': SearchDatastore_Task, 'duration_secs': 0.011831} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 980.252357] env[63515]: DEBUG oslo_concurrency.lockutils [None req-113af8fa-af69-464c-ac2e-e5cf96a3de06 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 980.252619] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-113af8fa-af69-464c-ac2e-e5cf96a3de06 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk to [datastore1] 61cf2f9c-5d35-46c4-a477-d2cef9fcd5ae/61cf2f9c-5d35-46c4-a477-d2cef9fcd5ae.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 980.252888] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e1caf73b-4d41-4f51-9aaa-254cd58df253 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.259437] env[63515]: DEBUG oslo_vmware.api [None req-113af8fa-af69-464c-ac2e-e5cf96a3de06 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Waiting for the task: (returnval){ [ 980.259437] env[63515]: value = "task-1111736" [ 980.259437] env[63515]: _type = "Task" [ 980.259437] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 980.271236] env[63515]: DEBUG oslo_vmware.api [None req-113af8fa-af69-464c-ac2e-e5cf96a3de06 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Task: {'id': task-1111736, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 980.281573] env[63515]: DEBUG oslo_concurrency.lockutils [None req-7c387ab3-2055-4517-9874-052fe02db033 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 980.309616] env[63515]: DEBUG nova.network.neutron [None req-adce163d-60b0-4e35-be4e-020840038d00 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: 724b05bc-ce4b-4f99-ad14-f3abea5567f2] Updating instance_info_cache with network_info: [{"id": "3dee43b6-2650-42e1-aa5a-4994a3ec7f05", "address": "fa:16:3e:89:06:d0", "network": {"id": "88dfb29d-41e0-4df7-8f70-e52cda05620c", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1355578400-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7d4ada9ec35f42b19c6480a9101d21a4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "357d2811-e990-4985-9f9e-b158d10d3699", "external-id": "nsx-vlan-transportzone-641", "segmentation_id": 641, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3dee43b6-26", "ovs_interfaceid": "3dee43b6-2650-42e1-aa5a-4994a3ec7f05", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 980.449029] env[63515]: DEBUG nova.scheduler.client.report [None req-978a15da-7dcc-413b-825d-e9c53aaf3163 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 980.519116] env[63515]: DEBUG oslo_concurrency.lockutils [None req-7d3c29b2-1cec-41c4-90f7-528df42289ec tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Lock "86b1ddcf-585c-41d9-84fe-17b794f1abbc" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.253s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 980.567186] env[63515]: DEBUG oslo_concurrency.lockutils [None req-e3adddd2-d173-427c-8a04-9143b135c23f tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Acquiring lock "86b1ddcf-585c-41d9-84fe-17b794f1abbc" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 980.567504] env[63515]: DEBUG oslo_concurrency.lockutils [None req-e3adddd2-d173-427c-8a04-9143b135c23f tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Lock "86b1ddcf-585c-41d9-84fe-17b794f1abbc" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.001s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 980.684170] env[63515]: INFO nova.compute.manager [None req-a5676663-6933-430d-aa5e-aeffa94052ee tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: 6a66dd9b-7b24-4b70-b58f-0830cf39376c] Took 19.84 seconds to build instance. [ 980.769226] env[63515]: DEBUG oslo_vmware.api [None req-113af8fa-af69-464c-ac2e-e5cf96a3de06 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Task: {'id': task-1111736, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.481865} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 980.769511] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-113af8fa-af69-464c-ac2e-e5cf96a3de06 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk to [datastore1] 61cf2f9c-5d35-46c4-a477-d2cef9fcd5ae/61cf2f9c-5d35-46c4-a477-d2cef9fcd5ae.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 980.769735] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-113af8fa-af69-464c-ac2e-e5cf96a3de06 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: 61cf2f9c-5d35-46c4-a477-d2cef9fcd5ae] Extending root virtual disk to 1048576 {{(pid=63515) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 980.770301] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-70bab428-c918-485c-824e-cbdb2a33ea91 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.778476] env[63515]: DEBUG oslo_vmware.api [None req-113af8fa-af69-464c-ac2e-e5cf96a3de06 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Waiting for the task: (returnval){ [ 980.778476] env[63515]: value = "task-1111737" [ 980.778476] env[63515]: _type = "Task" [ 980.778476] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 980.790117] env[63515]: DEBUG oslo_vmware.api [None req-113af8fa-af69-464c-ac2e-e5cf96a3de06 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Task: {'id': task-1111737, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 980.812897] env[63515]: DEBUG oslo_concurrency.lockutils [None req-adce163d-60b0-4e35-be4e-020840038d00 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Releasing lock "refresh_cache-724b05bc-ce4b-4f99-ad14-f3abea5567f2" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 980.813302] env[63515]: DEBUG nova.compute.manager [None req-adce163d-60b0-4e35-be4e-020840038d00 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: 724b05bc-ce4b-4f99-ad14-f3abea5567f2] Instance network_info: |[{"id": "3dee43b6-2650-42e1-aa5a-4994a3ec7f05", "address": "fa:16:3e:89:06:d0", "network": {"id": "88dfb29d-41e0-4df7-8f70-e52cda05620c", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1355578400-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7d4ada9ec35f42b19c6480a9101d21a4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "357d2811-e990-4985-9f9e-b158d10d3699", "external-id": "nsx-vlan-transportzone-641", "segmentation_id": 641, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3dee43b6-26", "ovs_interfaceid": "3dee43b6-2650-42e1-aa5a-4994a3ec7f05", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 980.813773] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-adce163d-60b0-4e35-be4e-020840038d00 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: 724b05bc-ce4b-4f99-ad14-f3abea5567f2] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:89:06:d0', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '357d2811-e990-4985-9f9e-b158d10d3699', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '3dee43b6-2650-42e1-aa5a-4994a3ec7f05', 'vif_model': 'vmxnet3'}] {{(pid=63515) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 980.822072] env[63515]: DEBUG oslo.service.loopingcall [None req-adce163d-60b0-4e35-be4e-020840038d00 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 980.822208] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 724b05bc-ce4b-4f99-ad14-f3abea5567f2] Creating VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 980.822424] env[63515]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-95375005-b046-4730-bb20-918cb3e07cdf {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.852938] env[63515]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 980.852938] env[63515]: value = "task-1111738" [ 980.852938] env[63515]: _type = "Task" [ 980.852938] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 980.865587] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1111738, 'name': CreateVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 980.953291] env[63515]: DEBUG oslo_concurrency.lockutils [None req-978a15da-7dcc-413b-825d-e9c53aaf3163 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.309s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 980.953910] env[63515]: DEBUG nova.compute.manager [None req-978a15da-7dcc-413b-825d-e9c53aaf3163 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: 461a0c4c-dfdb-4669-ac72-8913c0277298] Start building networks asynchronously for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 980.960707] env[63515]: DEBUG oslo_concurrency.lockutils [None req-16f0ea70-a973-4a3c-bc2e-292ffb4571ac tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 6.291s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 980.960707] env[63515]: DEBUG nova.objects.instance [None req-16f0ea70-a973-4a3c-bc2e-292ffb4571ac tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: 918d5b35-c46d-47c5-b59f-175831aed9d6] Trying to apply a migration context that does not seem to be set for this instance {{(pid=63515) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 981.071677] env[63515]: INFO nova.compute.manager [None req-e3adddd2-d173-427c-8a04-9143b135c23f tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] [instance: 86b1ddcf-585c-41d9-84fe-17b794f1abbc] Detaching volume faf85999-e3c6-4036-a18e-7d126676fad9 [ 981.115917] env[63515]: INFO nova.virt.block_device [None req-e3adddd2-d173-427c-8a04-9143b135c23f tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] [instance: 86b1ddcf-585c-41d9-84fe-17b794f1abbc] Attempting to driver detach volume faf85999-e3c6-4036-a18e-7d126676fad9 from mountpoint /dev/sdc [ 981.116195] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-e3adddd2-d173-427c-8a04-9143b135c23f tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] [instance: 86b1ddcf-585c-41d9-84fe-17b794f1abbc] Volume detach. Driver type: vmdk {{(pid=63515) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 981.116385] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-e3adddd2-d173-427c-8a04-9143b135c23f tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] [instance: 86b1ddcf-585c-41d9-84fe-17b794f1abbc] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-243550', 'volume_id': 'faf85999-e3c6-4036-a18e-7d126676fad9', 'name': 'volume-faf85999-e3c6-4036-a18e-7d126676fad9', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '86b1ddcf-585c-41d9-84fe-17b794f1abbc', 'attached_at': '', 'detached_at': '', 'volume_id': 'faf85999-e3c6-4036-a18e-7d126676fad9', 'serial': 'faf85999-e3c6-4036-a18e-7d126676fad9'} {{(pid=63515) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 981.117339] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25b8cedd-717a-467f-819d-889d1d211bf4 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.142876] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2aed5b47-4746-46f1-86f1-1c39e9bc0e3a {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.152264] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-627a234d-2b2f-49cd-84bd-7bfeec1a9a8f {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.173537] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0dbeca9-2a27-4ed8-9ccc-2c9d5240fa7e {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.192538] env[63515]: DEBUG oslo_concurrency.lockutils [None req-a5676663-6933-430d-aa5e-aeffa94052ee tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Lock "6a66dd9b-7b24-4b70-b58f-0830cf39376c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 21.358s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 981.193040] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-e3adddd2-d173-427c-8a04-9143b135c23f tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] The volume has not been displaced from its original location: [datastore1] volume-faf85999-e3c6-4036-a18e-7d126676fad9/volume-faf85999-e3c6-4036-a18e-7d126676fad9.vmdk. No consolidation needed. {{(pid=63515) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 981.199917] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-e3adddd2-d173-427c-8a04-9143b135c23f tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] [instance: 86b1ddcf-585c-41d9-84fe-17b794f1abbc] Reconfiguring VM instance instance-00000051 to detach disk 2002 {{(pid=63515) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 981.200413] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2d50d050-1f5d-440a-9e1a-1b1a97105a96 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.219720] env[63515]: DEBUG oslo_vmware.api [None req-e3adddd2-d173-427c-8a04-9143b135c23f tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Waiting for the task: (returnval){ [ 981.219720] env[63515]: value = "task-1111739" [ 981.219720] env[63515]: _type = "Task" [ 981.219720] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 981.228379] env[63515]: DEBUG oslo_vmware.api [None req-e3adddd2-d173-427c-8a04-9143b135c23f tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Task: {'id': task-1111739, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 981.287310] env[63515]: DEBUG oslo_vmware.api [None req-113af8fa-af69-464c-ac2e-e5cf96a3de06 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Task: {'id': task-1111737, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.069237} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 981.287598] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-113af8fa-af69-464c-ac2e-e5cf96a3de06 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: 61cf2f9c-5d35-46c4-a477-d2cef9fcd5ae] Extended root virtual disk {{(pid=63515) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 981.288385] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8cf80d7-ccf9-4332-8094-3ed8ef8b2bdd {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.309976] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-113af8fa-af69-464c-ac2e-e5cf96a3de06 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: 61cf2f9c-5d35-46c4-a477-d2cef9fcd5ae] Reconfiguring VM instance instance-0000005e to attach disk [datastore1] 61cf2f9c-5d35-46c4-a477-d2cef9fcd5ae/61cf2f9c-5d35-46c4-a477-d2cef9fcd5ae.vmdk or device None with type sparse {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 981.310290] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7ad9100f-a0b5-4659-965e-0d5a461ed235 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.329619] env[63515]: DEBUG oslo_vmware.api [None req-113af8fa-af69-464c-ac2e-e5cf96a3de06 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Waiting for the task: (returnval){ [ 981.329619] env[63515]: value = "task-1111740" [ 981.329619] env[63515]: _type = "Task" [ 981.329619] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 981.338460] env[63515]: DEBUG oslo_vmware.api [None req-113af8fa-af69-464c-ac2e-e5cf96a3de06 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Task: {'id': task-1111740, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 981.363048] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1111738, 'name': CreateVM_Task, 'duration_secs': 0.389589} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 981.363308] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 724b05bc-ce4b-4f99-ad14-f3abea5567f2] Created VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 981.364014] env[63515]: DEBUG oslo_concurrency.lockutils [None req-adce163d-60b0-4e35-be4e-020840038d00 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 981.364204] env[63515]: DEBUG oslo_concurrency.lockutils [None req-adce163d-60b0-4e35-be4e-020840038d00 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 981.364559] env[63515]: DEBUG oslo_concurrency.lockutils [None req-adce163d-60b0-4e35-be4e-020840038d00 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 981.364788] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-46812a94-81c3-4301-853d-90ed88daecca {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.369541] env[63515]: DEBUG oslo_vmware.api [None req-adce163d-60b0-4e35-be4e-020840038d00 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Waiting for the task: (returnval){ [ 981.369541] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]522825c6-2844-74a5-da85-0f125fefb068" [ 981.369541] env[63515]: _type = "Task" [ 981.369541] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 981.377951] env[63515]: DEBUG oslo_vmware.api [None req-adce163d-60b0-4e35-be4e-020840038d00 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]522825c6-2844-74a5-da85-0f125fefb068, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 981.467432] env[63515]: DEBUG nova.compute.utils [None req-978a15da-7dcc-413b-825d-e9c53aaf3163 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Using /dev/sd instead of None {{(pid=63515) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 981.471878] env[63515]: DEBUG nova.compute.manager [None req-978a15da-7dcc-413b-825d-e9c53aaf3163 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: 461a0c4c-dfdb-4669-ac72-8913c0277298] Allocating IP information in the background. {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 981.471878] env[63515]: DEBUG nova.network.neutron [None req-978a15da-7dcc-413b-825d-e9c53aaf3163 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: 461a0c4c-dfdb-4669-ac72-8913c0277298] allocate_for_instance() {{(pid=63515) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 981.542845] env[63515]: DEBUG nova.policy [None req-978a15da-7dcc-413b-825d-e9c53aaf3163 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9b19bb2d32d84c019541c3b2e711a202', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '299fdeff647f486390366d5bbf911518', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63515) authorize /opt/stack/nova/nova/policy.py:201}} [ 981.729285] env[63515]: DEBUG oslo_vmware.api [None req-e3adddd2-d173-427c-8a04-9143b135c23f tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Task: {'id': task-1111739, 'name': ReconfigVM_Task, 'duration_secs': 0.230201} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 981.729592] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-e3adddd2-d173-427c-8a04-9143b135c23f tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] [instance: 86b1ddcf-585c-41d9-84fe-17b794f1abbc] Reconfigured VM instance instance-00000051 to detach disk 2002 {{(pid=63515) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 981.734697] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3c56d288-e52f-434e-b392-04ce944ce317 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.751973] env[63515]: DEBUG oslo_vmware.api [None req-e3adddd2-d173-427c-8a04-9143b135c23f tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Waiting for the task: (returnval){ [ 981.751973] env[63515]: value = "task-1111741" [ 981.751973] env[63515]: _type = "Task" [ 981.751973] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 981.759689] env[63515]: DEBUG oslo_vmware.api [None req-e3adddd2-d173-427c-8a04-9143b135c23f tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Task: {'id': task-1111741, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 981.842370] env[63515]: DEBUG oslo_vmware.api [None req-113af8fa-af69-464c-ac2e-e5cf96a3de06 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Task: {'id': task-1111740, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 981.857150] env[63515]: DEBUG nova.network.neutron [None req-978a15da-7dcc-413b-825d-e9c53aaf3163 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: 461a0c4c-dfdb-4669-ac72-8913c0277298] Successfully created port: 2536d489-b430-4d50-ba2e-36edf5e42e17 {{(pid=63515) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 981.881918] env[63515]: DEBUG oslo_vmware.api [None req-adce163d-60b0-4e35-be4e-020840038d00 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]522825c6-2844-74a5-da85-0f125fefb068, 'name': SearchDatastore_Task, 'duration_secs': 0.009426} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 981.882461] env[63515]: DEBUG oslo_concurrency.lockutils [None req-adce163d-60b0-4e35-be4e-020840038d00 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 981.882523] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-adce163d-60b0-4e35-be4e-020840038d00 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: 724b05bc-ce4b-4f99-ad14-f3abea5567f2] Processing image 8a120570-cb06-4099-b262-554ca0ad15c5 {{(pid=63515) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 981.882713] env[63515]: DEBUG oslo_concurrency.lockutils [None req-adce163d-60b0-4e35-be4e-020840038d00 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 981.882863] env[63515]: DEBUG oslo_concurrency.lockutils [None req-adce163d-60b0-4e35-be4e-020840038d00 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 981.883051] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-adce163d-60b0-4e35-be4e-020840038d00 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 981.883321] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-56ace4f9-0598-4106-94e2-61a6c0d360be {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.891542] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-adce163d-60b0-4e35-be4e-020840038d00 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 981.891812] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-adce163d-60b0-4e35-be4e-020840038d00 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63515) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 981.892562] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4ffb9fc5-f18d-4169-a00a-f632e2c8a28e {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.897930] env[63515]: DEBUG oslo_vmware.api [None req-adce163d-60b0-4e35-be4e-020840038d00 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Waiting for the task: (returnval){ [ 981.897930] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]52bbd58b-3c09-9611-83f0-d33f51dd9964" [ 981.897930] env[63515]: _type = "Task" [ 981.897930] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 981.905937] env[63515]: DEBUG oslo_vmware.api [None req-adce163d-60b0-4e35-be4e-020840038d00 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52bbd58b-3c09-9611-83f0-d33f51dd9964, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 981.972749] env[63515]: DEBUG nova.compute.manager [None req-978a15da-7dcc-413b-825d-e9c53aaf3163 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: 461a0c4c-dfdb-4669-ac72-8913c0277298] Start building block device mappings for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 981.980380] env[63515]: DEBUG oslo_concurrency.lockutils [None req-16f0ea70-a973-4a3c-bc2e-292ffb4571ac tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.016s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 981.980380] env[63515]: DEBUG oslo_concurrency.lockutils [None req-e595b7dd-52b8-43f9-9fcf-6cb888cef16d tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 7.002s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 981.980380] env[63515]: DEBUG nova.objects.instance [None req-e595b7dd-52b8-43f9-9fcf-6cb888cef16d tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Lazy-loading 'resources' on Instance uuid 2a29f37d-e786-48d3-b126-4854d6e7c67c {{(pid=63515) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 982.262363] env[63515]: DEBUG oslo_vmware.api [None req-e3adddd2-d173-427c-8a04-9143b135c23f tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Task: {'id': task-1111741, 'name': ReconfigVM_Task, 'duration_secs': 0.145529} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 982.262650] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-e3adddd2-d173-427c-8a04-9143b135c23f tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] [instance: 86b1ddcf-585c-41d9-84fe-17b794f1abbc] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-243550', 'volume_id': 'faf85999-e3c6-4036-a18e-7d126676fad9', 'name': 'volume-faf85999-e3c6-4036-a18e-7d126676fad9', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '86b1ddcf-585c-41d9-84fe-17b794f1abbc', 'attached_at': '', 'detached_at': '', 'volume_id': 'faf85999-e3c6-4036-a18e-7d126676fad9', 'serial': 'faf85999-e3c6-4036-a18e-7d126676fad9'} {{(pid=63515) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 982.304058] env[63515]: DEBUG oslo_concurrency.lockutils [None req-505b9ae2-f7d6-4938-b4d4-0d6c86cd7e2f tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Acquiring lock "6a66dd9b-7b24-4b70-b58f-0830cf39376c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 982.304364] env[63515]: DEBUG oslo_concurrency.lockutils [None req-505b9ae2-f7d6-4938-b4d4-0d6c86cd7e2f tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Lock "6a66dd9b-7b24-4b70-b58f-0830cf39376c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 982.304586] env[63515]: DEBUG oslo_concurrency.lockutils [None req-505b9ae2-f7d6-4938-b4d4-0d6c86cd7e2f tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Acquiring lock "6a66dd9b-7b24-4b70-b58f-0830cf39376c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 982.304773] env[63515]: DEBUG oslo_concurrency.lockutils [None req-505b9ae2-f7d6-4938-b4d4-0d6c86cd7e2f tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Lock "6a66dd9b-7b24-4b70-b58f-0830cf39376c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 982.304946] env[63515]: DEBUG oslo_concurrency.lockutils [None req-505b9ae2-f7d6-4938-b4d4-0d6c86cd7e2f tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Lock "6a66dd9b-7b24-4b70-b58f-0830cf39376c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 982.307249] env[63515]: INFO nova.compute.manager [None req-505b9ae2-f7d6-4938-b4d4-0d6c86cd7e2f tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: 6a66dd9b-7b24-4b70-b58f-0830cf39376c] Terminating instance [ 982.309120] env[63515]: DEBUG nova.compute.manager [None req-505b9ae2-f7d6-4938-b4d4-0d6c86cd7e2f tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: 6a66dd9b-7b24-4b70-b58f-0830cf39376c] Start destroying the instance on the hypervisor. {{(pid=63515) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 982.309321] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-505b9ae2-f7d6-4938-b4d4-0d6c86cd7e2f tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: 6a66dd9b-7b24-4b70-b58f-0830cf39376c] Destroying instance {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 982.311635] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0231103-2ce9-4dcb-9413-fcba30c06af2 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.317425] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-505b9ae2-f7d6-4938-b4d4-0d6c86cd7e2f tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: 6a66dd9b-7b24-4b70-b58f-0830cf39376c] Powering off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 982.317715] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e134d80b-dfb2-477c-a085-a32f5eebdc38 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.323619] env[63515]: DEBUG oslo_vmware.api [None req-505b9ae2-f7d6-4938-b4d4-0d6c86cd7e2f tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Waiting for the task: (returnval){ [ 982.323619] env[63515]: value = "task-1111742" [ 982.323619] env[63515]: _type = "Task" [ 982.323619] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 982.331388] env[63515]: DEBUG oslo_vmware.api [None req-505b9ae2-f7d6-4938-b4d4-0d6c86cd7e2f tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Task: {'id': task-1111742, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 982.340257] env[63515]: DEBUG oslo_vmware.api [None req-113af8fa-af69-464c-ac2e-e5cf96a3de06 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Task: {'id': task-1111740, 'name': ReconfigVM_Task, 'duration_secs': 0.892624} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 982.340606] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-113af8fa-af69-464c-ac2e-e5cf96a3de06 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: 61cf2f9c-5d35-46c4-a477-d2cef9fcd5ae] Reconfigured VM instance instance-0000005e to attach disk [datastore1] 61cf2f9c-5d35-46c4-a477-d2cef9fcd5ae/61cf2f9c-5d35-46c4-a477-d2cef9fcd5ae.vmdk or device None with type sparse {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 982.341275] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f890819c-cac9-424d-9a55-5cb64128f222 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.346718] env[63515]: DEBUG oslo_vmware.api [None req-113af8fa-af69-464c-ac2e-e5cf96a3de06 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Waiting for the task: (returnval){ [ 982.346718] env[63515]: value = "task-1111743" [ 982.346718] env[63515]: _type = "Task" [ 982.346718] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 982.355253] env[63515]: DEBUG oslo_vmware.api [None req-113af8fa-af69-464c-ac2e-e5cf96a3de06 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Task: {'id': task-1111743, 'name': Rename_Task} progress is 5%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 982.409813] env[63515]: DEBUG oslo_vmware.api [None req-adce163d-60b0-4e35-be4e-020840038d00 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52bbd58b-3c09-9611-83f0-d33f51dd9964, 'name': SearchDatastore_Task, 'duration_secs': 0.008356} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 982.410651] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cd25eff9-9d00-494c-b985-9f84a990fd72 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.415982] env[63515]: DEBUG oslo_vmware.api [None req-adce163d-60b0-4e35-be4e-020840038d00 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Waiting for the task: (returnval){ [ 982.415982] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]52934635-7049-0f57-57db-52b1dfdfe601" [ 982.415982] env[63515]: _type = "Task" [ 982.415982] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 982.423761] env[63515]: DEBUG oslo_vmware.api [None req-adce163d-60b0-4e35-be4e-020840038d00 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52934635-7049-0f57-57db-52b1dfdfe601, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 982.687433] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8fec2564-89ad-444e-86c1-417473076fea {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.695085] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e06934a-8f18-48e7-8be8-4d62207fde6e {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.726376] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aedd3b7e-e8ed-4db9-92b3-dccb52ea0e37 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.733649] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6712ec5b-3c90-4a71-8577-b0db1f240cd3 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.746328] env[63515]: DEBUG nova.compute.provider_tree [None req-e595b7dd-52b8-43f9-9fcf-6cb888cef16d tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 982.813783] env[63515]: DEBUG nova.objects.instance [None req-e3adddd2-d173-427c-8a04-9143b135c23f tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Lazy-loading 'flavor' on Instance uuid 86b1ddcf-585c-41d9-84fe-17b794f1abbc {{(pid=63515) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 982.834576] env[63515]: DEBUG oslo_vmware.api [None req-505b9ae2-f7d6-4938-b4d4-0d6c86cd7e2f tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Task: {'id': task-1111742, 'name': PowerOffVM_Task, 'duration_secs': 0.205273} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 982.834849] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-505b9ae2-f7d6-4938-b4d4-0d6c86cd7e2f tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: 6a66dd9b-7b24-4b70-b58f-0830cf39376c] Powered off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 982.835055] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-505b9ae2-f7d6-4938-b4d4-0d6c86cd7e2f tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: 6a66dd9b-7b24-4b70-b58f-0830cf39376c] Unregistering the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 982.835304] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b5c83f75-3f23-45f6-99b3-7470a3c5e554 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.856519] env[63515]: DEBUG oslo_vmware.api [None req-113af8fa-af69-464c-ac2e-e5cf96a3de06 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Task: {'id': task-1111743, 'name': Rename_Task, 'duration_secs': 0.214078} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 982.856730] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-113af8fa-af69-464c-ac2e-e5cf96a3de06 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: 61cf2f9c-5d35-46c4-a477-d2cef9fcd5ae] Powering on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 982.856992] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0d29374d-f77f-41af-b846-ba48e386a70c {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.865134] env[63515]: DEBUG oslo_vmware.api [None req-113af8fa-af69-464c-ac2e-e5cf96a3de06 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Waiting for the task: (returnval){ [ 982.865134] env[63515]: value = "task-1111745" [ 982.865134] env[63515]: _type = "Task" [ 982.865134] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 982.873906] env[63515]: DEBUG oslo_vmware.api [None req-113af8fa-af69-464c-ac2e-e5cf96a3de06 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Task: {'id': task-1111745, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 982.901362] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-505b9ae2-f7d6-4938-b4d4-0d6c86cd7e2f tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: 6a66dd9b-7b24-4b70-b58f-0830cf39376c] Unregistered the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 982.901661] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-505b9ae2-f7d6-4938-b4d4-0d6c86cd7e2f tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: 6a66dd9b-7b24-4b70-b58f-0830cf39376c] Deleting contents of the VM from datastore datastore2 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 982.901889] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-505b9ae2-f7d6-4938-b4d4-0d6c86cd7e2f tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Deleting the datastore file [datastore2] 6a66dd9b-7b24-4b70-b58f-0830cf39376c {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 982.902182] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-48e00bd2-b9ea-4356-ba35-863237252f64 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.908299] env[63515]: DEBUG oslo_vmware.api [None req-505b9ae2-f7d6-4938-b4d4-0d6c86cd7e2f tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Waiting for the task: (returnval){ [ 982.908299] env[63515]: value = "task-1111746" [ 982.908299] env[63515]: _type = "Task" [ 982.908299] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 982.915909] env[63515]: DEBUG oslo_vmware.api [None req-505b9ae2-f7d6-4938-b4d4-0d6c86cd7e2f tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Task: {'id': task-1111746, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 982.924745] env[63515]: DEBUG oslo_vmware.api [None req-adce163d-60b0-4e35-be4e-020840038d00 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52934635-7049-0f57-57db-52b1dfdfe601, 'name': SearchDatastore_Task, 'duration_secs': 0.010557} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 982.925009] env[63515]: DEBUG oslo_concurrency.lockutils [None req-adce163d-60b0-4e35-be4e-020840038d00 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 982.925280] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-adce163d-60b0-4e35-be4e-020840038d00 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk to [datastore1] 724b05bc-ce4b-4f99-ad14-f3abea5567f2/724b05bc-ce4b-4f99-ad14-f3abea5567f2.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 982.925542] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2b48e2b2-0ecb-49e3-8d60-a3585fc13fc4 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.931186] env[63515]: DEBUG oslo_vmware.api [None req-adce163d-60b0-4e35-be4e-020840038d00 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Waiting for the task: (returnval){ [ 982.931186] env[63515]: value = "task-1111747" [ 982.931186] env[63515]: _type = "Task" [ 982.931186] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 982.938767] env[63515]: DEBUG oslo_vmware.api [None req-adce163d-60b0-4e35-be4e-020840038d00 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Task: {'id': task-1111747, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 982.986900] env[63515]: DEBUG nova.compute.manager [None req-978a15da-7dcc-413b-825d-e9c53aaf3163 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: 461a0c4c-dfdb-4669-ac72-8913c0277298] Start spawning the instance on the hypervisor. {{(pid=63515) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 983.019238] env[63515]: DEBUG nova.virt.hardware [None req-978a15da-7dcc-413b-825d-e9c53aaf3163 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T02:52:14Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T02:51:56Z,direct_url=,disk_format='vmdk',id=8a120570-cb06-4099-b262-554ca0ad15c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b569255cc43e42e7ae3f2b4ad37c6ef4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T02:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 983.019520] env[63515]: DEBUG nova.virt.hardware [None req-978a15da-7dcc-413b-825d-e9c53aaf3163 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Flavor limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 983.019725] env[63515]: DEBUG nova.virt.hardware [None req-978a15da-7dcc-413b-825d-e9c53aaf3163 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Image limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 983.019929] env[63515]: DEBUG nova.virt.hardware [None req-978a15da-7dcc-413b-825d-e9c53aaf3163 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Flavor pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 983.020095] env[63515]: DEBUG nova.virt.hardware [None req-978a15da-7dcc-413b-825d-e9c53aaf3163 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Image pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 983.020282] env[63515]: DEBUG nova.virt.hardware [None req-978a15da-7dcc-413b-825d-e9c53aaf3163 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 983.020503] env[63515]: DEBUG nova.virt.hardware [None req-978a15da-7dcc-413b-825d-e9c53aaf3163 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 983.020667] env[63515]: DEBUG nova.virt.hardware [None req-978a15da-7dcc-413b-825d-e9c53aaf3163 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 983.020857] env[63515]: DEBUG nova.virt.hardware [None req-978a15da-7dcc-413b-825d-e9c53aaf3163 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Got 1 possible topologies {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 983.021045] env[63515]: DEBUG nova.virt.hardware [None req-978a15da-7dcc-413b-825d-e9c53aaf3163 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 983.021230] env[63515]: DEBUG nova.virt.hardware [None req-978a15da-7dcc-413b-825d-e9c53aaf3163 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 983.022242] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-686ca638-5c65-49bd-a265-03bcb2888251 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.031397] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76e3de21-612f-4a82-b7b2-2c0bad48ccc0 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.250343] env[63515]: DEBUG nova.scheduler.client.report [None req-e595b7dd-52b8-43f9-9fcf-6cb888cef16d tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 983.383810] env[63515]: DEBUG oslo_vmware.api [None req-113af8fa-af69-464c-ac2e-e5cf96a3de06 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Task: {'id': task-1111745, 'name': PowerOnVM_Task} progress is 100%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.420495] env[63515]: DEBUG oslo_vmware.api [None req-505b9ae2-f7d6-4938-b4d4-0d6c86cd7e2f tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Task: {'id': task-1111746, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.154278} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 983.421099] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-505b9ae2-f7d6-4938-b4d4-0d6c86cd7e2f tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Deleted the datastore file {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 983.421406] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-505b9ae2-f7d6-4938-b4d4-0d6c86cd7e2f tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: 6a66dd9b-7b24-4b70-b58f-0830cf39376c] Deleted contents of the VM from datastore datastore2 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 983.421873] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-505b9ae2-f7d6-4938-b4d4-0d6c86cd7e2f tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: 6a66dd9b-7b24-4b70-b58f-0830cf39376c] Instance destroyed {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 983.422181] env[63515]: INFO nova.compute.manager [None req-505b9ae2-f7d6-4938-b4d4-0d6c86cd7e2f tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] [instance: 6a66dd9b-7b24-4b70-b58f-0830cf39376c] Took 1.11 seconds to destroy the instance on the hypervisor. [ 983.422647] env[63515]: DEBUG oslo.service.loopingcall [None req-505b9ae2-f7d6-4938-b4d4-0d6c86cd7e2f tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 983.423051] env[63515]: DEBUG nova.compute.manager [-] [instance: 6a66dd9b-7b24-4b70-b58f-0830cf39376c] Deallocating network for instance {{(pid=63515) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 983.423322] env[63515]: DEBUG nova.network.neutron [-] [instance: 6a66dd9b-7b24-4b70-b58f-0830cf39376c] deallocate_for_instance() {{(pid=63515) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 983.440652] env[63515]: DEBUG oslo_vmware.api [None req-adce163d-60b0-4e35-be4e-020840038d00 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Task: {'id': task-1111747, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.493613} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 983.441119] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-adce163d-60b0-4e35-be4e-020840038d00 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk to [datastore1] 724b05bc-ce4b-4f99-ad14-f3abea5567f2/724b05bc-ce4b-4f99-ad14-f3abea5567f2.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 983.441660] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-adce163d-60b0-4e35-be4e-020840038d00 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: 724b05bc-ce4b-4f99-ad14-f3abea5567f2] Extending root virtual disk to 1048576 {{(pid=63515) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 983.442255] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-8a69904a-5a5c-4e37-8e0d-15072f7cd152 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.451028] env[63515]: DEBUG oslo_vmware.api [None req-adce163d-60b0-4e35-be4e-020840038d00 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Waiting for the task: (returnval){ [ 983.451028] env[63515]: value = "task-1111748" [ 983.451028] env[63515]: _type = "Task" [ 983.451028] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 983.457930] env[63515]: DEBUG oslo_vmware.api [None req-adce163d-60b0-4e35-be4e-020840038d00 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Task: {'id': task-1111748, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.589980] env[63515]: DEBUG nova.network.neutron [None req-978a15da-7dcc-413b-825d-e9c53aaf3163 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: 461a0c4c-dfdb-4669-ac72-8913c0277298] Successfully updated port: 2536d489-b430-4d50-ba2e-36edf5e42e17 {{(pid=63515) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 983.677245] env[63515]: DEBUG nova.compute.manager [req-a90db913-6343-4a0f-9886-e9ad1d4d1e6e req-38d4f07f-7bb2-4d82-a608-94955cdcf3cc service nova] [instance: 461a0c4c-dfdb-4669-ac72-8913c0277298] Received event network-vif-plugged-2536d489-b430-4d50-ba2e-36edf5e42e17 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 983.677245] env[63515]: DEBUG oslo_concurrency.lockutils [req-a90db913-6343-4a0f-9886-e9ad1d4d1e6e req-38d4f07f-7bb2-4d82-a608-94955cdcf3cc service nova] Acquiring lock "461a0c4c-dfdb-4669-ac72-8913c0277298-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 983.678150] env[63515]: DEBUG oslo_concurrency.lockutils [req-a90db913-6343-4a0f-9886-e9ad1d4d1e6e req-38d4f07f-7bb2-4d82-a608-94955cdcf3cc service nova] Lock "461a0c4c-dfdb-4669-ac72-8913c0277298-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 983.678460] env[63515]: DEBUG oslo_concurrency.lockutils [req-a90db913-6343-4a0f-9886-e9ad1d4d1e6e req-38d4f07f-7bb2-4d82-a608-94955cdcf3cc service nova] Lock "461a0c4c-dfdb-4669-ac72-8913c0277298-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 983.678966] env[63515]: DEBUG nova.compute.manager [req-a90db913-6343-4a0f-9886-e9ad1d4d1e6e req-38d4f07f-7bb2-4d82-a608-94955cdcf3cc service nova] [instance: 461a0c4c-dfdb-4669-ac72-8913c0277298] No waiting events found dispatching network-vif-plugged-2536d489-b430-4d50-ba2e-36edf5e42e17 {{(pid=63515) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 983.679411] env[63515]: WARNING nova.compute.manager [req-a90db913-6343-4a0f-9886-e9ad1d4d1e6e req-38d4f07f-7bb2-4d82-a608-94955cdcf3cc service nova] [instance: 461a0c4c-dfdb-4669-ac72-8913c0277298] Received unexpected event network-vif-plugged-2536d489-b430-4d50-ba2e-36edf5e42e17 for instance with vm_state building and task_state spawning. [ 983.759020] env[63515]: DEBUG oslo_concurrency.lockutils [None req-e595b7dd-52b8-43f9-9fcf-6cb888cef16d tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.779s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 983.759020] env[63515]: DEBUG oslo_concurrency.lockutils [None req-ee146559-ead8-41ee-8bac-7272cc90bf4b tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 5.380s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 983.759733] env[63515]: DEBUG nova.objects.instance [None req-ee146559-ead8-41ee-8bac-7272cc90bf4b tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Lazy-loading 'resources' on Instance uuid 918d5b35-c46d-47c5-b59f-175831aed9d6 {{(pid=63515) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 983.789711] env[63515]: INFO nova.scheduler.client.report [None req-e595b7dd-52b8-43f9-9fcf-6cb888cef16d tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Deleted allocations for instance 2a29f37d-e786-48d3-b126-4854d6e7c67c [ 983.822515] env[63515]: DEBUG oslo_concurrency.lockutils [None req-e3adddd2-d173-427c-8a04-9143b135c23f tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Lock "86b1ddcf-585c-41d9-84fe-17b794f1abbc" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.255s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 983.875648] env[63515]: DEBUG nova.compute.manager [req-d9f65f22-14a7-4f7f-9511-dac0ea50c7f6 req-d30dfe0c-fedb-4ea1-8550-f1a604d2db93 service nova] [instance: 6a66dd9b-7b24-4b70-b58f-0830cf39376c] Received event network-vif-deleted-b190896f-9020-45f3-bb87-f1fe6401b017 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 983.877019] env[63515]: INFO nova.compute.manager [req-d9f65f22-14a7-4f7f-9511-dac0ea50c7f6 req-d30dfe0c-fedb-4ea1-8550-f1a604d2db93 service nova] [instance: 6a66dd9b-7b24-4b70-b58f-0830cf39376c] Neutron deleted interface b190896f-9020-45f3-bb87-f1fe6401b017; detaching it from the instance and deleting it from the info cache [ 983.877019] env[63515]: DEBUG nova.network.neutron [req-d9f65f22-14a7-4f7f-9511-dac0ea50c7f6 req-d30dfe0c-fedb-4ea1-8550-f1a604d2db93 service nova] [instance: 6a66dd9b-7b24-4b70-b58f-0830cf39376c] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 983.881968] env[63515]: DEBUG oslo_vmware.api [None req-113af8fa-af69-464c-ac2e-e5cf96a3de06 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Task: {'id': task-1111745, 'name': PowerOnVM_Task, 'duration_secs': 0.515634} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 983.882651] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-113af8fa-af69-464c-ac2e-e5cf96a3de06 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: 61cf2f9c-5d35-46c4-a477-d2cef9fcd5ae] Powered on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 983.882879] env[63515]: INFO nova.compute.manager [None req-113af8fa-af69-464c-ac2e-e5cf96a3de06 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: 61cf2f9c-5d35-46c4-a477-d2cef9fcd5ae] Took 9.22 seconds to spawn the instance on the hypervisor. [ 983.883083] env[63515]: DEBUG nova.compute.manager [None req-113af8fa-af69-464c-ac2e-e5cf96a3de06 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: 61cf2f9c-5d35-46c4-a477-d2cef9fcd5ae] Checking state {{(pid=63515) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 983.884079] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92faa37c-71fa-4652-9dca-7eb94c41cfb9 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.960396] env[63515]: DEBUG oslo_vmware.api [None req-adce163d-60b0-4e35-be4e-020840038d00 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Task: {'id': task-1111748, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.072836} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 983.960667] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-adce163d-60b0-4e35-be4e-020840038d00 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: 724b05bc-ce4b-4f99-ad14-f3abea5567f2] Extended root virtual disk {{(pid=63515) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 983.961430] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90d16b8e-88f0-4ffe-9b99-f97c9048d42f {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.988884] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-adce163d-60b0-4e35-be4e-020840038d00 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: 724b05bc-ce4b-4f99-ad14-f3abea5567f2] Reconfiguring VM instance instance-0000005f to attach disk [datastore1] 724b05bc-ce4b-4f99-ad14-f3abea5567f2/724b05bc-ce4b-4f99-ad14-f3abea5567f2.vmdk or device None with type sparse {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 983.989355] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cdd62f1d-50a3-4e41-9c80-19c7e3a939ed {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.015326] env[63515]: DEBUG oslo_vmware.api [None req-adce163d-60b0-4e35-be4e-020840038d00 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Waiting for the task: (returnval){ [ 984.015326] env[63515]: value = "task-1111749" [ 984.015326] env[63515]: _type = "Task" [ 984.015326] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 984.025314] env[63515]: DEBUG oslo_vmware.api [None req-adce163d-60b0-4e35-be4e-020840038d00 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Task: {'id': task-1111749, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.095265] env[63515]: DEBUG oslo_concurrency.lockutils [None req-978a15da-7dcc-413b-825d-e9c53aaf3163 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Acquiring lock "refresh_cache-461a0c4c-dfdb-4669-ac72-8913c0277298" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 984.095419] env[63515]: DEBUG oslo_concurrency.lockutils [None req-978a15da-7dcc-413b-825d-e9c53aaf3163 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Acquired lock "refresh_cache-461a0c4c-dfdb-4669-ac72-8913c0277298" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 984.095572] env[63515]: DEBUG nova.network.neutron [None req-978a15da-7dcc-413b-825d-e9c53aaf3163 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: 461a0c4c-dfdb-4669-ac72-8913c0277298] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 984.298174] env[63515]: DEBUG oslo_concurrency.lockutils [None req-e595b7dd-52b8-43f9-9fcf-6cb888cef16d tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Lock "2a29f37d-e786-48d3-b126-4854d6e7c67c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 13.332s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 984.348699] env[63515]: DEBUG oslo_concurrency.lockutils [None req-5f7eade2-bb20-4c09-b233-6856c972d63b tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Acquiring lock "86b1ddcf-585c-41d9-84fe-17b794f1abbc" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 984.349089] env[63515]: DEBUG oslo_concurrency.lockutils [None req-5f7eade2-bb20-4c09-b233-6856c972d63b tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Lock "86b1ddcf-585c-41d9-84fe-17b794f1abbc" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 984.349439] env[63515]: DEBUG oslo_concurrency.lockutils [None req-5f7eade2-bb20-4c09-b233-6856c972d63b tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Acquiring lock "86b1ddcf-585c-41d9-84fe-17b794f1abbc-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 984.349763] env[63515]: DEBUG oslo_concurrency.lockutils [None req-5f7eade2-bb20-4c09-b233-6856c972d63b tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Lock "86b1ddcf-585c-41d9-84fe-17b794f1abbc-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 984.350057] env[63515]: DEBUG oslo_concurrency.lockutils [None req-5f7eade2-bb20-4c09-b233-6856c972d63b tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Lock "86b1ddcf-585c-41d9-84fe-17b794f1abbc-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 984.357681] env[63515]: DEBUG nova.network.neutron [-] [instance: 6a66dd9b-7b24-4b70-b58f-0830cf39376c] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 984.359468] env[63515]: INFO nova.compute.manager [None req-5f7eade2-bb20-4c09-b233-6856c972d63b tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] [instance: 86b1ddcf-585c-41d9-84fe-17b794f1abbc] Terminating instance [ 984.361578] env[63515]: DEBUG nova.compute.manager [None req-5f7eade2-bb20-4c09-b233-6856c972d63b tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] [instance: 86b1ddcf-585c-41d9-84fe-17b794f1abbc] Start destroying the instance on the hypervisor. {{(pid=63515) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 984.361723] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-5f7eade2-bb20-4c09-b233-6856c972d63b tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] [instance: 86b1ddcf-585c-41d9-84fe-17b794f1abbc] Destroying instance {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 984.362927] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6a81ef1-bb5d-4a89-b39d-dbbf3bba49de {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.371661] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-5f7eade2-bb20-4c09-b233-6856c972d63b tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] [instance: 86b1ddcf-585c-41d9-84fe-17b794f1abbc] Powering off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 984.371897] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-26d5172e-ae1f-49fb-a0b0-45cc5d712d76 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.381668] env[63515]: DEBUG oslo_vmware.api [None req-5f7eade2-bb20-4c09-b233-6856c972d63b tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Waiting for the task: (returnval){ [ 984.381668] env[63515]: value = "task-1111750" [ 984.381668] env[63515]: _type = "Task" [ 984.381668] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 984.386178] env[63515]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-bfa620c6-3137-4ed6-984d-af79aa2257b3 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.393489] env[63515]: DEBUG oslo_vmware.api [None req-5f7eade2-bb20-4c09-b233-6856c972d63b tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Task: {'id': task-1111750, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.402382] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b250574-5e5b-4878-8b50-f2574c613218 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.417819] env[63515]: INFO nova.compute.manager [None req-113af8fa-af69-464c-ac2e-e5cf96a3de06 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: 61cf2f9c-5d35-46c4-a477-d2cef9fcd5ae] Took 19.95 seconds to build instance. [ 984.439282] env[63515]: DEBUG nova.compute.manager [req-d9f65f22-14a7-4f7f-9511-dac0ea50c7f6 req-d30dfe0c-fedb-4ea1-8550-f1a604d2db93 service nova] [instance: 6a66dd9b-7b24-4b70-b58f-0830cf39376c] Detach interface failed, port_id=b190896f-9020-45f3-bb87-f1fe6401b017, reason: Instance 6a66dd9b-7b24-4b70-b58f-0830cf39376c could not be found. {{(pid=63515) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 984.529038] env[63515]: DEBUG oslo_vmware.api [None req-adce163d-60b0-4e35-be4e-020840038d00 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Task: {'id': task-1111749, 'name': ReconfigVM_Task, 'duration_secs': 0.293756} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 984.530523] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-adce163d-60b0-4e35-be4e-020840038d00 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: 724b05bc-ce4b-4f99-ad14-f3abea5567f2] Reconfigured VM instance instance-0000005f to attach disk [datastore1] 724b05bc-ce4b-4f99-ad14-f3abea5567f2/724b05bc-ce4b-4f99-ad14-f3abea5567f2.vmdk or device None with type sparse {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 984.531862] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3320cc8e-ba6a-4888-a314-367b43a8a380 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.536017] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-3c907047-e2bf-48e1-8a7e-c339a6b6af56 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.546733] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-488c97b0-bf38-433b-9ace-8b163e6bec5a {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.551933] env[63515]: DEBUG oslo_vmware.api [None req-adce163d-60b0-4e35-be4e-020840038d00 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Waiting for the task: (returnval){ [ 984.551933] env[63515]: value = "task-1111751" [ 984.551933] env[63515]: _type = "Task" [ 984.551933] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 984.585078] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98268aa7-045c-4abf-98cd-bfd4bcc85762 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.592719] env[63515]: DEBUG oslo_vmware.api [None req-adce163d-60b0-4e35-be4e-020840038d00 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Task: {'id': task-1111751, 'name': Rename_Task} progress is 14%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.597731] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4ac3dc7-187f-43a8-897d-9166cdb77d5c {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.615410] env[63515]: DEBUG nova.compute.provider_tree [None req-ee146559-ead8-41ee-8bac-7272cc90bf4b tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Updating inventory in ProviderTree for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 984.648538] env[63515]: DEBUG nova.network.neutron [None req-978a15da-7dcc-413b-825d-e9c53aaf3163 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: 461a0c4c-dfdb-4669-ac72-8913c0277298] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 984.862435] env[63515]: INFO nova.compute.manager [-] [instance: 6a66dd9b-7b24-4b70-b58f-0830cf39376c] Took 1.44 seconds to deallocate network for instance. [ 984.863257] env[63515]: DEBUG nova.network.neutron [None req-978a15da-7dcc-413b-825d-e9c53aaf3163 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: 461a0c4c-dfdb-4669-ac72-8913c0277298] Updating instance_info_cache with network_info: [{"id": "2536d489-b430-4d50-ba2e-36edf5e42e17", "address": "fa:16:3e:e6:bc:df", "network": {"id": "581d448e-be1b-4d5b-8a6e-c27326aaffc5", "bridge": "br-int", "label": "tempest-ServersTestJSON-1769610073-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "299fdeff647f486390366d5bbf911518", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fb143ef7-8271-4a8a-a4aa-8eba9a89f6a1", "external-id": "nsx-vlan-transportzone-504", "segmentation_id": 504, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2536d489-b4", "ovs_interfaceid": "2536d489-b430-4d50-ba2e-36edf5e42e17", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 984.893619] env[63515]: DEBUG oslo_vmware.api [None req-5f7eade2-bb20-4c09-b233-6856c972d63b tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Task: {'id': task-1111750, 'name': PowerOffVM_Task, 'duration_secs': 0.299392} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 984.893928] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-5f7eade2-bb20-4c09-b233-6856c972d63b tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] [instance: 86b1ddcf-585c-41d9-84fe-17b794f1abbc] Powered off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 984.894120] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-5f7eade2-bb20-4c09-b233-6856c972d63b tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] [instance: 86b1ddcf-585c-41d9-84fe-17b794f1abbc] Unregistering the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 984.894384] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3f4ea850-411d-443d-a315-f06bd1e66608 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.920214] env[63515]: DEBUG oslo_concurrency.lockutils [None req-113af8fa-af69-464c-ac2e-e5cf96a3de06 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Lock "61cf2f9c-5d35-46c4-a477-d2cef9fcd5ae" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 21.468s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 984.961021] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-5f7eade2-bb20-4c09-b233-6856c972d63b tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] [instance: 86b1ddcf-585c-41d9-84fe-17b794f1abbc] Unregistered the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 984.961021] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-5f7eade2-bb20-4c09-b233-6856c972d63b tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] [instance: 86b1ddcf-585c-41d9-84fe-17b794f1abbc] Deleting contents of the VM from datastore datastore2 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 984.961021] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-5f7eade2-bb20-4c09-b233-6856c972d63b tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Deleting the datastore file [datastore2] 86b1ddcf-585c-41d9-84fe-17b794f1abbc {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 984.961021] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b7d1ff44-7f47-4d93-ba79-f24781f24d5c {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.970494] env[63515]: DEBUG oslo_vmware.api [None req-5f7eade2-bb20-4c09-b233-6856c972d63b tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Waiting for the task: (returnval){ [ 984.970494] env[63515]: value = "task-1111753" [ 984.970494] env[63515]: _type = "Task" [ 984.970494] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 984.978647] env[63515]: DEBUG oslo_vmware.api [None req-5f7eade2-bb20-4c09-b233-6856c972d63b tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Task: {'id': task-1111753, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.064329] env[63515]: DEBUG oslo_vmware.api [None req-adce163d-60b0-4e35-be4e-020840038d00 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Task: {'id': task-1111751, 'name': Rename_Task, 'duration_secs': 0.150641} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 985.064636] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-adce163d-60b0-4e35-be4e-020840038d00 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: 724b05bc-ce4b-4f99-ad14-f3abea5567f2] Powering on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 985.064923] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b6ad8e39-d931-41bb-a31d-70c2478360fd {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.071730] env[63515]: DEBUG oslo_vmware.api [None req-adce163d-60b0-4e35-be4e-020840038d00 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Waiting for the task: (returnval){ [ 985.071730] env[63515]: value = "task-1111754" [ 985.071730] env[63515]: _type = "Task" [ 985.071730] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 985.080637] env[63515]: DEBUG oslo_vmware.api [None req-adce163d-60b0-4e35-be4e-020840038d00 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Task: {'id': task-1111754, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.147312] env[63515]: ERROR nova.scheduler.client.report [None req-ee146559-ead8-41ee-8bac-7272cc90bf4b tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [req-a168b228-96f1-4ea8-9fb1-20cf7da23f6d] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 170, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 2bd86232-2b6f-44d5-9057-1a3a6b27185a. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-a168b228-96f1-4ea8-9fb1-20cf7da23f6d"}]} [ 985.161976] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2f334d7-cb30-4622-84e2-2d957fc494b1 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.165629] env[63515]: DEBUG nova.scheduler.client.report [None req-ee146559-ead8-41ee-8bac-7272cc90bf4b tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Refreshing inventories for resource provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 985.172874] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-57eda9b3-5939-4c13-a654-1f2f06887098 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: 61cf2f9c-5d35-46c4-a477-d2cef9fcd5ae] Suspending the VM {{(pid=63515) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1162}} [ 985.172874] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-ae33256a-6347-42eb-9deb-1f078ff99dfd {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.179042] env[63515]: DEBUG oslo_vmware.api [None req-57eda9b3-5939-4c13-a654-1f2f06887098 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Waiting for the task: (returnval){ [ 985.179042] env[63515]: value = "task-1111755" [ 985.179042] env[63515]: _type = "Task" [ 985.179042] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 985.187226] env[63515]: DEBUG oslo_vmware.api [None req-57eda9b3-5939-4c13-a654-1f2f06887098 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Task: {'id': task-1111755, 'name': SuspendVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.193111] env[63515]: DEBUG nova.scheduler.client.report [None req-ee146559-ead8-41ee-8bac-7272cc90bf4b tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Updating ProviderTree inventory for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 985.193332] env[63515]: DEBUG nova.compute.provider_tree [None req-ee146559-ead8-41ee-8bac-7272cc90bf4b tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Updating inventory in ProviderTree for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 985.211211] env[63515]: DEBUG nova.scheduler.client.report [None req-ee146559-ead8-41ee-8bac-7272cc90bf4b tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Refreshing aggregate associations for resource provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a, aggregates: None {{(pid=63515) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 985.239780] env[63515]: DEBUG nova.scheduler.client.report [None req-ee146559-ead8-41ee-8bac-7272cc90bf4b tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Refreshing trait associations for resource provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a, traits: COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=63515) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 985.366744] env[63515]: DEBUG oslo_concurrency.lockutils [None req-978a15da-7dcc-413b-825d-e9c53aaf3163 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Releasing lock "refresh_cache-461a0c4c-dfdb-4669-ac72-8913c0277298" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 985.367079] env[63515]: DEBUG nova.compute.manager [None req-978a15da-7dcc-413b-825d-e9c53aaf3163 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: 461a0c4c-dfdb-4669-ac72-8913c0277298] Instance network_info: |[{"id": "2536d489-b430-4d50-ba2e-36edf5e42e17", "address": "fa:16:3e:e6:bc:df", "network": {"id": "581d448e-be1b-4d5b-8a6e-c27326aaffc5", "bridge": "br-int", "label": "tempest-ServersTestJSON-1769610073-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "299fdeff647f486390366d5bbf911518", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fb143ef7-8271-4a8a-a4aa-8eba9a89f6a1", "external-id": "nsx-vlan-transportzone-504", "segmentation_id": 504, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2536d489-b4", "ovs_interfaceid": "2536d489-b430-4d50-ba2e-36edf5e42e17", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 985.367771] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-978a15da-7dcc-413b-825d-e9c53aaf3163 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: 461a0c4c-dfdb-4669-ac72-8913c0277298] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:e6:bc:df', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'fb143ef7-8271-4a8a-a4aa-8eba9a89f6a1', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '2536d489-b430-4d50-ba2e-36edf5e42e17', 'vif_model': 'vmxnet3'}] {{(pid=63515) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 985.377566] env[63515]: DEBUG oslo.service.loopingcall [None req-978a15da-7dcc-413b-825d-e9c53aaf3163 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 985.378873] env[63515]: DEBUG oslo_concurrency.lockutils [None req-505b9ae2-f7d6-4938-b4d4-0d6c86cd7e2f tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 985.378972] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 461a0c4c-dfdb-4669-ac72-8913c0277298] Creating VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 985.379298] env[63515]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-14fd050a-ecf8-40f7-9b66-1ae501f55ffb {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.401876] env[63515]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 985.401876] env[63515]: value = "task-1111756" [ 985.401876] env[63515]: _type = "Task" [ 985.401876] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 985.420663] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1111756, 'name': CreateVM_Task} progress is 6%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.483910] env[63515]: DEBUG oslo_vmware.api [None req-5f7eade2-bb20-4c09-b233-6856c972d63b tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Task: {'id': task-1111753, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.178041} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 985.485058] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-5f7eade2-bb20-4c09-b233-6856c972d63b tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Deleted the datastore file {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 985.485324] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-5f7eade2-bb20-4c09-b233-6856c972d63b tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] [instance: 86b1ddcf-585c-41d9-84fe-17b794f1abbc] Deleted contents of the VM from datastore datastore2 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 985.485510] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-5f7eade2-bb20-4c09-b233-6856c972d63b tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] [instance: 86b1ddcf-585c-41d9-84fe-17b794f1abbc] Instance destroyed {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 985.485692] env[63515]: INFO nova.compute.manager [None req-5f7eade2-bb20-4c09-b233-6856c972d63b tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] [instance: 86b1ddcf-585c-41d9-84fe-17b794f1abbc] Took 1.12 seconds to destroy the instance on the hypervisor. [ 985.485952] env[63515]: DEBUG oslo.service.loopingcall [None req-5f7eade2-bb20-4c09-b233-6856c972d63b tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 985.486198] env[63515]: DEBUG nova.compute.manager [-] [instance: 86b1ddcf-585c-41d9-84fe-17b794f1abbc] Deallocating network for instance {{(pid=63515) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 985.486301] env[63515]: DEBUG nova.network.neutron [-] [instance: 86b1ddcf-585c-41d9-84fe-17b794f1abbc] deallocate_for_instance() {{(pid=63515) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 985.493048] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4c666b8-46e9-4b34-a971-42241018aa28 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.499064] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ebf07503-d6ca-4c8c-b92c-60b00728ce5e {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.537681] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69a58e32-5312-4c31-ab67-f2b4a9181ec7 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.547027] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9236a361-d9ef-48d9-8396-8c6fd65f3f76 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.563232] env[63515]: DEBUG nova.compute.provider_tree [None req-ee146559-ead8-41ee-8bac-7272cc90bf4b tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 985.583088] env[63515]: DEBUG oslo_vmware.api [None req-adce163d-60b0-4e35-be4e-020840038d00 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Task: {'id': task-1111754, 'name': PowerOnVM_Task} progress is 78%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.688914] env[63515]: DEBUG oslo_vmware.api [None req-57eda9b3-5939-4c13-a654-1f2f06887098 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Task: {'id': task-1111755, 'name': SuspendVM_Task} progress is 54%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.709854] env[63515]: DEBUG nova.compute.manager [req-2b07d193-1ae8-4c4a-acab-39751c22ccef req-3e3b586a-91f8-46b8-bacb-92cf72da0bc5 service nova] [instance: 461a0c4c-dfdb-4669-ac72-8913c0277298] Received event network-changed-2536d489-b430-4d50-ba2e-36edf5e42e17 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 985.710062] env[63515]: DEBUG nova.compute.manager [req-2b07d193-1ae8-4c4a-acab-39751c22ccef req-3e3b586a-91f8-46b8-bacb-92cf72da0bc5 service nova] [instance: 461a0c4c-dfdb-4669-ac72-8913c0277298] Refreshing instance network info cache due to event network-changed-2536d489-b430-4d50-ba2e-36edf5e42e17. {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 985.710281] env[63515]: DEBUG oslo_concurrency.lockutils [req-2b07d193-1ae8-4c4a-acab-39751c22ccef req-3e3b586a-91f8-46b8-bacb-92cf72da0bc5 service nova] Acquiring lock "refresh_cache-461a0c4c-dfdb-4669-ac72-8913c0277298" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 985.710482] env[63515]: DEBUG oslo_concurrency.lockutils [req-2b07d193-1ae8-4c4a-acab-39751c22ccef req-3e3b586a-91f8-46b8-bacb-92cf72da0bc5 service nova] Acquired lock "refresh_cache-461a0c4c-dfdb-4669-ac72-8913c0277298" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 985.710595] env[63515]: DEBUG nova.network.neutron [req-2b07d193-1ae8-4c4a-acab-39751c22ccef req-3e3b586a-91f8-46b8-bacb-92cf72da0bc5 service nova] [instance: 461a0c4c-dfdb-4669-ac72-8913c0277298] Refreshing network info cache for port 2536d489-b430-4d50-ba2e-36edf5e42e17 {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 985.913924] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1111756, 'name': CreateVM_Task, 'duration_secs': 0.461617} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 985.914273] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 461a0c4c-dfdb-4669-ac72-8913c0277298] Created VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 985.915031] env[63515]: DEBUG oslo_concurrency.lockutils [None req-978a15da-7dcc-413b-825d-e9c53aaf3163 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 985.915228] env[63515]: DEBUG oslo_concurrency.lockutils [None req-978a15da-7dcc-413b-825d-e9c53aaf3163 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 985.916265] env[63515]: DEBUG oslo_concurrency.lockutils [None req-978a15da-7dcc-413b-825d-e9c53aaf3163 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 985.916566] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d9ea68c5-2b7c-4a50-a9a6-e6442fd04671 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.923269] env[63515]: DEBUG oslo_vmware.api [None req-978a15da-7dcc-413b-825d-e9c53aaf3163 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Waiting for the task: (returnval){ [ 985.923269] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]529d761d-d0db-ad89-dfd1-fb44c0acb9ef" [ 985.923269] env[63515]: _type = "Task" [ 985.923269] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 985.932930] env[63515]: DEBUG oslo_vmware.api [None req-978a15da-7dcc-413b-825d-e9c53aaf3163 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]529d761d-d0db-ad89-dfd1-fb44c0acb9ef, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.071419] env[63515]: DEBUG nova.scheduler.client.report [None req-ee146559-ead8-41ee-8bac-7272cc90bf4b tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 986.089545] env[63515]: DEBUG oslo_vmware.api [None req-adce163d-60b0-4e35-be4e-020840038d00 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Task: {'id': task-1111754, 'name': PowerOnVM_Task, 'duration_secs': 0.81847} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 986.090071] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-adce163d-60b0-4e35-be4e-020840038d00 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: 724b05bc-ce4b-4f99-ad14-f3abea5567f2] Powered on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 986.090424] env[63515]: INFO nova.compute.manager [None req-adce163d-60b0-4e35-be4e-020840038d00 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: 724b05bc-ce4b-4f99-ad14-f3abea5567f2] Took 9.09 seconds to spawn the instance on the hypervisor. [ 986.090715] env[63515]: DEBUG nova.compute.manager [None req-adce163d-60b0-4e35-be4e-020840038d00 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: 724b05bc-ce4b-4f99-ad14-f3abea5567f2] Checking state {{(pid=63515) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 986.091724] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9e55588-140a-4adc-8212-bccd4398b89e {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.193301] env[63515]: DEBUG oslo_vmware.api [None req-57eda9b3-5939-4c13-a654-1f2f06887098 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Task: {'id': task-1111755, 'name': SuspendVM_Task, 'duration_secs': 0.727596} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 986.193301] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-57eda9b3-5939-4c13-a654-1f2f06887098 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: 61cf2f9c-5d35-46c4-a477-d2cef9fcd5ae] Suspended the VM {{(pid=63515) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1166}} [ 986.193301] env[63515]: DEBUG nova.compute.manager [None req-57eda9b3-5939-4c13-a654-1f2f06887098 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: 61cf2f9c-5d35-46c4-a477-d2cef9fcd5ae] Checking state {{(pid=63515) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 986.193301] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57114a14-e686-4253-af4e-acbe2f58a3bc {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.270045] env[63515]: DEBUG nova.compute.manager [req-9f5a7dfc-3205-4d8a-9ab0-e1c07da1093d req-a7f8fea4-9011-41b0-a700-f2d433f02867 service nova] [instance: 86b1ddcf-585c-41d9-84fe-17b794f1abbc] Received event network-vif-deleted-43c4b108-fec8-46ea-942f-d3e734fe2be6 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 986.270346] env[63515]: INFO nova.compute.manager [req-9f5a7dfc-3205-4d8a-9ab0-e1c07da1093d req-a7f8fea4-9011-41b0-a700-f2d433f02867 service nova] [instance: 86b1ddcf-585c-41d9-84fe-17b794f1abbc] Neutron deleted interface 43c4b108-fec8-46ea-942f-d3e734fe2be6; detaching it from the instance and deleting it from the info cache [ 986.270573] env[63515]: DEBUG nova.network.neutron [req-9f5a7dfc-3205-4d8a-9ab0-e1c07da1093d req-a7f8fea4-9011-41b0-a700-f2d433f02867 service nova] [instance: 86b1ddcf-585c-41d9-84fe-17b794f1abbc] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 986.440101] env[63515]: DEBUG oslo_vmware.api [None req-978a15da-7dcc-413b-825d-e9c53aaf3163 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]529d761d-d0db-ad89-dfd1-fb44c0acb9ef, 'name': SearchDatastore_Task, 'duration_secs': 0.019137} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 986.440101] env[63515]: DEBUG oslo_concurrency.lockutils [None req-978a15da-7dcc-413b-825d-e9c53aaf3163 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 986.440101] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-978a15da-7dcc-413b-825d-e9c53aaf3163 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: 461a0c4c-dfdb-4669-ac72-8913c0277298] Processing image 8a120570-cb06-4099-b262-554ca0ad15c5 {{(pid=63515) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 986.441152] env[63515]: DEBUG oslo_concurrency.lockutils [None req-978a15da-7dcc-413b-825d-e9c53aaf3163 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 986.441355] env[63515]: DEBUG oslo_concurrency.lockutils [None req-978a15da-7dcc-413b-825d-e9c53aaf3163 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 986.441594] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-978a15da-7dcc-413b-825d-e9c53aaf3163 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 986.441917] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d1f1a831-b647-4ab0-8d60-954f5aeb337a {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.455071] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-978a15da-7dcc-413b-825d-e9c53aaf3163 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 986.458539] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-978a15da-7dcc-413b-825d-e9c53aaf3163 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63515) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 986.459393] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-81331e1b-10da-41cf-a6e1-9b2acf6e87cb {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.468442] env[63515]: DEBUG oslo_vmware.api [None req-978a15da-7dcc-413b-825d-e9c53aaf3163 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Waiting for the task: (returnval){ [ 986.468442] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]520b0aa0-f878-84d9-e50c-7fabc3a0b66c" [ 986.468442] env[63515]: _type = "Task" [ 986.468442] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 986.477740] env[63515]: DEBUG oslo_vmware.api [None req-978a15da-7dcc-413b-825d-e9c53aaf3163 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]520b0aa0-f878-84d9-e50c-7fabc3a0b66c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.579230] env[63515]: DEBUG oslo_concurrency.lockutils [None req-ee146559-ead8-41ee-8bac-7272cc90bf4b tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.820s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 986.583343] env[63515]: DEBUG oslo_concurrency.lockutils [None req-7c387ab3-2055-4517-9874-052fe02db033 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 6.302s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 986.614916] env[63515]: INFO nova.compute.manager [None req-adce163d-60b0-4e35-be4e-020840038d00 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: 724b05bc-ce4b-4f99-ad14-f3abea5567f2] Took 17.23 seconds to build instance. [ 986.620286] env[63515]: INFO nova.scheduler.client.report [None req-ee146559-ead8-41ee-8bac-7272cc90bf4b tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Deleted allocations for instance 918d5b35-c46d-47c5-b59f-175831aed9d6 [ 986.713430] env[63515]: DEBUG nova.network.neutron [req-2b07d193-1ae8-4c4a-acab-39751c22ccef req-3e3b586a-91f8-46b8-bacb-92cf72da0bc5 service nova] [instance: 461a0c4c-dfdb-4669-ac72-8913c0277298] Updated VIF entry in instance network info cache for port 2536d489-b430-4d50-ba2e-36edf5e42e17. {{(pid=63515) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 986.713430] env[63515]: DEBUG nova.network.neutron [req-2b07d193-1ae8-4c4a-acab-39751c22ccef req-3e3b586a-91f8-46b8-bacb-92cf72da0bc5 service nova] [instance: 461a0c4c-dfdb-4669-ac72-8913c0277298] Updating instance_info_cache with network_info: [{"id": "2536d489-b430-4d50-ba2e-36edf5e42e17", "address": "fa:16:3e:e6:bc:df", "network": {"id": "581d448e-be1b-4d5b-8a6e-c27326aaffc5", "bridge": "br-int", "label": "tempest-ServersTestJSON-1769610073-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "299fdeff647f486390366d5bbf911518", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fb143ef7-8271-4a8a-a4aa-8eba9a89f6a1", "external-id": "nsx-vlan-transportzone-504", "segmentation_id": 504, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2536d489-b4", "ovs_interfaceid": "2536d489-b430-4d50-ba2e-36edf5e42e17", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 986.742498] env[63515]: DEBUG nova.network.neutron [-] [instance: 86b1ddcf-585c-41d9-84fe-17b794f1abbc] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 986.775061] env[63515]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-cdd98664-1acd-4f9e-8099-ab3b23ca7114 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.785815] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc1d41e2-6cf8-4bcd-9914-1dcf4e0bec0e {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.815350] env[63515]: DEBUG nova.compute.manager [req-9f5a7dfc-3205-4d8a-9ab0-e1c07da1093d req-a7f8fea4-9011-41b0-a700-f2d433f02867 service nova] [instance: 86b1ddcf-585c-41d9-84fe-17b794f1abbc] Detach interface failed, port_id=43c4b108-fec8-46ea-942f-d3e734fe2be6, reason: Instance 86b1ddcf-585c-41d9-84fe-17b794f1abbc could not be found. {{(pid=63515) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 986.978332] env[63515]: DEBUG oslo_vmware.api [None req-978a15da-7dcc-413b-825d-e9c53aaf3163 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]520b0aa0-f878-84d9-e50c-7fabc3a0b66c, 'name': SearchDatastore_Task, 'duration_secs': 0.022807} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 986.979773] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-023a9398-c487-418e-98bd-3a682af19bf6 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.984955] env[63515]: DEBUG oslo_vmware.api [None req-978a15da-7dcc-413b-825d-e9c53aaf3163 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Waiting for the task: (returnval){ [ 986.984955] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]52affc1c-4f20-7ab1-01f7-db6656091cbb" [ 986.984955] env[63515]: _type = "Task" [ 986.984955] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 986.992925] env[63515]: DEBUG oslo_vmware.api [None req-978a15da-7dcc-413b-825d-e9c53aaf3163 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52affc1c-4f20-7ab1-01f7-db6656091cbb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.095799] env[63515]: INFO nova.compute.claims [None req-7c387ab3-2055-4517-9874-052fe02db033 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 68766cd4-84be-475b-8494-d7ab43a9e969] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 987.122396] env[63515]: DEBUG oslo_concurrency.lockutils [None req-adce163d-60b0-4e35-be4e-020840038d00 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Lock "724b05bc-ce4b-4f99-ad14-f3abea5567f2" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 18.755s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 987.129751] env[63515]: DEBUG oslo_concurrency.lockutils [None req-ee146559-ead8-41ee-8bac-7272cc90bf4b tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Lock "918d5b35-c46d-47c5-b59f-175831aed9d6" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 11.276s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 987.219192] env[63515]: DEBUG oslo_concurrency.lockutils [req-2b07d193-1ae8-4c4a-acab-39751c22ccef req-3e3b586a-91f8-46b8-bacb-92cf72da0bc5 service nova] Releasing lock "refresh_cache-461a0c4c-dfdb-4669-ac72-8913c0277298" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 987.247609] env[63515]: INFO nova.compute.manager [-] [instance: 86b1ddcf-585c-41d9-84fe-17b794f1abbc] Took 1.76 seconds to deallocate network for instance. [ 987.496379] env[63515]: DEBUG oslo_vmware.api [None req-978a15da-7dcc-413b-825d-e9c53aaf3163 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52affc1c-4f20-7ab1-01f7-db6656091cbb, 'name': SearchDatastore_Task, 'duration_secs': 0.035299} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 987.496900] env[63515]: DEBUG oslo_concurrency.lockutils [None req-978a15da-7dcc-413b-825d-e9c53aaf3163 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 987.497355] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-978a15da-7dcc-413b-825d-e9c53aaf3163 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk to [datastore1] 461a0c4c-dfdb-4669-ac72-8913c0277298/461a0c4c-dfdb-4669-ac72-8913c0277298.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 987.497646] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e74a5935-aa71-496c-8128-31d274f3c5af {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.504274] env[63515]: DEBUG oslo_vmware.api [None req-978a15da-7dcc-413b-825d-e9c53aaf3163 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Waiting for the task: (returnval){ [ 987.504274] env[63515]: value = "task-1111758" [ 987.504274] env[63515]: _type = "Task" [ 987.504274] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 987.513186] env[63515]: DEBUG oslo_vmware.api [None req-978a15da-7dcc-413b-825d-e9c53aaf3163 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Task: {'id': task-1111758, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.605377] env[63515]: INFO nova.compute.resource_tracker [None req-7c387ab3-2055-4517-9874-052fe02db033 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 68766cd4-84be-475b-8494-d7ab43a9e969] Updating resource usage from migration 0453c850-4f70-44d5-9d13-96d466406033 [ 987.756309] env[63515]: DEBUG oslo_concurrency.lockutils [None req-5f7eade2-bb20-4c09-b233-6856c972d63b tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 987.800963] env[63515]: DEBUG oslo_concurrency.lockutils [None req-2af70737-0f47-4d49-aa40-887a3044227c tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Acquiring lock "61cf2f9c-5d35-46c4-a477-d2cef9fcd5ae" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 987.801267] env[63515]: DEBUG oslo_concurrency.lockutils [None req-2af70737-0f47-4d49-aa40-887a3044227c tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Lock "61cf2f9c-5d35-46c4-a477-d2cef9fcd5ae" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 987.801490] env[63515]: DEBUG oslo_concurrency.lockutils [None req-2af70737-0f47-4d49-aa40-887a3044227c tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Acquiring lock "61cf2f9c-5d35-46c4-a477-d2cef9fcd5ae-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 987.801884] env[63515]: DEBUG oslo_concurrency.lockutils [None req-2af70737-0f47-4d49-aa40-887a3044227c tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Lock "61cf2f9c-5d35-46c4-a477-d2cef9fcd5ae-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 987.801996] env[63515]: DEBUG oslo_concurrency.lockutils [None req-2af70737-0f47-4d49-aa40-887a3044227c tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Lock "61cf2f9c-5d35-46c4-a477-d2cef9fcd5ae-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 987.806167] env[63515]: INFO nova.compute.manager [None req-2af70737-0f47-4d49-aa40-887a3044227c tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: 61cf2f9c-5d35-46c4-a477-d2cef9fcd5ae] Terminating instance [ 987.808788] env[63515]: DEBUG nova.compute.manager [None req-2af70737-0f47-4d49-aa40-887a3044227c tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: 61cf2f9c-5d35-46c4-a477-d2cef9fcd5ae] Start destroying the instance on the hypervisor. {{(pid=63515) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 987.809887] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-2af70737-0f47-4d49-aa40-887a3044227c tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: 61cf2f9c-5d35-46c4-a477-d2cef9fcd5ae] Destroying instance {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 987.810338] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ac6ceba-25d1-4828-9f20-456ab11dd555 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.824188] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-2af70737-0f47-4d49-aa40-887a3044227c tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: 61cf2f9c-5d35-46c4-a477-d2cef9fcd5ae] Unregistering the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 987.824539] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-673666b9-a66a-4582-a35a-c94266573998 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.871235] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67c5043b-3e58-4585-80c1-f55b0c9047b5 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.883668] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fcb0f29a-55bc-4ab8-94b0-89eef7b4523f {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.892608] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-2af70737-0f47-4d49-aa40-887a3044227c tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: 61cf2f9c-5d35-46c4-a477-d2cef9fcd5ae] Unregistered the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 987.892854] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-2af70737-0f47-4d49-aa40-887a3044227c tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: 61cf2f9c-5d35-46c4-a477-d2cef9fcd5ae] Deleting contents of the VM from datastore datastore1 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 987.893058] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-2af70737-0f47-4d49-aa40-887a3044227c tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Deleting the datastore file [datastore1] 61cf2f9c-5d35-46c4-a477-d2cef9fcd5ae {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 987.922303] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-680fd66d-fe76-44f0-a8e2-a3dc166bfc48 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.925515] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a805249-eed4-4f64-bbc8-fb0f3c6a1064 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.937823] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e99d6a9e-f958-48d2-bc42-3262448455b9 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.942701] env[63515]: DEBUG oslo_vmware.api [None req-2af70737-0f47-4d49-aa40-887a3044227c tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Waiting for the task: (returnval){ [ 987.942701] env[63515]: value = "task-1111760" [ 987.942701] env[63515]: _type = "Task" [ 987.942701] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 987.955870] env[63515]: DEBUG nova.compute.provider_tree [None req-7c387ab3-2055-4517-9874-052fe02db033 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 987.962067] env[63515]: DEBUG oslo_vmware.api [None req-2af70737-0f47-4d49-aa40-887a3044227c tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Task: {'id': task-1111760, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.016608] env[63515]: DEBUG oslo_vmware.api [None req-978a15da-7dcc-413b-825d-e9c53aaf3163 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Task: {'id': task-1111758, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.500338} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 988.016931] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-978a15da-7dcc-413b-825d-e9c53aaf3163 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk to [datastore1] 461a0c4c-dfdb-4669-ac72-8913c0277298/461a0c4c-dfdb-4669-ac72-8913c0277298.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 988.017131] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-978a15da-7dcc-413b-825d-e9c53aaf3163 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: 461a0c4c-dfdb-4669-ac72-8913c0277298] Extending root virtual disk to 1048576 {{(pid=63515) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 988.017400] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-3977f5f3-5ca8-4c0b-86cc-b2b3895fc8f2 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.026801] env[63515]: DEBUG oslo_vmware.api [None req-978a15da-7dcc-413b-825d-e9c53aaf3163 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Waiting for the task: (returnval){ [ 988.026801] env[63515]: value = "task-1111761" [ 988.026801] env[63515]: _type = "Task" [ 988.026801] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 988.034951] env[63515]: DEBUG oslo_vmware.api [None req-978a15da-7dcc-413b-825d-e9c53aaf3163 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Task: {'id': task-1111761, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.373514] env[63515]: DEBUG nova.compute.manager [req-7eaa16b6-9b97-42fd-a004-89afc6e0ec98 req-d6d03f00-4a08-4070-a3dd-059d05a5249b service nova] [instance: 724b05bc-ce4b-4f99-ad14-f3abea5567f2] Received event network-changed-3dee43b6-2650-42e1-aa5a-4994a3ec7f05 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 988.373789] env[63515]: DEBUG nova.compute.manager [req-7eaa16b6-9b97-42fd-a004-89afc6e0ec98 req-d6d03f00-4a08-4070-a3dd-059d05a5249b service nova] [instance: 724b05bc-ce4b-4f99-ad14-f3abea5567f2] Refreshing instance network info cache due to event network-changed-3dee43b6-2650-42e1-aa5a-4994a3ec7f05. {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 988.373942] env[63515]: DEBUG oslo_concurrency.lockutils [req-7eaa16b6-9b97-42fd-a004-89afc6e0ec98 req-d6d03f00-4a08-4070-a3dd-059d05a5249b service nova] Acquiring lock "refresh_cache-724b05bc-ce4b-4f99-ad14-f3abea5567f2" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 988.374103] env[63515]: DEBUG oslo_concurrency.lockutils [req-7eaa16b6-9b97-42fd-a004-89afc6e0ec98 req-d6d03f00-4a08-4070-a3dd-059d05a5249b service nova] Acquired lock "refresh_cache-724b05bc-ce4b-4f99-ad14-f3abea5567f2" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 988.374272] env[63515]: DEBUG nova.network.neutron [req-7eaa16b6-9b97-42fd-a004-89afc6e0ec98 req-d6d03f00-4a08-4070-a3dd-059d05a5249b service nova] [instance: 724b05bc-ce4b-4f99-ad14-f3abea5567f2] Refreshing network info cache for port 3dee43b6-2650-42e1-aa5a-4994a3ec7f05 {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 988.456880] env[63515]: DEBUG oslo_vmware.api [None req-2af70737-0f47-4d49-aa40-887a3044227c tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Task: {'id': task-1111760, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.219616} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 988.456880] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-2af70737-0f47-4d49-aa40-887a3044227c tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Deleted the datastore file {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 988.457044] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-2af70737-0f47-4d49-aa40-887a3044227c tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: 61cf2f9c-5d35-46c4-a477-d2cef9fcd5ae] Deleted contents of the VM from datastore datastore1 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 988.457578] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-2af70737-0f47-4d49-aa40-887a3044227c tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: 61cf2f9c-5d35-46c4-a477-d2cef9fcd5ae] Instance destroyed {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 988.457769] env[63515]: INFO nova.compute.manager [None req-2af70737-0f47-4d49-aa40-887a3044227c tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: 61cf2f9c-5d35-46c4-a477-d2cef9fcd5ae] Took 0.65 seconds to destroy the instance on the hypervisor. [ 988.458085] env[63515]: DEBUG oslo.service.loopingcall [None req-2af70737-0f47-4d49-aa40-887a3044227c tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 988.458486] env[63515]: DEBUG nova.compute.manager [-] [instance: 61cf2f9c-5d35-46c4-a477-d2cef9fcd5ae] Deallocating network for instance {{(pid=63515) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 988.458611] env[63515]: DEBUG nova.network.neutron [-] [instance: 61cf2f9c-5d35-46c4-a477-d2cef9fcd5ae] deallocate_for_instance() {{(pid=63515) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 988.464435] env[63515]: DEBUG nova.scheduler.client.report [None req-7c387ab3-2055-4517-9874-052fe02db033 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 988.535866] env[63515]: DEBUG oslo_vmware.api [None req-978a15da-7dcc-413b-825d-e9c53aaf3163 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Task: {'id': task-1111761, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.068122} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 988.536306] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-978a15da-7dcc-413b-825d-e9c53aaf3163 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: 461a0c4c-dfdb-4669-ac72-8913c0277298] Extended root virtual disk {{(pid=63515) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 988.537142] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71597fd7-547f-4a90-af74-9b010102a9db {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.560055] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-978a15da-7dcc-413b-825d-e9c53aaf3163 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: 461a0c4c-dfdb-4669-ac72-8913c0277298] Reconfiguring VM instance instance-00000060 to attach disk [datastore1] 461a0c4c-dfdb-4669-ac72-8913c0277298/461a0c4c-dfdb-4669-ac72-8913c0277298.vmdk or device None with type sparse {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 988.560355] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1e0bcd91-1e11-4feb-9b43-ce1e945f59e8 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.579560] env[63515]: DEBUG oslo_vmware.api [None req-978a15da-7dcc-413b-825d-e9c53aaf3163 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Waiting for the task: (returnval){ [ 988.579560] env[63515]: value = "task-1111762" [ 988.579560] env[63515]: _type = "Task" [ 988.579560] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 988.588484] env[63515]: DEBUG oslo_vmware.api [None req-978a15da-7dcc-413b-825d-e9c53aaf3163 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Task: {'id': task-1111762, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.779282] env[63515]: DEBUG oslo_concurrency.lockutils [None req-9d22472d-269a-4ade-805d-0c1138cb0261 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Acquiring lock "c6ac42a9-a233-48b4-bcb2-fb791c7446bb" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 988.779578] env[63515]: DEBUG oslo_concurrency.lockutils [None req-9d22472d-269a-4ade-805d-0c1138cb0261 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Lock "c6ac42a9-a233-48b4-bcb2-fb791c7446bb" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 988.814746] env[63515]: DEBUG nova.compute.manager [req-5cdf7dc6-41ed-4631-b5d0-4aafde3bf9e6 req-7997c7b0-94a7-4319-94f4-64f63bc367ed service nova] [instance: 61cf2f9c-5d35-46c4-a477-d2cef9fcd5ae] Received event network-vif-deleted-99e47951-e384-4aa7-a394-3d0a6bade8d7 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 988.815531] env[63515]: INFO nova.compute.manager [req-5cdf7dc6-41ed-4631-b5d0-4aafde3bf9e6 req-7997c7b0-94a7-4319-94f4-64f63bc367ed service nova] [instance: 61cf2f9c-5d35-46c4-a477-d2cef9fcd5ae] Neutron deleted interface 99e47951-e384-4aa7-a394-3d0a6bade8d7; detaching it from the instance and deleting it from the info cache [ 988.816591] env[63515]: DEBUG nova.network.neutron [req-5cdf7dc6-41ed-4631-b5d0-4aafde3bf9e6 req-7997c7b0-94a7-4319-94f4-64f63bc367ed service nova] [instance: 61cf2f9c-5d35-46c4-a477-d2cef9fcd5ae] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 988.971525] env[63515]: DEBUG oslo_concurrency.lockutils [None req-7c387ab3-2055-4517-9874-052fe02db033 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.388s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 988.972760] env[63515]: INFO nova.compute.manager [None req-7c387ab3-2055-4517-9874-052fe02db033 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 68766cd4-84be-475b-8494-d7ab43a9e969] Migrating [ 988.978531] env[63515]: DEBUG oslo_concurrency.lockutils [None req-505b9ae2-f7d6-4938-b4d4-0d6c86cd7e2f tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.600s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 988.978788] env[63515]: DEBUG nova.objects.instance [None req-505b9ae2-f7d6-4938-b4d4-0d6c86cd7e2f tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Lazy-loading 'resources' on Instance uuid 6a66dd9b-7b24-4b70-b58f-0830cf39376c {{(pid=63515) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 989.090187] env[63515]: DEBUG oslo_vmware.api [None req-978a15da-7dcc-413b-825d-e9c53aaf3163 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Task: {'id': task-1111762, 'name': ReconfigVM_Task, 'duration_secs': 0.410019} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 989.090488] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-978a15da-7dcc-413b-825d-e9c53aaf3163 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: 461a0c4c-dfdb-4669-ac72-8913c0277298] Reconfigured VM instance instance-00000060 to attach disk [datastore1] 461a0c4c-dfdb-4669-ac72-8913c0277298/461a0c4c-dfdb-4669-ac72-8913c0277298.vmdk or device None with type sparse {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 989.091156] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d213cecf-8015-4f0c-bea6-b936ad5de09d {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.097518] env[63515]: DEBUG oslo_vmware.api [None req-978a15da-7dcc-413b-825d-e9c53aaf3163 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Waiting for the task: (returnval){ [ 989.097518] env[63515]: value = "task-1111763" [ 989.097518] env[63515]: _type = "Task" [ 989.097518] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 989.105135] env[63515]: DEBUG oslo_vmware.api [None req-978a15da-7dcc-413b-825d-e9c53aaf3163 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Task: {'id': task-1111763, 'name': Rename_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.112785] env[63515]: DEBUG nova.network.neutron [req-7eaa16b6-9b97-42fd-a004-89afc6e0ec98 req-d6d03f00-4a08-4070-a3dd-059d05a5249b service nova] [instance: 724b05bc-ce4b-4f99-ad14-f3abea5567f2] Updated VIF entry in instance network info cache for port 3dee43b6-2650-42e1-aa5a-4994a3ec7f05. {{(pid=63515) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 989.112785] env[63515]: DEBUG nova.network.neutron [req-7eaa16b6-9b97-42fd-a004-89afc6e0ec98 req-d6d03f00-4a08-4070-a3dd-059d05a5249b service nova] [instance: 724b05bc-ce4b-4f99-ad14-f3abea5567f2] Updating instance_info_cache with network_info: [{"id": "3dee43b6-2650-42e1-aa5a-4994a3ec7f05", "address": "fa:16:3e:89:06:d0", "network": {"id": "88dfb29d-41e0-4df7-8f70-e52cda05620c", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1355578400-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.146", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7d4ada9ec35f42b19c6480a9101d21a4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "357d2811-e990-4985-9f9e-b158d10d3699", "external-id": "nsx-vlan-transportzone-641", "segmentation_id": 641, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3dee43b6-26", "ovs_interfaceid": "3dee43b6-2650-42e1-aa5a-4994a3ec7f05", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 989.238880] env[63515]: DEBUG oslo_concurrency.lockutils [None req-bea83ded-7b10-4210-b5be-b45b3e9e3126 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Acquiring lock "63ce797d-7180-4209-ac2c-81978bf7607a" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 989.239100] env[63515]: DEBUG oslo_concurrency.lockutils [None req-bea83ded-7b10-4210-b5be-b45b3e9e3126 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Lock "63ce797d-7180-4209-ac2c-81978bf7607a" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 989.239364] env[63515]: DEBUG oslo_concurrency.lockutils [None req-bea83ded-7b10-4210-b5be-b45b3e9e3126 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Acquiring lock "63ce797d-7180-4209-ac2c-81978bf7607a-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 989.239603] env[63515]: DEBUG oslo_concurrency.lockutils [None req-bea83ded-7b10-4210-b5be-b45b3e9e3126 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Lock "63ce797d-7180-4209-ac2c-81978bf7607a-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 989.239793] env[63515]: DEBUG oslo_concurrency.lockutils [None req-bea83ded-7b10-4210-b5be-b45b3e9e3126 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Lock "63ce797d-7180-4209-ac2c-81978bf7607a-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 989.242010] env[63515]: INFO nova.compute.manager [None req-bea83ded-7b10-4210-b5be-b45b3e9e3126 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] [instance: 63ce797d-7180-4209-ac2c-81978bf7607a] Terminating instance [ 989.243949] env[63515]: DEBUG nova.compute.manager [None req-bea83ded-7b10-4210-b5be-b45b3e9e3126 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] [instance: 63ce797d-7180-4209-ac2c-81978bf7607a] Start destroying the instance on the hypervisor. {{(pid=63515) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 989.244184] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-bea83ded-7b10-4210-b5be-b45b3e9e3126 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] [instance: 63ce797d-7180-4209-ac2c-81978bf7607a] Destroying instance {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 989.245077] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1960d773-1413-4496-9087-31f9888483e8 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.252812] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-bea83ded-7b10-4210-b5be-b45b3e9e3126 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] [instance: 63ce797d-7180-4209-ac2c-81978bf7607a] Powering off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 989.253094] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5ca3d24d-911f-456b-aefe-c86b08e52cd5 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.259714] env[63515]: DEBUG oslo_vmware.api [None req-bea83ded-7b10-4210-b5be-b45b3e9e3126 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Waiting for the task: (returnval){ [ 989.259714] env[63515]: value = "task-1111764" [ 989.259714] env[63515]: _type = "Task" [ 989.259714] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 989.267687] env[63515]: DEBUG oslo_vmware.api [None req-bea83ded-7b10-4210-b5be-b45b3e9e3126 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Task: {'id': task-1111764, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.282215] env[63515]: DEBUG nova.compute.manager [None req-9d22472d-269a-4ade-805d-0c1138cb0261 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: c6ac42a9-a233-48b4-bcb2-fb791c7446bb] Starting instance... {{(pid=63515) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 989.291725] env[63515]: DEBUG nova.network.neutron [-] [instance: 61cf2f9c-5d35-46c4-a477-d2cef9fcd5ae] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 989.306337] env[63515]: DEBUG oslo_concurrency.lockutils [None req-655527ff-c992-4371-875f-a71b38db6213 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Acquiring lock "d6c5a9f9-5e5c-4b14-9b9a-7825bbf526a5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 989.306749] env[63515]: DEBUG oslo_concurrency.lockutils [None req-655527ff-c992-4371-875f-a71b38db6213 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Lock "d6c5a9f9-5e5c-4b14-9b9a-7825bbf526a5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 989.319424] env[63515]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-dab555f5-a90c-4ce1-ad25-6c0d5c156a04 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.329185] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2bab7df1-91d9-4e63-82ba-0a2016324dd7 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.363668] env[63515]: DEBUG nova.compute.manager [req-5cdf7dc6-41ed-4631-b5d0-4aafde3bf9e6 req-7997c7b0-94a7-4319-94f4-64f63bc367ed service nova] [instance: 61cf2f9c-5d35-46c4-a477-d2cef9fcd5ae] Detach interface failed, port_id=99e47951-e384-4aa7-a394-3d0a6bade8d7, reason: Instance 61cf2f9c-5d35-46c4-a477-d2cef9fcd5ae could not be found. {{(pid=63515) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 989.492510] env[63515]: DEBUG oslo_concurrency.lockutils [None req-7c387ab3-2055-4517-9874-052fe02db033 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Acquiring lock "refresh_cache-68766cd4-84be-475b-8494-d7ab43a9e969" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 989.492770] env[63515]: DEBUG oslo_concurrency.lockutils [None req-7c387ab3-2055-4517-9874-052fe02db033 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Acquired lock "refresh_cache-68766cd4-84be-475b-8494-d7ab43a9e969" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 989.492961] env[63515]: DEBUG nova.network.neutron [None req-7c387ab3-2055-4517-9874-052fe02db033 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 68766cd4-84be-475b-8494-d7ab43a9e969] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 989.609587] env[63515]: DEBUG oslo_vmware.api [None req-978a15da-7dcc-413b-825d-e9c53aaf3163 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Task: {'id': task-1111763, 'name': Rename_Task, 'duration_secs': 0.209792} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 989.609893] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-978a15da-7dcc-413b-825d-e9c53aaf3163 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: 461a0c4c-dfdb-4669-ac72-8913c0277298] Powering on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 989.610174] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-4dab233f-31d5-432f-8b1f-676fe16589a0 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.616508] env[63515]: DEBUG oslo_concurrency.lockutils [req-7eaa16b6-9b97-42fd-a004-89afc6e0ec98 req-d6d03f00-4a08-4070-a3dd-059d05a5249b service nova] Releasing lock "refresh_cache-724b05bc-ce4b-4f99-ad14-f3abea5567f2" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 989.618155] env[63515]: DEBUG oslo_vmware.api [None req-978a15da-7dcc-413b-825d-e9c53aaf3163 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Waiting for the task: (returnval){ [ 989.618155] env[63515]: value = "task-1111765" [ 989.618155] env[63515]: _type = "Task" [ 989.618155] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 989.625956] env[63515]: DEBUG oslo_vmware.api [None req-978a15da-7dcc-413b-825d-e9c53aaf3163 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Task: {'id': task-1111765, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.676922] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0b39e31-f742-4f15-aa27-ea17448833cd {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.684742] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5f6f100-11d3-4eda-b178-7a1077717bd9 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.713905] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-751ee2be-3c0d-4405-9ab7-d3ad33a85275 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.720879] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0aae548-9d63-4539-bae4-90bf455d31c9 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.733394] env[63515]: DEBUG nova.compute.provider_tree [None req-505b9ae2-f7d6-4938-b4d4-0d6c86cd7e2f tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 989.768903] env[63515]: DEBUG oslo_vmware.api [None req-bea83ded-7b10-4210-b5be-b45b3e9e3126 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Task: {'id': task-1111764, 'name': PowerOffVM_Task, 'duration_secs': 0.258308} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 989.769178] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-bea83ded-7b10-4210-b5be-b45b3e9e3126 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] [instance: 63ce797d-7180-4209-ac2c-81978bf7607a] Powered off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 989.769355] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-bea83ded-7b10-4210-b5be-b45b3e9e3126 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] [instance: 63ce797d-7180-4209-ac2c-81978bf7607a] Unregistering the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 989.769585] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-dbeeaa3c-3496-495c-b256-41a028f476d4 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.792442] env[63515]: INFO nova.compute.manager [-] [instance: 61cf2f9c-5d35-46c4-a477-d2cef9fcd5ae] Took 1.33 seconds to deallocate network for instance. [ 989.800519] env[63515]: DEBUG oslo_concurrency.lockutils [None req-9d22472d-269a-4ade-805d-0c1138cb0261 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 989.810805] env[63515]: DEBUG nova.compute.manager [None req-655527ff-c992-4371-875f-a71b38db6213 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: d6c5a9f9-5e5c-4b14-9b9a-7825bbf526a5] Starting instance... {{(pid=63515) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 989.830246] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-bea83ded-7b10-4210-b5be-b45b3e9e3126 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] [instance: 63ce797d-7180-4209-ac2c-81978bf7607a] Unregistered the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 989.830479] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-bea83ded-7b10-4210-b5be-b45b3e9e3126 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] [instance: 63ce797d-7180-4209-ac2c-81978bf7607a] Deleting contents of the VM from datastore datastore1 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 989.830668] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-bea83ded-7b10-4210-b5be-b45b3e9e3126 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Deleting the datastore file [datastore1] 63ce797d-7180-4209-ac2c-81978bf7607a {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 989.830938] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-2676c822-01dd-4fa9-8f34-a3c808b7ad63 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.837907] env[63515]: DEBUG oslo_vmware.api [None req-bea83ded-7b10-4210-b5be-b45b3e9e3126 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Waiting for the task: (returnval){ [ 989.837907] env[63515]: value = "task-1111767" [ 989.837907] env[63515]: _type = "Task" [ 989.837907] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 989.848367] env[63515]: DEBUG oslo_vmware.api [None req-bea83ded-7b10-4210-b5be-b45b3e9e3126 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Task: {'id': task-1111767, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.127872] env[63515]: DEBUG oslo_vmware.api [None req-978a15da-7dcc-413b-825d-e9c53aaf3163 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Task: {'id': task-1111765, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.174268] env[63515]: DEBUG oslo_service.periodic_task [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=63515) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 990.174509] env[63515]: DEBUG oslo_service.periodic_task [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=63515) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 990.232380] env[63515]: DEBUG nova.network.neutron [None req-7c387ab3-2055-4517-9874-052fe02db033 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 68766cd4-84be-475b-8494-d7ab43a9e969] Updating instance_info_cache with network_info: [{"id": "28b59076-140f-46b5-b4a9-0bf530a1d43d", "address": "fa:16:3e:a3:55:0c", "network": {"id": "5758c885-e1c9-45c8-86aa-27a802101f52", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-734458138-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.204", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b18d4daacbc84e758f9f9ae4f3f09d28", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69351262-8d39-441a-85ba-3a78df436d17", "external-id": "nsx-vlan-transportzone-205", "segmentation_id": 205, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap28b59076-14", "ovs_interfaceid": "28b59076-140f-46b5-b4a9-0bf530a1d43d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 990.236680] env[63515]: DEBUG nova.scheduler.client.report [None req-505b9ae2-f7d6-4938-b4d4-0d6c86cd7e2f tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 990.299584] env[63515]: DEBUG oslo_concurrency.lockutils [None req-2af70737-0f47-4d49-aa40-887a3044227c tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 990.331242] env[63515]: DEBUG oslo_concurrency.lockutils [None req-655527ff-c992-4371-875f-a71b38db6213 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 990.348327] env[63515]: DEBUG oslo_vmware.api [None req-bea83ded-7b10-4210-b5be-b45b3e9e3126 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Task: {'id': task-1111767, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.632204] env[63515]: DEBUG oslo_vmware.api [None req-978a15da-7dcc-413b-825d-e9c53aaf3163 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Task: {'id': task-1111765, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.681925] env[63515]: DEBUG oslo_service.periodic_task [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=63515) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 990.682124] env[63515]: DEBUG nova.compute.manager [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Starting heal instance info cache {{(pid=63515) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9926}} [ 990.682324] env[63515]: DEBUG nova.compute.manager [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Rebuilding the list of instances to heal {{(pid=63515) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9930}} [ 990.734864] env[63515]: DEBUG oslo_concurrency.lockutils [None req-7c387ab3-2055-4517-9874-052fe02db033 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Releasing lock "refresh_cache-68766cd4-84be-475b-8494-d7ab43a9e969" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 990.741066] env[63515]: DEBUG oslo_concurrency.lockutils [None req-505b9ae2-f7d6-4938-b4d4-0d6c86cd7e2f tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.762s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 990.743591] env[63515]: DEBUG oslo_concurrency.lockutils [None req-5f7eade2-bb20-4c09-b233-6856c972d63b tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.987s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 990.743894] env[63515]: DEBUG nova.objects.instance [None req-5f7eade2-bb20-4c09-b233-6856c972d63b tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Lazy-loading 'resources' on Instance uuid 86b1ddcf-585c-41d9-84fe-17b794f1abbc {{(pid=63515) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 990.765385] env[63515]: INFO nova.scheduler.client.report [None req-505b9ae2-f7d6-4938-b4d4-0d6c86cd7e2f tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Deleted allocations for instance 6a66dd9b-7b24-4b70-b58f-0830cf39376c [ 990.848689] env[63515]: DEBUG oslo_vmware.api [None req-bea83ded-7b10-4210-b5be-b45b3e9e3126 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Task: {'id': task-1111767, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.66599} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 990.849065] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-bea83ded-7b10-4210-b5be-b45b3e9e3126 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Deleted the datastore file {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 990.849276] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-bea83ded-7b10-4210-b5be-b45b3e9e3126 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] [instance: 63ce797d-7180-4209-ac2c-81978bf7607a] Deleted contents of the VM from datastore datastore1 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 990.849458] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-bea83ded-7b10-4210-b5be-b45b3e9e3126 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] [instance: 63ce797d-7180-4209-ac2c-81978bf7607a] Instance destroyed {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 990.849641] env[63515]: INFO nova.compute.manager [None req-bea83ded-7b10-4210-b5be-b45b3e9e3126 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] [instance: 63ce797d-7180-4209-ac2c-81978bf7607a] Took 1.61 seconds to destroy the instance on the hypervisor. [ 990.849890] env[63515]: DEBUG oslo.service.loopingcall [None req-bea83ded-7b10-4210-b5be-b45b3e9e3126 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 990.850110] env[63515]: DEBUG nova.compute.manager [-] [instance: 63ce797d-7180-4209-ac2c-81978bf7607a] Deallocating network for instance {{(pid=63515) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 990.850215] env[63515]: DEBUG nova.network.neutron [-] [instance: 63ce797d-7180-4209-ac2c-81978bf7607a] deallocate_for_instance() {{(pid=63515) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 991.128120] env[63515]: DEBUG nova.compute.manager [req-f32f72f1-9604-4d54-9b25-43ba80a5b40d req-6a4543dd-1613-4ad3-b830-8516d8e55018 service nova] [instance: 63ce797d-7180-4209-ac2c-81978bf7607a] Received event network-vif-deleted-25ea9d0d-3405-42da-a576-a2a1f37fa75a {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 991.128396] env[63515]: INFO nova.compute.manager [req-f32f72f1-9604-4d54-9b25-43ba80a5b40d req-6a4543dd-1613-4ad3-b830-8516d8e55018 service nova] [instance: 63ce797d-7180-4209-ac2c-81978bf7607a] Neutron deleted interface 25ea9d0d-3405-42da-a576-a2a1f37fa75a; detaching it from the instance and deleting it from the info cache [ 991.128519] env[63515]: DEBUG nova.network.neutron [req-f32f72f1-9604-4d54-9b25-43ba80a5b40d req-6a4543dd-1613-4ad3-b830-8516d8e55018 service nova] [instance: 63ce797d-7180-4209-ac2c-81978bf7607a] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 991.134661] env[63515]: DEBUG oslo_vmware.api [None req-978a15da-7dcc-413b-825d-e9c53aaf3163 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Task: {'id': task-1111765, 'name': PowerOnVM_Task, 'duration_secs': 1.102376} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 991.135132] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-978a15da-7dcc-413b-825d-e9c53aaf3163 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: 461a0c4c-dfdb-4669-ac72-8913c0277298] Powered on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 991.135132] env[63515]: INFO nova.compute.manager [None req-978a15da-7dcc-413b-825d-e9c53aaf3163 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: 461a0c4c-dfdb-4669-ac72-8913c0277298] Took 8.15 seconds to spawn the instance on the hypervisor. [ 991.135239] env[63515]: DEBUG nova.compute.manager [None req-978a15da-7dcc-413b-825d-e9c53aaf3163 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: 461a0c4c-dfdb-4669-ac72-8913c0277298] Checking state {{(pid=63515) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 991.136099] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e726551-0799-43ee-9f06-9918c8e9cd51 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.188766] env[63515]: DEBUG nova.compute.manager [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] [instance: 63ce797d-7180-4209-ac2c-81978bf7607a] Skipping network cache update for instance because it is being deleted. {{(pid=63515) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9943}} [ 991.189125] env[63515]: DEBUG nova.compute.manager [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] [instance: 461a0c4c-dfdb-4669-ac72-8913c0277298] Skipping network cache update for instance because it is Building. {{(pid=63515) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 991.203240] env[63515]: DEBUG oslo_concurrency.lockutils [None req-642b03cc-2e02-4a86-bca6-75c3316e966c tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Acquiring lock "7d572ac0-f6af-4622-96cc-e75983420222" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 991.203698] env[63515]: DEBUG oslo_concurrency.lockutils [None req-642b03cc-2e02-4a86-bca6-75c3316e966c tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Lock "7d572ac0-f6af-4622-96cc-e75983420222" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.001s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 991.204090] env[63515]: INFO nova.compute.manager [None req-642b03cc-2e02-4a86-bca6-75c3316e966c tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] [instance: 7d572ac0-f6af-4622-96cc-e75983420222] Shelving [ 991.223397] env[63515]: DEBUG oslo_concurrency.lockutils [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Acquiring lock "refresh_cache-5cbce760-0163-4b27-8ae3-e46c926c8916" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 991.223631] env[63515]: DEBUG oslo_concurrency.lockutils [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Acquired lock "refresh_cache-5cbce760-0163-4b27-8ae3-e46c926c8916" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 991.224012] env[63515]: DEBUG nova.network.neutron [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] [instance: 5cbce760-0163-4b27-8ae3-e46c926c8916] Forcefully refreshing network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2004}} [ 991.224191] env[63515]: DEBUG nova.objects.instance [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Lazy-loading 'info_cache' on Instance uuid 5cbce760-0163-4b27-8ae3-e46c926c8916 {{(pid=63515) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 991.272851] env[63515]: DEBUG oslo_concurrency.lockutils [None req-505b9ae2-f7d6-4938-b4d4-0d6c86cd7e2f tempest-ServerDiskConfigTestJSON-1137995976 tempest-ServerDiskConfigTestJSON-1137995976-project-member] Lock "6a66dd9b-7b24-4b70-b58f-0830cf39376c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.968s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 991.421736] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7053b0e-3e75-4702-aecb-f989e521aebb {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.429866] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-497b8706-4bd8-4389-9362-6a8be1c84f99 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.459809] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5e5f626-e581-4ef4-9e7a-156007659dae {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.467717] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20e9eaed-c069-4bdf-9672-e84d8df8ed12 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.480597] env[63515]: DEBUG nova.compute.provider_tree [None req-5f7eade2-bb20-4c09-b233-6856c972d63b tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 991.609301] env[63515]: DEBUG nova.network.neutron [-] [instance: 63ce797d-7180-4209-ac2c-81978bf7607a] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 991.631930] env[63515]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-82438edd-8e95-4208-9dbd-a433a6faef13 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.641356] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd797d3b-5845-45d7-9b5f-254d8f0a03c0 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.661532] env[63515]: INFO nova.compute.manager [None req-978a15da-7dcc-413b-825d-e9c53aaf3163 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: 461a0c4c-dfdb-4669-ac72-8913c0277298] Took 17.78 seconds to build instance. [ 991.677816] env[63515]: DEBUG nova.compute.manager [req-f32f72f1-9604-4d54-9b25-43ba80a5b40d req-6a4543dd-1613-4ad3-b830-8516d8e55018 service nova] [instance: 63ce797d-7180-4209-ac2c-81978bf7607a] Detach interface failed, port_id=25ea9d0d-3405-42da-a576-a2a1f37fa75a, reason: Instance 63ce797d-7180-4209-ac2c-81978bf7607a could not be found. {{(pid=63515) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 991.710736] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-642b03cc-2e02-4a86-bca6-75c3316e966c tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] [instance: 7d572ac0-f6af-4622-96cc-e75983420222] Powering off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 991.711071] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b2f8bb1c-e7ac-41d6-8a6a-da0ebd52f275 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.717944] env[63515]: DEBUG oslo_vmware.api [None req-642b03cc-2e02-4a86-bca6-75c3316e966c tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Waiting for the task: (returnval){ [ 991.717944] env[63515]: value = "task-1111768" [ 991.717944] env[63515]: _type = "Task" [ 991.717944] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 991.727908] env[63515]: DEBUG oslo_vmware.api [None req-642b03cc-2e02-4a86-bca6-75c3316e966c tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Task: {'id': task-1111768, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.984361] env[63515]: DEBUG nova.scheduler.client.report [None req-5f7eade2-bb20-4c09-b233-6856c972d63b tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 992.111698] env[63515]: INFO nova.compute.manager [-] [instance: 63ce797d-7180-4209-ac2c-81978bf7607a] Took 1.26 seconds to deallocate network for instance. [ 992.164124] env[63515]: DEBUG oslo_concurrency.lockutils [None req-978a15da-7dcc-413b-825d-e9c53aaf3163 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Lock "461a0c4c-dfdb-4669-ac72-8913c0277298" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 19.291s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 992.228593] env[63515]: DEBUG oslo_vmware.api [None req-642b03cc-2e02-4a86-bca6-75c3316e966c tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Task: {'id': task-1111768, 'name': PowerOffVM_Task, 'duration_secs': 0.21391} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 992.228849] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-642b03cc-2e02-4a86-bca6-75c3316e966c tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] [instance: 7d572ac0-f6af-4622-96cc-e75983420222] Powered off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 992.229670] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-428a6785-d1a9-4b58-8fd6-3bba444c2a3f {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.254621] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c717d18-f95c-4598-aec8-ef51bed4e7e0 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.257700] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e03b6206-29dd-4eaf-8701-f74dd6fbc1ef {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.286819] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-7c387ab3-2055-4517-9874-052fe02db033 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 68766cd4-84be-475b-8494-d7ab43a9e969] Updating instance '68766cd4-84be-475b-8494-d7ab43a9e969' progress to 0 {{(pid=63515) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 992.493686] env[63515]: DEBUG oslo_concurrency.lockutils [None req-5f7eade2-bb20-4c09-b233-6856c972d63b tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.750s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 992.496386] env[63515]: DEBUG oslo_concurrency.lockutils [None req-9d22472d-269a-4ade-805d-0c1138cb0261 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.696s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 992.498012] env[63515]: INFO nova.compute.claims [None req-9d22472d-269a-4ade-805d-0c1138cb0261 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: c6ac42a9-a233-48b4-bcb2-fb791c7446bb] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 992.519769] env[63515]: INFO nova.scheduler.client.report [None req-5f7eade2-bb20-4c09-b233-6856c972d63b tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Deleted allocations for instance 86b1ddcf-585c-41d9-84fe-17b794f1abbc [ 992.562204] env[63515]: DEBUG oslo_concurrency.lockutils [None req-1cf33b4a-1ed6-417e-b050-ff4dbd167e75 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Acquiring lock "461a0c4c-dfdb-4669-ac72-8913c0277298" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 992.562503] env[63515]: DEBUG oslo_concurrency.lockutils [None req-1cf33b4a-1ed6-417e-b050-ff4dbd167e75 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Lock "461a0c4c-dfdb-4669-ac72-8913c0277298" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 992.562782] env[63515]: DEBUG oslo_concurrency.lockutils [None req-1cf33b4a-1ed6-417e-b050-ff4dbd167e75 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Acquiring lock "461a0c4c-dfdb-4669-ac72-8913c0277298-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 992.563030] env[63515]: DEBUG oslo_concurrency.lockutils [None req-1cf33b4a-1ed6-417e-b050-ff4dbd167e75 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Lock "461a0c4c-dfdb-4669-ac72-8913c0277298-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 992.563272] env[63515]: DEBUG oslo_concurrency.lockutils [None req-1cf33b4a-1ed6-417e-b050-ff4dbd167e75 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Lock "461a0c4c-dfdb-4669-ac72-8913c0277298-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 992.565367] env[63515]: INFO nova.compute.manager [None req-1cf33b4a-1ed6-417e-b050-ff4dbd167e75 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: 461a0c4c-dfdb-4669-ac72-8913c0277298] Terminating instance [ 992.567213] env[63515]: DEBUG nova.compute.manager [None req-1cf33b4a-1ed6-417e-b050-ff4dbd167e75 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: 461a0c4c-dfdb-4669-ac72-8913c0277298] Start destroying the instance on the hypervisor. {{(pid=63515) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 992.567407] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-1cf33b4a-1ed6-417e-b050-ff4dbd167e75 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: 461a0c4c-dfdb-4669-ac72-8913c0277298] Destroying instance {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 992.568299] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ead6c84-9871-4cb5-be86-d6bb1edac6c5 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.576938] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-1cf33b4a-1ed6-417e-b050-ff4dbd167e75 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: 461a0c4c-dfdb-4669-ac72-8913c0277298] Powering off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 992.577233] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9d22a9c0-8263-4a24-a2a5-b015f249c1a4 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.583830] env[63515]: DEBUG oslo_vmware.api [None req-1cf33b4a-1ed6-417e-b050-ff4dbd167e75 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Waiting for the task: (returnval){ [ 992.583830] env[63515]: value = "task-1111769" [ 992.583830] env[63515]: _type = "Task" [ 992.583830] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 992.592232] env[63515]: DEBUG oslo_vmware.api [None req-1cf33b4a-1ed6-417e-b050-ff4dbd167e75 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Task: {'id': task-1111769, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.618196] env[63515]: DEBUG oslo_concurrency.lockutils [None req-bea83ded-7b10-4210-b5be-b45b3e9e3126 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 992.771732] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-642b03cc-2e02-4a86-bca6-75c3316e966c tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] [instance: 7d572ac0-f6af-4622-96cc-e75983420222] Creating Snapshot of the VM instance {{(pid=63515) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 992.774121] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-f8612ffc-c1bc-4e75-937a-a982cf1cdfef {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.781257] env[63515]: DEBUG oslo_vmware.api [None req-642b03cc-2e02-4a86-bca6-75c3316e966c tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Waiting for the task: (returnval){ [ 992.781257] env[63515]: value = "task-1111770" [ 992.781257] env[63515]: _type = "Task" [ 992.781257] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 992.794865] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-7c387ab3-2055-4517-9874-052fe02db033 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 68766cd4-84be-475b-8494-d7ab43a9e969] Powering off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 992.795208] env[63515]: DEBUG oslo_vmware.api [None req-642b03cc-2e02-4a86-bca6-75c3316e966c tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Task: {'id': task-1111770, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.795438] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f2a52293-f876-43dd-a025-84bc1482fea2 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.802452] env[63515]: DEBUG oslo_vmware.api [None req-7c387ab3-2055-4517-9874-052fe02db033 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Waiting for the task: (returnval){ [ 992.802452] env[63515]: value = "task-1111771" [ 992.802452] env[63515]: _type = "Task" [ 992.802452] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 992.810760] env[63515]: DEBUG oslo_vmware.api [None req-7c387ab3-2055-4517-9874-052fe02db033 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Task: {'id': task-1111771, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.029371] env[63515]: DEBUG oslo_concurrency.lockutils [None req-5f7eade2-bb20-4c09-b233-6856c972d63b tempest-AttachVolumeTestJSON-1776504699 tempest-AttachVolumeTestJSON-1776504699-project-member] Lock "86b1ddcf-585c-41d9-84fe-17b794f1abbc" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.680s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 993.057687] env[63515]: DEBUG nova.network.neutron [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] [instance: 5cbce760-0163-4b27-8ae3-e46c926c8916] Updating instance_info_cache with network_info: [{"id": "5fa807c4-5ce1-492a-bea3-d8f7a65d898f", "address": "fa:16:3e:23:bb:83", "network": {"id": "850bfa95-3df6-4019-8260-ade20ec58799", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-832000501-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "08c9aa51b1864410809f8877c7cd49c4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3cf748a8-7ae0-4dca-817d-e727c30d72f4", "external-id": "nsx-vlan-transportzone-853", "segmentation_id": 853, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5fa807c4-5c", "ovs_interfaceid": "5fa807c4-5ce1-492a-bea3-d8f7a65d898f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 993.093785] env[63515]: DEBUG oslo_vmware.api [None req-1cf33b4a-1ed6-417e-b050-ff4dbd167e75 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Task: {'id': task-1111769, 'name': PowerOffVM_Task, 'duration_secs': 0.181178} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 993.094185] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-1cf33b4a-1ed6-417e-b050-ff4dbd167e75 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: 461a0c4c-dfdb-4669-ac72-8913c0277298] Powered off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 993.095346] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-1cf33b4a-1ed6-417e-b050-ff4dbd167e75 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: 461a0c4c-dfdb-4669-ac72-8913c0277298] Unregistering the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 993.095346] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a66eb378-3045-4121-a364-781fa37884de {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.261760] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-1cf33b4a-1ed6-417e-b050-ff4dbd167e75 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: 461a0c4c-dfdb-4669-ac72-8913c0277298] Unregistered the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 993.262128] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-1cf33b4a-1ed6-417e-b050-ff4dbd167e75 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: 461a0c4c-dfdb-4669-ac72-8913c0277298] Deleting contents of the VM from datastore datastore1 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 993.262837] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-1cf33b4a-1ed6-417e-b050-ff4dbd167e75 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Deleting the datastore file [datastore1] 461a0c4c-dfdb-4669-ac72-8913c0277298 {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 993.262837] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4ca0eae5-09a0-4576-9dcc-cd851f7e7980 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.268786] env[63515]: DEBUG oslo_vmware.api [None req-1cf33b4a-1ed6-417e-b050-ff4dbd167e75 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Waiting for the task: (returnval){ [ 993.268786] env[63515]: value = "task-1111773" [ 993.268786] env[63515]: _type = "Task" [ 993.268786] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 993.277573] env[63515]: DEBUG oslo_vmware.api [None req-1cf33b4a-1ed6-417e-b050-ff4dbd167e75 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Task: {'id': task-1111773, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.292892] env[63515]: DEBUG oslo_vmware.api [None req-642b03cc-2e02-4a86-bca6-75c3316e966c tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Task: {'id': task-1111770, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.312657] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-7c387ab3-2055-4517-9874-052fe02db033 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 68766cd4-84be-475b-8494-d7ab43a9e969] VM already powered off {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 993.312944] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-7c387ab3-2055-4517-9874-052fe02db033 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 68766cd4-84be-475b-8494-d7ab43a9e969] Updating instance '68766cd4-84be-475b-8494-d7ab43a9e969' progress to 17 {{(pid=63515) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 993.560656] env[63515]: DEBUG oslo_concurrency.lockutils [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Releasing lock "refresh_cache-5cbce760-0163-4b27-8ae3-e46c926c8916" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 993.561014] env[63515]: DEBUG nova.compute.manager [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] [instance: 5cbce760-0163-4b27-8ae3-e46c926c8916] Updated the network info_cache for instance {{(pid=63515) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9997}} [ 993.561360] env[63515]: DEBUG oslo_service.periodic_task [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=63515) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 993.565358] env[63515]: DEBUG oslo_service.periodic_task [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=63515) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 993.565962] env[63515]: DEBUG oslo_service.periodic_task [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=63515) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 993.566265] env[63515]: DEBUG oslo_service.periodic_task [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=63515) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 993.566531] env[63515]: DEBUG oslo_service.periodic_task [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=63515) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 993.566815] env[63515]: DEBUG oslo_service.periodic_task [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=63515) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 993.567059] env[63515]: DEBUG nova.compute.manager [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=63515) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10545}} [ 993.567318] env[63515]: DEBUG oslo_service.periodic_task [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Running periodic task ComputeManager.update_available_resource {{(pid=63515) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 993.700645] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4d61c4a-9506-4931-b0ef-21a97b556eca {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.708418] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a6546dc-5f8f-4074-85b0-99581e97469d {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.739224] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74e889a6-fecc-47dc-bc26-67103191f68a {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.746955] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9879502-819f-4d2f-8ee0-a32cb93ec717 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.761019] env[63515]: DEBUG nova.compute.provider_tree [None req-9d22472d-269a-4ade-805d-0c1138cb0261 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 993.778715] env[63515]: DEBUG oslo_vmware.api [None req-1cf33b4a-1ed6-417e-b050-ff4dbd167e75 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Task: {'id': task-1111773, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.172361} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 993.779014] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-1cf33b4a-1ed6-417e-b050-ff4dbd167e75 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Deleted the datastore file {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 993.779196] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-1cf33b4a-1ed6-417e-b050-ff4dbd167e75 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: 461a0c4c-dfdb-4669-ac72-8913c0277298] Deleted contents of the VM from datastore datastore1 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 993.779452] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-1cf33b4a-1ed6-417e-b050-ff4dbd167e75 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: 461a0c4c-dfdb-4669-ac72-8913c0277298] Instance destroyed {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 993.779759] env[63515]: INFO nova.compute.manager [None req-1cf33b4a-1ed6-417e-b050-ff4dbd167e75 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: 461a0c4c-dfdb-4669-ac72-8913c0277298] Took 1.21 seconds to destroy the instance on the hypervisor. [ 993.779839] env[63515]: DEBUG oslo.service.loopingcall [None req-1cf33b4a-1ed6-417e-b050-ff4dbd167e75 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 993.780100] env[63515]: DEBUG nova.compute.manager [-] [instance: 461a0c4c-dfdb-4669-ac72-8913c0277298] Deallocating network for instance {{(pid=63515) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 993.780170] env[63515]: DEBUG nova.network.neutron [-] [instance: 461a0c4c-dfdb-4669-ac72-8913c0277298] deallocate_for_instance() {{(pid=63515) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 993.790817] env[63515]: DEBUG oslo_vmware.api [None req-642b03cc-2e02-4a86-bca6-75c3316e966c tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Task: {'id': task-1111770, 'name': CreateSnapshot_Task, 'duration_secs': 0.583597} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 993.791860] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-642b03cc-2e02-4a86-bca6-75c3316e966c tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] [instance: 7d572ac0-f6af-4622-96cc-e75983420222] Created Snapshot of the VM instance {{(pid=63515) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 993.792842] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e56ba34-00dc-4812-9c36-f5e8e93b2579 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.820550] env[63515]: DEBUG nova.virt.hardware [None req-7c387ab3-2055-4517-9874-052fe02db033 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T02:52:16Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=8a120570-cb06-4099-b262-554ca0ad15c5,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 993.820755] env[63515]: DEBUG nova.virt.hardware [None req-7c387ab3-2055-4517-9874-052fe02db033 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Flavor limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 993.820918] env[63515]: DEBUG nova.virt.hardware [None req-7c387ab3-2055-4517-9874-052fe02db033 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Image limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 993.821116] env[63515]: DEBUG nova.virt.hardware [None req-7c387ab3-2055-4517-9874-052fe02db033 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Flavor pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 993.821270] env[63515]: DEBUG nova.virt.hardware [None req-7c387ab3-2055-4517-9874-052fe02db033 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Image pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 993.821426] env[63515]: DEBUG nova.virt.hardware [None req-7c387ab3-2055-4517-9874-052fe02db033 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 993.821633] env[63515]: DEBUG nova.virt.hardware [None req-7c387ab3-2055-4517-9874-052fe02db033 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 993.821813] env[63515]: DEBUG nova.virt.hardware [None req-7c387ab3-2055-4517-9874-052fe02db033 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 993.821964] env[63515]: DEBUG nova.virt.hardware [None req-7c387ab3-2055-4517-9874-052fe02db033 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Got 1 possible topologies {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 993.822144] env[63515]: DEBUG nova.virt.hardware [None req-7c387ab3-2055-4517-9874-052fe02db033 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 993.822321] env[63515]: DEBUG nova.virt.hardware [None req-7c387ab3-2055-4517-9874-052fe02db033 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 993.827838] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-82a75961-6f4c-4610-a030-a5b4b4c231df {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.846284] env[63515]: DEBUG oslo_vmware.api [None req-7c387ab3-2055-4517-9874-052fe02db033 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Waiting for the task: (returnval){ [ 993.846284] env[63515]: value = "task-1111775" [ 993.846284] env[63515]: _type = "Task" [ 993.846284] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 993.856225] env[63515]: DEBUG oslo_vmware.api [None req-7c387ab3-2055-4517-9874-052fe02db033 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Task: {'id': task-1111775, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.070608] env[63515]: DEBUG oslo_concurrency.lockutils [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 994.086362] env[63515]: DEBUG nova.compute.manager [req-1d29be2a-ff75-4e98-8ff2-fe0501475876 req-2312de21-8dfa-4a39-9fd7-97ad8481dd41 service nova] [instance: 461a0c4c-dfdb-4669-ac72-8913c0277298] Received event network-vif-deleted-2536d489-b430-4d50-ba2e-36edf5e42e17 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 994.086660] env[63515]: INFO nova.compute.manager [req-1d29be2a-ff75-4e98-8ff2-fe0501475876 req-2312de21-8dfa-4a39-9fd7-97ad8481dd41 service nova] [instance: 461a0c4c-dfdb-4669-ac72-8913c0277298] Neutron deleted interface 2536d489-b430-4d50-ba2e-36edf5e42e17; detaching it from the instance and deleting it from the info cache [ 994.086882] env[63515]: DEBUG nova.network.neutron [req-1d29be2a-ff75-4e98-8ff2-fe0501475876 req-2312de21-8dfa-4a39-9fd7-97ad8481dd41 service nova] [instance: 461a0c4c-dfdb-4669-ac72-8913c0277298] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 994.264068] env[63515]: DEBUG nova.scheduler.client.report [None req-9d22472d-269a-4ade-805d-0c1138cb0261 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 994.313432] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-642b03cc-2e02-4a86-bca6-75c3316e966c tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] [instance: 7d572ac0-f6af-4622-96cc-e75983420222] Creating linked-clone VM from snapshot {{(pid=63515) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 994.314091] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-ead5afb7-ad42-415c-ab26-f854ff87d25d {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.323257] env[63515]: DEBUG oslo_vmware.api [None req-642b03cc-2e02-4a86-bca6-75c3316e966c tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Waiting for the task: (returnval){ [ 994.323257] env[63515]: value = "task-1111776" [ 994.323257] env[63515]: _type = "Task" [ 994.323257] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 994.337862] env[63515]: DEBUG oslo_vmware.api [None req-642b03cc-2e02-4a86-bca6-75c3316e966c tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Task: {'id': task-1111776, 'name': CloneVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.356237] env[63515]: DEBUG oslo_vmware.api [None req-7c387ab3-2055-4517-9874-052fe02db033 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Task: {'id': task-1111775, 'name': ReconfigVM_Task, 'duration_secs': 0.159037} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 994.356602] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-7c387ab3-2055-4517-9874-052fe02db033 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 68766cd4-84be-475b-8494-d7ab43a9e969] Updating instance '68766cd4-84be-475b-8494-d7ab43a9e969' progress to 33 {{(pid=63515) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 994.559391] env[63515]: DEBUG nova.network.neutron [-] [instance: 461a0c4c-dfdb-4669-ac72-8913c0277298] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 994.593086] env[63515]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c8b4103a-91da-46ed-85fe-03cc6a4dcd8b {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.603026] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-524dd28f-e146-47a4-84a9-6167015bc5cd {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.631033] env[63515]: DEBUG nova.compute.manager [req-1d29be2a-ff75-4e98-8ff2-fe0501475876 req-2312de21-8dfa-4a39-9fd7-97ad8481dd41 service nova] [instance: 461a0c4c-dfdb-4669-ac72-8913c0277298] Detach interface failed, port_id=2536d489-b430-4d50-ba2e-36edf5e42e17, reason: Instance 461a0c4c-dfdb-4669-ac72-8913c0277298 could not be found. {{(pid=63515) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 994.769497] env[63515]: DEBUG oslo_concurrency.lockutils [None req-9d22472d-269a-4ade-805d-0c1138cb0261 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.273s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 994.770041] env[63515]: DEBUG nova.compute.manager [None req-9d22472d-269a-4ade-805d-0c1138cb0261 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: c6ac42a9-a233-48b4-bcb2-fb791c7446bb] Start building networks asynchronously for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 994.772786] env[63515]: DEBUG oslo_concurrency.lockutils [None req-2af70737-0f47-4d49-aa40-887a3044227c tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 4.473s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 994.773514] env[63515]: DEBUG nova.objects.instance [None req-2af70737-0f47-4d49-aa40-887a3044227c tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Lazy-loading 'resources' on Instance uuid 61cf2f9c-5d35-46c4-a477-d2cef9fcd5ae {{(pid=63515) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 994.835501] env[63515]: DEBUG oslo_vmware.api [None req-642b03cc-2e02-4a86-bca6-75c3316e966c tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Task: {'id': task-1111776, 'name': CloneVM_Task} progress is 94%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.863323] env[63515]: DEBUG nova.virt.hardware [None req-7c387ab3-2055-4517-9874-052fe02db033 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T02:52:14Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=8a120570-cb06-4099-b262-554ca0ad15c5,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 994.863635] env[63515]: DEBUG nova.virt.hardware [None req-7c387ab3-2055-4517-9874-052fe02db033 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Flavor limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 994.863843] env[63515]: DEBUG nova.virt.hardware [None req-7c387ab3-2055-4517-9874-052fe02db033 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Image limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 994.864094] env[63515]: DEBUG nova.virt.hardware [None req-7c387ab3-2055-4517-9874-052fe02db033 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Flavor pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 994.864299] env[63515]: DEBUG nova.virt.hardware [None req-7c387ab3-2055-4517-9874-052fe02db033 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Image pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 994.864680] env[63515]: DEBUG nova.virt.hardware [None req-7c387ab3-2055-4517-9874-052fe02db033 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 994.864931] env[63515]: DEBUG nova.virt.hardware [None req-7c387ab3-2055-4517-9874-052fe02db033 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 994.865124] env[63515]: DEBUG nova.virt.hardware [None req-7c387ab3-2055-4517-9874-052fe02db033 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 994.865304] env[63515]: DEBUG nova.virt.hardware [None req-7c387ab3-2055-4517-9874-052fe02db033 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Got 1 possible topologies {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 994.865475] env[63515]: DEBUG nova.virt.hardware [None req-7c387ab3-2055-4517-9874-052fe02db033 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 994.865657] env[63515]: DEBUG nova.virt.hardware [None req-7c387ab3-2055-4517-9874-052fe02db033 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 994.871290] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-7c387ab3-2055-4517-9874-052fe02db033 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 68766cd4-84be-475b-8494-d7ab43a9e969] Reconfiguring VM instance instance-00000049 to detach disk 2000 {{(pid=63515) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 994.871592] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d367d6d9-e0f2-42be-9e46-362dfcb949ef {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.891154] env[63515]: DEBUG oslo_vmware.api [None req-7c387ab3-2055-4517-9874-052fe02db033 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Waiting for the task: (returnval){ [ 994.891154] env[63515]: value = "task-1111777" [ 994.891154] env[63515]: _type = "Task" [ 994.891154] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 994.899360] env[63515]: DEBUG oslo_vmware.api [None req-7c387ab3-2055-4517-9874-052fe02db033 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Task: {'id': task-1111777, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.062783] env[63515]: INFO nova.compute.manager [-] [instance: 461a0c4c-dfdb-4669-ac72-8913c0277298] Took 1.28 seconds to deallocate network for instance. [ 995.276724] env[63515]: DEBUG nova.compute.utils [None req-9d22472d-269a-4ade-805d-0c1138cb0261 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Using /dev/sd instead of None {{(pid=63515) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 995.278485] env[63515]: DEBUG nova.compute.manager [None req-9d22472d-269a-4ade-805d-0c1138cb0261 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: c6ac42a9-a233-48b4-bcb2-fb791c7446bb] Allocating IP information in the background. {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 995.282015] env[63515]: DEBUG nova.network.neutron [None req-9d22472d-269a-4ade-805d-0c1138cb0261 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: c6ac42a9-a233-48b4-bcb2-fb791c7446bb] allocate_for_instance() {{(pid=63515) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 995.344186] env[63515]: DEBUG oslo_vmware.api [None req-642b03cc-2e02-4a86-bca6-75c3316e966c tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Task: {'id': task-1111776, 'name': CloneVM_Task} progress is 94%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.393019] env[63515]: DEBUG nova.policy [None req-9d22472d-269a-4ade-805d-0c1138cb0261 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '16880abf1c5b4341800c94ada2c756c7', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7d4ada9ec35f42b19c6480a9101d21a4', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63515) authorize /opt/stack/nova/nova/policy.py:201}} [ 995.402704] env[63515]: DEBUG oslo_vmware.api [None req-7c387ab3-2055-4517-9874-052fe02db033 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Task: {'id': task-1111777, 'name': ReconfigVM_Task} progress is 99%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.516324] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4cf7bc6-3cec-4a92-9491-d95658d92403 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.524238] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e757fdc-0b33-44b2-ae62-cf8a90da09b5 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.561915] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5605c35-34dc-4221-afc8-e6d5567b1ee6 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.570138] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6702159d-51d7-470c-9043-0033fc3ecef2 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.575735] env[63515]: DEBUG oslo_concurrency.lockutils [None req-1cf33b4a-1ed6-417e-b050-ff4dbd167e75 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 995.593787] env[63515]: DEBUG nova.compute.provider_tree [None req-2af70737-0f47-4d49-aa40-887a3044227c tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 995.785965] env[63515]: DEBUG nova.compute.manager [None req-9d22472d-269a-4ade-805d-0c1138cb0261 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: c6ac42a9-a233-48b4-bcb2-fb791c7446bb] Start building block device mappings for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 995.836384] env[63515]: DEBUG nova.network.neutron [None req-9d22472d-269a-4ade-805d-0c1138cb0261 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: c6ac42a9-a233-48b4-bcb2-fb791c7446bb] Successfully created port: 161e8482-76a0-4c82-8354-af132110794d {{(pid=63515) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 995.841387] env[63515]: DEBUG oslo_vmware.api [None req-642b03cc-2e02-4a86-bca6-75c3316e966c tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Task: {'id': task-1111776, 'name': CloneVM_Task, 'duration_secs': 1.395481} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 995.842207] env[63515]: INFO nova.virt.vmwareapi.vmops [None req-642b03cc-2e02-4a86-bca6-75c3316e966c tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] [instance: 7d572ac0-f6af-4622-96cc-e75983420222] Created linked-clone VM from snapshot [ 995.847553] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e81e582-2c1b-4f44-ba02-37779637b407 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.856905] env[63515]: DEBUG nova.virt.vmwareapi.images [None req-642b03cc-2e02-4a86-bca6-75c3316e966c tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] [instance: 7d572ac0-f6af-4622-96cc-e75983420222] Uploading image cb0243fe-58ca-4459-930c-548f47818dfc {{(pid=63515) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 995.883800] env[63515]: DEBUG oslo_vmware.rw_handles [None req-642b03cc-2e02-4a86-bca6-75c3316e966c tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 995.883800] env[63515]: value = "vm-243556" [ 995.883800] env[63515]: _type = "VirtualMachine" [ 995.883800] env[63515]: }. {{(pid=63515) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 995.883800] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-192dd13b-9fff-4645-baba-443f4eec805b {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.890681] env[63515]: DEBUG oslo_vmware.rw_handles [None req-642b03cc-2e02-4a86-bca6-75c3316e966c tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Lease: (returnval){ [ 995.890681] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]5225ebe6-351e-c809-34e7-4ece59037432" [ 995.890681] env[63515]: _type = "HttpNfcLease" [ 995.890681] env[63515]: } obtained for exporting VM: (result){ [ 995.890681] env[63515]: value = "vm-243556" [ 995.890681] env[63515]: _type = "VirtualMachine" [ 995.890681] env[63515]: }. {{(pid=63515) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 995.890921] env[63515]: DEBUG oslo_vmware.api [None req-642b03cc-2e02-4a86-bca6-75c3316e966c tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Waiting for the lease: (returnval){ [ 995.890921] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]5225ebe6-351e-c809-34e7-4ece59037432" [ 995.890921] env[63515]: _type = "HttpNfcLease" [ 995.890921] env[63515]: } to be ready. {{(pid=63515) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 995.901301] env[63515]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 995.901301] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]5225ebe6-351e-c809-34e7-4ece59037432" [ 995.901301] env[63515]: _type = "HttpNfcLease" [ 995.901301] env[63515]: } is initializing. {{(pid=63515) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 995.904568] env[63515]: DEBUG oslo_vmware.api [None req-7c387ab3-2055-4517-9874-052fe02db033 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Task: {'id': task-1111777, 'name': ReconfigVM_Task} progress is 99%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.100929] env[63515]: DEBUG nova.scheduler.client.report [None req-2af70737-0f47-4d49-aa40-887a3044227c tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 996.408640] env[63515]: DEBUG oslo_vmware.api [None req-7c387ab3-2055-4517-9874-052fe02db033 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Task: {'id': task-1111777, 'name': ReconfigVM_Task, 'duration_secs': 1.223623} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 996.409708] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-7c387ab3-2055-4517-9874-052fe02db033 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 68766cd4-84be-475b-8494-d7ab43a9e969] Reconfigured VM instance instance-00000049 to detach disk 2000 {{(pid=63515) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 996.410011] env[63515]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 996.410011] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]5225ebe6-351e-c809-34e7-4ece59037432" [ 996.410011] env[63515]: _type = "HttpNfcLease" [ 996.410011] env[63515]: } is ready. {{(pid=63515) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 996.410697] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f89fa94-a5ec-4f1f-9da8-e6e5d55669cc {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.416113] env[63515]: DEBUG oslo_vmware.rw_handles [None req-642b03cc-2e02-4a86-bca6-75c3316e966c tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 996.416113] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]5225ebe6-351e-c809-34e7-4ece59037432" [ 996.416113] env[63515]: _type = "HttpNfcLease" [ 996.416113] env[63515]: }. {{(pid=63515) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 996.416113] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dcfe4a77-caa6-48fb-afbd-7210a3ea8cb4 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.424028] env[63515]: DEBUG oslo_vmware.rw_handles [None req-642b03cc-2e02-4a86-bca6-75c3316e966c tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52441ff2-014b-c0b9-8ef9-e606e9ff7e5f/disk-0.vmdk from lease info. {{(pid=63515) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 996.424259] env[63515]: DEBUG oslo_vmware.rw_handles [None req-642b03cc-2e02-4a86-bca6-75c3316e966c tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52441ff2-014b-c0b9-8ef9-e606e9ff7e5f/disk-0.vmdk for reading. {{(pid=63515) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 996.445974] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-7c387ab3-2055-4517-9874-052fe02db033 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 68766cd4-84be-475b-8494-d7ab43a9e969] Reconfiguring VM instance instance-00000049 to attach disk [datastore2] 68766cd4-84be-475b-8494-d7ab43a9e969/68766cd4-84be-475b-8494-d7ab43a9e969.vmdk or device None with type thin {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 996.448034] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-53bd3441-e0b5-4265-8270-48c1a946b1fa {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.523823] env[63515]: DEBUG oslo_vmware.api [None req-7c387ab3-2055-4517-9874-052fe02db033 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Waiting for the task: (returnval){ [ 996.523823] env[63515]: value = "task-1111779" [ 996.523823] env[63515]: _type = "Task" [ 996.523823] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 996.532738] env[63515]: DEBUG oslo_vmware.api [None req-7c387ab3-2055-4517-9874-052fe02db033 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Task: {'id': task-1111779, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.552196] env[63515]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-20396872-9a33-418b-816a-d810ec610d54 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.606489] env[63515]: DEBUG oslo_concurrency.lockutils [None req-2af70737-0f47-4d49-aa40-887a3044227c tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.833s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 996.611270] env[63515]: DEBUG oslo_concurrency.lockutils [None req-655527ff-c992-4371-875f-a71b38db6213 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 6.278s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 996.611270] env[63515]: INFO nova.compute.claims [None req-655527ff-c992-4371-875f-a71b38db6213 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: d6c5a9f9-5e5c-4b14-9b9a-7825bbf526a5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 996.646587] env[63515]: INFO nova.scheduler.client.report [None req-2af70737-0f47-4d49-aa40-887a3044227c tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Deleted allocations for instance 61cf2f9c-5d35-46c4-a477-d2cef9fcd5ae [ 996.795101] env[63515]: DEBUG nova.compute.manager [None req-9d22472d-269a-4ade-805d-0c1138cb0261 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: c6ac42a9-a233-48b4-bcb2-fb791c7446bb] Start spawning the instance on the hypervisor. {{(pid=63515) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 996.818244] env[63515]: DEBUG nova.virt.hardware [None req-9d22472d-269a-4ade-805d-0c1138cb0261 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T02:52:14Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T02:51:56Z,direct_url=,disk_format='vmdk',id=8a120570-cb06-4099-b262-554ca0ad15c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b569255cc43e42e7ae3f2b4ad37c6ef4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T02:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 996.818531] env[63515]: DEBUG nova.virt.hardware [None req-9d22472d-269a-4ade-805d-0c1138cb0261 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Flavor limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 996.818700] env[63515]: DEBUG nova.virt.hardware [None req-9d22472d-269a-4ade-805d-0c1138cb0261 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Image limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 996.818943] env[63515]: DEBUG nova.virt.hardware [None req-9d22472d-269a-4ade-805d-0c1138cb0261 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Flavor pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 996.819468] env[63515]: DEBUG nova.virt.hardware [None req-9d22472d-269a-4ade-805d-0c1138cb0261 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Image pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 996.819669] env[63515]: DEBUG nova.virt.hardware [None req-9d22472d-269a-4ade-805d-0c1138cb0261 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 996.819893] env[63515]: DEBUG nova.virt.hardware [None req-9d22472d-269a-4ade-805d-0c1138cb0261 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 996.820225] env[63515]: DEBUG nova.virt.hardware [None req-9d22472d-269a-4ade-805d-0c1138cb0261 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 996.820415] env[63515]: DEBUG nova.virt.hardware [None req-9d22472d-269a-4ade-805d-0c1138cb0261 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Got 1 possible topologies {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 996.820588] env[63515]: DEBUG nova.virt.hardware [None req-9d22472d-269a-4ade-805d-0c1138cb0261 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 996.820768] env[63515]: DEBUG nova.virt.hardware [None req-9d22472d-269a-4ade-805d-0c1138cb0261 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 996.821652] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c640880b-00a5-4f0a-8ddc-0fc4b7f5d929 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.829853] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b257e8e-d8e2-44db-944d-39323e972095 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.035421] env[63515]: DEBUG oslo_vmware.api [None req-7c387ab3-2055-4517-9874-052fe02db033 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Task: {'id': task-1111779, 'name': ReconfigVM_Task, 'duration_secs': 0.265784} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 997.037555] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-7c387ab3-2055-4517-9874-052fe02db033 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 68766cd4-84be-475b-8494-d7ab43a9e969] Reconfigured VM instance instance-00000049 to attach disk [datastore2] 68766cd4-84be-475b-8494-d7ab43a9e969/68766cd4-84be-475b-8494-d7ab43a9e969.vmdk or device None with type thin {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 997.037930] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-7c387ab3-2055-4517-9874-052fe02db033 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 68766cd4-84be-475b-8494-d7ab43a9e969] Updating instance '68766cd4-84be-475b-8494-d7ab43a9e969' progress to 50 {{(pid=63515) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 997.048392] env[63515]: DEBUG oslo_concurrency.lockutils [None req-b36e84df-7edf-4fd3-971f-5ed772c5647a tempest-ServerRescueTestJSONUnderV235-800430211 tempest-ServerRescueTestJSONUnderV235-800430211-project-member] Acquiring lock "fcfe3376-8dfa-4189-a267-ff6402713c1b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 997.048715] env[63515]: DEBUG oslo_concurrency.lockutils [None req-b36e84df-7edf-4fd3-971f-5ed772c5647a tempest-ServerRescueTestJSONUnderV235-800430211 tempest-ServerRescueTestJSONUnderV235-800430211-project-member] Lock "fcfe3376-8dfa-4189-a267-ff6402713c1b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 997.158250] env[63515]: DEBUG oslo_concurrency.lockutils [None req-2af70737-0f47-4d49-aa40-887a3044227c tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Lock "61cf2f9c-5d35-46c4-a477-d2cef9fcd5ae" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 9.357s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 997.551022] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-260e3bcf-0e4f-4f17-8ebc-e2ffd7f3ffe9 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.559883] env[63515]: DEBUG nova.compute.manager [None req-b36e84df-7edf-4fd3-971f-5ed772c5647a tempest-ServerRescueTestJSONUnderV235-800430211 tempest-ServerRescueTestJSONUnderV235-800430211-project-member] [instance: fcfe3376-8dfa-4189-a267-ff6402713c1b] Starting instance... {{(pid=63515) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 997.584880] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7f6add9-286c-48bd-87d3-3339afcf5486 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.603086] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-7c387ab3-2055-4517-9874-052fe02db033 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 68766cd4-84be-475b-8494-d7ab43a9e969] Updating instance '68766cd4-84be-475b-8494-d7ab43a9e969' progress to 67 {{(pid=63515) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 997.665262] env[63515]: DEBUG nova.compute.manager [req-aab85752-cf03-43d3-acfd-679a15cbeb8a req-a07d51a8-4f16-4646-8b18-5fd09658f14b service nova] [instance: c6ac42a9-a233-48b4-bcb2-fb791c7446bb] Received event network-vif-plugged-161e8482-76a0-4c82-8354-af132110794d {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 997.665899] env[63515]: DEBUG oslo_concurrency.lockutils [req-aab85752-cf03-43d3-acfd-679a15cbeb8a req-a07d51a8-4f16-4646-8b18-5fd09658f14b service nova] Acquiring lock "c6ac42a9-a233-48b4-bcb2-fb791c7446bb-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 997.666117] env[63515]: DEBUG oslo_concurrency.lockutils [req-aab85752-cf03-43d3-acfd-679a15cbeb8a req-a07d51a8-4f16-4646-8b18-5fd09658f14b service nova] Lock "c6ac42a9-a233-48b4-bcb2-fb791c7446bb-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 997.666311] env[63515]: DEBUG oslo_concurrency.lockutils [req-aab85752-cf03-43d3-acfd-679a15cbeb8a req-a07d51a8-4f16-4646-8b18-5fd09658f14b service nova] Lock "c6ac42a9-a233-48b4-bcb2-fb791c7446bb-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 997.666491] env[63515]: DEBUG nova.compute.manager [req-aab85752-cf03-43d3-acfd-679a15cbeb8a req-a07d51a8-4f16-4646-8b18-5fd09658f14b service nova] [instance: c6ac42a9-a233-48b4-bcb2-fb791c7446bb] No waiting events found dispatching network-vif-plugged-161e8482-76a0-4c82-8354-af132110794d {{(pid=63515) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 997.666721] env[63515]: WARNING nova.compute.manager [req-aab85752-cf03-43d3-acfd-679a15cbeb8a req-a07d51a8-4f16-4646-8b18-5fd09658f14b service nova] [instance: c6ac42a9-a233-48b4-bcb2-fb791c7446bb] Received unexpected event network-vif-plugged-161e8482-76a0-4c82-8354-af132110794d for instance with vm_state building and task_state spawning. [ 997.753201] env[63515]: DEBUG nova.network.neutron [None req-9d22472d-269a-4ade-805d-0c1138cb0261 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: c6ac42a9-a233-48b4-bcb2-fb791c7446bb] Successfully updated port: 161e8482-76a0-4c82-8354-af132110794d {{(pid=63515) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 997.841419] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9e7c3f3-8c7a-416d-8ab4-686ab2799746 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.850065] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7006dcd-2026-4d28-a9c5-b00aea374eb3 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.882483] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4ed0753-5ca8-4303-bd59-afc6217057c1 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.890471] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf5b451d-84e4-4872-84b0-a45838c441d4 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.904622] env[63515]: DEBUG nova.compute.provider_tree [None req-655527ff-c992-4371-875f-a71b38db6213 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 998.080019] env[63515]: DEBUG oslo_concurrency.lockutils [None req-b36e84df-7edf-4fd3-971f-5ed772c5647a tempest-ServerRescueTestJSONUnderV235-800430211 tempest-ServerRescueTestJSONUnderV235-800430211-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 998.177554] env[63515]: DEBUG nova.network.neutron [None req-7c387ab3-2055-4517-9874-052fe02db033 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 68766cd4-84be-475b-8494-d7ab43a9e969] Port 28b59076-140f-46b5-b4a9-0bf530a1d43d binding to destination host cpu-1 is already ACTIVE {{(pid=63515) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 998.183034] env[63515]: DEBUG oslo_concurrency.lockutils [None req-96531a62-98f3-4a7c-9020-152fd796dc38 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Acquiring lock "b76cf3dd-7b7e-4e5d-bf3c-e15bd37069ac" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 998.183374] env[63515]: DEBUG oslo_concurrency.lockutils [None req-96531a62-98f3-4a7c-9020-152fd796dc38 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Lock "b76cf3dd-7b7e-4e5d-bf3c-e15bd37069ac" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 998.256334] env[63515]: DEBUG oslo_concurrency.lockutils [None req-9d22472d-269a-4ade-805d-0c1138cb0261 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Acquiring lock "refresh_cache-c6ac42a9-a233-48b4-bcb2-fb791c7446bb" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 998.256491] env[63515]: DEBUG oslo_concurrency.lockutils [None req-9d22472d-269a-4ade-805d-0c1138cb0261 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Acquired lock "refresh_cache-c6ac42a9-a233-48b4-bcb2-fb791c7446bb" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 998.256829] env[63515]: DEBUG nova.network.neutron [None req-9d22472d-269a-4ade-805d-0c1138cb0261 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: c6ac42a9-a233-48b4-bcb2-fb791c7446bb] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 998.408261] env[63515]: DEBUG nova.scheduler.client.report [None req-655527ff-c992-4371-875f-a71b38db6213 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 998.688299] env[63515]: DEBUG nova.compute.manager [None req-96531a62-98f3-4a7c-9020-152fd796dc38 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: b76cf3dd-7b7e-4e5d-bf3c-e15bd37069ac] Starting instance... {{(pid=63515) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 998.797906] env[63515]: DEBUG nova.network.neutron [None req-9d22472d-269a-4ade-805d-0c1138cb0261 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: c6ac42a9-a233-48b4-bcb2-fb791c7446bb] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 998.914310] env[63515]: DEBUG oslo_concurrency.lockutils [None req-655527ff-c992-4371-875f-a71b38db6213 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.305s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 998.915016] env[63515]: DEBUG nova.compute.manager [None req-655527ff-c992-4371-875f-a71b38db6213 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: d6c5a9f9-5e5c-4b14-9b9a-7825bbf526a5] Start building networks asynchronously for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 998.918582] env[63515]: DEBUG oslo_concurrency.lockutils [None req-bea83ded-7b10-4210-b5be-b45b3e9e3126 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 6.300s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 998.918840] env[63515]: DEBUG nova.objects.instance [None req-bea83ded-7b10-4210-b5be-b45b3e9e3126 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Lazy-loading 'resources' on Instance uuid 63ce797d-7180-4209-ac2c-81978bf7607a {{(pid=63515) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 998.960177] env[63515]: DEBUG nova.network.neutron [None req-9d22472d-269a-4ade-805d-0c1138cb0261 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: c6ac42a9-a233-48b4-bcb2-fb791c7446bb] Updating instance_info_cache with network_info: [{"id": "161e8482-76a0-4c82-8354-af132110794d", "address": "fa:16:3e:3e:77:7f", "network": {"id": "88dfb29d-41e0-4df7-8f70-e52cda05620c", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1355578400-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7d4ada9ec35f42b19c6480a9101d21a4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "357d2811-e990-4985-9f9e-b158d10d3699", "external-id": "nsx-vlan-transportzone-641", "segmentation_id": 641, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap161e8482-76", "ovs_interfaceid": "161e8482-76a0-4c82-8354-af132110794d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 999.207844] env[63515]: DEBUG oslo_concurrency.lockutils [None req-7c387ab3-2055-4517-9874-052fe02db033 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Acquiring lock "68766cd4-84be-475b-8494-d7ab43a9e969-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 999.208112] env[63515]: DEBUG oslo_concurrency.lockutils [None req-7c387ab3-2055-4517-9874-052fe02db033 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Lock "68766cd4-84be-475b-8494-d7ab43a9e969-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 999.208247] env[63515]: DEBUG oslo_concurrency.lockutils [None req-7c387ab3-2055-4517-9874-052fe02db033 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Lock "68766cd4-84be-475b-8494-d7ab43a9e969-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 999.225184] env[63515]: DEBUG oslo_concurrency.lockutils [None req-96531a62-98f3-4a7c-9020-152fd796dc38 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 999.422409] env[63515]: DEBUG nova.compute.utils [None req-655527ff-c992-4371-875f-a71b38db6213 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Using /dev/sd instead of None {{(pid=63515) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 999.424070] env[63515]: DEBUG nova.compute.manager [None req-655527ff-c992-4371-875f-a71b38db6213 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: d6c5a9f9-5e5c-4b14-9b9a-7825bbf526a5] Allocating IP information in the background. {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 999.424269] env[63515]: DEBUG nova.network.neutron [None req-655527ff-c992-4371-875f-a71b38db6213 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: d6c5a9f9-5e5c-4b14-9b9a-7825bbf526a5] allocate_for_instance() {{(pid=63515) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 999.462954] env[63515]: DEBUG oslo_concurrency.lockutils [None req-9d22472d-269a-4ade-805d-0c1138cb0261 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Releasing lock "refresh_cache-c6ac42a9-a233-48b4-bcb2-fb791c7446bb" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 999.463351] env[63515]: DEBUG nova.compute.manager [None req-9d22472d-269a-4ade-805d-0c1138cb0261 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: c6ac42a9-a233-48b4-bcb2-fb791c7446bb] Instance network_info: |[{"id": "161e8482-76a0-4c82-8354-af132110794d", "address": "fa:16:3e:3e:77:7f", "network": {"id": "88dfb29d-41e0-4df7-8f70-e52cda05620c", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1355578400-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7d4ada9ec35f42b19c6480a9101d21a4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "357d2811-e990-4985-9f9e-b158d10d3699", "external-id": "nsx-vlan-transportzone-641", "segmentation_id": 641, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap161e8482-76", "ovs_interfaceid": "161e8482-76a0-4c82-8354-af132110794d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 999.463811] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-9d22472d-269a-4ade-805d-0c1138cb0261 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: c6ac42a9-a233-48b4-bcb2-fb791c7446bb] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:3e:77:7f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '357d2811-e990-4985-9f9e-b158d10d3699', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '161e8482-76a0-4c82-8354-af132110794d', 'vif_model': 'vmxnet3'}] {{(pid=63515) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 999.472206] env[63515]: DEBUG oslo.service.loopingcall [None req-9d22472d-269a-4ade-805d-0c1138cb0261 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 999.473784] env[63515]: DEBUG nova.policy [None req-655527ff-c992-4371-875f-a71b38db6213 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ed295ef8f0134bcaa668149ac9f4dd74', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'aef947bfaf7d4180b46b4be662e3d903', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63515) authorize /opt/stack/nova/nova/policy.py:201}} [ 999.478369] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c6ac42a9-a233-48b4-bcb2-fb791c7446bb] Creating VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 999.479059] env[63515]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3acedf50-e0ed-40bd-9dce-242c84d1574e {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.503711] env[63515]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 999.503711] env[63515]: value = "task-1111782" [ 999.503711] env[63515]: _type = "Task" [ 999.503711] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 999.515795] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1111782, 'name': CreateVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.644866] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39af97f5-e4be-4d01-9291-8cee9b0d1501 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.653351] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a118389-929e-4675-9f7a-2d831897e8fc {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.688892] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53cf3511-6fa0-49c8-b619-a9e6c0943ec8 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.692380] env[63515]: DEBUG nova.compute.manager [req-ad689873-1781-41ef-92dc-c379242bb405 req-7e3dcdbc-6e72-45ab-a1cf-259bc9561b09 service nova] [instance: c6ac42a9-a233-48b4-bcb2-fb791c7446bb] Received event network-changed-161e8482-76a0-4c82-8354-af132110794d {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 999.692572] env[63515]: DEBUG nova.compute.manager [req-ad689873-1781-41ef-92dc-c379242bb405 req-7e3dcdbc-6e72-45ab-a1cf-259bc9561b09 service nova] [instance: c6ac42a9-a233-48b4-bcb2-fb791c7446bb] Refreshing instance network info cache due to event network-changed-161e8482-76a0-4c82-8354-af132110794d. {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 999.692789] env[63515]: DEBUG oslo_concurrency.lockutils [req-ad689873-1781-41ef-92dc-c379242bb405 req-7e3dcdbc-6e72-45ab-a1cf-259bc9561b09 service nova] Acquiring lock "refresh_cache-c6ac42a9-a233-48b4-bcb2-fb791c7446bb" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 999.692933] env[63515]: DEBUG oslo_concurrency.lockutils [req-ad689873-1781-41ef-92dc-c379242bb405 req-7e3dcdbc-6e72-45ab-a1cf-259bc9561b09 service nova] Acquired lock "refresh_cache-c6ac42a9-a233-48b4-bcb2-fb791c7446bb" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 999.693106] env[63515]: DEBUG nova.network.neutron [req-ad689873-1781-41ef-92dc-c379242bb405 req-7e3dcdbc-6e72-45ab-a1cf-259bc9561b09 service nova] [instance: c6ac42a9-a233-48b4-bcb2-fb791c7446bb] Refreshing network info cache for port 161e8482-76a0-4c82-8354-af132110794d {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 999.700095] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45d1d411-7edf-4cbd-a26e-a1c4026aabb5 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.716525] env[63515]: DEBUG nova.compute.provider_tree [None req-bea83ded-7b10-4210-b5be-b45b3e9e3126 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 999.856024] env[63515]: DEBUG nova.network.neutron [None req-655527ff-c992-4371-875f-a71b38db6213 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: d6c5a9f9-5e5c-4b14-9b9a-7825bbf526a5] Successfully created port: ee82711d-2d77-4127-b804-7db65e838617 {{(pid=63515) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 999.932692] env[63515]: DEBUG nova.compute.manager [None req-655527ff-c992-4371-875f-a71b38db6213 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: d6c5a9f9-5e5c-4b14-9b9a-7825bbf526a5] Start building block device mappings for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1000.019029] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1111782, 'name': CreateVM_Task, 'duration_secs': 0.370269} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1000.019029] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c6ac42a9-a233-48b4-bcb2-fb791c7446bb] Created VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1000.019029] env[63515]: DEBUG oslo_concurrency.lockutils [None req-9d22472d-269a-4ade-805d-0c1138cb0261 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1000.019029] env[63515]: DEBUG oslo_concurrency.lockutils [None req-9d22472d-269a-4ade-805d-0c1138cb0261 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1000.019029] env[63515]: DEBUG oslo_concurrency.lockutils [None req-9d22472d-269a-4ade-805d-0c1138cb0261 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1000.019029] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-120e34b5-0ac7-443f-82ef-cac06eb53789 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.023457] env[63515]: DEBUG oslo_vmware.api [None req-9d22472d-269a-4ade-805d-0c1138cb0261 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Waiting for the task: (returnval){ [ 1000.023457] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]526fce7e-1779-2b3d-87dd-ed940c705c01" [ 1000.023457] env[63515]: _type = "Task" [ 1000.023457] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1000.031944] env[63515]: DEBUG oslo_vmware.api [None req-9d22472d-269a-4ade-805d-0c1138cb0261 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]526fce7e-1779-2b3d-87dd-ed940c705c01, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.222602] env[63515]: DEBUG nova.scheduler.client.report [None req-bea83ded-7b10-4210-b5be-b45b3e9e3126 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1000.256287] env[63515]: DEBUG oslo_concurrency.lockutils [None req-7c387ab3-2055-4517-9874-052fe02db033 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Acquiring lock "refresh_cache-68766cd4-84be-475b-8494-d7ab43a9e969" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1000.256480] env[63515]: DEBUG oslo_concurrency.lockutils [None req-7c387ab3-2055-4517-9874-052fe02db033 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Acquired lock "refresh_cache-68766cd4-84be-475b-8494-d7ab43a9e969" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1000.256680] env[63515]: DEBUG nova.network.neutron [None req-7c387ab3-2055-4517-9874-052fe02db033 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 68766cd4-84be-475b-8494-d7ab43a9e969] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1000.418702] env[63515]: DEBUG nova.network.neutron [req-ad689873-1781-41ef-92dc-c379242bb405 req-7e3dcdbc-6e72-45ab-a1cf-259bc9561b09 service nova] [instance: c6ac42a9-a233-48b4-bcb2-fb791c7446bb] Updated VIF entry in instance network info cache for port 161e8482-76a0-4c82-8354-af132110794d. {{(pid=63515) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1000.419199] env[63515]: DEBUG nova.network.neutron [req-ad689873-1781-41ef-92dc-c379242bb405 req-7e3dcdbc-6e72-45ab-a1cf-259bc9561b09 service nova] [instance: c6ac42a9-a233-48b4-bcb2-fb791c7446bb] Updating instance_info_cache with network_info: [{"id": "161e8482-76a0-4c82-8354-af132110794d", "address": "fa:16:3e:3e:77:7f", "network": {"id": "88dfb29d-41e0-4df7-8f70-e52cda05620c", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1355578400-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7d4ada9ec35f42b19c6480a9101d21a4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "357d2811-e990-4985-9f9e-b158d10d3699", "external-id": "nsx-vlan-transportzone-641", "segmentation_id": 641, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap161e8482-76", "ovs_interfaceid": "161e8482-76a0-4c82-8354-af132110794d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1000.533880] env[63515]: DEBUG oslo_vmware.api [None req-9d22472d-269a-4ade-805d-0c1138cb0261 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]526fce7e-1779-2b3d-87dd-ed940c705c01, 'name': SearchDatastore_Task, 'duration_secs': 0.009996} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1000.534231] env[63515]: DEBUG oslo_concurrency.lockutils [None req-9d22472d-269a-4ade-805d-0c1138cb0261 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1000.534468] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-9d22472d-269a-4ade-805d-0c1138cb0261 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: c6ac42a9-a233-48b4-bcb2-fb791c7446bb] Processing image 8a120570-cb06-4099-b262-554ca0ad15c5 {{(pid=63515) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1000.534711] env[63515]: DEBUG oslo_concurrency.lockutils [None req-9d22472d-269a-4ade-805d-0c1138cb0261 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1000.534852] env[63515]: DEBUG oslo_concurrency.lockutils [None req-9d22472d-269a-4ade-805d-0c1138cb0261 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1000.535047] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-9d22472d-269a-4ade-805d-0c1138cb0261 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1000.535321] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a0b49df0-a73d-4ad6-a59d-040493f4b6ed {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.543685] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-9d22472d-269a-4ade-805d-0c1138cb0261 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1000.543877] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-9d22472d-269a-4ade-805d-0c1138cb0261 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63515) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1000.544614] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f1671d27-5eed-4950-adf4-fc7e09496716 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.549924] env[63515]: DEBUG oslo_vmware.api [None req-9d22472d-269a-4ade-805d-0c1138cb0261 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Waiting for the task: (returnval){ [ 1000.549924] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]5225c4fe-2e1c-c6f2-8481-1ca3cc96632e" [ 1000.549924] env[63515]: _type = "Task" [ 1000.549924] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1000.557324] env[63515]: DEBUG oslo_vmware.api [None req-9d22472d-269a-4ade-805d-0c1138cb0261 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]5225c4fe-2e1c-c6f2-8481-1ca3cc96632e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.731049] env[63515]: DEBUG oslo_concurrency.lockutils [None req-bea83ded-7b10-4210-b5be-b45b3e9e3126 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.813s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1000.733796] env[63515]: DEBUG oslo_concurrency.lockutils [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 6.663s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1000.734030] env[63515]: DEBUG oslo_concurrency.lockutils [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1000.734375] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=63515) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1000.734581] env[63515]: DEBUG oslo_concurrency.lockutils [None req-1cf33b4a-1ed6-417e-b050-ff4dbd167e75 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 5.159s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1000.734868] env[63515]: DEBUG nova.objects.instance [None req-1cf33b4a-1ed6-417e-b050-ff4dbd167e75 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Lazy-loading 'resources' on Instance uuid 461a0c4c-dfdb-4669-ac72-8913c0277298 {{(pid=63515) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1000.737117] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-909c319e-f96e-4bd4-b9a2-9f183f53942f {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.745896] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2345586-b121-42bd-b734-84e9e2716711 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.763317] env[63515]: INFO nova.scheduler.client.report [None req-bea83ded-7b10-4210-b5be-b45b3e9e3126 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Deleted allocations for instance 63ce797d-7180-4209-ac2c-81978bf7607a [ 1000.764718] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7865287b-ebce-48f7-994f-62bc4cc07440 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.774861] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c168c995-c60c-4b33-b61f-e7d9a06bd165 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.806320] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=179811MB free_disk=171GB free_vcpus=48 pci_devices=None {{(pid=63515) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1000.806320] env[63515]: DEBUG oslo_concurrency.lockutils [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1000.921760] env[63515]: DEBUG oslo_concurrency.lockutils [req-ad689873-1781-41ef-92dc-c379242bb405 req-7e3dcdbc-6e72-45ab-a1cf-259bc9561b09 service nova] Releasing lock "refresh_cache-c6ac42a9-a233-48b4-bcb2-fb791c7446bb" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1000.940035] env[63515]: DEBUG nova.compute.manager [None req-655527ff-c992-4371-875f-a71b38db6213 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: d6c5a9f9-5e5c-4b14-9b9a-7825bbf526a5] Start spawning the instance on the hypervisor. {{(pid=63515) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1000.972157] env[63515]: DEBUG nova.virt.hardware [None req-655527ff-c992-4371-875f-a71b38db6213 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T02:52:14Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T02:51:56Z,direct_url=,disk_format='vmdk',id=8a120570-cb06-4099-b262-554ca0ad15c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b569255cc43e42e7ae3f2b4ad37c6ef4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T02:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1000.972413] env[63515]: DEBUG nova.virt.hardware [None req-655527ff-c992-4371-875f-a71b38db6213 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Flavor limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1000.972569] env[63515]: DEBUG nova.virt.hardware [None req-655527ff-c992-4371-875f-a71b38db6213 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Image limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1000.972749] env[63515]: DEBUG nova.virt.hardware [None req-655527ff-c992-4371-875f-a71b38db6213 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Flavor pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1000.976014] env[63515]: DEBUG nova.virt.hardware [None req-655527ff-c992-4371-875f-a71b38db6213 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Image pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1000.976014] env[63515]: DEBUG nova.virt.hardware [None req-655527ff-c992-4371-875f-a71b38db6213 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1000.976014] env[63515]: DEBUG nova.virt.hardware [None req-655527ff-c992-4371-875f-a71b38db6213 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1000.976014] env[63515]: DEBUG nova.virt.hardware [None req-655527ff-c992-4371-875f-a71b38db6213 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1000.976014] env[63515]: DEBUG nova.virt.hardware [None req-655527ff-c992-4371-875f-a71b38db6213 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Got 1 possible topologies {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1000.976014] env[63515]: DEBUG nova.virt.hardware [None req-655527ff-c992-4371-875f-a71b38db6213 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1000.976014] env[63515]: DEBUG nova.virt.hardware [None req-655527ff-c992-4371-875f-a71b38db6213 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1000.976014] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c775bc9e-d79a-4cf3-9b31-929203292656 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.984273] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4804790a-c0ea-499a-88b4-3225c88fbd64 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.065688] env[63515]: DEBUG oslo_vmware.api [None req-9d22472d-269a-4ade-805d-0c1138cb0261 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]5225c4fe-2e1c-c6f2-8481-1ca3cc96632e, 'name': SearchDatastore_Task, 'duration_secs': 0.010401} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1001.066799] env[63515]: DEBUG nova.network.neutron [None req-7c387ab3-2055-4517-9874-052fe02db033 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 68766cd4-84be-475b-8494-d7ab43a9e969] Updating instance_info_cache with network_info: [{"id": "28b59076-140f-46b5-b4a9-0bf530a1d43d", "address": "fa:16:3e:a3:55:0c", "network": {"id": "5758c885-e1c9-45c8-86aa-27a802101f52", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-734458138-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.204", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b18d4daacbc84e758f9f9ae4f3f09d28", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69351262-8d39-441a-85ba-3a78df436d17", "external-id": "nsx-vlan-transportzone-205", "segmentation_id": 205, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap28b59076-14", "ovs_interfaceid": "28b59076-140f-46b5-b4a9-0bf530a1d43d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1001.073022] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-146aa0b0-4d94-403f-9670-12815a359718 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.076943] env[63515]: DEBUG oslo_vmware.api [None req-9d22472d-269a-4ade-805d-0c1138cb0261 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Waiting for the task: (returnval){ [ 1001.076943] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]5208fa9c-03fa-0135-5bb6-e5f77a614dcc" [ 1001.076943] env[63515]: _type = "Task" [ 1001.076943] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1001.088097] env[63515]: DEBUG oslo_vmware.api [None req-9d22472d-269a-4ade-805d-0c1138cb0261 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]5208fa9c-03fa-0135-5bb6-e5f77a614dcc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.276541] env[63515]: DEBUG oslo_concurrency.lockutils [None req-bea83ded-7b10-4210-b5be-b45b3e9e3126 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Lock "63ce797d-7180-4209-ac2c-81978bf7607a" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 12.037s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1001.435837] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87d415b5-cb30-4601-8a01-4acfafdfde3b {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.444160] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f217858f-442b-43a7-9f02-a9525ee18d68 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.473971] env[63515]: DEBUG oslo_concurrency.lockutils [None req-92f0089c-1858-4533-be29-3b628a87ca62 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Acquiring lock "5cbce760-0163-4b27-8ae3-e46c926c8916" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1001.474269] env[63515]: DEBUG oslo_concurrency.lockutils [None req-92f0089c-1858-4533-be29-3b628a87ca62 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Lock "5cbce760-0163-4b27-8ae3-e46c926c8916" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1001.474700] env[63515]: DEBUG oslo_concurrency.lockutils [None req-92f0089c-1858-4533-be29-3b628a87ca62 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Acquiring lock "5cbce760-0163-4b27-8ae3-e46c926c8916-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1001.474700] env[63515]: DEBUG oslo_concurrency.lockutils [None req-92f0089c-1858-4533-be29-3b628a87ca62 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Lock "5cbce760-0163-4b27-8ae3-e46c926c8916-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1001.474871] env[63515]: DEBUG oslo_concurrency.lockutils [None req-92f0089c-1858-4533-be29-3b628a87ca62 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Lock "5cbce760-0163-4b27-8ae3-e46c926c8916-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1001.477758] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad8da43d-d4b1-41c5-b90d-b82d5aeed782 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.480863] env[63515]: INFO nova.compute.manager [None req-92f0089c-1858-4533-be29-3b628a87ca62 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] [instance: 5cbce760-0163-4b27-8ae3-e46c926c8916] Terminating instance [ 1001.483102] env[63515]: DEBUG nova.compute.manager [None req-92f0089c-1858-4533-be29-3b628a87ca62 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] [instance: 5cbce760-0163-4b27-8ae3-e46c926c8916] Start destroying the instance on the hypervisor. {{(pid=63515) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1001.483332] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-92f0089c-1858-4533-be29-3b628a87ca62 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] [instance: 5cbce760-0163-4b27-8ae3-e46c926c8916] Destroying instance {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1001.484225] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-671ce848-3d0d-407c-a4e9-f48b9cc4bd75 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.491997] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8466bed-a15c-44f9-a583-0171b6149818 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.497692] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-92f0089c-1858-4533-be29-3b628a87ca62 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] [instance: 5cbce760-0163-4b27-8ae3-e46c926c8916] Powering off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1001.498224] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-95f83a9e-7e2e-4890-8957-617c87f0bcc7 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.507818] env[63515]: DEBUG nova.compute.provider_tree [None req-1cf33b4a-1ed6-417e-b050-ff4dbd167e75 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1001.510940] env[63515]: DEBUG oslo_vmware.api [None req-92f0089c-1858-4533-be29-3b628a87ca62 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Waiting for the task: (returnval){ [ 1001.510940] env[63515]: value = "task-1111783" [ 1001.510940] env[63515]: _type = "Task" [ 1001.510940] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1001.519840] env[63515]: DEBUG oslo_vmware.api [None req-92f0089c-1858-4533-be29-3b628a87ca62 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Task: {'id': task-1111783, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.574836] env[63515]: DEBUG oslo_concurrency.lockutils [None req-7c387ab3-2055-4517-9874-052fe02db033 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Releasing lock "refresh_cache-68766cd4-84be-475b-8494-d7ab43a9e969" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1001.591933] env[63515]: DEBUG oslo_vmware.api [None req-9d22472d-269a-4ade-805d-0c1138cb0261 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]5208fa9c-03fa-0135-5bb6-e5f77a614dcc, 'name': SearchDatastore_Task, 'duration_secs': 0.01308} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1001.592310] env[63515]: DEBUG oslo_concurrency.lockutils [None req-9d22472d-269a-4ade-805d-0c1138cb0261 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1001.592680] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-9d22472d-269a-4ade-805d-0c1138cb0261 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk to [datastore2] c6ac42a9-a233-48b4-bcb2-fb791c7446bb/c6ac42a9-a233-48b4-bcb2-fb791c7446bb.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1001.592992] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3beffe98-9006-45e3-9283-6c87ddb15206 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.609640] env[63515]: DEBUG oslo_vmware.api [None req-9d22472d-269a-4ade-805d-0c1138cb0261 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Waiting for the task: (returnval){ [ 1001.609640] env[63515]: value = "task-1111784" [ 1001.609640] env[63515]: _type = "Task" [ 1001.609640] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1001.621493] env[63515]: DEBUG oslo_vmware.api [None req-9d22472d-269a-4ade-805d-0c1138cb0261 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Task: {'id': task-1111784, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.723462] env[63515]: DEBUG nova.compute.manager [req-bdab84b6-4611-42ab-b8db-7f4b260f7885 req-453d99ed-94a1-4ef1-af08-c9bdd706e326 service nova] [instance: d6c5a9f9-5e5c-4b14-9b9a-7825bbf526a5] Received event network-vif-plugged-ee82711d-2d77-4127-b804-7db65e838617 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1001.723727] env[63515]: DEBUG oslo_concurrency.lockutils [req-bdab84b6-4611-42ab-b8db-7f4b260f7885 req-453d99ed-94a1-4ef1-af08-c9bdd706e326 service nova] Acquiring lock "d6c5a9f9-5e5c-4b14-9b9a-7825bbf526a5-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1001.723954] env[63515]: DEBUG oslo_concurrency.lockutils [req-bdab84b6-4611-42ab-b8db-7f4b260f7885 req-453d99ed-94a1-4ef1-af08-c9bdd706e326 service nova] Lock "d6c5a9f9-5e5c-4b14-9b9a-7825bbf526a5-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1001.724347] env[63515]: DEBUG oslo_concurrency.lockutils [req-bdab84b6-4611-42ab-b8db-7f4b260f7885 req-453d99ed-94a1-4ef1-af08-c9bdd706e326 service nova] Lock "d6c5a9f9-5e5c-4b14-9b9a-7825bbf526a5-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1001.724581] env[63515]: DEBUG nova.compute.manager [req-bdab84b6-4611-42ab-b8db-7f4b260f7885 req-453d99ed-94a1-4ef1-af08-c9bdd706e326 service nova] [instance: d6c5a9f9-5e5c-4b14-9b9a-7825bbf526a5] No waiting events found dispatching network-vif-plugged-ee82711d-2d77-4127-b804-7db65e838617 {{(pid=63515) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1001.724752] env[63515]: WARNING nova.compute.manager [req-bdab84b6-4611-42ab-b8db-7f4b260f7885 req-453d99ed-94a1-4ef1-af08-c9bdd706e326 service nova] [instance: d6c5a9f9-5e5c-4b14-9b9a-7825bbf526a5] Received unexpected event network-vif-plugged-ee82711d-2d77-4127-b804-7db65e838617 for instance with vm_state building and task_state spawning. [ 1001.735676] env[63515]: DEBUG nova.network.neutron [None req-655527ff-c992-4371-875f-a71b38db6213 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: d6c5a9f9-5e5c-4b14-9b9a-7825bbf526a5] Successfully updated port: ee82711d-2d77-4127-b804-7db65e838617 {{(pid=63515) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1002.012840] env[63515]: DEBUG nova.scheduler.client.report [None req-1cf33b4a-1ed6-417e-b050-ff4dbd167e75 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1002.027171] env[63515]: DEBUG oslo_vmware.api [None req-92f0089c-1858-4533-be29-3b628a87ca62 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Task: {'id': task-1111783, 'name': PowerOffVM_Task, 'duration_secs': 0.188561} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1002.027403] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-92f0089c-1858-4533-be29-3b628a87ca62 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] [instance: 5cbce760-0163-4b27-8ae3-e46c926c8916] Powered off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1002.027612] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-92f0089c-1858-4533-be29-3b628a87ca62 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] [instance: 5cbce760-0163-4b27-8ae3-e46c926c8916] Unregistering the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1002.027910] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-0244413d-669a-4294-88d2-6eb18bfa16b1 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.096138] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-92f0089c-1858-4533-be29-3b628a87ca62 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] [instance: 5cbce760-0163-4b27-8ae3-e46c926c8916] Unregistered the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1002.096395] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-92f0089c-1858-4533-be29-3b628a87ca62 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] [instance: 5cbce760-0163-4b27-8ae3-e46c926c8916] Deleting contents of the VM from datastore datastore1 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1002.096607] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-92f0089c-1858-4533-be29-3b628a87ca62 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Deleting the datastore file [datastore1] 5cbce760-0163-4b27-8ae3-e46c926c8916 {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1002.096983] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-44e9c611-6ce3-4a89-b8fd-0ef162ca6ffd {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.104169] env[63515]: DEBUG oslo_vmware.api [None req-92f0089c-1858-4533-be29-3b628a87ca62 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Waiting for the task: (returnval){ [ 1002.104169] env[63515]: value = "task-1111786" [ 1002.104169] env[63515]: _type = "Task" [ 1002.104169] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1002.109190] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bef2f700-a03a-44a3-848a-a2fd4a4a1ffa {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.118912] env[63515]: DEBUG oslo_vmware.api [None req-92f0089c-1858-4533-be29-3b628a87ca62 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Task: {'id': task-1111786, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.137100] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff51a535-63de-4186-8e83-8efdc528ffc6 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.143645] env[63515]: DEBUG oslo_vmware.api [None req-9d22472d-269a-4ade-805d-0c1138cb0261 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Task: {'id': task-1111784, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.152035] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-7c387ab3-2055-4517-9874-052fe02db033 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 68766cd4-84be-475b-8494-d7ab43a9e969] Updating instance '68766cd4-84be-475b-8494-d7ab43a9e969' progress to 83 {{(pid=63515) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1002.240391] env[63515]: DEBUG oslo_concurrency.lockutils [None req-655527ff-c992-4371-875f-a71b38db6213 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Acquiring lock "refresh_cache-d6c5a9f9-5e5c-4b14-9b9a-7825bbf526a5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1002.240391] env[63515]: DEBUG oslo_concurrency.lockutils [None req-655527ff-c992-4371-875f-a71b38db6213 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Acquired lock "refresh_cache-d6c5a9f9-5e5c-4b14-9b9a-7825bbf526a5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1002.240391] env[63515]: DEBUG nova.network.neutron [None req-655527ff-c992-4371-875f-a71b38db6213 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: d6c5a9f9-5e5c-4b14-9b9a-7825bbf526a5] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1002.521124] env[63515]: DEBUG oslo_concurrency.lockutils [None req-1cf33b4a-1ed6-417e-b050-ff4dbd167e75 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.786s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1002.523961] env[63515]: DEBUG oslo_concurrency.lockutils [None req-b36e84df-7edf-4fd3-971f-5ed772c5647a tempest-ServerRescueTestJSONUnderV235-800430211 tempest-ServerRescueTestJSONUnderV235-800430211-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.444s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1002.525537] env[63515]: INFO nova.compute.claims [None req-b36e84df-7edf-4fd3-971f-5ed772c5647a tempest-ServerRescueTestJSONUnderV235-800430211 tempest-ServerRescueTestJSONUnderV235-800430211-project-member] [instance: fcfe3376-8dfa-4189-a267-ff6402713c1b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1002.542413] env[63515]: INFO nova.scheduler.client.report [None req-1cf33b4a-1ed6-417e-b050-ff4dbd167e75 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Deleted allocations for instance 461a0c4c-dfdb-4669-ac72-8913c0277298 [ 1002.615879] env[63515]: DEBUG oslo_vmware.api [None req-92f0089c-1858-4533-be29-3b628a87ca62 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Task: {'id': task-1111786, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.388886} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1002.619603] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-92f0089c-1858-4533-be29-3b628a87ca62 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Deleted the datastore file {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1002.619866] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-92f0089c-1858-4533-be29-3b628a87ca62 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] [instance: 5cbce760-0163-4b27-8ae3-e46c926c8916] Deleted contents of the VM from datastore datastore1 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1002.620098] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-92f0089c-1858-4533-be29-3b628a87ca62 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] [instance: 5cbce760-0163-4b27-8ae3-e46c926c8916] Instance destroyed {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1002.620817] env[63515]: INFO nova.compute.manager [None req-92f0089c-1858-4533-be29-3b628a87ca62 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] [instance: 5cbce760-0163-4b27-8ae3-e46c926c8916] Took 1.14 seconds to destroy the instance on the hypervisor. [ 1002.620817] env[63515]: DEBUG oslo.service.loopingcall [None req-92f0089c-1858-4533-be29-3b628a87ca62 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1002.620817] env[63515]: DEBUG nova.compute.manager [-] [instance: 5cbce760-0163-4b27-8ae3-e46c926c8916] Deallocating network for instance {{(pid=63515) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1002.620981] env[63515]: DEBUG nova.network.neutron [-] [instance: 5cbce760-0163-4b27-8ae3-e46c926c8916] deallocate_for_instance() {{(pid=63515) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1002.639017] env[63515]: DEBUG oslo_vmware.api [None req-9d22472d-269a-4ade-805d-0c1138cb0261 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Task: {'id': task-1111784, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.73818} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1002.639017] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-9d22472d-269a-4ade-805d-0c1138cb0261 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk to [datastore2] c6ac42a9-a233-48b4-bcb2-fb791c7446bb/c6ac42a9-a233-48b4-bcb2-fb791c7446bb.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1002.639017] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-9d22472d-269a-4ade-805d-0c1138cb0261 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: c6ac42a9-a233-48b4-bcb2-fb791c7446bb] Extending root virtual disk to 1048576 {{(pid=63515) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1002.639017] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-97370627-0098-4da8-94ab-fd476d3d31c5 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.643375] env[63515]: DEBUG oslo_vmware.api [None req-9d22472d-269a-4ade-805d-0c1138cb0261 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Waiting for the task: (returnval){ [ 1002.643375] env[63515]: value = "task-1111787" [ 1002.643375] env[63515]: _type = "Task" [ 1002.643375] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1002.652636] env[63515]: DEBUG oslo_vmware.api [None req-9d22472d-269a-4ade-805d-0c1138cb0261 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Task: {'id': task-1111787, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.663151] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-7c387ab3-2055-4517-9874-052fe02db033 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 68766cd4-84be-475b-8494-d7ab43a9e969] Updating instance '68766cd4-84be-475b-8494-d7ab43a9e969' progress to 100 {{(pid=63515) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1002.780899] env[63515]: DEBUG nova.network.neutron [None req-655527ff-c992-4371-875f-a71b38db6213 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: d6c5a9f9-5e5c-4b14-9b9a-7825bbf526a5] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1003.058614] env[63515]: DEBUG oslo_concurrency.lockutils [None req-1cf33b4a-1ed6-417e-b050-ff4dbd167e75 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Lock "461a0c4c-dfdb-4669-ac72-8913c0277298" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 10.496s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1003.095485] env[63515]: DEBUG nova.network.neutron [None req-655527ff-c992-4371-875f-a71b38db6213 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: d6c5a9f9-5e5c-4b14-9b9a-7825bbf526a5] Updating instance_info_cache with network_info: [{"id": "ee82711d-2d77-4127-b804-7db65e838617", "address": "fa:16:3e:21:f3:e4", "network": {"id": "cec614cb-7a5b-4b16-a83d-061ff62d6677", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1501468300-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "aef947bfaf7d4180b46b4be662e3d903", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3a0ddd7d-c321-4187-bdd8-b19044ea2c4a", "external-id": "nsx-vlan-transportzone-747", "segmentation_id": 747, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapee82711d-2d", "ovs_interfaceid": "ee82711d-2d77-4127-b804-7db65e838617", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1003.153229] env[63515]: DEBUG oslo_vmware.api [None req-9d22472d-269a-4ade-805d-0c1138cb0261 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Task: {'id': task-1111787, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.088652} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1003.153514] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-9d22472d-269a-4ade-805d-0c1138cb0261 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: c6ac42a9-a233-48b4-bcb2-fb791c7446bb] Extended root virtual disk {{(pid=63515) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1003.154379] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2289f92-cc3e-49c8-ae4e-a58124450475 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.179331] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-9d22472d-269a-4ade-805d-0c1138cb0261 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: c6ac42a9-a233-48b4-bcb2-fb791c7446bb] Reconfiguring VM instance instance-00000061 to attach disk [datastore2] c6ac42a9-a233-48b4-bcb2-fb791c7446bb/c6ac42a9-a233-48b4-bcb2-fb791c7446bb.vmdk or device None with type sparse {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1003.180698] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9711a5f0-349a-4f55-a649-b322d859bc92 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.206076] env[63515]: DEBUG oslo_vmware.api [None req-9d22472d-269a-4ade-805d-0c1138cb0261 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Waiting for the task: (returnval){ [ 1003.206076] env[63515]: value = "task-1111788" [ 1003.206076] env[63515]: _type = "Task" [ 1003.206076] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1003.215215] env[63515]: DEBUG oslo_vmware.api [None req-9d22472d-269a-4ade-805d-0c1138cb0261 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Task: {'id': task-1111788, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.561447] env[63515]: DEBUG nova.network.neutron [-] [instance: 5cbce760-0163-4b27-8ae3-e46c926c8916] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1003.583395] env[63515]: DEBUG oslo_concurrency.lockutils [None req-e5acb69c-ccc9-4835-a62a-46a52c0ee313 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Acquiring lock "d024b6b2-dd10-4112-89e8-ced57efe8208" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1003.583650] env[63515]: DEBUG oslo_concurrency.lockutils [None req-e5acb69c-ccc9-4835-a62a-46a52c0ee313 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Lock "d024b6b2-dd10-4112-89e8-ced57efe8208" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1003.583860] env[63515]: DEBUG oslo_concurrency.lockutils [None req-e5acb69c-ccc9-4835-a62a-46a52c0ee313 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Acquiring lock "d024b6b2-dd10-4112-89e8-ced57efe8208-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1003.584079] env[63515]: DEBUG oslo_concurrency.lockutils [None req-e5acb69c-ccc9-4835-a62a-46a52c0ee313 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Lock "d024b6b2-dd10-4112-89e8-ced57efe8208-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1003.584608] env[63515]: DEBUG oslo_concurrency.lockutils [None req-e5acb69c-ccc9-4835-a62a-46a52c0ee313 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Lock "d024b6b2-dd10-4112-89e8-ced57efe8208-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1003.586863] env[63515]: INFO nova.compute.manager [None req-e5acb69c-ccc9-4835-a62a-46a52c0ee313 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: d024b6b2-dd10-4112-89e8-ced57efe8208] Terminating instance [ 1003.589311] env[63515]: DEBUG nova.compute.manager [None req-e5acb69c-ccc9-4835-a62a-46a52c0ee313 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: d024b6b2-dd10-4112-89e8-ced57efe8208] Start destroying the instance on the hypervisor. {{(pid=63515) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1003.589517] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-e5acb69c-ccc9-4835-a62a-46a52c0ee313 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: d024b6b2-dd10-4112-89e8-ced57efe8208] Destroying instance {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1003.590361] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-efc7ad8a-41b1-4e57-8d17-0ff7165c7ace {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.598679] env[63515]: DEBUG oslo_concurrency.lockutils [None req-655527ff-c992-4371-875f-a71b38db6213 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Releasing lock "refresh_cache-d6c5a9f9-5e5c-4b14-9b9a-7825bbf526a5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1003.599018] env[63515]: DEBUG nova.compute.manager [None req-655527ff-c992-4371-875f-a71b38db6213 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: d6c5a9f9-5e5c-4b14-9b9a-7825bbf526a5] Instance network_info: |[{"id": "ee82711d-2d77-4127-b804-7db65e838617", "address": "fa:16:3e:21:f3:e4", "network": {"id": "cec614cb-7a5b-4b16-a83d-061ff62d6677", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1501468300-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "aef947bfaf7d4180b46b4be662e3d903", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3a0ddd7d-c321-4187-bdd8-b19044ea2c4a", "external-id": "nsx-vlan-transportzone-747", "segmentation_id": 747, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapee82711d-2d", "ovs_interfaceid": "ee82711d-2d77-4127-b804-7db65e838617", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1003.599355] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-e5acb69c-ccc9-4835-a62a-46a52c0ee313 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: d024b6b2-dd10-4112-89e8-ced57efe8208] Powering off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1003.602165] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-655527ff-c992-4371-875f-a71b38db6213 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: d6c5a9f9-5e5c-4b14-9b9a-7825bbf526a5] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:21:f3:e4', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '3a0ddd7d-c321-4187-bdd8-b19044ea2c4a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ee82711d-2d77-4127-b804-7db65e838617', 'vif_model': 'vmxnet3'}] {{(pid=63515) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1003.610153] env[63515]: DEBUG oslo.service.loopingcall [None req-655527ff-c992-4371-875f-a71b38db6213 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1003.610780] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-09b318f3-4e55-4ac1-a1ee-4af16ca45f23 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.613015] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d6c5a9f9-5e5c-4b14-9b9a-7825bbf526a5] Creating VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1003.613249] env[63515]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ccb66799-5631-43bb-ae54-92746a523ebe {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.634297] env[63515]: DEBUG oslo_vmware.api [None req-e5acb69c-ccc9-4835-a62a-46a52c0ee313 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Waiting for the task: (returnval){ [ 1003.634297] env[63515]: value = "task-1111789" [ 1003.634297] env[63515]: _type = "Task" [ 1003.634297] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1003.638428] env[63515]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1003.638428] env[63515]: value = "task-1111790" [ 1003.638428] env[63515]: _type = "Task" [ 1003.638428] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1003.645369] env[63515]: DEBUG oslo_vmware.api [None req-e5acb69c-ccc9-4835-a62a-46a52c0ee313 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Task: {'id': task-1111789, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.651162] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1111790, 'name': CreateVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.720324] env[63515]: DEBUG oslo_vmware.api [None req-9d22472d-269a-4ade-805d-0c1138cb0261 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Task: {'id': task-1111788, 'name': ReconfigVM_Task, 'duration_secs': 0.322781} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1003.720479] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-9d22472d-269a-4ade-805d-0c1138cb0261 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: c6ac42a9-a233-48b4-bcb2-fb791c7446bb] Reconfigured VM instance instance-00000061 to attach disk [datastore2] c6ac42a9-a233-48b4-bcb2-fb791c7446bb/c6ac42a9-a233-48b4-bcb2-fb791c7446bb.vmdk or device None with type sparse {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1003.721204] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-7a381cb2-78b2-4a65-a66c-12637a601f30 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.730581] env[63515]: DEBUG oslo_vmware.api [None req-9d22472d-269a-4ade-805d-0c1138cb0261 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Waiting for the task: (returnval){ [ 1003.730581] env[63515]: value = "task-1111791" [ 1003.730581] env[63515]: _type = "Task" [ 1003.730581] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1003.741088] env[63515]: DEBUG oslo_vmware.api [None req-9d22472d-269a-4ade-805d-0c1138cb0261 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Task: {'id': task-1111791, 'name': Rename_Task} progress is 6%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.756406] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe0c40e2-b2d5-410b-abbb-5e4b2ef8c333 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.762120] env[63515]: DEBUG nova.compute.manager [req-4cc9af7f-2859-4faa-aae0-9693a9f8aa2f req-0b17df13-b1c7-4155-8a23-1ccfb2206f7e service nova] [instance: d6c5a9f9-5e5c-4b14-9b9a-7825bbf526a5] Received event network-changed-ee82711d-2d77-4127-b804-7db65e838617 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1003.762364] env[63515]: DEBUG nova.compute.manager [req-4cc9af7f-2859-4faa-aae0-9693a9f8aa2f req-0b17df13-b1c7-4155-8a23-1ccfb2206f7e service nova] [instance: d6c5a9f9-5e5c-4b14-9b9a-7825bbf526a5] Refreshing instance network info cache due to event network-changed-ee82711d-2d77-4127-b804-7db65e838617. {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1003.762585] env[63515]: DEBUG oslo_concurrency.lockutils [req-4cc9af7f-2859-4faa-aae0-9693a9f8aa2f req-0b17df13-b1c7-4155-8a23-1ccfb2206f7e service nova] Acquiring lock "refresh_cache-d6c5a9f9-5e5c-4b14-9b9a-7825bbf526a5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1003.762786] env[63515]: DEBUG oslo_concurrency.lockutils [req-4cc9af7f-2859-4faa-aae0-9693a9f8aa2f req-0b17df13-b1c7-4155-8a23-1ccfb2206f7e service nova] Acquired lock "refresh_cache-d6c5a9f9-5e5c-4b14-9b9a-7825bbf526a5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1003.762979] env[63515]: DEBUG nova.network.neutron [req-4cc9af7f-2859-4faa-aae0-9693a9f8aa2f req-0b17df13-b1c7-4155-8a23-1ccfb2206f7e service nova] [instance: d6c5a9f9-5e5c-4b14-9b9a-7825bbf526a5] Refreshing network info cache for port ee82711d-2d77-4127-b804-7db65e838617 {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1003.769630] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-064a9015-d5df-4dd7-9204-ad0c70adedcc {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.803500] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4291dc17-72c9-404a-81ab-cba2ed762d3a {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.811903] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6399ad01-bf95-45ca-895d-2787cefb520d {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.827027] env[63515]: DEBUG nova.compute.provider_tree [None req-b36e84df-7edf-4fd3-971f-5ed772c5647a tempest-ServerRescueTestJSONUnderV235-800430211 tempest-ServerRescueTestJSONUnderV235-800430211-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1004.064233] env[63515]: INFO nova.compute.manager [-] [instance: 5cbce760-0163-4b27-8ae3-e46c926c8916] Took 1.44 seconds to deallocate network for instance. [ 1004.145290] env[63515]: DEBUG oslo_vmware.api [None req-e5acb69c-ccc9-4835-a62a-46a52c0ee313 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Task: {'id': task-1111789, 'name': PowerOffVM_Task, 'duration_secs': 0.20335} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1004.145972] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-e5acb69c-ccc9-4835-a62a-46a52c0ee313 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: d024b6b2-dd10-4112-89e8-ced57efe8208] Powered off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1004.146234] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-e5acb69c-ccc9-4835-a62a-46a52c0ee313 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: d024b6b2-dd10-4112-89e8-ced57efe8208] Unregistering the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1004.146809] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-8ccc9db7-5769-4ce2-9200-535917a5f60a {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.151270] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1111790, 'name': CreateVM_Task, 'duration_secs': 0.403882} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1004.151730] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d6c5a9f9-5e5c-4b14-9b9a-7825bbf526a5] Created VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1004.152478] env[63515]: DEBUG oslo_concurrency.lockutils [None req-655527ff-c992-4371-875f-a71b38db6213 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1004.152643] env[63515]: DEBUG oslo_concurrency.lockutils [None req-655527ff-c992-4371-875f-a71b38db6213 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1004.152960] env[63515]: DEBUG oslo_concurrency.lockutils [None req-655527ff-c992-4371-875f-a71b38db6213 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1004.153218] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9ec1a67a-6e24-4c38-b28d-30ead3f96b7a {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.157548] env[63515]: DEBUG oslo_vmware.api [None req-655527ff-c992-4371-875f-a71b38db6213 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Waiting for the task: (returnval){ [ 1004.157548] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]5216df79-e5cd-6cd6-a673-f601135111d5" [ 1004.157548] env[63515]: _type = "Task" [ 1004.157548] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1004.165674] env[63515]: DEBUG oslo_vmware.api [None req-655527ff-c992-4371-875f-a71b38db6213 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]5216df79-e5cd-6cd6-a673-f601135111d5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.214904] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-e5acb69c-ccc9-4835-a62a-46a52c0ee313 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: d024b6b2-dd10-4112-89e8-ced57efe8208] Unregistered the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1004.215178] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-e5acb69c-ccc9-4835-a62a-46a52c0ee313 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: d024b6b2-dd10-4112-89e8-ced57efe8208] Deleting contents of the VM from datastore datastore1 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1004.215369] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-e5acb69c-ccc9-4835-a62a-46a52c0ee313 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Deleting the datastore file [datastore1] d024b6b2-dd10-4112-89e8-ced57efe8208 {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1004.215621] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-85d2ef21-8f5f-457c-bdf4-742b57bc89f2 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.222265] env[63515]: DEBUG oslo_vmware.api [None req-e5acb69c-ccc9-4835-a62a-46a52c0ee313 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Waiting for the task: (returnval){ [ 1004.222265] env[63515]: value = "task-1111793" [ 1004.222265] env[63515]: _type = "Task" [ 1004.222265] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1004.230806] env[63515]: DEBUG oslo_vmware.api [None req-e5acb69c-ccc9-4835-a62a-46a52c0ee313 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Task: {'id': task-1111793, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.240177] env[63515]: DEBUG oslo_vmware.api [None req-9d22472d-269a-4ade-805d-0c1138cb0261 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Task: {'id': task-1111791, 'name': Rename_Task, 'duration_secs': 0.171355} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1004.240332] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-9d22472d-269a-4ade-805d-0c1138cb0261 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: c6ac42a9-a233-48b4-bcb2-fb791c7446bb] Powering on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1004.240477] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-352e2d5c-b3b4-4f0b-a6ba-165fb75c2e51 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.246618] env[63515]: DEBUG oslo_vmware.api [None req-9d22472d-269a-4ade-805d-0c1138cb0261 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Waiting for the task: (returnval){ [ 1004.246618] env[63515]: value = "task-1111794" [ 1004.246618] env[63515]: _type = "Task" [ 1004.246618] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1004.254550] env[63515]: DEBUG oslo_vmware.api [None req-9d22472d-269a-4ade-805d-0c1138cb0261 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Task: {'id': task-1111794, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.329829] env[63515]: DEBUG nova.scheduler.client.report [None req-b36e84df-7edf-4fd3-971f-5ed772c5647a tempest-ServerRescueTestJSONUnderV235-800430211 tempest-ServerRescueTestJSONUnderV235-800430211-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1004.481967] env[63515]: DEBUG nova.network.neutron [req-4cc9af7f-2859-4faa-aae0-9693a9f8aa2f req-0b17df13-b1c7-4155-8a23-1ccfb2206f7e service nova] [instance: d6c5a9f9-5e5c-4b14-9b9a-7825bbf526a5] Updated VIF entry in instance network info cache for port ee82711d-2d77-4127-b804-7db65e838617. {{(pid=63515) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1004.482360] env[63515]: DEBUG nova.network.neutron [req-4cc9af7f-2859-4faa-aae0-9693a9f8aa2f req-0b17df13-b1c7-4155-8a23-1ccfb2206f7e service nova] [instance: d6c5a9f9-5e5c-4b14-9b9a-7825bbf526a5] Updating instance_info_cache with network_info: [{"id": "ee82711d-2d77-4127-b804-7db65e838617", "address": "fa:16:3e:21:f3:e4", "network": {"id": "cec614cb-7a5b-4b16-a83d-061ff62d6677", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1501468300-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "aef947bfaf7d4180b46b4be662e3d903", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3a0ddd7d-c321-4187-bdd8-b19044ea2c4a", "external-id": "nsx-vlan-transportzone-747", "segmentation_id": 747, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapee82711d-2d", "ovs_interfaceid": "ee82711d-2d77-4127-b804-7db65e838617", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1004.571455] env[63515]: DEBUG oslo_concurrency.lockutils [None req-92f0089c-1858-4533-be29-3b628a87ca62 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1004.614492] env[63515]: DEBUG oslo_concurrency.lockutils [None req-b6bf67af-7b53-4121-830a-6dc78e8dcba3 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Acquiring lock "68766cd4-84be-475b-8494-d7ab43a9e969" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1004.614722] env[63515]: DEBUG oslo_concurrency.lockutils [None req-b6bf67af-7b53-4121-830a-6dc78e8dcba3 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Lock "68766cd4-84be-475b-8494-d7ab43a9e969" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.001s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1004.614783] env[63515]: DEBUG nova.compute.manager [None req-b6bf67af-7b53-4121-830a-6dc78e8dcba3 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 68766cd4-84be-475b-8494-d7ab43a9e969] Going to confirm migration 4 {{(pid=63515) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 1004.668114] env[63515]: DEBUG oslo_vmware.api [None req-655527ff-c992-4371-875f-a71b38db6213 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]5216df79-e5cd-6cd6-a673-f601135111d5, 'name': SearchDatastore_Task, 'duration_secs': 0.010267} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1004.668502] env[63515]: DEBUG oslo_concurrency.lockutils [None req-655527ff-c992-4371-875f-a71b38db6213 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1004.668766] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-655527ff-c992-4371-875f-a71b38db6213 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: d6c5a9f9-5e5c-4b14-9b9a-7825bbf526a5] Processing image 8a120570-cb06-4099-b262-554ca0ad15c5 {{(pid=63515) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1004.669091] env[63515]: DEBUG oslo_concurrency.lockutils [None req-655527ff-c992-4371-875f-a71b38db6213 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1004.669290] env[63515]: DEBUG oslo_concurrency.lockutils [None req-655527ff-c992-4371-875f-a71b38db6213 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1004.669507] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-655527ff-c992-4371-875f-a71b38db6213 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1004.669836] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-326c0128-2575-46c1-9102-57b00e784e94 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.679916] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-655527ff-c992-4371-875f-a71b38db6213 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1004.680239] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-655527ff-c992-4371-875f-a71b38db6213 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63515) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1004.681271] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fa4c1453-355e-481e-93b6-423cec48ed5f {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.686998] env[63515]: DEBUG oslo_vmware.api [None req-655527ff-c992-4371-875f-a71b38db6213 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Waiting for the task: (returnval){ [ 1004.686998] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]5212f5dd-3da7-07b0-a0e0-74109aa1e18f" [ 1004.686998] env[63515]: _type = "Task" [ 1004.686998] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1004.694866] env[63515]: DEBUG oslo_vmware.api [None req-655527ff-c992-4371-875f-a71b38db6213 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]5212f5dd-3da7-07b0-a0e0-74109aa1e18f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.731690] env[63515]: DEBUG oslo_vmware.api [None req-e5acb69c-ccc9-4835-a62a-46a52c0ee313 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Task: {'id': task-1111793, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.190921} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1004.731959] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-e5acb69c-ccc9-4835-a62a-46a52c0ee313 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Deleted the datastore file {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1004.732168] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-e5acb69c-ccc9-4835-a62a-46a52c0ee313 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: d024b6b2-dd10-4112-89e8-ced57efe8208] Deleted contents of the VM from datastore datastore1 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1004.732360] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-e5acb69c-ccc9-4835-a62a-46a52c0ee313 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: d024b6b2-dd10-4112-89e8-ced57efe8208] Instance destroyed {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1004.732537] env[63515]: INFO nova.compute.manager [None req-e5acb69c-ccc9-4835-a62a-46a52c0ee313 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: d024b6b2-dd10-4112-89e8-ced57efe8208] Took 1.14 seconds to destroy the instance on the hypervisor. [ 1004.732875] env[63515]: DEBUG oslo.service.loopingcall [None req-e5acb69c-ccc9-4835-a62a-46a52c0ee313 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1004.732978] env[63515]: DEBUG nova.compute.manager [-] [instance: d024b6b2-dd10-4112-89e8-ced57efe8208] Deallocating network for instance {{(pid=63515) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1004.733086] env[63515]: DEBUG nova.network.neutron [-] [instance: d024b6b2-dd10-4112-89e8-ced57efe8208] deallocate_for_instance() {{(pid=63515) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1004.757548] env[63515]: DEBUG oslo_vmware.api [None req-9d22472d-269a-4ade-805d-0c1138cb0261 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Task: {'id': task-1111794, 'name': PowerOnVM_Task} progress is 98%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.820776] env[63515]: DEBUG oslo_vmware.rw_handles [None req-642b03cc-2e02-4a86-bca6-75c3316e966c tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52441ff2-014b-c0b9-8ef9-e606e9ff7e5f/disk-0.vmdk. {{(pid=63515) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1004.821716] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a268aad5-ec63-4a78-9044-eeb0bf624fda {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.827711] env[63515]: DEBUG oslo_vmware.rw_handles [None req-642b03cc-2e02-4a86-bca6-75c3316e966c tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52441ff2-014b-c0b9-8ef9-e606e9ff7e5f/disk-0.vmdk is in state: ready. {{(pid=63515) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1004.827880] env[63515]: ERROR oslo_vmware.rw_handles [None req-642b03cc-2e02-4a86-bca6-75c3316e966c tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52441ff2-014b-c0b9-8ef9-e606e9ff7e5f/disk-0.vmdk due to incomplete transfer. [ 1004.828105] env[63515]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-b9e55f87-a4aa-4b48-bc89-ecbaec65c5c4 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.834972] env[63515]: DEBUG oslo_vmware.rw_handles [None req-642b03cc-2e02-4a86-bca6-75c3316e966c tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52441ff2-014b-c0b9-8ef9-e606e9ff7e5f/disk-0.vmdk. {{(pid=63515) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 1004.835196] env[63515]: DEBUG nova.virt.vmwareapi.images [None req-642b03cc-2e02-4a86-bca6-75c3316e966c tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] [instance: 7d572ac0-f6af-4622-96cc-e75983420222] Uploaded image cb0243fe-58ca-4459-930c-548f47818dfc to the Glance image server {{(pid=63515) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 1004.837521] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-642b03cc-2e02-4a86-bca6-75c3316e966c tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] [instance: 7d572ac0-f6af-4622-96cc-e75983420222] Destroying the VM {{(pid=63515) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 1004.838330] env[63515]: DEBUG oslo_concurrency.lockutils [None req-b36e84df-7edf-4fd3-971f-5ed772c5647a tempest-ServerRescueTestJSONUnderV235-800430211 tempest-ServerRescueTestJSONUnderV235-800430211-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.314s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1004.838836] env[63515]: DEBUG nova.compute.manager [None req-b36e84df-7edf-4fd3-971f-5ed772c5647a tempest-ServerRescueTestJSONUnderV235-800430211 tempest-ServerRescueTestJSONUnderV235-800430211-project-member] [instance: fcfe3376-8dfa-4189-a267-ff6402713c1b] Start building networks asynchronously for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1004.841643] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-5c73af7e-db32-4e25-af5b-a8766e3fc323 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.843907] env[63515]: DEBUG oslo_concurrency.lockutils [None req-96531a62-98f3-4a7c-9020-152fd796dc38 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.619s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1004.845425] env[63515]: INFO nova.compute.claims [None req-96531a62-98f3-4a7c-9020-152fd796dc38 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: b76cf3dd-7b7e-4e5d-bf3c-e15bd37069ac] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1004.852557] env[63515]: DEBUG oslo_vmware.api [None req-642b03cc-2e02-4a86-bca6-75c3316e966c tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Waiting for the task: (returnval){ [ 1004.852557] env[63515]: value = "task-1111795" [ 1004.852557] env[63515]: _type = "Task" [ 1004.852557] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1004.861700] env[63515]: DEBUG oslo_vmware.api [None req-642b03cc-2e02-4a86-bca6-75c3316e966c tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Task: {'id': task-1111795, 'name': Destroy_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.989818] env[63515]: DEBUG oslo_concurrency.lockutils [req-4cc9af7f-2859-4faa-aae0-9693a9f8aa2f req-0b17df13-b1c7-4155-8a23-1ccfb2206f7e service nova] Releasing lock "refresh_cache-d6c5a9f9-5e5c-4b14-9b9a-7825bbf526a5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1004.989818] env[63515]: DEBUG nova.compute.manager [req-4cc9af7f-2859-4faa-aae0-9693a9f8aa2f req-0b17df13-b1c7-4155-8a23-1ccfb2206f7e service nova] [instance: 5cbce760-0163-4b27-8ae3-e46c926c8916] Received event network-vif-deleted-5fa807c4-5ce1-492a-bea3-d8f7a65d898f {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1005.014707] env[63515]: DEBUG nova.compute.manager [req-4441a9cb-9434-45b4-a79a-84990b13a94e req-e4e53fd0-9a49-453e-8354-26460c1b8711 service nova] [instance: d024b6b2-dd10-4112-89e8-ced57efe8208] Received event network-vif-deleted-06e24fbd-1278-4395-8108-f966a61c7b6c {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1005.014983] env[63515]: INFO nova.compute.manager [req-4441a9cb-9434-45b4-a79a-84990b13a94e req-e4e53fd0-9a49-453e-8354-26460c1b8711 service nova] [instance: d024b6b2-dd10-4112-89e8-ced57efe8208] Neutron deleted interface 06e24fbd-1278-4395-8108-f966a61c7b6c; detaching it from the instance and deleting it from the info cache [ 1005.015198] env[63515]: DEBUG nova.network.neutron [req-4441a9cb-9434-45b4-a79a-84990b13a94e req-e4e53fd0-9a49-453e-8354-26460c1b8711 service nova] [instance: d024b6b2-dd10-4112-89e8-ced57efe8208] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1005.152272] env[63515]: DEBUG oslo_concurrency.lockutils [None req-b6bf67af-7b53-4121-830a-6dc78e8dcba3 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Acquiring lock "refresh_cache-68766cd4-84be-475b-8494-d7ab43a9e969" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1005.152471] env[63515]: DEBUG oslo_concurrency.lockutils [None req-b6bf67af-7b53-4121-830a-6dc78e8dcba3 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Acquired lock "refresh_cache-68766cd4-84be-475b-8494-d7ab43a9e969" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1005.152655] env[63515]: DEBUG nova.network.neutron [None req-b6bf67af-7b53-4121-830a-6dc78e8dcba3 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 68766cd4-84be-475b-8494-d7ab43a9e969] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1005.152851] env[63515]: DEBUG nova.objects.instance [None req-b6bf67af-7b53-4121-830a-6dc78e8dcba3 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Lazy-loading 'info_cache' on Instance uuid 68766cd4-84be-475b-8494-d7ab43a9e969 {{(pid=63515) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1005.196234] env[63515]: DEBUG oslo_vmware.api [None req-655527ff-c992-4371-875f-a71b38db6213 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]5212f5dd-3da7-07b0-a0e0-74109aa1e18f, 'name': SearchDatastore_Task, 'duration_secs': 0.012088} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1005.197236] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4bf5608d-77a1-485d-93a3-ac427f33f9ee {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.201932] env[63515]: DEBUG oslo_vmware.api [None req-655527ff-c992-4371-875f-a71b38db6213 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Waiting for the task: (returnval){ [ 1005.201932] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]52a304e1-6efd-0ed8-1c36-fd8f4e1a9173" [ 1005.201932] env[63515]: _type = "Task" [ 1005.201932] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1005.208903] env[63515]: DEBUG oslo_vmware.api [None req-655527ff-c992-4371-875f-a71b38db6213 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52a304e1-6efd-0ed8-1c36-fd8f4e1a9173, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.255851] env[63515]: DEBUG oslo_vmware.api [None req-9d22472d-269a-4ade-805d-0c1138cb0261 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Task: {'id': task-1111794, 'name': PowerOnVM_Task, 'duration_secs': 0.552836} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1005.256146] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-9d22472d-269a-4ade-805d-0c1138cb0261 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: c6ac42a9-a233-48b4-bcb2-fb791c7446bb] Powered on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1005.256357] env[63515]: INFO nova.compute.manager [None req-9d22472d-269a-4ade-805d-0c1138cb0261 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: c6ac42a9-a233-48b4-bcb2-fb791c7446bb] Took 8.46 seconds to spawn the instance on the hypervisor. [ 1005.256541] env[63515]: DEBUG nova.compute.manager [None req-9d22472d-269a-4ade-805d-0c1138cb0261 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: c6ac42a9-a233-48b4-bcb2-fb791c7446bb] Checking state {{(pid=63515) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1005.257332] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-211861d1-1ac9-47a3-bde9-9181fd44b7fd {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.346403] env[63515]: DEBUG nova.compute.utils [None req-b36e84df-7edf-4fd3-971f-5ed772c5647a tempest-ServerRescueTestJSONUnderV235-800430211 tempest-ServerRescueTestJSONUnderV235-800430211-project-member] Using /dev/sd instead of None {{(pid=63515) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1005.347610] env[63515]: DEBUG nova.compute.manager [None req-b36e84df-7edf-4fd3-971f-5ed772c5647a tempest-ServerRescueTestJSONUnderV235-800430211 tempest-ServerRescueTestJSONUnderV235-800430211-project-member] [instance: fcfe3376-8dfa-4189-a267-ff6402713c1b] Allocating IP information in the background. {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1005.347809] env[63515]: DEBUG nova.network.neutron [None req-b36e84df-7edf-4fd3-971f-5ed772c5647a tempest-ServerRescueTestJSONUnderV235-800430211 tempest-ServerRescueTestJSONUnderV235-800430211-project-member] [instance: fcfe3376-8dfa-4189-a267-ff6402713c1b] allocate_for_instance() {{(pid=63515) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1005.362703] env[63515]: DEBUG oslo_vmware.api [None req-642b03cc-2e02-4a86-bca6-75c3316e966c tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Task: {'id': task-1111795, 'name': Destroy_Task, 'duration_secs': 0.311958} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1005.364147] env[63515]: INFO nova.virt.vmwareapi.vm_util [None req-642b03cc-2e02-4a86-bca6-75c3316e966c tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] [instance: 7d572ac0-f6af-4622-96cc-e75983420222] Destroyed the VM [ 1005.364147] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-642b03cc-2e02-4a86-bca6-75c3316e966c tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] [instance: 7d572ac0-f6af-4622-96cc-e75983420222] Deleting Snapshot of the VM instance {{(pid=63515) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1005.364147] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-bfa60ae7-1a7a-4a65-8b47-021dff36bfad {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.370883] env[63515]: DEBUG oslo_vmware.api [None req-642b03cc-2e02-4a86-bca6-75c3316e966c tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Waiting for the task: (returnval){ [ 1005.370883] env[63515]: value = "task-1111796" [ 1005.370883] env[63515]: _type = "Task" [ 1005.370883] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1005.379517] env[63515]: DEBUG oslo_vmware.api [None req-642b03cc-2e02-4a86-bca6-75c3316e966c tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Task: {'id': task-1111796, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.390217] env[63515]: DEBUG nova.policy [None req-b36e84df-7edf-4fd3-971f-5ed772c5647a tempest-ServerRescueTestJSONUnderV235-800430211 tempest-ServerRescueTestJSONUnderV235-800430211-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'df99da7f98d346ab8278a91a32658de0', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0be5ebcf766b4da5b33be3a12387e737', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63515) authorize /opt/stack/nova/nova/policy.py:201}} [ 1005.489887] env[63515]: DEBUG nova.network.neutron [-] [instance: d024b6b2-dd10-4112-89e8-ced57efe8208] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1005.518698] env[63515]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-67b27066-6005-4f4d-96ca-de273e54f468 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.527927] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1da068a-cb4c-40a6-acd0-043b3ca73c6b {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.554068] env[63515]: DEBUG nova.compute.manager [req-4441a9cb-9434-45b4-a79a-84990b13a94e req-e4e53fd0-9a49-453e-8354-26460c1b8711 service nova] [instance: d024b6b2-dd10-4112-89e8-ced57efe8208] Detach interface failed, port_id=06e24fbd-1278-4395-8108-f966a61c7b6c, reason: Instance d024b6b2-dd10-4112-89e8-ced57efe8208 could not be found. {{(pid=63515) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1005.638709] env[63515]: DEBUG nova.network.neutron [None req-b36e84df-7edf-4fd3-971f-5ed772c5647a tempest-ServerRescueTestJSONUnderV235-800430211 tempest-ServerRescueTestJSONUnderV235-800430211-project-member] [instance: fcfe3376-8dfa-4189-a267-ff6402713c1b] Successfully created port: 4462b1aa-72ea-41cc-878a-68c38920c8cc {{(pid=63515) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1005.713549] env[63515]: DEBUG oslo_vmware.api [None req-655527ff-c992-4371-875f-a71b38db6213 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52a304e1-6efd-0ed8-1c36-fd8f4e1a9173, 'name': SearchDatastore_Task, 'duration_secs': 0.048967} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1005.713831] env[63515]: DEBUG oslo_concurrency.lockutils [None req-655527ff-c992-4371-875f-a71b38db6213 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1005.714105] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-655527ff-c992-4371-875f-a71b38db6213 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk to [datastore2] d6c5a9f9-5e5c-4b14-9b9a-7825bbf526a5/d6c5a9f9-5e5c-4b14-9b9a-7825bbf526a5.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1005.714374] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-dbbaf593-3d0e-445d-963a-ffe1c92cbec5 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.720443] env[63515]: DEBUG oslo_vmware.api [None req-655527ff-c992-4371-875f-a71b38db6213 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Waiting for the task: (returnval){ [ 1005.720443] env[63515]: value = "task-1111797" [ 1005.720443] env[63515]: _type = "Task" [ 1005.720443] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1005.728937] env[63515]: DEBUG oslo_vmware.api [None req-655527ff-c992-4371-875f-a71b38db6213 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Task: {'id': task-1111797, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.774907] env[63515]: INFO nova.compute.manager [None req-9d22472d-269a-4ade-805d-0c1138cb0261 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: c6ac42a9-a233-48b4-bcb2-fb791c7446bb] Took 15.99 seconds to build instance. [ 1005.850957] env[63515]: DEBUG nova.compute.manager [None req-b36e84df-7edf-4fd3-971f-5ed772c5647a tempest-ServerRescueTestJSONUnderV235-800430211 tempest-ServerRescueTestJSONUnderV235-800430211-project-member] [instance: fcfe3376-8dfa-4189-a267-ff6402713c1b] Start building block device mappings for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1005.883800] env[63515]: DEBUG oslo_vmware.api [None req-642b03cc-2e02-4a86-bca6-75c3316e966c tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Task: {'id': task-1111796, 'name': RemoveSnapshot_Task, 'duration_secs': 0.33042} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1005.884637] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-642b03cc-2e02-4a86-bca6-75c3316e966c tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] [instance: 7d572ac0-f6af-4622-96cc-e75983420222] Deleted Snapshot of the VM instance {{(pid=63515) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1005.885010] env[63515]: DEBUG nova.compute.manager [None req-642b03cc-2e02-4a86-bca6-75c3316e966c tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] [instance: 7d572ac0-f6af-4622-96cc-e75983420222] Checking state {{(pid=63515) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1005.885855] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a1ebebc-f240-4161-b74d-6ce5c1a44e9a {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.993407] env[63515]: INFO nova.compute.manager [-] [instance: d024b6b2-dd10-4112-89e8-ced57efe8208] Took 1.26 seconds to deallocate network for instance. [ 1006.044050] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-834c194e-984f-4daf-a1c4-2904394b7a6b {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.053270] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-028d231d-9d62-45be-b2df-a98fd8e84506 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.084437] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e87d4627-62af-479b-b1c6-e0ec67257edf {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.092043] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73b80599-3a8c-453a-8396-6ffaa641fa9d {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.105533] env[63515]: DEBUG nova.compute.provider_tree [None req-96531a62-98f3-4a7c-9020-152fd796dc38 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1006.231075] env[63515]: DEBUG oslo_vmware.api [None req-655527ff-c992-4371-875f-a71b38db6213 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Task: {'id': task-1111797, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1006.276964] env[63515]: DEBUG oslo_concurrency.lockutils [None req-9d22472d-269a-4ade-805d-0c1138cb0261 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Lock "c6ac42a9-a233-48b4-bcb2-fb791c7446bb" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 17.497s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1006.369018] env[63515]: DEBUG nova.network.neutron [None req-b6bf67af-7b53-4121-830a-6dc78e8dcba3 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 68766cd4-84be-475b-8494-d7ab43a9e969] Updating instance_info_cache with network_info: [{"id": "28b59076-140f-46b5-b4a9-0bf530a1d43d", "address": "fa:16:3e:a3:55:0c", "network": {"id": "5758c885-e1c9-45c8-86aa-27a802101f52", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-734458138-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.204", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b18d4daacbc84e758f9f9ae4f3f09d28", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69351262-8d39-441a-85ba-3a78df436d17", "external-id": "nsx-vlan-transportzone-205", "segmentation_id": 205, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap28b59076-14", "ovs_interfaceid": "28b59076-140f-46b5-b4a9-0bf530a1d43d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1006.399744] env[63515]: INFO nova.compute.manager [None req-642b03cc-2e02-4a86-bca6-75c3316e966c tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] [instance: 7d572ac0-f6af-4622-96cc-e75983420222] Shelve offloading [ 1006.402190] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-642b03cc-2e02-4a86-bca6-75c3316e966c tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] [instance: 7d572ac0-f6af-4622-96cc-e75983420222] Powering off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1006.402440] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9a0c2934-b61a-4b72-ba05-3f1d4c48bd1b {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.409114] env[63515]: DEBUG oslo_vmware.api [None req-642b03cc-2e02-4a86-bca6-75c3316e966c tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Waiting for the task: (returnval){ [ 1006.409114] env[63515]: value = "task-1111798" [ 1006.409114] env[63515]: _type = "Task" [ 1006.409114] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1006.417812] env[63515]: DEBUG oslo_vmware.api [None req-642b03cc-2e02-4a86-bca6-75c3316e966c tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Task: {'id': task-1111798, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1006.501907] env[63515]: DEBUG oslo_concurrency.lockutils [None req-e5acb69c-ccc9-4835-a62a-46a52c0ee313 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1006.609106] env[63515]: DEBUG nova.scheduler.client.report [None req-96531a62-98f3-4a7c-9020-152fd796dc38 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1006.733723] env[63515]: DEBUG oslo_vmware.api [None req-655527ff-c992-4371-875f-a71b38db6213 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Task: {'id': task-1111797, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1006.863634] env[63515]: DEBUG nova.compute.manager [None req-b36e84df-7edf-4fd3-971f-5ed772c5647a tempest-ServerRescueTestJSONUnderV235-800430211 tempest-ServerRescueTestJSONUnderV235-800430211-project-member] [instance: fcfe3376-8dfa-4189-a267-ff6402713c1b] Start spawning the instance on the hypervisor. {{(pid=63515) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1006.872325] env[63515]: DEBUG oslo_concurrency.lockutils [None req-b6bf67af-7b53-4121-830a-6dc78e8dcba3 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Releasing lock "refresh_cache-68766cd4-84be-475b-8494-d7ab43a9e969" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1006.872573] env[63515]: DEBUG nova.objects.instance [None req-b6bf67af-7b53-4121-830a-6dc78e8dcba3 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Lazy-loading 'migration_context' on Instance uuid 68766cd4-84be-475b-8494-d7ab43a9e969 {{(pid=63515) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1006.899116] env[63515]: DEBUG nova.virt.hardware [None req-b36e84df-7edf-4fd3-971f-5ed772c5647a tempest-ServerRescueTestJSONUnderV235-800430211 tempest-ServerRescueTestJSONUnderV235-800430211-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T02:52:14Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T02:51:56Z,direct_url=,disk_format='vmdk',id=8a120570-cb06-4099-b262-554ca0ad15c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b569255cc43e42e7ae3f2b4ad37c6ef4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T02:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1006.899522] env[63515]: DEBUG nova.virt.hardware [None req-b36e84df-7edf-4fd3-971f-5ed772c5647a tempest-ServerRescueTestJSONUnderV235-800430211 tempest-ServerRescueTestJSONUnderV235-800430211-project-member] Flavor limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1006.899753] env[63515]: DEBUG nova.virt.hardware [None req-b36e84df-7edf-4fd3-971f-5ed772c5647a tempest-ServerRescueTestJSONUnderV235-800430211 tempest-ServerRescueTestJSONUnderV235-800430211-project-member] Image limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1006.900027] env[63515]: DEBUG nova.virt.hardware [None req-b36e84df-7edf-4fd3-971f-5ed772c5647a tempest-ServerRescueTestJSONUnderV235-800430211 tempest-ServerRescueTestJSONUnderV235-800430211-project-member] Flavor pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1006.900231] env[63515]: DEBUG nova.virt.hardware [None req-b36e84df-7edf-4fd3-971f-5ed772c5647a tempest-ServerRescueTestJSONUnderV235-800430211 tempest-ServerRescueTestJSONUnderV235-800430211-project-member] Image pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1006.900423] env[63515]: DEBUG nova.virt.hardware [None req-b36e84df-7edf-4fd3-971f-5ed772c5647a tempest-ServerRescueTestJSONUnderV235-800430211 tempest-ServerRescueTestJSONUnderV235-800430211-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1006.900673] env[63515]: DEBUG nova.virt.hardware [None req-b36e84df-7edf-4fd3-971f-5ed772c5647a tempest-ServerRescueTestJSONUnderV235-800430211 tempest-ServerRescueTestJSONUnderV235-800430211-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1006.900883] env[63515]: DEBUG nova.virt.hardware [None req-b36e84df-7edf-4fd3-971f-5ed772c5647a tempest-ServerRescueTestJSONUnderV235-800430211 tempest-ServerRescueTestJSONUnderV235-800430211-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1006.901185] env[63515]: DEBUG nova.virt.hardware [None req-b36e84df-7edf-4fd3-971f-5ed772c5647a tempest-ServerRescueTestJSONUnderV235-800430211 tempest-ServerRescueTestJSONUnderV235-800430211-project-member] Got 1 possible topologies {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1006.901468] env[63515]: DEBUG nova.virt.hardware [None req-b36e84df-7edf-4fd3-971f-5ed772c5647a tempest-ServerRescueTestJSONUnderV235-800430211 tempest-ServerRescueTestJSONUnderV235-800430211-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1006.902099] env[63515]: DEBUG nova.virt.hardware [None req-b36e84df-7edf-4fd3-971f-5ed772c5647a tempest-ServerRescueTestJSONUnderV235-800430211 tempest-ServerRescueTestJSONUnderV235-800430211-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1006.903221] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1152ce3-cd60-4429-8b63-e7712a626dd9 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.914154] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6bc7fc46-cd2b-4e33-9fb8-815a30f1ff81 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.931244] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-642b03cc-2e02-4a86-bca6-75c3316e966c tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] [instance: 7d572ac0-f6af-4622-96cc-e75983420222] VM already powered off {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1006.931530] env[63515]: DEBUG nova.compute.manager [None req-642b03cc-2e02-4a86-bca6-75c3316e966c tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] [instance: 7d572ac0-f6af-4622-96cc-e75983420222] Checking state {{(pid=63515) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1006.932393] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b90be935-10c3-45d9-8201-2ed29eec2d5f {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.938643] env[63515]: DEBUG oslo_concurrency.lockutils [None req-642b03cc-2e02-4a86-bca6-75c3316e966c tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Acquiring lock "refresh_cache-7d572ac0-f6af-4622-96cc-e75983420222" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1006.938879] env[63515]: DEBUG oslo_concurrency.lockutils [None req-642b03cc-2e02-4a86-bca6-75c3316e966c tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Acquired lock "refresh_cache-7d572ac0-f6af-4622-96cc-e75983420222" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1006.939188] env[63515]: DEBUG nova.network.neutron [None req-642b03cc-2e02-4a86-bca6-75c3316e966c tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] [instance: 7d572ac0-f6af-4622-96cc-e75983420222] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1007.068511] env[63515]: DEBUG nova.compute.manager [req-a2ba92a9-d3ca-431d-ae5b-819a90b50c65 req-4b00e14d-f4f3-48a8-a848-7368a80ccd95 service nova] [instance: 724b05bc-ce4b-4f99-ad14-f3abea5567f2] Received event network-changed-3dee43b6-2650-42e1-aa5a-4994a3ec7f05 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1007.068721] env[63515]: DEBUG nova.compute.manager [req-a2ba92a9-d3ca-431d-ae5b-819a90b50c65 req-4b00e14d-f4f3-48a8-a848-7368a80ccd95 service nova] [instance: 724b05bc-ce4b-4f99-ad14-f3abea5567f2] Refreshing instance network info cache due to event network-changed-3dee43b6-2650-42e1-aa5a-4994a3ec7f05. {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1007.069121] env[63515]: DEBUG oslo_concurrency.lockutils [req-a2ba92a9-d3ca-431d-ae5b-819a90b50c65 req-4b00e14d-f4f3-48a8-a848-7368a80ccd95 service nova] Acquiring lock "refresh_cache-724b05bc-ce4b-4f99-ad14-f3abea5567f2" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1007.069430] env[63515]: DEBUG oslo_concurrency.lockutils [req-a2ba92a9-d3ca-431d-ae5b-819a90b50c65 req-4b00e14d-f4f3-48a8-a848-7368a80ccd95 service nova] Acquired lock "refresh_cache-724b05bc-ce4b-4f99-ad14-f3abea5567f2" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1007.071316] env[63515]: DEBUG nova.network.neutron [req-a2ba92a9-d3ca-431d-ae5b-819a90b50c65 req-4b00e14d-f4f3-48a8-a848-7368a80ccd95 service nova] [instance: 724b05bc-ce4b-4f99-ad14-f3abea5567f2] Refreshing network info cache for port 3dee43b6-2650-42e1-aa5a-4994a3ec7f05 {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1007.116876] env[63515]: DEBUG oslo_concurrency.lockutils [None req-96531a62-98f3-4a7c-9020-152fd796dc38 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.273s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1007.117606] env[63515]: DEBUG nova.compute.manager [None req-96531a62-98f3-4a7c-9020-152fd796dc38 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: b76cf3dd-7b7e-4e5d-bf3c-e15bd37069ac] Start building networks asynchronously for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1007.121838] env[63515]: DEBUG oslo_concurrency.lockutils [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 6.316s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1007.233437] env[63515]: DEBUG oslo_vmware.api [None req-655527ff-c992-4371-875f-a71b38db6213 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Task: {'id': task-1111797, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.107314} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1007.233698] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-655527ff-c992-4371-875f-a71b38db6213 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk to [datastore2] d6c5a9f9-5e5c-4b14-9b9a-7825bbf526a5/d6c5a9f9-5e5c-4b14-9b9a-7825bbf526a5.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1007.233919] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-655527ff-c992-4371-875f-a71b38db6213 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: d6c5a9f9-5e5c-4b14-9b9a-7825bbf526a5] Extending root virtual disk to 1048576 {{(pid=63515) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1007.234219] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c0b2335f-559f-4ec3-9b21-bdcb85ac5583 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.242394] env[63515]: DEBUG oslo_vmware.api [None req-655527ff-c992-4371-875f-a71b38db6213 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Waiting for the task: (returnval){ [ 1007.242394] env[63515]: value = "task-1111799" [ 1007.242394] env[63515]: _type = "Task" [ 1007.242394] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1007.251959] env[63515]: DEBUG oslo_vmware.api [None req-655527ff-c992-4371-875f-a71b38db6213 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Task: {'id': task-1111799, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1007.285346] env[63515]: DEBUG nova.network.neutron [None req-b36e84df-7edf-4fd3-971f-5ed772c5647a tempest-ServerRescueTestJSONUnderV235-800430211 tempest-ServerRescueTestJSONUnderV235-800430211-project-member] [instance: fcfe3376-8dfa-4189-a267-ff6402713c1b] Successfully updated port: 4462b1aa-72ea-41cc-878a-68c38920c8cc {{(pid=63515) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1007.375698] env[63515]: DEBUG nova.objects.base [None req-b6bf67af-7b53-4121-830a-6dc78e8dcba3 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Object Instance<68766cd4-84be-475b-8494-d7ab43a9e969> lazy-loaded attributes: info_cache,migration_context {{(pid=63515) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1007.376766] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-823ead94-f1fc-4592-9bdf-ecfdc482351b {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.397442] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ea62e206-a909-427f-8e02-25c80c02e4ba {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.403882] env[63515]: DEBUG oslo_vmware.api [None req-b6bf67af-7b53-4121-830a-6dc78e8dcba3 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Waiting for the task: (returnval){ [ 1007.403882] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]525fbc51-447b-1bba-57ee-c1beb9fa6deb" [ 1007.403882] env[63515]: _type = "Task" [ 1007.403882] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1007.414684] env[63515]: DEBUG oslo_vmware.api [None req-b6bf67af-7b53-4121-830a-6dc78e8dcba3 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]525fbc51-447b-1bba-57ee-c1beb9fa6deb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1007.624189] env[63515]: DEBUG nova.compute.utils [None req-96531a62-98f3-4a7c-9020-152fd796dc38 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Using /dev/sd instead of None {{(pid=63515) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1007.627020] env[63515]: DEBUG nova.compute.manager [None req-96531a62-98f3-4a7c-9020-152fd796dc38 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: b76cf3dd-7b7e-4e5d-bf3c-e15bd37069ac] Allocating IP information in the background. {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1007.627020] env[63515]: DEBUG nova.network.neutron [None req-96531a62-98f3-4a7c-9020-152fd796dc38 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: b76cf3dd-7b7e-4e5d-bf3c-e15bd37069ac] allocate_for_instance() {{(pid=63515) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1007.706476] env[63515]: DEBUG nova.network.neutron [None req-642b03cc-2e02-4a86-bca6-75c3316e966c tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] [instance: 7d572ac0-f6af-4622-96cc-e75983420222] Updating instance_info_cache with network_info: [{"id": "dbfd09c2-9ec4-4821-8613-43c6d6cf4724", "address": "fa:16:3e:6d:92:3d", "network": {"id": "09fb20ca-2b17-4d46-b915-6aff85b862d8", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1815529427-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.244", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "19ce3b05eec64475b685faad1fd801f6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ded18042-834c-4792-b3e8-b1c377446432", "external-id": "nsx-vlan-transportzone-293", "segmentation_id": 293, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdbfd09c2-9e", "ovs_interfaceid": "dbfd09c2-9ec4-4821-8613-43c6d6cf4724", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1007.762292] env[63515]: DEBUG oslo_vmware.api [None req-655527ff-c992-4371-875f-a71b38db6213 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Task: {'id': task-1111799, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.082592} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1007.762604] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-655527ff-c992-4371-875f-a71b38db6213 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: d6c5a9f9-5e5c-4b14-9b9a-7825bbf526a5] Extended root virtual disk {{(pid=63515) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1007.763468] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb45064a-1b08-4c13-bc3a-99c71990eb3e {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.786758] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-655527ff-c992-4371-875f-a71b38db6213 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: d6c5a9f9-5e5c-4b14-9b9a-7825bbf526a5] Reconfiguring VM instance instance-00000062 to attach disk [datastore2] d6c5a9f9-5e5c-4b14-9b9a-7825bbf526a5/d6c5a9f9-5e5c-4b14-9b9a-7825bbf526a5.vmdk or device None with type sparse {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1007.787096] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-696d98a9-0ced-4139-9b62-9e588734ee4f {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.804606] env[63515]: DEBUG oslo_concurrency.lockutils [None req-b36e84df-7edf-4fd3-971f-5ed772c5647a tempest-ServerRescueTestJSONUnderV235-800430211 tempest-ServerRescueTestJSONUnderV235-800430211-project-member] Acquiring lock "refresh_cache-fcfe3376-8dfa-4189-a267-ff6402713c1b" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1007.804606] env[63515]: DEBUG oslo_concurrency.lockutils [None req-b36e84df-7edf-4fd3-971f-5ed772c5647a tempest-ServerRescueTestJSONUnderV235-800430211 tempest-ServerRescueTestJSONUnderV235-800430211-project-member] Acquired lock "refresh_cache-fcfe3376-8dfa-4189-a267-ff6402713c1b" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1007.804606] env[63515]: DEBUG nova.network.neutron [None req-b36e84df-7edf-4fd3-971f-5ed772c5647a tempest-ServerRescueTestJSONUnderV235-800430211 tempest-ServerRescueTestJSONUnderV235-800430211-project-member] [instance: fcfe3376-8dfa-4189-a267-ff6402713c1b] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1007.813017] env[63515]: DEBUG oslo_vmware.api [None req-655527ff-c992-4371-875f-a71b38db6213 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Waiting for the task: (returnval){ [ 1007.813017] env[63515]: value = "task-1111800" [ 1007.813017] env[63515]: _type = "Task" [ 1007.813017] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1007.823328] env[63515]: DEBUG oslo_vmware.api [None req-655527ff-c992-4371-875f-a71b38db6213 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Task: {'id': task-1111800, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1007.915725] env[63515]: DEBUG oslo_vmware.api [None req-b6bf67af-7b53-4121-830a-6dc78e8dcba3 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]525fbc51-447b-1bba-57ee-c1beb9fa6deb, 'name': SearchDatastore_Task, 'duration_secs': 0.00823} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1007.915725] env[63515]: DEBUG oslo_concurrency.lockutils [None req-b6bf67af-7b53-4121-830a-6dc78e8dcba3 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1007.973900] env[63515]: DEBUG nova.policy [None req-96531a62-98f3-4a7c-9020-152fd796dc38 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f58ac812970845dc825202b842feb5d0', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ea099ce07bfb4a8da014d9303cf552eb', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63515) authorize /opt/stack/nova/nova/policy.py:201}} [ 1008.136831] env[63515]: DEBUG nova.compute.manager [None req-96531a62-98f3-4a7c-9020-152fd796dc38 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: b76cf3dd-7b7e-4e5d-bf3c-e15bd37069ac] Start building block device mappings for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1008.140133] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Applying migration context for instance 68766cd4-84be-475b-8494-d7ab43a9e969 as it has an incoming, in-progress migration 0453c850-4f70-44d5-9d13-96d466406033. Migration status is confirming {{(pid=63515) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1016}} [ 1008.141481] env[63515]: INFO nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] [instance: 68766cd4-84be-475b-8494-d7ab43a9e969] Updating resource usage from migration 0453c850-4f70-44d5-9d13-96d466406033 [ 1008.166283] env[63515]: WARNING nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Instance 5cbce760-0163-4b27-8ae3-e46c926c8916 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 1008.166447] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Instance 19475235-dbec-40a5-bd8f-8070616b7ddb actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63515) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1008.166573] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Instance e13da90c-28e6-43d2-99b7-19c5095954ca actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63515) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1008.166761] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Instance 7d572ac0-f6af-4622-96cc-e75983420222 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63515) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1008.166917] env[63515]: WARNING nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Instance d024b6b2-dd10-4112-89e8-ced57efe8208 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 1008.167237] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Instance 724b05bc-ce4b-4f99-ad14-f3abea5567f2 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63515) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1008.167237] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Migration 0453c850-4f70-44d5-9d13-96d466406033 is active on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63515) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1712}} [ 1008.167367] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Instance 68766cd4-84be-475b-8494-d7ab43a9e969 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=63515) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1008.167405] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Instance c6ac42a9-a233-48b4-bcb2-fb791c7446bb actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63515) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1008.167519] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Instance d6c5a9f9-5e5c-4b14-9b9a-7825bbf526a5 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63515) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1008.167671] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Instance fcfe3376-8dfa-4189-a267-ff6402713c1b actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63515) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1008.167763] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Instance b76cf3dd-7b7e-4e5d-bf3c-e15bd37069ac actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63515) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1008.168051] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Total usable vcpus: 48, total allocated vcpus: 10 {{(pid=63515) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1008.168206] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=2496MB phys_disk=200GB used_disk=10GB total_vcpus=48 used_vcpus=10 pci_stats=[] {{(pid=63515) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1008.209696] env[63515]: DEBUG oslo_concurrency.lockutils [None req-642b03cc-2e02-4a86-bca6-75c3316e966c tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Releasing lock "refresh_cache-7d572ac0-f6af-4622-96cc-e75983420222" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1008.325632] env[63515]: DEBUG oslo_vmware.api [None req-655527ff-c992-4371-875f-a71b38db6213 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Task: {'id': task-1111800, 'name': ReconfigVM_Task, 'duration_secs': 0.268567} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1008.325926] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-655527ff-c992-4371-875f-a71b38db6213 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: d6c5a9f9-5e5c-4b14-9b9a-7825bbf526a5] Reconfigured VM instance instance-00000062 to attach disk [datastore2] d6c5a9f9-5e5c-4b14-9b9a-7825bbf526a5/d6c5a9f9-5e5c-4b14-9b9a-7825bbf526a5.vmdk or device None with type sparse {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1008.326577] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-afe00f13-c942-4b25-a6d4-65239b15c4a7 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.336336] env[63515]: DEBUG oslo_vmware.api [None req-655527ff-c992-4371-875f-a71b38db6213 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Waiting for the task: (returnval){ [ 1008.336336] env[63515]: value = "task-1111801" [ 1008.336336] env[63515]: _type = "Task" [ 1008.336336] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1008.347240] env[63515]: DEBUG oslo_vmware.api [None req-655527ff-c992-4371-875f-a71b38db6213 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Task: {'id': task-1111801, 'name': Rename_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1008.374435] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26ffd7ec-fa62-43b2-aa33-bbcdc18e5100 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.383773] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b716cfa9-f849-4a78-a6ae-56ce0939d29f {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.416646] env[63515]: DEBUG nova.network.neutron [None req-b36e84df-7edf-4fd3-971f-5ed772c5647a tempest-ServerRescueTestJSONUnderV235-800430211 tempest-ServerRescueTestJSONUnderV235-800430211-project-member] [instance: fcfe3376-8dfa-4189-a267-ff6402713c1b] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1008.419128] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0bbe21a-2e9d-4d8c-9567-6bd51f8a52c2 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.427448] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c352cf13-6383-4d50-b6c0-9e312f6014b6 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.444020] env[63515]: DEBUG nova.compute.provider_tree [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1008.461116] env[63515]: DEBUG nova.network.neutron [req-a2ba92a9-d3ca-431d-ae5b-819a90b50c65 req-4b00e14d-f4f3-48a8-a848-7368a80ccd95 service nova] [instance: 724b05bc-ce4b-4f99-ad14-f3abea5567f2] Updated VIF entry in instance network info cache for port 3dee43b6-2650-42e1-aa5a-4994a3ec7f05. {{(pid=63515) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1008.461518] env[63515]: DEBUG nova.network.neutron [req-a2ba92a9-d3ca-431d-ae5b-819a90b50c65 req-4b00e14d-f4f3-48a8-a848-7368a80ccd95 service nova] [instance: 724b05bc-ce4b-4f99-ad14-f3abea5567f2] Updating instance_info_cache with network_info: [{"id": "3dee43b6-2650-42e1-aa5a-4994a3ec7f05", "address": "fa:16:3e:89:06:d0", "network": {"id": "88dfb29d-41e0-4df7-8f70-e52cda05620c", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1355578400-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.146", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7d4ada9ec35f42b19c6480a9101d21a4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "357d2811-e990-4985-9f9e-b158d10d3699", "external-id": "nsx-vlan-transportzone-641", "segmentation_id": 641, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3dee43b6-26", "ovs_interfaceid": "3dee43b6-2650-42e1-aa5a-4994a3ec7f05", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1008.705468] env[63515]: DEBUG nova.compute.manager [req-8bf1a21f-3483-48dd-bc61-405be478d6b2 req-d385ec4e-3018-470f-ae89-9475dc5946b4 service nova] [instance: c6ac42a9-a233-48b4-bcb2-fb791c7446bb] Received event network-changed-161e8482-76a0-4c82-8354-af132110794d {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1008.705468] env[63515]: DEBUG nova.compute.manager [req-8bf1a21f-3483-48dd-bc61-405be478d6b2 req-d385ec4e-3018-470f-ae89-9475dc5946b4 service nova] [instance: c6ac42a9-a233-48b4-bcb2-fb791c7446bb] Refreshing instance network info cache due to event network-changed-161e8482-76a0-4c82-8354-af132110794d. {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1008.705468] env[63515]: DEBUG oslo_concurrency.lockutils [req-8bf1a21f-3483-48dd-bc61-405be478d6b2 req-d385ec4e-3018-470f-ae89-9475dc5946b4 service nova] Acquiring lock "refresh_cache-c6ac42a9-a233-48b4-bcb2-fb791c7446bb" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1008.705626] env[63515]: DEBUG oslo_concurrency.lockutils [req-8bf1a21f-3483-48dd-bc61-405be478d6b2 req-d385ec4e-3018-470f-ae89-9475dc5946b4 service nova] Acquired lock "refresh_cache-c6ac42a9-a233-48b4-bcb2-fb791c7446bb" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1008.705750] env[63515]: DEBUG nova.network.neutron [req-8bf1a21f-3483-48dd-bc61-405be478d6b2 req-d385ec4e-3018-470f-ae89-9475dc5946b4 service nova] [instance: c6ac42a9-a233-48b4-bcb2-fb791c7446bb] Refreshing network info cache for port 161e8482-76a0-4c82-8354-af132110794d {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1008.727965] env[63515]: DEBUG nova.network.neutron [None req-96531a62-98f3-4a7c-9020-152fd796dc38 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: b76cf3dd-7b7e-4e5d-bf3c-e15bd37069ac] Successfully created port: acea8927-7e43-427c-bed2-2420d4708f3b {{(pid=63515) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1008.731121] env[63515]: DEBUG nova.network.neutron [None req-b36e84df-7edf-4fd3-971f-5ed772c5647a tempest-ServerRescueTestJSONUnderV235-800430211 tempest-ServerRescueTestJSONUnderV235-800430211-project-member] [instance: fcfe3376-8dfa-4189-a267-ff6402713c1b] Updating instance_info_cache with network_info: [{"id": "4462b1aa-72ea-41cc-878a-68c38920c8cc", "address": "fa:16:3e:4f:bd:04", "network": {"id": "9fc2b9c9-6afc-4510-bf13-f87f1e1bd53b", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-513162532-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "0be5ebcf766b4da5b33be3a12387e737", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0248a27a-1d7f-4195-987b-06bfc8467347", "external-id": "nsx-vlan-transportzone-26", "segmentation_id": 26, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4462b1aa-72", "ovs_interfaceid": "4462b1aa-72ea-41cc-878a-68c38920c8cc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1008.850551] env[63515]: DEBUG oslo_vmware.api [None req-655527ff-c992-4371-875f-a71b38db6213 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Task: {'id': task-1111801, 'name': Rename_Task} progress is 99%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1008.914648] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-642b03cc-2e02-4a86-bca6-75c3316e966c tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] [instance: 7d572ac0-f6af-4622-96cc-e75983420222] Destroying instance {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1008.914648] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5339dde8-dd11-487d-a435-a8212f6e998f {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.923338] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-642b03cc-2e02-4a86-bca6-75c3316e966c tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] [instance: 7d572ac0-f6af-4622-96cc-e75983420222] Unregistering the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1008.923338] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-6d104702-0d3b-4980-917a-1a4714c60348 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.946193] env[63515]: DEBUG nova.scheduler.client.report [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1008.964413] env[63515]: DEBUG oslo_concurrency.lockutils [req-a2ba92a9-d3ca-431d-ae5b-819a90b50c65 req-4b00e14d-f4f3-48a8-a848-7368a80ccd95 service nova] Releasing lock "refresh_cache-724b05bc-ce4b-4f99-ad14-f3abea5567f2" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1008.967050] env[63515]: DEBUG nova.compute.manager [req-a2ba92a9-d3ca-431d-ae5b-819a90b50c65 req-4b00e14d-f4f3-48a8-a848-7368a80ccd95 service nova] [instance: c6ac42a9-a233-48b4-bcb2-fb791c7446bb] Received event network-changed-161e8482-76a0-4c82-8354-af132110794d {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1008.967050] env[63515]: DEBUG nova.compute.manager [req-a2ba92a9-d3ca-431d-ae5b-819a90b50c65 req-4b00e14d-f4f3-48a8-a848-7368a80ccd95 service nova] [instance: c6ac42a9-a233-48b4-bcb2-fb791c7446bb] Refreshing instance network info cache due to event network-changed-161e8482-76a0-4c82-8354-af132110794d. {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1008.967050] env[63515]: DEBUG oslo_concurrency.lockutils [req-a2ba92a9-d3ca-431d-ae5b-819a90b50c65 req-4b00e14d-f4f3-48a8-a848-7368a80ccd95 service nova] Acquiring lock "refresh_cache-c6ac42a9-a233-48b4-bcb2-fb791c7446bb" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1008.990890] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-642b03cc-2e02-4a86-bca6-75c3316e966c tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] [instance: 7d572ac0-f6af-4622-96cc-e75983420222] Unregistered the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1008.990890] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-642b03cc-2e02-4a86-bca6-75c3316e966c tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] [instance: 7d572ac0-f6af-4622-96cc-e75983420222] Deleting contents of the VM from datastore datastore2 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1008.991053] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-642b03cc-2e02-4a86-bca6-75c3316e966c tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Deleting the datastore file [datastore2] 7d572ac0-f6af-4622-96cc-e75983420222 {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1008.991334] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-fe071016-7446-4937-a50c-3713f699b2e2 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.999543] env[63515]: DEBUG oslo_vmware.api [None req-642b03cc-2e02-4a86-bca6-75c3316e966c tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Waiting for the task: (returnval){ [ 1008.999543] env[63515]: value = "task-1111803" [ 1008.999543] env[63515]: _type = "Task" [ 1008.999543] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1009.007738] env[63515]: DEBUG oslo_vmware.api [None req-642b03cc-2e02-4a86-bca6-75c3316e966c tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Task: {'id': task-1111803, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1009.101142] env[63515]: DEBUG nova.compute.manager [req-2c79e385-285c-4d37-87ec-72d7af62ca9e req-a6e85153-c431-444a-920b-65823615d53b service nova] [instance: fcfe3376-8dfa-4189-a267-ff6402713c1b] Received event network-vif-plugged-4462b1aa-72ea-41cc-878a-68c38920c8cc {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1009.101454] env[63515]: DEBUG oslo_concurrency.lockutils [req-2c79e385-285c-4d37-87ec-72d7af62ca9e req-a6e85153-c431-444a-920b-65823615d53b service nova] Acquiring lock "fcfe3376-8dfa-4189-a267-ff6402713c1b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1009.101720] env[63515]: DEBUG oslo_concurrency.lockutils [req-2c79e385-285c-4d37-87ec-72d7af62ca9e req-a6e85153-c431-444a-920b-65823615d53b service nova] Lock "fcfe3376-8dfa-4189-a267-ff6402713c1b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1009.101957] env[63515]: DEBUG oslo_concurrency.lockutils [req-2c79e385-285c-4d37-87ec-72d7af62ca9e req-a6e85153-c431-444a-920b-65823615d53b service nova] Lock "fcfe3376-8dfa-4189-a267-ff6402713c1b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1009.102208] env[63515]: DEBUG nova.compute.manager [req-2c79e385-285c-4d37-87ec-72d7af62ca9e req-a6e85153-c431-444a-920b-65823615d53b service nova] [instance: fcfe3376-8dfa-4189-a267-ff6402713c1b] No waiting events found dispatching network-vif-plugged-4462b1aa-72ea-41cc-878a-68c38920c8cc {{(pid=63515) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1009.102463] env[63515]: WARNING nova.compute.manager [req-2c79e385-285c-4d37-87ec-72d7af62ca9e req-a6e85153-c431-444a-920b-65823615d53b service nova] [instance: fcfe3376-8dfa-4189-a267-ff6402713c1b] Received unexpected event network-vif-plugged-4462b1aa-72ea-41cc-878a-68c38920c8cc for instance with vm_state building and task_state spawning. [ 1009.102693] env[63515]: DEBUG nova.compute.manager [req-2c79e385-285c-4d37-87ec-72d7af62ca9e req-a6e85153-c431-444a-920b-65823615d53b service nova] [instance: fcfe3376-8dfa-4189-a267-ff6402713c1b] Received event network-changed-4462b1aa-72ea-41cc-878a-68c38920c8cc {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1009.102919] env[63515]: DEBUG nova.compute.manager [req-2c79e385-285c-4d37-87ec-72d7af62ca9e req-a6e85153-c431-444a-920b-65823615d53b service nova] [instance: fcfe3376-8dfa-4189-a267-ff6402713c1b] Refreshing instance network info cache due to event network-changed-4462b1aa-72ea-41cc-878a-68c38920c8cc. {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1009.103157] env[63515]: DEBUG oslo_concurrency.lockutils [req-2c79e385-285c-4d37-87ec-72d7af62ca9e req-a6e85153-c431-444a-920b-65823615d53b service nova] Acquiring lock "refresh_cache-fcfe3376-8dfa-4189-a267-ff6402713c1b" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1009.150437] env[63515]: DEBUG nova.compute.manager [None req-96531a62-98f3-4a7c-9020-152fd796dc38 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: b76cf3dd-7b7e-4e5d-bf3c-e15bd37069ac] Start spawning the instance on the hypervisor. {{(pid=63515) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1009.176915] env[63515]: DEBUG nova.virt.hardware [None req-96531a62-98f3-4a7c-9020-152fd796dc38 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T02:52:14Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T02:51:56Z,direct_url=,disk_format='vmdk',id=8a120570-cb06-4099-b262-554ca0ad15c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b569255cc43e42e7ae3f2b4ad37c6ef4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T02:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1009.177226] env[63515]: DEBUG nova.virt.hardware [None req-96531a62-98f3-4a7c-9020-152fd796dc38 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Flavor limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1009.177396] env[63515]: DEBUG nova.virt.hardware [None req-96531a62-98f3-4a7c-9020-152fd796dc38 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Image limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1009.177725] env[63515]: DEBUG nova.virt.hardware [None req-96531a62-98f3-4a7c-9020-152fd796dc38 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Flavor pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1009.177934] env[63515]: DEBUG nova.virt.hardware [None req-96531a62-98f3-4a7c-9020-152fd796dc38 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Image pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1009.178145] env[63515]: DEBUG nova.virt.hardware [None req-96531a62-98f3-4a7c-9020-152fd796dc38 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1009.178417] env[63515]: DEBUG nova.virt.hardware [None req-96531a62-98f3-4a7c-9020-152fd796dc38 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1009.178612] env[63515]: DEBUG nova.virt.hardware [None req-96531a62-98f3-4a7c-9020-152fd796dc38 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1009.178813] env[63515]: DEBUG nova.virt.hardware [None req-96531a62-98f3-4a7c-9020-152fd796dc38 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Got 1 possible topologies {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1009.179047] env[63515]: DEBUG nova.virt.hardware [None req-96531a62-98f3-4a7c-9020-152fd796dc38 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1009.179282] env[63515]: DEBUG nova.virt.hardware [None req-96531a62-98f3-4a7c-9020-152fd796dc38 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1009.180187] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51f03b78-7396-4f7a-ab79-28d6543c493d {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.189478] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4f5b312-71f8-4088-b55c-19d7609fa9a2 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.233929] env[63515]: DEBUG oslo_concurrency.lockutils [None req-b36e84df-7edf-4fd3-971f-5ed772c5647a tempest-ServerRescueTestJSONUnderV235-800430211 tempest-ServerRescueTestJSONUnderV235-800430211-project-member] Releasing lock "refresh_cache-fcfe3376-8dfa-4189-a267-ff6402713c1b" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1009.234279] env[63515]: DEBUG nova.compute.manager [None req-b36e84df-7edf-4fd3-971f-5ed772c5647a tempest-ServerRescueTestJSONUnderV235-800430211 tempest-ServerRescueTestJSONUnderV235-800430211-project-member] [instance: fcfe3376-8dfa-4189-a267-ff6402713c1b] Instance network_info: |[{"id": "4462b1aa-72ea-41cc-878a-68c38920c8cc", "address": "fa:16:3e:4f:bd:04", "network": {"id": "9fc2b9c9-6afc-4510-bf13-f87f1e1bd53b", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-513162532-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "0be5ebcf766b4da5b33be3a12387e737", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0248a27a-1d7f-4195-987b-06bfc8467347", "external-id": "nsx-vlan-transportzone-26", "segmentation_id": 26, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4462b1aa-72", "ovs_interfaceid": "4462b1aa-72ea-41cc-878a-68c38920c8cc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1009.234706] env[63515]: DEBUG oslo_concurrency.lockutils [req-2c79e385-285c-4d37-87ec-72d7af62ca9e req-a6e85153-c431-444a-920b-65823615d53b service nova] Acquired lock "refresh_cache-fcfe3376-8dfa-4189-a267-ff6402713c1b" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1009.234789] env[63515]: DEBUG nova.network.neutron [req-2c79e385-285c-4d37-87ec-72d7af62ca9e req-a6e85153-c431-444a-920b-65823615d53b service nova] [instance: fcfe3376-8dfa-4189-a267-ff6402713c1b] Refreshing network info cache for port 4462b1aa-72ea-41cc-878a-68c38920c8cc {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1009.236236] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-b36e84df-7edf-4fd3-971f-5ed772c5647a tempest-ServerRescueTestJSONUnderV235-800430211 tempest-ServerRescueTestJSONUnderV235-800430211-project-member] [instance: fcfe3376-8dfa-4189-a267-ff6402713c1b] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:4f:bd:04', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '0248a27a-1d7f-4195-987b-06bfc8467347', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '4462b1aa-72ea-41cc-878a-68c38920c8cc', 'vif_model': 'vmxnet3'}] {{(pid=63515) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1009.244267] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-b36e84df-7edf-4fd3-971f-5ed772c5647a tempest-ServerRescueTestJSONUnderV235-800430211 tempest-ServerRescueTestJSONUnderV235-800430211-project-member] Creating folder: Project (0be5ebcf766b4da5b33be3a12387e737). Parent ref: group-v243370. {{(pid=63515) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1009.247053] env[63515]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-7a686c10-5c91-4c0b-a39c-b9b69dff47dd {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.259308] env[63515]: INFO nova.virt.vmwareapi.vm_util [None req-b36e84df-7edf-4fd3-971f-5ed772c5647a tempest-ServerRescueTestJSONUnderV235-800430211 tempest-ServerRescueTestJSONUnderV235-800430211-project-member] Created folder: Project (0be5ebcf766b4da5b33be3a12387e737) in parent group-v243370. [ 1009.259514] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-b36e84df-7edf-4fd3-971f-5ed772c5647a tempest-ServerRescueTestJSONUnderV235-800430211 tempest-ServerRescueTestJSONUnderV235-800430211-project-member] Creating folder: Instances. Parent ref: group-v243559. {{(pid=63515) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1009.259756] env[63515]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-87628bba-4293-4a74-bbe9-66e8e30fff41 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.269607] env[63515]: INFO nova.virt.vmwareapi.vm_util [None req-b36e84df-7edf-4fd3-971f-5ed772c5647a tempest-ServerRescueTestJSONUnderV235-800430211 tempest-ServerRescueTestJSONUnderV235-800430211-project-member] Created folder: Instances in parent group-v243559. [ 1009.269868] env[63515]: DEBUG oslo.service.loopingcall [None req-b36e84df-7edf-4fd3-971f-5ed772c5647a tempest-ServerRescueTestJSONUnderV235-800430211 tempest-ServerRescueTestJSONUnderV235-800430211-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1009.270889] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fcfe3376-8dfa-4189-a267-ff6402713c1b] Creating VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1009.270889] env[63515]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-5b750c31-f932-495e-a78f-92939727b1f1 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.294179] env[63515]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1009.294179] env[63515]: value = "task-1111806" [ 1009.294179] env[63515]: _type = "Task" [ 1009.294179] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1009.303094] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1111806, 'name': CreateVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1009.347828] env[63515]: DEBUG oslo_vmware.api [None req-655527ff-c992-4371-875f-a71b38db6213 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Task: {'id': task-1111801, 'name': Rename_Task} progress is 99%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1009.451769] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=63515) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1009.452224] env[63515]: DEBUG oslo_concurrency.lockutils [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.330s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1009.452289] env[63515]: DEBUG oslo_concurrency.lockutils [None req-92f0089c-1858-4533-be29-3b628a87ca62 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 4.881s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1009.452482] env[63515]: DEBUG oslo_concurrency.lockutils [None req-92f0089c-1858-4533-be29-3b628a87ca62 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1009.454707] env[63515]: DEBUG oslo_concurrency.lockutils [None req-e5acb69c-ccc9-4835-a62a-46a52c0ee313 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.953s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1009.454903] env[63515]: DEBUG oslo_concurrency.lockutils [None req-e5acb69c-ccc9-4835-a62a-46a52c0ee313 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1009.456595] env[63515]: DEBUG oslo_concurrency.lockutils [None req-b6bf67af-7b53-4121-830a-6dc78e8dcba3 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 1.541s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1009.461762] env[63515]: DEBUG nova.network.neutron [req-8bf1a21f-3483-48dd-bc61-405be478d6b2 req-d385ec4e-3018-470f-ae89-9475dc5946b4 service nova] [instance: c6ac42a9-a233-48b4-bcb2-fb791c7446bb] Updated VIF entry in instance network info cache for port 161e8482-76a0-4c82-8354-af132110794d. {{(pid=63515) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1009.461762] env[63515]: DEBUG nova.network.neutron [req-8bf1a21f-3483-48dd-bc61-405be478d6b2 req-d385ec4e-3018-470f-ae89-9475dc5946b4 service nova] [instance: c6ac42a9-a233-48b4-bcb2-fb791c7446bb] Updating instance_info_cache with network_info: [{"id": "161e8482-76a0-4c82-8354-af132110794d", "address": "fa:16:3e:3e:77:7f", "network": {"id": "88dfb29d-41e0-4df7-8f70-e52cda05620c", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1355578400-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7d4ada9ec35f42b19c6480a9101d21a4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "357d2811-e990-4985-9f9e-b158d10d3699", "external-id": "nsx-vlan-transportzone-641", "segmentation_id": 641, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap161e8482-76", "ovs_interfaceid": "161e8482-76a0-4c82-8354-af132110794d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1009.485283] env[63515]: INFO nova.scheduler.client.report [None req-92f0089c-1858-4533-be29-3b628a87ca62 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Deleted allocations for instance 5cbce760-0163-4b27-8ae3-e46c926c8916 [ 1009.487846] env[63515]: INFO nova.scheduler.client.report [None req-e5acb69c-ccc9-4835-a62a-46a52c0ee313 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Deleted allocations for instance d024b6b2-dd10-4112-89e8-ced57efe8208 [ 1009.508964] env[63515]: DEBUG oslo_vmware.api [None req-642b03cc-2e02-4a86-bca6-75c3316e966c tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Task: {'id': task-1111803, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.200099} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1009.509436] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-642b03cc-2e02-4a86-bca6-75c3316e966c tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Deleted the datastore file {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1009.509654] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-642b03cc-2e02-4a86-bca6-75c3316e966c tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] [instance: 7d572ac0-f6af-4622-96cc-e75983420222] Deleted contents of the VM from datastore datastore2 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1009.509843] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-642b03cc-2e02-4a86-bca6-75c3316e966c tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] [instance: 7d572ac0-f6af-4622-96cc-e75983420222] Instance destroyed {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1009.528425] env[63515]: DEBUG nova.network.neutron [req-2c79e385-285c-4d37-87ec-72d7af62ca9e req-a6e85153-c431-444a-920b-65823615d53b service nova] [instance: fcfe3376-8dfa-4189-a267-ff6402713c1b] Updated VIF entry in instance network info cache for port 4462b1aa-72ea-41cc-878a-68c38920c8cc. {{(pid=63515) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1009.528983] env[63515]: DEBUG nova.network.neutron [req-2c79e385-285c-4d37-87ec-72d7af62ca9e req-a6e85153-c431-444a-920b-65823615d53b service nova] [instance: fcfe3376-8dfa-4189-a267-ff6402713c1b] Updating instance_info_cache with network_info: [{"id": "4462b1aa-72ea-41cc-878a-68c38920c8cc", "address": "fa:16:3e:4f:bd:04", "network": {"id": "9fc2b9c9-6afc-4510-bf13-f87f1e1bd53b", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-513162532-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "0be5ebcf766b4da5b33be3a12387e737", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0248a27a-1d7f-4195-987b-06bfc8467347", "external-id": "nsx-vlan-transportzone-26", "segmentation_id": 26, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4462b1aa-72", "ovs_interfaceid": "4462b1aa-72ea-41cc-878a-68c38920c8cc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1009.534209] env[63515]: INFO nova.scheduler.client.report [None req-642b03cc-2e02-4a86-bca6-75c3316e966c tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Deleted allocations for instance 7d572ac0-f6af-4622-96cc-e75983420222 [ 1009.805092] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1111806, 'name': CreateVM_Task, 'duration_secs': 0.386356} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1009.805247] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fcfe3376-8dfa-4189-a267-ff6402713c1b] Created VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1009.805969] env[63515]: DEBUG oslo_concurrency.lockutils [None req-b36e84df-7edf-4fd3-971f-5ed772c5647a tempest-ServerRescueTestJSONUnderV235-800430211 tempest-ServerRescueTestJSONUnderV235-800430211-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1009.806162] env[63515]: DEBUG oslo_concurrency.lockutils [None req-b36e84df-7edf-4fd3-971f-5ed772c5647a tempest-ServerRescueTestJSONUnderV235-800430211 tempest-ServerRescueTestJSONUnderV235-800430211-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1009.806477] env[63515]: DEBUG oslo_concurrency.lockutils [None req-b36e84df-7edf-4fd3-971f-5ed772c5647a tempest-ServerRescueTestJSONUnderV235-800430211 tempest-ServerRescueTestJSONUnderV235-800430211-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1009.806746] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c3f91112-0b0a-4d53-9ea9-ea4f03aa32fa {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.811252] env[63515]: DEBUG oslo_vmware.api [None req-b36e84df-7edf-4fd3-971f-5ed772c5647a tempest-ServerRescueTestJSONUnderV235-800430211 tempest-ServerRescueTestJSONUnderV235-800430211-project-member] Waiting for the task: (returnval){ [ 1009.811252] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]52f30855-f849-5cf9-0fce-2e6598dedf55" [ 1009.811252] env[63515]: _type = "Task" [ 1009.811252] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1009.820130] env[63515]: DEBUG oslo_vmware.api [None req-b36e84df-7edf-4fd3-971f-5ed772c5647a tempest-ServerRescueTestJSONUnderV235-800430211 tempest-ServerRescueTestJSONUnderV235-800430211-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52f30855-f849-5cf9-0fce-2e6598dedf55, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1009.845643] env[63515]: DEBUG oslo_vmware.api [None req-655527ff-c992-4371-875f-a71b38db6213 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Task: {'id': task-1111801, 'name': Rename_Task, 'duration_secs': 1.171947} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1009.845893] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-655527ff-c992-4371-875f-a71b38db6213 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: d6c5a9f9-5e5c-4b14-9b9a-7825bbf526a5] Powering on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1009.846148] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9c69ca42-cdb3-4965-b829-edadb451f9ae {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.853075] env[63515]: DEBUG oslo_vmware.api [None req-655527ff-c992-4371-875f-a71b38db6213 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Waiting for the task: (returnval){ [ 1009.853075] env[63515]: value = "task-1111807" [ 1009.853075] env[63515]: _type = "Task" [ 1009.853075] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1009.860818] env[63515]: DEBUG oslo_vmware.api [None req-655527ff-c992-4371-875f-a71b38db6213 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Task: {'id': task-1111807, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1009.969572] env[63515]: DEBUG oslo_concurrency.lockutils [req-8bf1a21f-3483-48dd-bc61-405be478d6b2 req-d385ec4e-3018-470f-ae89-9475dc5946b4 service nova] Releasing lock "refresh_cache-c6ac42a9-a233-48b4-bcb2-fb791c7446bb" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1009.970131] env[63515]: DEBUG oslo_concurrency.lockutils [req-a2ba92a9-d3ca-431d-ae5b-819a90b50c65 req-4b00e14d-f4f3-48a8-a848-7368a80ccd95 service nova] Acquired lock "refresh_cache-c6ac42a9-a233-48b4-bcb2-fb791c7446bb" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1009.970341] env[63515]: DEBUG nova.network.neutron [req-a2ba92a9-d3ca-431d-ae5b-819a90b50c65 req-4b00e14d-f4f3-48a8-a848-7368a80ccd95 service nova] [instance: c6ac42a9-a233-48b4-bcb2-fb791c7446bb] Refreshing network info cache for port 161e8482-76a0-4c82-8354-af132110794d {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1009.999287] env[63515]: DEBUG oslo_concurrency.lockutils [None req-92f0089c-1858-4533-be29-3b628a87ca62 tempest-ServerRescueNegativeTestJSON-1502871851 tempest-ServerRescueNegativeTestJSON-1502871851-project-member] Lock "5cbce760-0163-4b27-8ae3-e46c926c8916" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.525s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1010.000746] env[63515]: DEBUG oslo_concurrency.lockutils [None req-e5acb69c-ccc9-4835-a62a-46a52c0ee313 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Lock "d024b6b2-dd10-4112-89e8-ced57efe8208" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.417s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1010.032240] env[63515]: DEBUG oslo_concurrency.lockutils [req-2c79e385-285c-4d37-87ec-72d7af62ca9e req-a6e85153-c431-444a-920b-65823615d53b service nova] Releasing lock "refresh_cache-fcfe3376-8dfa-4189-a267-ff6402713c1b" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1010.032619] env[63515]: DEBUG nova.compute.manager [req-2c79e385-285c-4d37-87ec-72d7af62ca9e req-a6e85153-c431-444a-920b-65823615d53b service nova] [instance: 7d572ac0-f6af-4622-96cc-e75983420222] Received event network-vif-unplugged-dbfd09c2-9ec4-4821-8613-43c6d6cf4724 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1010.032912] env[63515]: DEBUG oslo_concurrency.lockutils [req-2c79e385-285c-4d37-87ec-72d7af62ca9e req-a6e85153-c431-444a-920b-65823615d53b service nova] Acquiring lock "7d572ac0-f6af-4622-96cc-e75983420222-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1010.033279] env[63515]: DEBUG oslo_concurrency.lockutils [req-2c79e385-285c-4d37-87ec-72d7af62ca9e req-a6e85153-c431-444a-920b-65823615d53b service nova] Lock "7d572ac0-f6af-4622-96cc-e75983420222-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1010.033401] env[63515]: DEBUG oslo_concurrency.lockutils [req-2c79e385-285c-4d37-87ec-72d7af62ca9e req-a6e85153-c431-444a-920b-65823615d53b service nova] Lock "7d572ac0-f6af-4622-96cc-e75983420222-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1010.033628] env[63515]: DEBUG nova.compute.manager [req-2c79e385-285c-4d37-87ec-72d7af62ca9e req-a6e85153-c431-444a-920b-65823615d53b service nova] [instance: 7d572ac0-f6af-4622-96cc-e75983420222] No waiting events found dispatching network-vif-unplugged-dbfd09c2-9ec4-4821-8613-43c6d6cf4724 {{(pid=63515) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1010.033858] env[63515]: WARNING nova.compute.manager [req-2c79e385-285c-4d37-87ec-72d7af62ca9e req-a6e85153-c431-444a-920b-65823615d53b service nova] [instance: 7d572ac0-f6af-4622-96cc-e75983420222] Received unexpected event network-vif-unplugged-dbfd09c2-9ec4-4821-8613-43c6d6cf4724 for instance with vm_state shelved and task_state shelving_offloading. [ 1010.034066] env[63515]: DEBUG nova.compute.manager [req-2c79e385-285c-4d37-87ec-72d7af62ca9e req-a6e85153-c431-444a-920b-65823615d53b service nova] [instance: 7d572ac0-f6af-4622-96cc-e75983420222] Received event network-changed-dbfd09c2-9ec4-4821-8613-43c6d6cf4724 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1010.034241] env[63515]: DEBUG nova.compute.manager [req-2c79e385-285c-4d37-87ec-72d7af62ca9e req-a6e85153-c431-444a-920b-65823615d53b service nova] [instance: 7d572ac0-f6af-4622-96cc-e75983420222] Refreshing instance network info cache due to event network-changed-dbfd09c2-9ec4-4821-8613-43c6d6cf4724. {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1010.034440] env[63515]: DEBUG oslo_concurrency.lockutils [req-2c79e385-285c-4d37-87ec-72d7af62ca9e req-a6e85153-c431-444a-920b-65823615d53b service nova] Acquiring lock "refresh_cache-7d572ac0-f6af-4622-96cc-e75983420222" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1010.034600] env[63515]: DEBUG oslo_concurrency.lockutils [req-2c79e385-285c-4d37-87ec-72d7af62ca9e req-a6e85153-c431-444a-920b-65823615d53b service nova] Acquired lock "refresh_cache-7d572ac0-f6af-4622-96cc-e75983420222" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1010.034920] env[63515]: DEBUG nova.network.neutron [req-2c79e385-285c-4d37-87ec-72d7af62ca9e req-a6e85153-c431-444a-920b-65823615d53b service nova] [instance: 7d572ac0-f6af-4622-96cc-e75983420222] Refreshing network info cache for port dbfd09c2-9ec4-4821-8613-43c6d6cf4724 {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1010.039733] env[63515]: DEBUG oslo_concurrency.lockutils [None req-642b03cc-2e02-4a86-bca6-75c3316e966c tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1010.111463] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6fe6987-c45b-4918-95ed-32846835e075 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.121214] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a5f6ab3-267a-4773-b7fd-502c0d4a1897 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.152714] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e0d5637-73fe-4b41-bd49-cec8c1019281 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.161834] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20d7c25e-8f42-43a7-a0a7-89a395698721 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.176477] env[63515]: DEBUG nova.compute.provider_tree [None req-b6bf67af-7b53-4121-830a-6dc78e8dcba3 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1010.322957] env[63515]: DEBUG oslo_vmware.api [None req-b36e84df-7edf-4fd3-971f-5ed772c5647a tempest-ServerRescueTestJSONUnderV235-800430211 tempest-ServerRescueTestJSONUnderV235-800430211-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52f30855-f849-5cf9-0fce-2e6598dedf55, 'name': SearchDatastore_Task, 'duration_secs': 0.01217} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1010.323332] env[63515]: DEBUG oslo_concurrency.lockutils [None req-b36e84df-7edf-4fd3-971f-5ed772c5647a tempest-ServerRescueTestJSONUnderV235-800430211 tempest-ServerRescueTestJSONUnderV235-800430211-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1010.323633] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-b36e84df-7edf-4fd3-971f-5ed772c5647a tempest-ServerRescueTestJSONUnderV235-800430211 tempest-ServerRescueTestJSONUnderV235-800430211-project-member] [instance: fcfe3376-8dfa-4189-a267-ff6402713c1b] Processing image 8a120570-cb06-4099-b262-554ca0ad15c5 {{(pid=63515) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1010.323886] env[63515]: DEBUG oslo_concurrency.lockutils [None req-b36e84df-7edf-4fd3-971f-5ed772c5647a tempest-ServerRescueTestJSONUnderV235-800430211 tempest-ServerRescueTestJSONUnderV235-800430211-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1010.324050] env[63515]: DEBUG oslo_concurrency.lockutils [None req-b36e84df-7edf-4fd3-971f-5ed772c5647a tempest-ServerRescueTestJSONUnderV235-800430211 tempest-ServerRescueTestJSONUnderV235-800430211-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1010.324242] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-b36e84df-7edf-4fd3-971f-5ed772c5647a tempest-ServerRescueTestJSONUnderV235-800430211 tempest-ServerRescueTestJSONUnderV235-800430211-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1010.324523] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-9fafb169-0263-4b1a-a9ee-176801a4f17e {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.336454] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-b36e84df-7edf-4fd3-971f-5ed772c5647a tempest-ServerRescueTestJSONUnderV235-800430211 tempest-ServerRescueTestJSONUnderV235-800430211-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1010.336454] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-b36e84df-7edf-4fd3-971f-5ed772c5647a tempest-ServerRescueTestJSONUnderV235-800430211 tempest-ServerRescueTestJSONUnderV235-800430211-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63515) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1010.336454] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6a9bb759-3cf8-4755-ba50-56e0029d704e {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.344739] env[63515]: DEBUG oslo_vmware.api [None req-b36e84df-7edf-4fd3-971f-5ed772c5647a tempest-ServerRescueTestJSONUnderV235-800430211 tempest-ServerRescueTestJSONUnderV235-800430211-project-member] Waiting for the task: (returnval){ [ 1010.344739] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]526516e0-206c-6dcb-c62b-72da20d2fd69" [ 1010.344739] env[63515]: _type = "Task" [ 1010.344739] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1010.354115] env[63515]: DEBUG oslo_vmware.api [None req-b36e84df-7edf-4fd3-971f-5ed772c5647a tempest-ServerRescueTestJSONUnderV235-800430211 tempest-ServerRescueTestJSONUnderV235-800430211-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]526516e0-206c-6dcb-c62b-72da20d2fd69, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1010.364195] env[63515]: DEBUG oslo_vmware.api [None req-655527ff-c992-4371-875f-a71b38db6213 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Task: {'id': task-1111807, 'name': PowerOnVM_Task, 'duration_secs': 0.493649} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1010.364195] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-655527ff-c992-4371-875f-a71b38db6213 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: d6c5a9f9-5e5c-4b14-9b9a-7825bbf526a5] Powered on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1010.364263] env[63515]: INFO nova.compute.manager [None req-655527ff-c992-4371-875f-a71b38db6213 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: d6c5a9f9-5e5c-4b14-9b9a-7825bbf526a5] Took 9.42 seconds to spawn the instance on the hypervisor. [ 1010.365180] env[63515]: DEBUG nova.compute.manager [None req-655527ff-c992-4371-875f-a71b38db6213 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: d6c5a9f9-5e5c-4b14-9b9a-7825bbf526a5] Checking state {{(pid=63515) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1010.365480] env[63515]: DEBUG nova.compute.manager [req-c8d4c74e-33f6-48bc-a27d-ff211819d4ad req-860221ee-b4b7-47fd-b6ed-1c21a44c0f2b service nova] [instance: b76cf3dd-7b7e-4e5d-bf3c-e15bd37069ac] Received event network-vif-plugged-acea8927-7e43-427c-bed2-2420d4708f3b {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1010.365673] env[63515]: DEBUG oslo_concurrency.lockutils [req-c8d4c74e-33f6-48bc-a27d-ff211819d4ad req-860221ee-b4b7-47fd-b6ed-1c21a44c0f2b service nova] Acquiring lock "b76cf3dd-7b7e-4e5d-bf3c-e15bd37069ac-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1010.365885] env[63515]: DEBUG oslo_concurrency.lockutils [req-c8d4c74e-33f6-48bc-a27d-ff211819d4ad req-860221ee-b4b7-47fd-b6ed-1c21a44c0f2b service nova] Lock "b76cf3dd-7b7e-4e5d-bf3c-e15bd37069ac-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1010.366060] env[63515]: DEBUG oslo_concurrency.lockutils [req-c8d4c74e-33f6-48bc-a27d-ff211819d4ad req-860221ee-b4b7-47fd-b6ed-1c21a44c0f2b service nova] Lock "b76cf3dd-7b7e-4e5d-bf3c-e15bd37069ac-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1010.366226] env[63515]: DEBUG nova.compute.manager [req-c8d4c74e-33f6-48bc-a27d-ff211819d4ad req-860221ee-b4b7-47fd-b6ed-1c21a44c0f2b service nova] [instance: b76cf3dd-7b7e-4e5d-bf3c-e15bd37069ac] No waiting events found dispatching network-vif-plugged-acea8927-7e43-427c-bed2-2420d4708f3b {{(pid=63515) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1010.366385] env[63515]: WARNING nova.compute.manager [req-c8d4c74e-33f6-48bc-a27d-ff211819d4ad req-860221ee-b4b7-47fd-b6ed-1c21a44c0f2b service nova] [instance: b76cf3dd-7b7e-4e5d-bf3c-e15bd37069ac] Received unexpected event network-vif-plugged-acea8927-7e43-427c-bed2-2420d4708f3b for instance with vm_state building and task_state spawning. [ 1010.367202] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b66cca13-82b2-4588-b43a-7d4f7e09a9e2 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.682372] env[63515]: DEBUG nova.scheduler.client.report [None req-b6bf67af-7b53-4121-830a-6dc78e8dcba3 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1010.782218] env[63515]: DEBUG nova.compute.manager [req-04be140f-01da-46fc-b757-156fbf471d04 req-f4f7d2af-8a70-4f83-9ef9-e1ebc37c6684 service nova] [instance: 724b05bc-ce4b-4f99-ad14-f3abea5567f2] Received event network-changed-3dee43b6-2650-42e1-aa5a-4994a3ec7f05 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1010.782491] env[63515]: DEBUG nova.compute.manager [req-04be140f-01da-46fc-b757-156fbf471d04 req-f4f7d2af-8a70-4f83-9ef9-e1ebc37c6684 service nova] [instance: 724b05bc-ce4b-4f99-ad14-f3abea5567f2] Refreshing instance network info cache due to event network-changed-3dee43b6-2650-42e1-aa5a-4994a3ec7f05. {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1010.782719] env[63515]: DEBUG oslo_concurrency.lockutils [req-04be140f-01da-46fc-b757-156fbf471d04 req-f4f7d2af-8a70-4f83-9ef9-e1ebc37c6684 service nova] Acquiring lock "refresh_cache-724b05bc-ce4b-4f99-ad14-f3abea5567f2" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1010.782896] env[63515]: DEBUG oslo_concurrency.lockutils [req-04be140f-01da-46fc-b757-156fbf471d04 req-f4f7d2af-8a70-4f83-9ef9-e1ebc37c6684 service nova] Acquired lock "refresh_cache-724b05bc-ce4b-4f99-ad14-f3abea5567f2" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1010.783102] env[63515]: DEBUG nova.network.neutron [req-04be140f-01da-46fc-b757-156fbf471d04 req-f4f7d2af-8a70-4f83-9ef9-e1ebc37c6684 service nova] [instance: 724b05bc-ce4b-4f99-ad14-f3abea5567f2] Refreshing network info cache for port 3dee43b6-2650-42e1-aa5a-4994a3ec7f05 {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1010.828950] env[63515]: DEBUG nova.network.neutron [req-2c79e385-285c-4d37-87ec-72d7af62ca9e req-a6e85153-c431-444a-920b-65823615d53b service nova] [instance: 7d572ac0-f6af-4622-96cc-e75983420222] Updated VIF entry in instance network info cache for port dbfd09c2-9ec4-4821-8613-43c6d6cf4724. {{(pid=63515) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1010.829920] env[63515]: DEBUG nova.network.neutron [req-2c79e385-285c-4d37-87ec-72d7af62ca9e req-a6e85153-c431-444a-920b-65823615d53b service nova] [instance: 7d572ac0-f6af-4622-96cc-e75983420222] Updating instance_info_cache with network_info: [{"id": "dbfd09c2-9ec4-4821-8613-43c6d6cf4724", "address": "fa:16:3e:6d:92:3d", "network": {"id": "09fb20ca-2b17-4d46-b915-6aff85b862d8", "bridge": null, "label": "tempest-AttachVolumeShelveTestJSON-1815529427-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.244", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "19ce3b05eec64475b685faad1fd801f6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tapdbfd09c2-9e", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1010.831774] env[63515]: DEBUG nova.network.neutron [req-a2ba92a9-d3ca-431d-ae5b-819a90b50c65 req-4b00e14d-f4f3-48a8-a848-7368a80ccd95 service nova] [instance: c6ac42a9-a233-48b4-bcb2-fb791c7446bb] Updated VIF entry in instance network info cache for port 161e8482-76a0-4c82-8354-af132110794d. {{(pid=63515) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1010.832112] env[63515]: DEBUG nova.network.neutron [req-a2ba92a9-d3ca-431d-ae5b-819a90b50c65 req-4b00e14d-f4f3-48a8-a848-7368a80ccd95 service nova] [instance: c6ac42a9-a233-48b4-bcb2-fb791c7446bb] Updating instance_info_cache with network_info: [{"id": "161e8482-76a0-4c82-8354-af132110794d", "address": "fa:16:3e:3e:77:7f", "network": {"id": "88dfb29d-41e0-4df7-8f70-e52cda05620c", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1355578400-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7d4ada9ec35f42b19c6480a9101d21a4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "357d2811-e990-4985-9f9e-b158d10d3699", "external-id": "nsx-vlan-transportzone-641", "segmentation_id": 641, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap161e8482-76", "ovs_interfaceid": "161e8482-76a0-4c82-8354-af132110794d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1010.857438] env[63515]: DEBUG oslo_vmware.api [None req-b36e84df-7edf-4fd3-971f-5ed772c5647a tempest-ServerRescueTestJSONUnderV235-800430211 tempest-ServerRescueTestJSONUnderV235-800430211-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]526516e0-206c-6dcb-c62b-72da20d2fd69, 'name': SearchDatastore_Task, 'duration_secs': 0.013292} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1010.858376] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5ef32d68-1362-4cd7-a110-d1297766e848 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.867208] env[63515]: DEBUG oslo_vmware.api [None req-b36e84df-7edf-4fd3-971f-5ed772c5647a tempest-ServerRescueTestJSONUnderV235-800430211 tempest-ServerRescueTestJSONUnderV235-800430211-project-member] Waiting for the task: (returnval){ [ 1010.867208] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]52d0069b-5b80-da50-4773-52d15c61719c" [ 1010.867208] env[63515]: _type = "Task" [ 1010.867208] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1010.882968] env[63515]: DEBUG oslo_vmware.api [None req-b36e84df-7edf-4fd3-971f-5ed772c5647a tempest-ServerRescueTestJSONUnderV235-800430211 tempest-ServerRescueTestJSONUnderV235-800430211-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52d0069b-5b80-da50-4773-52d15c61719c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1010.886308] env[63515]: INFO nova.compute.manager [None req-655527ff-c992-4371-875f-a71b38db6213 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: d6c5a9f9-5e5c-4b14-9b9a-7825bbf526a5] Took 20.57 seconds to build instance. [ 1010.907247] env[63515]: DEBUG oslo_concurrency.lockutils [None req-042daf1f-2fe3-4b1d-bb04-2f7a0479fcba tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Acquiring lock "e925fc93-a731-498a-984c-9b1f3eabf353" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1010.907477] env[63515]: DEBUG oslo_concurrency.lockutils [None req-042daf1f-2fe3-4b1d-bb04-2f7a0479fcba tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Lock "e925fc93-a731-498a-984c-9b1f3eabf353" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1010.909896] env[63515]: DEBUG nova.network.neutron [None req-96531a62-98f3-4a7c-9020-152fd796dc38 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: b76cf3dd-7b7e-4e5d-bf3c-e15bd37069ac] Successfully updated port: acea8927-7e43-427c-bed2-2420d4708f3b {{(pid=63515) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1011.334171] env[63515]: DEBUG oslo_concurrency.lockutils [req-2c79e385-285c-4d37-87ec-72d7af62ca9e req-a6e85153-c431-444a-920b-65823615d53b service nova] Releasing lock "refresh_cache-7d572ac0-f6af-4622-96cc-e75983420222" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1011.335130] env[63515]: DEBUG oslo_concurrency.lockutils [req-a2ba92a9-d3ca-431d-ae5b-819a90b50c65 req-4b00e14d-f4f3-48a8-a848-7368a80ccd95 service nova] Releasing lock "refresh_cache-c6ac42a9-a233-48b4-bcb2-fb791c7446bb" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1011.381061] env[63515]: DEBUG oslo_vmware.api [None req-b36e84df-7edf-4fd3-971f-5ed772c5647a tempest-ServerRescueTestJSONUnderV235-800430211 tempest-ServerRescueTestJSONUnderV235-800430211-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52d0069b-5b80-da50-4773-52d15c61719c, 'name': SearchDatastore_Task, 'duration_secs': 0.018569} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1011.383429] env[63515]: DEBUG oslo_concurrency.lockutils [None req-b36e84df-7edf-4fd3-971f-5ed772c5647a tempest-ServerRescueTestJSONUnderV235-800430211 tempest-ServerRescueTestJSONUnderV235-800430211-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1011.383705] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-b36e84df-7edf-4fd3-971f-5ed772c5647a tempest-ServerRescueTestJSONUnderV235-800430211 tempest-ServerRescueTestJSONUnderV235-800430211-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk to [datastore1] fcfe3376-8dfa-4189-a267-ff6402713c1b/fcfe3376-8dfa-4189-a267-ff6402713c1b.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1011.384014] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-103ecd17-1933-4864-897c-11eb9cea3025 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.390306] env[63515]: DEBUG oslo_concurrency.lockutils [None req-655527ff-c992-4371-875f-a71b38db6213 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Lock "d6c5a9f9-5e5c-4b14-9b9a-7825bbf526a5" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 22.084s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1011.395156] env[63515]: DEBUG oslo_vmware.api [None req-b36e84df-7edf-4fd3-971f-5ed772c5647a tempest-ServerRescueTestJSONUnderV235-800430211 tempest-ServerRescueTestJSONUnderV235-800430211-project-member] Waiting for the task: (returnval){ [ 1011.395156] env[63515]: value = "task-1111808" [ 1011.395156] env[63515]: _type = "Task" [ 1011.395156] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1011.402760] env[63515]: DEBUG oslo_vmware.api [None req-b36e84df-7edf-4fd3-971f-5ed772c5647a tempest-ServerRescueTestJSONUnderV235-800430211 tempest-ServerRescueTestJSONUnderV235-800430211-project-member] Task: {'id': task-1111808, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1011.412317] env[63515]: DEBUG nova.compute.manager [None req-042daf1f-2fe3-4b1d-bb04-2f7a0479fcba tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: e925fc93-a731-498a-984c-9b1f3eabf353] Starting instance... {{(pid=63515) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1011.415991] env[63515]: DEBUG oslo_concurrency.lockutils [None req-96531a62-98f3-4a7c-9020-152fd796dc38 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Acquiring lock "refresh_cache-b76cf3dd-7b7e-4e5d-bf3c-e15bd37069ac" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1011.416151] env[63515]: DEBUG oslo_concurrency.lockutils [None req-96531a62-98f3-4a7c-9020-152fd796dc38 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Acquired lock "refresh_cache-b76cf3dd-7b7e-4e5d-bf3c-e15bd37069ac" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1011.416312] env[63515]: DEBUG nova.network.neutron [None req-96531a62-98f3-4a7c-9020-152fd796dc38 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: b76cf3dd-7b7e-4e5d-bf3c-e15bd37069ac] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1011.692772] env[63515]: DEBUG nova.network.neutron [req-04be140f-01da-46fc-b757-156fbf471d04 req-f4f7d2af-8a70-4f83-9ef9-e1ebc37c6684 service nova] [instance: 724b05bc-ce4b-4f99-ad14-f3abea5567f2] Updated VIF entry in instance network info cache for port 3dee43b6-2650-42e1-aa5a-4994a3ec7f05. {{(pid=63515) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1011.693186] env[63515]: DEBUG nova.network.neutron [req-04be140f-01da-46fc-b757-156fbf471d04 req-f4f7d2af-8a70-4f83-9ef9-e1ebc37c6684 service nova] [instance: 724b05bc-ce4b-4f99-ad14-f3abea5567f2] Updating instance_info_cache with network_info: [{"id": "3dee43b6-2650-42e1-aa5a-4994a3ec7f05", "address": "fa:16:3e:89:06:d0", "network": {"id": "88dfb29d-41e0-4df7-8f70-e52cda05620c", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1355578400-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.146", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7d4ada9ec35f42b19c6480a9101d21a4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "357d2811-e990-4985-9f9e-b158d10d3699", "external-id": "nsx-vlan-transportzone-641", "segmentation_id": 641, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3dee43b6-26", "ovs_interfaceid": "3dee43b6-2650-42e1-aa5a-4994a3ec7f05", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1011.698035] env[63515]: DEBUG oslo_concurrency.lockutils [None req-b6bf67af-7b53-4121-830a-6dc78e8dcba3 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.240s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1011.698035] env[63515]: DEBUG nova.compute.manager [None req-b6bf67af-7b53-4121-830a-6dc78e8dcba3 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 68766cd4-84be-475b-8494-d7ab43a9e969] Resized/migrated instance is powered off. Setting vm_state to 'stopped'. {{(pid=63515) _confirm_resize /opt/stack/nova/nova/compute/manager.py:4909}} [ 1011.703506] env[63515]: DEBUG oslo_concurrency.lockutils [None req-642b03cc-2e02-4a86-bca6-75c3316e966c tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.664s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1011.703851] env[63515]: DEBUG nova.objects.instance [None req-642b03cc-2e02-4a86-bca6-75c3316e966c tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Lazy-loading 'resources' on Instance uuid 7d572ac0-f6af-4622-96cc-e75983420222 {{(pid=63515) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1011.904409] env[63515]: DEBUG oslo_vmware.api [None req-b36e84df-7edf-4fd3-971f-5ed772c5647a tempest-ServerRescueTestJSONUnderV235-800430211 tempest-ServerRescueTestJSONUnderV235-800430211-project-member] Task: {'id': task-1111808, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.484871} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1011.904725] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-b36e84df-7edf-4fd3-971f-5ed772c5647a tempest-ServerRescueTestJSONUnderV235-800430211 tempest-ServerRescueTestJSONUnderV235-800430211-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk to [datastore1] fcfe3376-8dfa-4189-a267-ff6402713c1b/fcfe3376-8dfa-4189-a267-ff6402713c1b.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1011.905307] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-b36e84df-7edf-4fd3-971f-5ed772c5647a tempest-ServerRescueTestJSONUnderV235-800430211 tempest-ServerRescueTestJSONUnderV235-800430211-project-member] [instance: fcfe3376-8dfa-4189-a267-ff6402713c1b] Extending root virtual disk to 1048576 {{(pid=63515) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1011.905374] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-f71e542a-0bb8-484b-8d8e-5598dc7a93af {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.912403] env[63515]: DEBUG oslo_vmware.api [None req-b36e84df-7edf-4fd3-971f-5ed772c5647a tempest-ServerRescueTestJSONUnderV235-800430211 tempest-ServerRescueTestJSONUnderV235-800430211-project-member] Waiting for the task: (returnval){ [ 1011.912403] env[63515]: value = "task-1111809" [ 1011.912403] env[63515]: _type = "Task" [ 1011.912403] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1011.924967] env[63515]: DEBUG oslo_vmware.api [None req-b36e84df-7edf-4fd3-971f-5ed772c5647a tempest-ServerRescueTestJSONUnderV235-800430211 tempest-ServerRescueTestJSONUnderV235-800430211-project-member] Task: {'id': task-1111809, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1011.933335] env[63515]: DEBUG nova.compute.manager [req-e83e1da2-06a6-4b6a-9bec-2e879c55ee9f req-c28a57da-26d9-476f-9396-3e342f499cde service nova] [instance: d6c5a9f9-5e5c-4b14-9b9a-7825bbf526a5] Received event network-changed-ee82711d-2d77-4127-b804-7db65e838617 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1011.933538] env[63515]: DEBUG nova.compute.manager [req-e83e1da2-06a6-4b6a-9bec-2e879c55ee9f req-c28a57da-26d9-476f-9396-3e342f499cde service nova] [instance: d6c5a9f9-5e5c-4b14-9b9a-7825bbf526a5] Refreshing instance network info cache due to event network-changed-ee82711d-2d77-4127-b804-7db65e838617. {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1011.933820] env[63515]: DEBUG oslo_concurrency.lockutils [req-e83e1da2-06a6-4b6a-9bec-2e879c55ee9f req-c28a57da-26d9-476f-9396-3e342f499cde service nova] Acquiring lock "refresh_cache-d6c5a9f9-5e5c-4b14-9b9a-7825bbf526a5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1011.934105] env[63515]: DEBUG oslo_concurrency.lockutils [req-e83e1da2-06a6-4b6a-9bec-2e879c55ee9f req-c28a57da-26d9-476f-9396-3e342f499cde service nova] Acquired lock "refresh_cache-d6c5a9f9-5e5c-4b14-9b9a-7825bbf526a5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1011.934238] env[63515]: DEBUG nova.network.neutron [req-e83e1da2-06a6-4b6a-9bec-2e879c55ee9f req-c28a57da-26d9-476f-9396-3e342f499cde service nova] [instance: d6c5a9f9-5e5c-4b14-9b9a-7825bbf526a5] Refreshing network info cache for port ee82711d-2d77-4127-b804-7db65e838617 {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1011.944330] env[63515]: DEBUG oslo_concurrency.lockutils [None req-042daf1f-2fe3-4b1d-bb04-2f7a0479fcba tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1011.977632] env[63515]: DEBUG nova.network.neutron [None req-96531a62-98f3-4a7c-9020-152fd796dc38 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: b76cf3dd-7b7e-4e5d-bf3c-e15bd37069ac] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1012.139664] env[63515]: DEBUG nova.network.neutron [None req-96531a62-98f3-4a7c-9020-152fd796dc38 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: b76cf3dd-7b7e-4e5d-bf3c-e15bd37069ac] Updating instance_info_cache with network_info: [{"id": "acea8927-7e43-427c-bed2-2420d4708f3b", "address": "fa:16:3e:21:d3:df", "network": {"id": "3e5914c8-0cd4-49ee-81b4-1f3e280440bb", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-641632277-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ea099ce07bfb4a8da014d9303cf552eb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "225b6979-9329-403b-91fa-138bd41f6e83", "external-id": "nsx-vlan-transportzone-38", "segmentation_id": 38, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapacea8927-7e", "ovs_interfaceid": "acea8927-7e43-427c-bed2-2420d4708f3b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1012.205229] env[63515]: DEBUG oslo_concurrency.lockutils [req-04be140f-01da-46fc-b757-156fbf471d04 req-f4f7d2af-8a70-4f83-9ef9-e1ebc37c6684 service nova] Releasing lock "refresh_cache-724b05bc-ce4b-4f99-ad14-f3abea5567f2" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1012.209099] env[63515]: DEBUG nova.objects.instance [None req-642b03cc-2e02-4a86-bca6-75c3316e966c tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Lazy-loading 'numa_topology' on Instance uuid 7d572ac0-f6af-4622-96cc-e75983420222 {{(pid=63515) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1012.258901] env[63515]: INFO nova.scheduler.client.report [None req-b6bf67af-7b53-4121-830a-6dc78e8dcba3 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Deleted allocation for migration 0453c850-4f70-44d5-9d13-96d466406033 [ 1012.422666] env[63515]: DEBUG oslo_vmware.api [None req-b36e84df-7edf-4fd3-971f-5ed772c5647a tempest-ServerRescueTestJSONUnderV235-800430211 tempest-ServerRescueTestJSONUnderV235-800430211-project-member] Task: {'id': task-1111809, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.075369} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1012.423043] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-b36e84df-7edf-4fd3-971f-5ed772c5647a tempest-ServerRescueTestJSONUnderV235-800430211 tempest-ServerRescueTestJSONUnderV235-800430211-project-member] [instance: fcfe3376-8dfa-4189-a267-ff6402713c1b] Extended root virtual disk {{(pid=63515) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1012.423785] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0fec81e3-1ac5-4bf0-aaa1-d81c85fe14ed {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.447306] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-b36e84df-7edf-4fd3-971f-5ed772c5647a tempest-ServerRescueTestJSONUnderV235-800430211 tempest-ServerRescueTestJSONUnderV235-800430211-project-member] [instance: fcfe3376-8dfa-4189-a267-ff6402713c1b] Reconfiguring VM instance instance-00000063 to attach disk [datastore1] fcfe3376-8dfa-4189-a267-ff6402713c1b/fcfe3376-8dfa-4189-a267-ff6402713c1b.vmdk or device None with type sparse {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1012.449662] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e2666493-e44e-4bf4-9498-58b02fc9dc7d {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.471559] env[63515]: DEBUG oslo_vmware.api [None req-b36e84df-7edf-4fd3-971f-5ed772c5647a tempest-ServerRescueTestJSONUnderV235-800430211 tempest-ServerRescueTestJSONUnderV235-800430211-project-member] Waiting for the task: (returnval){ [ 1012.471559] env[63515]: value = "task-1111810" [ 1012.471559] env[63515]: _type = "Task" [ 1012.471559] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1012.482618] env[63515]: DEBUG oslo_vmware.api [None req-b36e84df-7edf-4fd3-971f-5ed772c5647a tempest-ServerRescueTestJSONUnderV235-800430211 tempest-ServerRescueTestJSONUnderV235-800430211-project-member] Task: {'id': task-1111810, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1012.642608] env[63515]: DEBUG oslo_concurrency.lockutils [None req-96531a62-98f3-4a7c-9020-152fd796dc38 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Releasing lock "refresh_cache-b76cf3dd-7b7e-4e5d-bf3c-e15bd37069ac" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1012.642954] env[63515]: DEBUG nova.compute.manager [None req-96531a62-98f3-4a7c-9020-152fd796dc38 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: b76cf3dd-7b7e-4e5d-bf3c-e15bd37069ac] Instance network_info: |[{"id": "acea8927-7e43-427c-bed2-2420d4708f3b", "address": "fa:16:3e:21:d3:df", "network": {"id": "3e5914c8-0cd4-49ee-81b4-1f3e280440bb", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-641632277-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ea099ce07bfb4a8da014d9303cf552eb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "225b6979-9329-403b-91fa-138bd41f6e83", "external-id": "nsx-vlan-transportzone-38", "segmentation_id": 38, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapacea8927-7e", "ovs_interfaceid": "acea8927-7e43-427c-bed2-2420d4708f3b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1012.643608] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-96531a62-98f3-4a7c-9020-152fd796dc38 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: b76cf3dd-7b7e-4e5d-bf3c-e15bd37069ac] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:21:d3:df', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '225b6979-9329-403b-91fa-138bd41f6e83', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'acea8927-7e43-427c-bed2-2420d4708f3b', 'vif_model': 'vmxnet3'}] {{(pid=63515) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1012.651637] env[63515]: DEBUG oslo.service.loopingcall [None req-96531a62-98f3-4a7c-9020-152fd796dc38 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1012.652289] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b76cf3dd-7b7e-4e5d-bf3c-e15bd37069ac] Creating VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1012.652573] env[63515]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c089e875-d245-4cab-9683-4ac4563d3f7c {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.675214] env[63515]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1012.675214] env[63515]: value = "task-1111811" [ 1012.675214] env[63515]: _type = "Task" [ 1012.675214] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1012.683503] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1111811, 'name': CreateVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1012.713268] env[63515]: DEBUG nova.objects.base [None req-642b03cc-2e02-4a86-bca6-75c3316e966c tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Object Instance<7d572ac0-f6af-4622-96cc-e75983420222> lazy-loaded attributes: resources,numa_topology {{(pid=63515) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1012.742162] env[63515]: DEBUG nova.network.neutron [req-e83e1da2-06a6-4b6a-9bec-2e879c55ee9f req-c28a57da-26d9-476f-9396-3e342f499cde service nova] [instance: d6c5a9f9-5e5c-4b14-9b9a-7825bbf526a5] Updated VIF entry in instance network info cache for port ee82711d-2d77-4127-b804-7db65e838617. {{(pid=63515) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1012.743753] env[63515]: DEBUG nova.network.neutron [req-e83e1da2-06a6-4b6a-9bec-2e879c55ee9f req-c28a57da-26d9-476f-9396-3e342f499cde service nova] [instance: d6c5a9f9-5e5c-4b14-9b9a-7825bbf526a5] Updating instance_info_cache with network_info: [{"id": "ee82711d-2d77-4127-b804-7db65e838617", "address": "fa:16:3e:21:f3:e4", "network": {"id": "cec614cb-7a5b-4b16-a83d-061ff62d6677", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1501468300-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.220", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "aef947bfaf7d4180b46b4be662e3d903", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3a0ddd7d-c321-4187-bdd8-b19044ea2c4a", "external-id": "nsx-vlan-transportzone-747", "segmentation_id": 747, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapee82711d-2d", "ovs_interfaceid": "ee82711d-2d77-4127-b804-7db65e838617", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1012.768185] env[63515]: DEBUG oslo_concurrency.lockutils [None req-b6bf67af-7b53-4121-830a-6dc78e8dcba3 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Lock "68766cd4-84be-475b-8494-d7ab43a9e969" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 8.153s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1012.807859] env[63515]: DEBUG nova.compute.manager [req-7d016afb-fe94-4d74-936a-13487b6578e0 req-3c7a6e8f-6682-47af-aabc-f3d20b319f2f service nova] [instance: b76cf3dd-7b7e-4e5d-bf3c-e15bd37069ac] Received event network-changed-acea8927-7e43-427c-bed2-2420d4708f3b {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1012.808137] env[63515]: DEBUG nova.compute.manager [req-7d016afb-fe94-4d74-936a-13487b6578e0 req-3c7a6e8f-6682-47af-aabc-f3d20b319f2f service nova] [instance: b76cf3dd-7b7e-4e5d-bf3c-e15bd37069ac] Refreshing instance network info cache due to event network-changed-acea8927-7e43-427c-bed2-2420d4708f3b. {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1012.808388] env[63515]: DEBUG oslo_concurrency.lockutils [req-7d016afb-fe94-4d74-936a-13487b6578e0 req-3c7a6e8f-6682-47af-aabc-f3d20b319f2f service nova] Acquiring lock "refresh_cache-b76cf3dd-7b7e-4e5d-bf3c-e15bd37069ac" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1012.808565] env[63515]: DEBUG oslo_concurrency.lockutils [req-7d016afb-fe94-4d74-936a-13487b6578e0 req-3c7a6e8f-6682-47af-aabc-f3d20b319f2f service nova] Acquired lock "refresh_cache-b76cf3dd-7b7e-4e5d-bf3c-e15bd37069ac" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1012.809200] env[63515]: DEBUG nova.network.neutron [req-7d016afb-fe94-4d74-936a-13487b6578e0 req-3c7a6e8f-6682-47af-aabc-f3d20b319f2f service nova] [instance: b76cf3dd-7b7e-4e5d-bf3c-e15bd37069ac] Refreshing network info cache for port acea8927-7e43-427c-bed2-2420d4708f3b {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1012.869267] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bdb99b11-5501-49ee-834f-0dda9071ac8a {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.876867] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e2a32a2-c022-46e8-b850-6e56da16a70f {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.909071] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-357d2550-18b2-4776-9049-69bfc225c2c8 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.917484] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-649e1422-0c70-498f-8c3e-0f8eb5aec586 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.933603] env[63515]: DEBUG nova.compute.provider_tree [None req-642b03cc-2e02-4a86-bca6-75c3316e966c tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1012.985967] env[63515]: DEBUG oslo_vmware.api [None req-b36e84df-7edf-4fd3-971f-5ed772c5647a tempest-ServerRescueTestJSONUnderV235-800430211 tempest-ServerRescueTestJSONUnderV235-800430211-project-member] Task: {'id': task-1111810, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.003520] env[63515]: DEBUG oslo_concurrency.lockutils [None req-878f8fb3-6690-40c9-9f6f-d6953e3612e1 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Acquiring lock "7d572ac0-f6af-4622-96cc-e75983420222" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1013.186978] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1111811, 'name': CreateVM_Task, 'duration_secs': 0.440541} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1013.187268] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b76cf3dd-7b7e-4e5d-bf3c-e15bd37069ac] Created VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1013.188314] env[63515]: DEBUG oslo_concurrency.lockutils [None req-96531a62-98f3-4a7c-9020-152fd796dc38 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1013.188546] env[63515]: DEBUG oslo_concurrency.lockutils [None req-96531a62-98f3-4a7c-9020-152fd796dc38 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1013.189021] env[63515]: DEBUG oslo_concurrency.lockutils [None req-96531a62-98f3-4a7c-9020-152fd796dc38 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1013.189397] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-08fccf93-d99d-4743-ae25-e0334b6b1d9c {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.195117] env[63515]: DEBUG oslo_vmware.api [None req-96531a62-98f3-4a7c-9020-152fd796dc38 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Waiting for the task: (returnval){ [ 1013.195117] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]529775ae-bd6a-f3fa-e8c1-76d8a840bae6" [ 1013.195117] env[63515]: _type = "Task" [ 1013.195117] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1013.203570] env[63515]: DEBUG oslo_vmware.api [None req-96531a62-98f3-4a7c-9020-152fd796dc38 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]529775ae-bd6a-f3fa-e8c1-76d8a840bae6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.247298] env[63515]: DEBUG oslo_concurrency.lockutils [req-e83e1da2-06a6-4b6a-9bec-2e879c55ee9f req-c28a57da-26d9-476f-9396-3e342f499cde service nova] Releasing lock "refresh_cache-d6c5a9f9-5e5c-4b14-9b9a-7825bbf526a5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1013.437527] env[63515]: DEBUG nova.scheduler.client.report [None req-642b03cc-2e02-4a86-bca6-75c3316e966c tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1013.446491] env[63515]: DEBUG nova.objects.instance [None req-e978376f-06cf-4dbb-9f2d-8124b426de89 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Lazy-loading 'flavor' on Instance uuid 68766cd4-84be-475b-8494-d7ab43a9e969 {{(pid=63515) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1013.484297] env[63515]: DEBUG oslo_vmware.api [None req-b36e84df-7edf-4fd3-971f-5ed772c5647a tempest-ServerRescueTestJSONUnderV235-800430211 tempest-ServerRescueTestJSONUnderV235-800430211-project-member] Task: {'id': task-1111810, 'name': ReconfigVM_Task, 'duration_secs': 0.699725} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1013.484572] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-b36e84df-7edf-4fd3-971f-5ed772c5647a tempest-ServerRescueTestJSONUnderV235-800430211 tempest-ServerRescueTestJSONUnderV235-800430211-project-member] [instance: fcfe3376-8dfa-4189-a267-ff6402713c1b] Reconfigured VM instance instance-00000063 to attach disk [datastore1] fcfe3376-8dfa-4189-a267-ff6402713c1b/fcfe3376-8dfa-4189-a267-ff6402713c1b.vmdk or device None with type sparse {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1013.485208] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e26c82c0-caec-414d-9b30-b1eb50a5ec75 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.491616] env[63515]: DEBUG oslo_vmware.api [None req-b36e84df-7edf-4fd3-971f-5ed772c5647a tempest-ServerRescueTestJSONUnderV235-800430211 tempest-ServerRescueTestJSONUnderV235-800430211-project-member] Waiting for the task: (returnval){ [ 1013.491616] env[63515]: value = "task-1111812" [ 1013.491616] env[63515]: _type = "Task" [ 1013.491616] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1013.500105] env[63515]: DEBUG oslo_vmware.api [None req-b36e84df-7edf-4fd3-971f-5ed772c5647a tempest-ServerRescueTestJSONUnderV235-800430211 tempest-ServerRescueTestJSONUnderV235-800430211-project-member] Task: {'id': task-1111812, 'name': Rename_Task} progress is 5%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.518052] env[63515]: DEBUG nova.network.neutron [req-7d016afb-fe94-4d74-936a-13487b6578e0 req-3c7a6e8f-6682-47af-aabc-f3d20b319f2f service nova] [instance: b76cf3dd-7b7e-4e5d-bf3c-e15bd37069ac] Updated VIF entry in instance network info cache for port acea8927-7e43-427c-bed2-2420d4708f3b. {{(pid=63515) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1013.518202] env[63515]: DEBUG nova.network.neutron [req-7d016afb-fe94-4d74-936a-13487b6578e0 req-3c7a6e8f-6682-47af-aabc-f3d20b319f2f service nova] [instance: b76cf3dd-7b7e-4e5d-bf3c-e15bd37069ac] Updating instance_info_cache with network_info: [{"id": "acea8927-7e43-427c-bed2-2420d4708f3b", "address": "fa:16:3e:21:d3:df", "network": {"id": "3e5914c8-0cd4-49ee-81b4-1f3e280440bb", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-641632277-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ea099ce07bfb4a8da014d9303cf552eb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "225b6979-9329-403b-91fa-138bd41f6e83", "external-id": "nsx-vlan-transportzone-38", "segmentation_id": 38, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapacea8927-7e", "ovs_interfaceid": "acea8927-7e43-427c-bed2-2420d4708f3b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1013.706111] env[63515]: DEBUG oslo_vmware.api [None req-96531a62-98f3-4a7c-9020-152fd796dc38 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]529775ae-bd6a-f3fa-e8c1-76d8a840bae6, 'name': SearchDatastore_Task, 'duration_secs': 0.010738} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1013.706356] env[63515]: DEBUG oslo_concurrency.lockutils [None req-96531a62-98f3-4a7c-9020-152fd796dc38 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1013.706600] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-96531a62-98f3-4a7c-9020-152fd796dc38 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: b76cf3dd-7b7e-4e5d-bf3c-e15bd37069ac] Processing image 8a120570-cb06-4099-b262-554ca0ad15c5 {{(pid=63515) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1013.706866] env[63515]: DEBUG oslo_concurrency.lockutils [None req-96531a62-98f3-4a7c-9020-152fd796dc38 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1013.707031] env[63515]: DEBUG oslo_concurrency.lockutils [None req-96531a62-98f3-4a7c-9020-152fd796dc38 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1013.707224] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-96531a62-98f3-4a7c-9020-152fd796dc38 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1013.707479] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-bc76e2f1-797f-4a28-bad8-eb799c7b230f {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.715291] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-96531a62-98f3-4a7c-9020-152fd796dc38 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1013.715476] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-96531a62-98f3-4a7c-9020-152fd796dc38 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63515) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1013.716197] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9688164f-ce58-488f-81a1-d08a262166c6 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.721290] env[63515]: DEBUG oslo_vmware.api [None req-96531a62-98f3-4a7c-9020-152fd796dc38 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Waiting for the task: (returnval){ [ 1013.721290] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]5282008d-dd48-8e98-8fbb-ef566df5e4b6" [ 1013.721290] env[63515]: _type = "Task" [ 1013.721290] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1013.728496] env[63515]: DEBUG oslo_vmware.api [None req-96531a62-98f3-4a7c-9020-152fd796dc38 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]5282008d-dd48-8e98-8fbb-ef566df5e4b6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.946165] env[63515]: DEBUG oslo_concurrency.lockutils [None req-642b03cc-2e02-4a86-bca6-75c3316e966c tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.243s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1013.948775] env[63515]: DEBUG oslo_concurrency.lockutils [None req-042daf1f-2fe3-4b1d-bb04-2f7a0479fcba tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.004s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1013.950365] env[63515]: INFO nova.compute.claims [None req-042daf1f-2fe3-4b1d-bb04-2f7a0479fcba tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: e925fc93-a731-498a-984c-9b1f3eabf353] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1013.955308] env[63515]: DEBUG oslo_concurrency.lockutils [None req-e978376f-06cf-4dbb-9f2d-8124b426de89 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Acquiring lock "refresh_cache-68766cd4-84be-475b-8494-d7ab43a9e969" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1013.955489] env[63515]: DEBUG oslo_concurrency.lockutils [None req-e978376f-06cf-4dbb-9f2d-8124b426de89 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Acquired lock "refresh_cache-68766cd4-84be-475b-8494-d7ab43a9e969" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1013.955676] env[63515]: DEBUG nova.network.neutron [None req-e978376f-06cf-4dbb-9f2d-8124b426de89 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 68766cd4-84be-475b-8494-d7ab43a9e969] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1013.955855] env[63515]: DEBUG nova.objects.instance [None req-e978376f-06cf-4dbb-9f2d-8124b426de89 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Lazy-loading 'info_cache' on Instance uuid 68766cd4-84be-475b-8494-d7ab43a9e969 {{(pid=63515) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1014.001313] env[63515]: DEBUG oslo_vmware.api [None req-b36e84df-7edf-4fd3-971f-5ed772c5647a tempest-ServerRescueTestJSONUnderV235-800430211 tempest-ServerRescueTestJSONUnderV235-800430211-project-member] Task: {'id': task-1111812, 'name': Rename_Task, 'duration_secs': 0.149375} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1014.001668] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-b36e84df-7edf-4fd3-971f-5ed772c5647a tempest-ServerRescueTestJSONUnderV235-800430211 tempest-ServerRescueTestJSONUnderV235-800430211-project-member] [instance: fcfe3376-8dfa-4189-a267-ff6402713c1b] Powering on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1014.001914] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5f9a2648-1932-4eac-9e86-9cb758283b66 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.008090] env[63515]: DEBUG oslo_vmware.api [None req-b36e84df-7edf-4fd3-971f-5ed772c5647a tempest-ServerRescueTestJSONUnderV235-800430211 tempest-ServerRescueTestJSONUnderV235-800430211-project-member] Waiting for the task: (returnval){ [ 1014.008090] env[63515]: value = "task-1111813" [ 1014.008090] env[63515]: _type = "Task" [ 1014.008090] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1014.016116] env[63515]: DEBUG oslo_vmware.api [None req-b36e84df-7edf-4fd3-971f-5ed772c5647a tempest-ServerRescueTestJSONUnderV235-800430211 tempest-ServerRescueTestJSONUnderV235-800430211-project-member] Task: {'id': task-1111813, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.021673] env[63515]: DEBUG oslo_concurrency.lockutils [req-7d016afb-fe94-4d74-936a-13487b6578e0 req-3c7a6e8f-6682-47af-aabc-f3d20b319f2f service nova] Releasing lock "refresh_cache-b76cf3dd-7b7e-4e5d-bf3c-e15bd37069ac" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1014.233419] env[63515]: DEBUG oslo_vmware.api [None req-96531a62-98f3-4a7c-9020-152fd796dc38 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]5282008d-dd48-8e98-8fbb-ef566df5e4b6, 'name': SearchDatastore_Task, 'duration_secs': 0.00794} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1014.234329] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-09159e3a-c8df-45f8-977c-b0caa217dbaa {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.239562] env[63515]: DEBUG oslo_vmware.api [None req-96531a62-98f3-4a7c-9020-152fd796dc38 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Waiting for the task: (returnval){ [ 1014.239562] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]52442114-96d6-7cd1-4fc4-1d8c888a1b03" [ 1014.239562] env[63515]: _type = "Task" [ 1014.239562] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1014.247177] env[63515]: DEBUG oslo_vmware.api [None req-96531a62-98f3-4a7c-9020-152fd796dc38 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52442114-96d6-7cd1-4fc4-1d8c888a1b03, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.459677] env[63515]: DEBUG nova.objects.base [None req-e978376f-06cf-4dbb-9f2d-8124b426de89 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Object Instance<68766cd4-84be-475b-8494-d7ab43a9e969> lazy-loaded attributes: flavor,info_cache {{(pid=63515) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1014.463367] env[63515]: DEBUG oslo_concurrency.lockutils [None req-642b03cc-2e02-4a86-bca6-75c3316e966c tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Lock "7d572ac0-f6af-4622-96cc-e75983420222" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 23.260s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1014.464311] env[63515]: DEBUG oslo_concurrency.lockutils [None req-878f8fb3-6690-40c9-9f6f-d6953e3612e1 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Lock "7d572ac0-f6af-4622-96cc-e75983420222" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 1.461s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1014.464550] env[63515]: INFO nova.compute.manager [None req-878f8fb3-6690-40c9-9f6f-d6953e3612e1 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] [instance: 7d572ac0-f6af-4622-96cc-e75983420222] Unshelving [ 1014.518726] env[63515]: DEBUG oslo_vmware.api [None req-b36e84df-7edf-4fd3-971f-5ed772c5647a tempest-ServerRescueTestJSONUnderV235-800430211 tempest-ServerRescueTestJSONUnderV235-800430211-project-member] Task: {'id': task-1111813, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.750959] env[63515]: DEBUG oslo_vmware.api [None req-96531a62-98f3-4a7c-9020-152fd796dc38 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52442114-96d6-7cd1-4fc4-1d8c888a1b03, 'name': SearchDatastore_Task, 'duration_secs': 0.03667} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1014.751283] env[63515]: DEBUG oslo_concurrency.lockutils [None req-96531a62-98f3-4a7c-9020-152fd796dc38 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1014.751607] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-96531a62-98f3-4a7c-9020-152fd796dc38 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk to [datastore1] b76cf3dd-7b7e-4e5d-bf3c-e15bd37069ac/b76cf3dd-7b7e-4e5d-bf3c-e15bd37069ac.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1014.751952] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-8bc0cba0-5c20-4483-a89c-6371001f908e {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.758659] env[63515]: DEBUG oslo_vmware.api [None req-96531a62-98f3-4a7c-9020-152fd796dc38 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Waiting for the task: (returnval){ [ 1014.758659] env[63515]: value = "task-1111814" [ 1014.758659] env[63515]: _type = "Task" [ 1014.758659] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1014.767021] env[63515]: DEBUG oslo_vmware.api [None req-96531a62-98f3-4a7c-9020-152fd796dc38 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Task: {'id': task-1111814, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.023694] env[63515]: DEBUG oslo_vmware.api [None req-b36e84df-7edf-4fd3-971f-5ed772c5647a tempest-ServerRescueTestJSONUnderV235-800430211 tempest-ServerRescueTestJSONUnderV235-800430211-project-member] Task: {'id': task-1111813, 'name': PowerOnVM_Task, 'duration_secs': 0.737598} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1015.024171] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-b36e84df-7edf-4fd3-971f-5ed772c5647a tempest-ServerRescueTestJSONUnderV235-800430211 tempest-ServerRescueTestJSONUnderV235-800430211-project-member] [instance: fcfe3376-8dfa-4189-a267-ff6402713c1b] Powered on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1015.024520] env[63515]: INFO nova.compute.manager [None req-b36e84df-7edf-4fd3-971f-5ed772c5647a tempest-ServerRescueTestJSONUnderV235-800430211 tempest-ServerRescueTestJSONUnderV235-800430211-project-member] [instance: fcfe3376-8dfa-4189-a267-ff6402713c1b] Took 8.16 seconds to spawn the instance on the hypervisor. [ 1015.024847] env[63515]: DEBUG nova.compute.manager [None req-b36e84df-7edf-4fd3-971f-5ed772c5647a tempest-ServerRescueTestJSONUnderV235-800430211 tempest-ServerRescueTestJSONUnderV235-800430211-project-member] [instance: fcfe3376-8dfa-4189-a267-ff6402713c1b] Checking state {{(pid=63515) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1015.026146] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0b3f95d-5155-4378-827a-da295e5ad698 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.122485] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db8c5ced-3b20-4008-8a5b-e159430cd59a {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.130974] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8aada19-8826-4bbc-ac8b-63f0503af8f0 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.165863] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-117a6cd8-1d6a-42e5-894a-62c2b8cfc62d {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.174264] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6269e810-dc91-4c03-89bc-47217975c87a {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.188038] env[63515]: DEBUG nova.compute.provider_tree [None req-042daf1f-2fe3-4b1d-bb04-2f7a0479fcba tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1015.268534] env[63515]: DEBUG oslo_vmware.api [None req-96531a62-98f3-4a7c-9020-152fd796dc38 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Task: {'id': task-1111814, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.46322} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1015.268791] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-96531a62-98f3-4a7c-9020-152fd796dc38 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk to [datastore1] b76cf3dd-7b7e-4e5d-bf3c-e15bd37069ac/b76cf3dd-7b7e-4e5d-bf3c-e15bd37069ac.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1015.269011] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-96531a62-98f3-4a7c-9020-152fd796dc38 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: b76cf3dd-7b7e-4e5d-bf3c-e15bd37069ac] Extending root virtual disk to 1048576 {{(pid=63515) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1015.269303] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-8c51fc96-d369-4f20-8d6c-971b7abdf5a6 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.275265] env[63515]: DEBUG oslo_vmware.api [None req-96531a62-98f3-4a7c-9020-152fd796dc38 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Waiting for the task: (returnval){ [ 1015.275265] env[63515]: value = "task-1111815" [ 1015.275265] env[63515]: _type = "Task" [ 1015.275265] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1015.279257] env[63515]: DEBUG nova.network.neutron [None req-e978376f-06cf-4dbb-9f2d-8124b426de89 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 68766cd4-84be-475b-8494-d7ab43a9e969] Updating instance_info_cache with network_info: [{"id": "28b59076-140f-46b5-b4a9-0bf530a1d43d", "address": "fa:16:3e:a3:55:0c", "network": {"id": "5758c885-e1c9-45c8-86aa-27a802101f52", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-734458138-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.204", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b18d4daacbc84e758f9f9ae4f3f09d28", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69351262-8d39-441a-85ba-3a78df436d17", "external-id": "nsx-vlan-transportzone-205", "segmentation_id": 205, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap28b59076-14", "ovs_interfaceid": "28b59076-140f-46b5-b4a9-0bf530a1d43d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1015.285709] env[63515]: DEBUG oslo_vmware.api [None req-96531a62-98f3-4a7c-9020-152fd796dc38 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Task: {'id': task-1111815, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.486059] env[63515]: DEBUG oslo_concurrency.lockutils [None req-878f8fb3-6690-40c9-9f6f-d6953e3612e1 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1015.551232] env[63515]: INFO nova.compute.manager [None req-b36e84df-7edf-4fd3-971f-5ed772c5647a tempest-ServerRescueTestJSONUnderV235-800430211 tempest-ServerRescueTestJSONUnderV235-800430211-project-member] [instance: fcfe3376-8dfa-4189-a267-ff6402713c1b] Took 17.49 seconds to build instance. [ 1015.691642] env[63515]: DEBUG nova.scheduler.client.report [None req-042daf1f-2fe3-4b1d-bb04-2f7a0479fcba tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1015.782187] env[63515]: DEBUG oslo_concurrency.lockutils [None req-e978376f-06cf-4dbb-9f2d-8124b426de89 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Releasing lock "refresh_cache-68766cd4-84be-475b-8494-d7ab43a9e969" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1015.786256] env[63515]: DEBUG oslo_vmware.api [None req-96531a62-98f3-4a7c-9020-152fd796dc38 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Task: {'id': task-1111815, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.062705} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1015.786691] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-96531a62-98f3-4a7c-9020-152fd796dc38 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: b76cf3dd-7b7e-4e5d-bf3c-e15bd37069ac] Extended root virtual disk {{(pid=63515) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1015.787475] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d857123e-fd65-47de-86d1-356a719e18cb {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.809541] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-96531a62-98f3-4a7c-9020-152fd796dc38 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: b76cf3dd-7b7e-4e5d-bf3c-e15bd37069ac] Reconfiguring VM instance instance-00000064 to attach disk [datastore1] b76cf3dd-7b7e-4e5d-bf3c-e15bd37069ac/b76cf3dd-7b7e-4e5d-bf3c-e15bd37069ac.vmdk or device None with type sparse {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1015.809994] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-538cb413-8fdb-4fe9-803f-5747fb28149a {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.833134] env[63515]: DEBUG oslo_vmware.api [None req-96531a62-98f3-4a7c-9020-152fd796dc38 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Waiting for the task: (returnval){ [ 1015.833134] env[63515]: value = "task-1111816" [ 1015.833134] env[63515]: _type = "Task" [ 1015.833134] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1015.840790] env[63515]: DEBUG oslo_vmware.api [None req-96531a62-98f3-4a7c-9020-152fd796dc38 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Task: {'id': task-1111816, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1016.020325] env[63515]: INFO nova.compute.manager [None req-33c1b962-9710-4b72-a8de-b30d0a17bad1 tempest-ServerRescueTestJSONUnderV235-800430211 tempest-ServerRescueTestJSONUnderV235-800430211-project-member] [instance: fcfe3376-8dfa-4189-a267-ff6402713c1b] Rescuing [ 1016.020606] env[63515]: DEBUG oslo_concurrency.lockutils [None req-33c1b962-9710-4b72-a8de-b30d0a17bad1 tempest-ServerRescueTestJSONUnderV235-800430211 tempest-ServerRescueTestJSONUnderV235-800430211-project-member] Acquiring lock "refresh_cache-fcfe3376-8dfa-4189-a267-ff6402713c1b" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1016.020792] env[63515]: DEBUG oslo_concurrency.lockutils [None req-33c1b962-9710-4b72-a8de-b30d0a17bad1 tempest-ServerRescueTestJSONUnderV235-800430211 tempest-ServerRescueTestJSONUnderV235-800430211-project-member] Acquired lock "refresh_cache-fcfe3376-8dfa-4189-a267-ff6402713c1b" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1016.021017] env[63515]: DEBUG nova.network.neutron [None req-33c1b962-9710-4b72-a8de-b30d0a17bad1 tempest-ServerRescueTestJSONUnderV235-800430211 tempest-ServerRescueTestJSONUnderV235-800430211-project-member] [instance: fcfe3376-8dfa-4189-a267-ff6402713c1b] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1016.052763] env[63515]: DEBUG oslo_concurrency.lockutils [None req-b36e84df-7edf-4fd3-971f-5ed772c5647a tempest-ServerRescueTestJSONUnderV235-800430211 tempest-ServerRescueTestJSONUnderV235-800430211-project-member] Lock "fcfe3376-8dfa-4189-a267-ff6402713c1b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 19.004s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1016.196304] env[63515]: DEBUG oslo_concurrency.lockutils [None req-042daf1f-2fe3-4b1d-bb04-2f7a0479fcba tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.247s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1016.196975] env[63515]: DEBUG nova.compute.manager [None req-042daf1f-2fe3-4b1d-bb04-2f7a0479fcba tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: e925fc93-a731-498a-984c-9b1f3eabf353] Start building networks asynchronously for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1016.199629] env[63515]: DEBUG oslo_concurrency.lockutils [None req-878f8fb3-6690-40c9-9f6f-d6953e3612e1 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.714s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1016.199849] env[63515]: DEBUG nova.objects.instance [None req-878f8fb3-6690-40c9-9f6f-d6953e3612e1 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Lazy-loading 'pci_requests' on Instance uuid 7d572ac0-f6af-4622-96cc-e75983420222 {{(pid=63515) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1016.288917] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-e978376f-06cf-4dbb-9f2d-8124b426de89 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 68766cd4-84be-475b-8494-d7ab43a9e969] Powering on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1016.289108] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-587f4026-124d-452a-b267-8c3593b78dee {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.296609] env[63515]: DEBUG oslo_vmware.api [None req-e978376f-06cf-4dbb-9f2d-8124b426de89 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Waiting for the task: (returnval){ [ 1016.296609] env[63515]: value = "task-1111817" [ 1016.296609] env[63515]: _type = "Task" [ 1016.296609] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1016.304655] env[63515]: DEBUG oslo_vmware.api [None req-e978376f-06cf-4dbb-9f2d-8124b426de89 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Task: {'id': task-1111817, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1016.343132] env[63515]: DEBUG oslo_vmware.api [None req-96531a62-98f3-4a7c-9020-152fd796dc38 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Task: {'id': task-1111816, 'name': ReconfigVM_Task, 'duration_secs': 0.293071} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1016.343419] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-96531a62-98f3-4a7c-9020-152fd796dc38 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: b76cf3dd-7b7e-4e5d-bf3c-e15bd37069ac] Reconfigured VM instance instance-00000064 to attach disk [datastore1] b76cf3dd-7b7e-4e5d-bf3c-e15bd37069ac/b76cf3dd-7b7e-4e5d-bf3c-e15bd37069ac.vmdk or device None with type sparse {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1016.344049] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ddc8c7ae-c4c7-4386-9268-da1ee9e0dae3 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.350211] env[63515]: DEBUG oslo_vmware.api [None req-96531a62-98f3-4a7c-9020-152fd796dc38 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Waiting for the task: (returnval){ [ 1016.350211] env[63515]: value = "task-1111818" [ 1016.350211] env[63515]: _type = "Task" [ 1016.350211] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1016.358575] env[63515]: DEBUG oslo_vmware.api [None req-96531a62-98f3-4a7c-9020-152fd796dc38 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Task: {'id': task-1111818, 'name': Rename_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1016.701790] env[63515]: DEBUG nova.network.neutron [None req-33c1b962-9710-4b72-a8de-b30d0a17bad1 tempest-ServerRescueTestJSONUnderV235-800430211 tempest-ServerRescueTestJSONUnderV235-800430211-project-member] [instance: fcfe3376-8dfa-4189-a267-ff6402713c1b] Updating instance_info_cache with network_info: [{"id": "4462b1aa-72ea-41cc-878a-68c38920c8cc", "address": "fa:16:3e:4f:bd:04", "network": {"id": "9fc2b9c9-6afc-4510-bf13-f87f1e1bd53b", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-513162532-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "0be5ebcf766b4da5b33be3a12387e737", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0248a27a-1d7f-4195-987b-06bfc8467347", "external-id": "nsx-vlan-transportzone-26", "segmentation_id": 26, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4462b1aa-72", "ovs_interfaceid": "4462b1aa-72ea-41cc-878a-68c38920c8cc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1016.703882] env[63515]: DEBUG nova.compute.utils [None req-042daf1f-2fe3-4b1d-bb04-2f7a0479fcba tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Using /dev/sd instead of None {{(pid=63515) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1016.706996] env[63515]: DEBUG nova.objects.instance [None req-878f8fb3-6690-40c9-9f6f-d6953e3612e1 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Lazy-loading 'numa_topology' on Instance uuid 7d572ac0-f6af-4622-96cc-e75983420222 {{(pid=63515) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1016.707935] env[63515]: DEBUG nova.compute.manager [None req-042daf1f-2fe3-4b1d-bb04-2f7a0479fcba tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: e925fc93-a731-498a-984c-9b1f3eabf353] Allocating IP information in the background. {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1016.708108] env[63515]: DEBUG nova.network.neutron [None req-042daf1f-2fe3-4b1d-bb04-2f7a0479fcba tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: e925fc93-a731-498a-984c-9b1f3eabf353] allocate_for_instance() {{(pid=63515) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1016.747342] env[63515]: DEBUG nova.policy [None req-042daf1f-2fe3-4b1d-bb04-2f7a0479fcba tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9b19bb2d32d84c019541c3b2e711a202', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '299fdeff647f486390366d5bbf911518', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63515) authorize /opt/stack/nova/nova/policy.py:201}} [ 1016.807178] env[63515]: DEBUG oslo_vmware.api [None req-e978376f-06cf-4dbb-9f2d-8124b426de89 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Task: {'id': task-1111817, 'name': PowerOnVM_Task, 'duration_secs': 0.366718} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1016.807524] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-e978376f-06cf-4dbb-9f2d-8124b426de89 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 68766cd4-84be-475b-8494-d7ab43a9e969] Powered on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1016.807760] env[63515]: DEBUG nova.compute.manager [None req-e978376f-06cf-4dbb-9f2d-8124b426de89 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 68766cd4-84be-475b-8494-d7ab43a9e969] Checking state {{(pid=63515) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1016.808603] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3849d7cf-c78b-420c-a5a5-1673f66566f0 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.861269] env[63515]: DEBUG oslo_vmware.api [None req-96531a62-98f3-4a7c-9020-152fd796dc38 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Task: {'id': task-1111818, 'name': Rename_Task, 'duration_secs': 0.166801} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1016.861599] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-96531a62-98f3-4a7c-9020-152fd796dc38 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: b76cf3dd-7b7e-4e5d-bf3c-e15bd37069ac] Powering on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1016.861868] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-01500e93-df96-4057-bed9-64c0521b3eb1 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.868225] env[63515]: DEBUG oslo_vmware.api [None req-96531a62-98f3-4a7c-9020-152fd796dc38 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Waiting for the task: (returnval){ [ 1016.868225] env[63515]: value = "task-1111819" [ 1016.868225] env[63515]: _type = "Task" [ 1016.868225] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1016.876255] env[63515]: DEBUG oslo_vmware.api [None req-96531a62-98f3-4a7c-9020-152fd796dc38 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Task: {'id': task-1111819, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.194222] env[63515]: DEBUG nova.network.neutron [None req-042daf1f-2fe3-4b1d-bb04-2f7a0479fcba tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: e925fc93-a731-498a-984c-9b1f3eabf353] Successfully created port: 80493aae-217e-4846-84b0-b60116b4afda {{(pid=63515) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1017.206273] env[63515]: DEBUG nova.compute.manager [None req-042daf1f-2fe3-4b1d-bb04-2f7a0479fcba tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: e925fc93-a731-498a-984c-9b1f3eabf353] Start building block device mappings for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1017.209098] env[63515]: INFO nova.compute.claims [None req-878f8fb3-6690-40c9-9f6f-d6953e3612e1 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] [instance: 7d572ac0-f6af-4622-96cc-e75983420222] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1017.211475] env[63515]: DEBUG oslo_concurrency.lockutils [None req-33c1b962-9710-4b72-a8de-b30d0a17bad1 tempest-ServerRescueTestJSONUnderV235-800430211 tempest-ServerRescueTestJSONUnderV235-800430211-project-member] Releasing lock "refresh_cache-fcfe3376-8dfa-4189-a267-ff6402713c1b" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1017.379988] env[63515]: DEBUG oslo_vmware.api [None req-96531a62-98f3-4a7c-9020-152fd796dc38 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Task: {'id': task-1111819, 'name': PowerOnVM_Task, 'duration_secs': 0.486449} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1017.380282] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-96531a62-98f3-4a7c-9020-152fd796dc38 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: b76cf3dd-7b7e-4e5d-bf3c-e15bd37069ac] Powered on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1017.380662] env[63515]: INFO nova.compute.manager [None req-96531a62-98f3-4a7c-9020-152fd796dc38 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: b76cf3dd-7b7e-4e5d-bf3c-e15bd37069ac] Took 8.23 seconds to spawn the instance on the hypervisor. [ 1017.380852] env[63515]: DEBUG nova.compute.manager [None req-96531a62-98f3-4a7c-9020-152fd796dc38 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: b76cf3dd-7b7e-4e5d-bf3c-e15bd37069ac] Checking state {{(pid=63515) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1017.381635] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cfb22a47-d5f6-4e05-ab12-a9d1cae10dab {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.756562] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-33c1b962-9710-4b72-a8de-b30d0a17bad1 tempest-ServerRescueTestJSONUnderV235-800430211 tempest-ServerRescueTestJSONUnderV235-800430211-project-member] [instance: fcfe3376-8dfa-4189-a267-ff6402713c1b] Powering off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1017.756940] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b7c4dbb4-c612-453c-b991-e9798f34fb7f {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.765352] env[63515]: DEBUG oslo_vmware.api [None req-33c1b962-9710-4b72-a8de-b30d0a17bad1 tempest-ServerRescueTestJSONUnderV235-800430211 tempest-ServerRescueTestJSONUnderV235-800430211-project-member] Waiting for the task: (returnval){ [ 1017.765352] env[63515]: value = "task-1111820" [ 1017.765352] env[63515]: _type = "Task" [ 1017.765352] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1017.778592] env[63515]: DEBUG oslo_vmware.api [None req-33c1b962-9710-4b72-a8de-b30d0a17bad1 tempest-ServerRescueTestJSONUnderV235-800430211 tempest-ServerRescueTestJSONUnderV235-800430211-project-member] Task: {'id': task-1111820, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.898919] env[63515]: INFO nova.compute.manager [None req-96531a62-98f3-4a7c-9020-152fd796dc38 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: b76cf3dd-7b7e-4e5d-bf3c-e15bd37069ac] Took 18.70 seconds to build instance. [ 1018.141799] env[63515]: DEBUG oslo_concurrency.lockutils [None req-e8c24bdc-d1c1-4d98-8d1d-53bd436f7627 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Acquiring lock "68766cd4-84be-475b-8494-d7ab43a9e969" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1018.142176] env[63515]: DEBUG oslo_concurrency.lockutils [None req-e8c24bdc-d1c1-4d98-8d1d-53bd436f7627 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Lock "68766cd4-84be-475b-8494-d7ab43a9e969" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1018.142452] env[63515]: DEBUG oslo_concurrency.lockutils [None req-e8c24bdc-d1c1-4d98-8d1d-53bd436f7627 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Acquiring lock "68766cd4-84be-475b-8494-d7ab43a9e969-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1018.142693] env[63515]: DEBUG oslo_concurrency.lockutils [None req-e8c24bdc-d1c1-4d98-8d1d-53bd436f7627 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Lock "68766cd4-84be-475b-8494-d7ab43a9e969-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1018.142918] env[63515]: DEBUG oslo_concurrency.lockutils [None req-e8c24bdc-d1c1-4d98-8d1d-53bd436f7627 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Lock "68766cd4-84be-475b-8494-d7ab43a9e969-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1018.145437] env[63515]: INFO nova.compute.manager [None req-e8c24bdc-d1c1-4d98-8d1d-53bd436f7627 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 68766cd4-84be-475b-8494-d7ab43a9e969] Terminating instance [ 1018.147993] env[63515]: DEBUG nova.compute.manager [None req-e8c24bdc-d1c1-4d98-8d1d-53bd436f7627 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 68766cd4-84be-475b-8494-d7ab43a9e969] Start destroying the instance on the hypervisor. {{(pid=63515) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1018.148212] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-e8c24bdc-d1c1-4d98-8d1d-53bd436f7627 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 68766cd4-84be-475b-8494-d7ab43a9e969] Destroying instance {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1018.149085] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1f03f8b-a581-4b3b-bacd-2eaafec05789 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.157432] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-e8c24bdc-d1c1-4d98-8d1d-53bd436f7627 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 68766cd4-84be-475b-8494-d7ab43a9e969] Powering off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1018.157664] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e442279f-05c1-45b1-b86f-736c876975f6 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.163596] env[63515]: DEBUG oslo_vmware.api [None req-e8c24bdc-d1c1-4d98-8d1d-53bd436f7627 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Waiting for the task: (returnval){ [ 1018.163596] env[63515]: value = "task-1111821" [ 1018.163596] env[63515]: _type = "Task" [ 1018.163596] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1018.171940] env[63515]: DEBUG oslo_vmware.api [None req-e8c24bdc-d1c1-4d98-8d1d-53bd436f7627 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Task: {'id': task-1111821, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.224908] env[63515]: DEBUG nova.compute.manager [None req-042daf1f-2fe3-4b1d-bb04-2f7a0479fcba tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: e925fc93-a731-498a-984c-9b1f3eabf353] Start spawning the instance on the hypervisor. {{(pid=63515) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1018.256929] env[63515]: DEBUG nova.virt.hardware [None req-042daf1f-2fe3-4b1d-bb04-2f7a0479fcba tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T02:52:14Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T02:51:56Z,direct_url=,disk_format='vmdk',id=8a120570-cb06-4099-b262-554ca0ad15c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b569255cc43e42e7ae3f2b4ad37c6ef4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T02:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1018.257192] env[63515]: DEBUG nova.virt.hardware [None req-042daf1f-2fe3-4b1d-bb04-2f7a0479fcba tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Flavor limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1018.257358] env[63515]: DEBUG nova.virt.hardware [None req-042daf1f-2fe3-4b1d-bb04-2f7a0479fcba tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Image limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1018.257542] env[63515]: DEBUG nova.virt.hardware [None req-042daf1f-2fe3-4b1d-bb04-2f7a0479fcba tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Flavor pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1018.257690] env[63515]: DEBUG nova.virt.hardware [None req-042daf1f-2fe3-4b1d-bb04-2f7a0479fcba tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Image pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1018.257840] env[63515]: DEBUG nova.virt.hardware [None req-042daf1f-2fe3-4b1d-bb04-2f7a0479fcba tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1018.258058] env[63515]: DEBUG nova.virt.hardware [None req-042daf1f-2fe3-4b1d-bb04-2f7a0479fcba tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1018.258224] env[63515]: DEBUG nova.virt.hardware [None req-042daf1f-2fe3-4b1d-bb04-2f7a0479fcba tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1018.258392] env[63515]: DEBUG nova.virt.hardware [None req-042daf1f-2fe3-4b1d-bb04-2f7a0479fcba tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Got 1 possible topologies {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1018.258574] env[63515]: DEBUG nova.virt.hardware [None req-042daf1f-2fe3-4b1d-bb04-2f7a0479fcba tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1018.258729] env[63515]: DEBUG nova.virt.hardware [None req-042daf1f-2fe3-4b1d-bb04-2f7a0479fcba tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1018.259672] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2e257aa-8601-4614-898d-b40573165663 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.272252] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c982279-0fde-44cd-9c78-f8a7cbdc2173 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.280740] env[63515]: DEBUG oslo_vmware.api [None req-33c1b962-9710-4b72-a8de-b30d0a17bad1 tempest-ServerRescueTestJSONUnderV235-800430211 tempest-ServerRescueTestJSONUnderV235-800430211-project-member] Task: {'id': task-1111820, 'name': PowerOffVM_Task, 'duration_secs': 0.24279} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1018.288098] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-33c1b962-9710-4b72-a8de-b30d0a17bad1 tempest-ServerRescueTestJSONUnderV235-800430211 tempest-ServerRescueTestJSONUnderV235-800430211-project-member] [instance: fcfe3376-8dfa-4189-a267-ff6402713c1b] Powered off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1018.291048] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0686edb-2559-43b0-991b-a5260445d174 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.312320] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fde06111-8c5d-4f42-8fe2-2667a519b932 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.339859] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-33c1b962-9710-4b72-a8de-b30d0a17bad1 tempest-ServerRescueTestJSONUnderV235-800430211 tempest-ServerRescueTestJSONUnderV235-800430211-project-member] [instance: fcfe3376-8dfa-4189-a267-ff6402713c1b] Powering off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1018.339859] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-33723fc2-ea17-4bee-a87e-1d5c6d19e594 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.345384] env[63515]: DEBUG oslo_vmware.api [None req-33c1b962-9710-4b72-a8de-b30d0a17bad1 tempest-ServerRescueTestJSONUnderV235-800430211 tempest-ServerRescueTestJSONUnderV235-800430211-project-member] Waiting for the task: (returnval){ [ 1018.345384] env[63515]: value = "task-1111822" [ 1018.345384] env[63515]: _type = "Task" [ 1018.345384] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1018.355908] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-33c1b962-9710-4b72-a8de-b30d0a17bad1 tempest-ServerRescueTestJSONUnderV235-800430211 tempest-ServerRescueTestJSONUnderV235-800430211-project-member] [instance: fcfe3376-8dfa-4189-a267-ff6402713c1b] VM already powered off {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1018.356103] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-33c1b962-9710-4b72-a8de-b30d0a17bad1 tempest-ServerRescueTestJSONUnderV235-800430211 tempest-ServerRescueTestJSONUnderV235-800430211-project-member] [instance: fcfe3376-8dfa-4189-a267-ff6402713c1b] Processing image 8a120570-cb06-4099-b262-554ca0ad15c5 {{(pid=63515) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1018.356329] env[63515]: DEBUG oslo_concurrency.lockutils [None req-33c1b962-9710-4b72-a8de-b30d0a17bad1 tempest-ServerRescueTestJSONUnderV235-800430211 tempest-ServerRescueTestJSONUnderV235-800430211-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1018.356486] env[63515]: DEBUG oslo_concurrency.lockutils [None req-33c1b962-9710-4b72-a8de-b30d0a17bad1 tempest-ServerRescueTestJSONUnderV235-800430211 tempest-ServerRescueTestJSONUnderV235-800430211-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1018.356665] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-33c1b962-9710-4b72-a8de-b30d0a17bad1 tempest-ServerRescueTestJSONUnderV235-800430211 tempest-ServerRescueTestJSONUnderV235-800430211-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1018.358856] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-3f93d268-9807-42c2-88df-76eb9442f0f3 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.367022] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-33c1b962-9710-4b72-a8de-b30d0a17bad1 tempest-ServerRescueTestJSONUnderV235-800430211 tempest-ServerRescueTestJSONUnderV235-800430211-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1018.367022] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-33c1b962-9710-4b72-a8de-b30d0a17bad1 tempest-ServerRescueTestJSONUnderV235-800430211 tempest-ServerRescueTestJSONUnderV235-800430211-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63515) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1018.367502] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a4d891b9-9c16-4826-8d02-027c32914809 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.373976] env[63515]: DEBUG oslo_vmware.api [None req-33c1b962-9710-4b72-a8de-b30d0a17bad1 tempest-ServerRescueTestJSONUnderV235-800430211 tempest-ServerRescueTestJSONUnderV235-800430211-project-member] Waiting for the task: (returnval){ [ 1018.373976] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]52bc79bf-4994-20ee-06ba-706a0d82a2e0" [ 1018.373976] env[63515]: _type = "Task" [ 1018.373976] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1018.381272] env[63515]: DEBUG oslo_vmware.api [None req-33c1b962-9710-4b72-a8de-b30d0a17bad1 tempest-ServerRescueTestJSONUnderV235-800430211 tempest-ServerRescueTestJSONUnderV235-800430211-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52bc79bf-4994-20ee-06ba-706a0d82a2e0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.393980] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4da72217-368f-493e-ab4d-138a82596a7c {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.400912] env[63515]: DEBUG oslo_concurrency.lockutils [None req-96531a62-98f3-4a7c-9020-152fd796dc38 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Lock "b76cf3dd-7b7e-4e5d-bf3c-e15bd37069ac" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 20.217s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1018.402218] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b985803e-34a3-46ef-b95d-e8d982537cdd {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.432873] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a057dbb0-801c-4b91-b526-3dd5eb7d1a95 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.440099] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9531f8d2-c2d1-442b-9f9a-9cd9acc76321 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.453011] env[63515]: DEBUG nova.compute.provider_tree [None req-878f8fb3-6690-40c9-9f6f-d6953e3612e1 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1018.604042] env[63515]: DEBUG nova.compute.manager [req-07dcf618-ad31-403c-9cb4-ca4f00908293 req-64f01de0-1680-4e09-af02-e3d04cf66dd0 service nova] [instance: e925fc93-a731-498a-984c-9b1f3eabf353] Received event network-vif-plugged-80493aae-217e-4846-84b0-b60116b4afda {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1018.604306] env[63515]: DEBUG oslo_concurrency.lockutils [req-07dcf618-ad31-403c-9cb4-ca4f00908293 req-64f01de0-1680-4e09-af02-e3d04cf66dd0 service nova] Acquiring lock "e925fc93-a731-498a-984c-9b1f3eabf353-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1018.604957] env[63515]: DEBUG oslo_concurrency.lockutils [req-07dcf618-ad31-403c-9cb4-ca4f00908293 req-64f01de0-1680-4e09-af02-e3d04cf66dd0 service nova] Lock "e925fc93-a731-498a-984c-9b1f3eabf353-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1018.606630] env[63515]: DEBUG oslo_concurrency.lockutils [req-07dcf618-ad31-403c-9cb4-ca4f00908293 req-64f01de0-1680-4e09-af02-e3d04cf66dd0 service nova] Lock "e925fc93-a731-498a-984c-9b1f3eabf353-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1018.606858] env[63515]: DEBUG nova.compute.manager [req-07dcf618-ad31-403c-9cb4-ca4f00908293 req-64f01de0-1680-4e09-af02-e3d04cf66dd0 service nova] [instance: e925fc93-a731-498a-984c-9b1f3eabf353] No waiting events found dispatching network-vif-plugged-80493aae-217e-4846-84b0-b60116b4afda {{(pid=63515) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1018.607081] env[63515]: WARNING nova.compute.manager [req-07dcf618-ad31-403c-9cb4-ca4f00908293 req-64f01de0-1680-4e09-af02-e3d04cf66dd0 service nova] [instance: e925fc93-a731-498a-984c-9b1f3eabf353] Received unexpected event network-vif-plugged-80493aae-217e-4846-84b0-b60116b4afda for instance with vm_state building and task_state spawning. [ 1018.678555] env[63515]: DEBUG oslo_vmware.api [None req-e8c24bdc-d1c1-4d98-8d1d-53bd436f7627 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Task: {'id': task-1111821, 'name': PowerOffVM_Task, 'duration_secs': 0.346467} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1018.678865] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-e8c24bdc-d1c1-4d98-8d1d-53bd436f7627 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 68766cd4-84be-475b-8494-d7ab43a9e969] Powered off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1018.679054] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-e8c24bdc-d1c1-4d98-8d1d-53bd436f7627 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 68766cd4-84be-475b-8494-d7ab43a9e969] Unregistering the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1018.679322] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-6df0b8bd-8472-4427-89bc-3882ca45f508 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.752955] env[63515]: DEBUG nova.network.neutron [None req-042daf1f-2fe3-4b1d-bb04-2f7a0479fcba tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: e925fc93-a731-498a-984c-9b1f3eabf353] Successfully updated port: 80493aae-217e-4846-84b0-b60116b4afda {{(pid=63515) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1018.758177] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-e8c24bdc-d1c1-4d98-8d1d-53bd436f7627 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 68766cd4-84be-475b-8494-d7ab43a9e969] Unregistered the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1018.758177] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-e8c24bdc-d1c1-4d98-8d1d-53bd436f7627 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 68766cd4-84be-475b-8494-d7ab43a9e969] Deleting contents of the VM from datastore datastore2 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1018.758177] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-e8c24bdc-d1c1-4d98-8d1d-53bd436f7627 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Deleting the datastore file [datastore2] 68766cd4-84be-475b-8494-d7ab43a9e969 {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1018.758576] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-12912383-ab88-47fa-9256-73ba56dffea1 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.765456] env[63515]: DEBUG oslo_vmware.api [None req-e8c24bdc-d1c1-4d98-8d1d-53bd436f7627 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Waiting for the task: (returnval){ [ 1018.765456] env[63515]: value = "task-1111824" [ 1018.765456] env[63515]: _type = "Task" [ 1018.765456] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1018.774446] env[63515]: DEBUG oslo_vmware.api [None req-e8c24bdc-d1c1-4d98-8d1d-53bd436f7627 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Task: {'id': task-1111824, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.884652] env[63515]: DEBUG oslo_vmware.api [None req-33c1b962-9710-4b72-a8de-b30d0a17bad1 tempest-ServerRescueTestJSONUnderV235-800430211 tempest-ServerRescueTestJSONUnderV235-800430211-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52bc79bf-4994-20ee-06ba-706a0d82a2e0, 'name': SearchDatastore_Task, 'duration_secs': 0.009097} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1018.885120] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-526fa876-3464-45e5-bbc3-50f58a4eb8a6 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.891393] env[63515]: DEBUG oslo_vmware.api [None req-33c1b962-9710-4b72-a8de-b30d0a17bad1 tempest-ServerRescueTestJSONUnderV235-800430211 tempest-ServerRescueTestJSONUnderV235-800430211-project-member] Waiting for the task: (returnval){ [ 1018.891393] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]526ad660-6e25-27fb-e3a6-bba475c0e514" [ 1018.891393] env[63515]: _type = "Task" [ 1018.891393] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1018.899463] env[63515]: DEBUG oslo_vmware.api [None req-33c1b962-9710-4b72-a8de-b30d0a17bad1 tempest-ServerRescueTestJSONUnderV235-800430211 tempest-ServerRescueTestJSONUnderV235-800430211-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]526ad660-6e25-27fb-e3a6-bba475c0e514, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.955901] env[63515]: DEBUG nova.scheduler.client.report [None req-878f8fb3-6690-40c9-9f6f-d6953e3612e1 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1019.257202] env[63515]: DEBUG oslo_concurrency.lockutils [None req-042daf1f-2fe3-4b1d-bb04-2f7a0479fcba tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Acquiring lock "refresh_cache-e925fc93-a731-498a-984c-9b1f3eabf353" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1019.257202] env[63515]: DEBUG oslo_concurrency.lockutils [None req-042daf1f-2fe3-4b1d-bb04-2f7a0479fcba tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Acquired lock "refresh_cache-e925fc93-a731-498a-984c-9b1f3eabf353" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1019.257202] env[63515]: DEBUG nova.network.neutron [None req-042daf1f-2fe3-4b1d-bb04-2f7a0479fcba tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: e925fc93-a731-498a-984c-9b1f3eabf353] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1019.277047] env[63515]: DEBUG oslo_vmware.api [None req-e8c24bdc-d1c1-4d98-8d1d-53bd436f7627 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Task: {'id': task-1111824, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.179996} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1019.277047] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-e8c24bdc-d1c1-4d98-8d1d-53bd436f7627 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Deleted the datastore file {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1019.277047] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-e8c24bdc-d1c1-4d98-8d1d-53bd436f7627 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 68766cd4-84be-475b-8494-d7ab43a9e969] Deleted contents of the VM from datastore datastore2 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1019.277047] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-e8c24bdc-d1c1-4d98-8d1d-53bd436f7627 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 68766cd4-84be-475b-8494-d7ab43a9e969] Instance destroyed {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1019.277249] env[63515]: INFO nova.compute.manager [None req-e8c24bdc-d1c1-4d98-8d1d-53bd436f7627 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 68766cd4-84be-475b-8494-d7ab43a9e969] Took 1.13 seconds to destroy the instance on the hypervisor. [ 1019.277528] env[63515]: DEBUG oslo.service.loopingcall [None req-e8c24bdc-d1c1-4d98-8d1d-53bd436f7627 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1019.278057] env[63515]: DEBUG nova.compute.manager [-] [instance: 68766cd4-84be-475b-8494-d7ab43a9e969] Deallocating network for instance {{(pid=63515) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1019.278057] env[63515]: DEBUG nova.network.neutron [-] [instance: 68766cd4-84be-475b-8494-d7ab43a9e969] deallocate_for_instance() {{(pid=63515) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1019.401066] env[63515]: DEBUG oslo_vmware.api [None req-33c1b962-9710-4b72-a8de-b30d0a17bad1 tempest-ServerRescueTestJSONUnderV235-800430211 tempest-ServerRescueTestJSONUnderV235-800430211-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]526ad660-6e25-27fb-e3a6-bba475c0e514, 'name': SearchDatastore_Task, 'duration_secs': 0.030076} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1019.401395] env[63515]: DEBUG oslo_concurrency.lockutils [None req-33c1b962-9710-4b72-a8de-b30d0a17bad1 tempest-ServerRescueTestJSONUnderV235-800430211 tempest-ServerRescueTestJSONUnderV235-800430211-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1019.401664] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-33c1b962-9710-4b72-a8de-b30d0a17bad1 tempest-ServerRescueTestJSONUnderV235-800430211 tempest-ServerRescueTestJSONUnderV235-800430211-project-member] Copying virtual disk from [datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk to [datastore1] fcfe3376-8dfa-4189-a267-ff6402713c1b/8a120570-cb06-4099-b262-554ca0ad15c5-rescue.vmdk. {{(pid=63515) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 1019.401934] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7fbaf2da-e0f3-441f-822d-8a44aefae0e3 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.408701] env[63515]: DEBUG oslo_vmware.api [None req-33c1b962-9710-4b72-a8de-b30d0a17bad1 tempest-ServerRescueTestJSONUnderV235-800430211 tempest-ServerRescueTestJSONUnderV235-800430211-project-member] Waiting for the task: (returnval){ [ 1019.408701] env[63515]: value = "task-1111825" [ 1019.408701] env[63515]: _type = "Task" [ 1019.408701] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1019.416510] env[63515]: DEBUG oslo_vmware.api [None req-33c1b962-9710-4b72-a8de-b30d0a17bad1 tempest-ServerRescueTestJSONUnderV235-800430211 tempest-ServerRescueTestJSONUnderV235-800430211-project-member] Task: {'id': task-1111825, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1019.461685] env[63515]: DEBUG oslo_concurrency.lockutils [None req-878f8fb3-6690-40c9-9f6f-d6953e3612e1 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.262s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1019.508785] env[63515]: INFO nova.network.neutron [None req-878f8fb3-6690-40c9-9f6f-d6953e3612e1 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] [instance: 7d572ac0-f6af-4622-96cc-e75983420222] Updating port dbfd09c2-9ec4-4821-8613-43c6d6cf4724 with attributes {'binding:host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 1019.791554] env[63515]: DEBUG nova.network.neutron [None req-042daf1f-2fe3-4b1d-bb04-2f7a0479fcba tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: e925fc93-a731-498a-984c-9b1f3eabf353] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1019.919336] env[63515]: DEBUG oslo_vmware.api [None req-33c1b962-9710-4b72-a8de-b30d0a17bad1 tempest-ServerRescueTestJSONUnderV235-800430211 tempest-ServerRescueTestJSONUnderV235-800430211-project-member] Task: {'id': task-1111825, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.481243} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1019.919794] env[63515]: INFO nova.virt.vmwareapi.ds_util [None req-33c1b962-9710-4b72-a8de-b30d0a17bad1 tempest-ServerRescueTestJSONUnderV235-800430211 tempest-ServerRescueTestJSONUnderV235-800430211-project-member] Copied virtual disk from [datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk to [datastore1] fcfe3376-8dfa-4189-a267-ff6402713c1b/8a120570-cb06-4099-b262-554ca0ad15c5-rescue.vmdk. [ 1019.921196] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63aee046-68fc-4b55-8a12-1b9745dc9d75 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.958941] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-33c1b962-9710-4b72-a8de-b30d0a17bad1 tempest-ServerRescueTestJSONUnderV235-800430211 tempest-ServerRescueTestJSONUnderV235-800430211-project-member] [instance: fcfe3376-8dfa-4189-a267-ff6402713c1b] Reconfiguring VM instance instance-00000063 to attach disk [datastore1] fcfe3376-8dfa-4189-a267-ff6402713c1b/8a120570-cb06-4099-b262-554ca0ad15c5-rescue.vmdk or device None with type thin {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1019.962691] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f63e3588-78e1-4ed0-8191-22faf5767c00 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.987022] env[63515]: DEBUG oslo_vmware.api [None req-33c1b962-9710-4b72-a8de-b30d0a17bad1 tempest-ServerRescueTestJSONUnderV235-800430211 tempest-ServerRescueTestJSONUnderV235-800430211-project-member] Waiting for the task: (returnval){ [ 1019.987022] env[63515]: value = "task-1111826" [ 1019.987022] env[63515]: _type = "Task" [ 1019.987022] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1019.997375] env[63515]: DEBUG oslo_vmware.api [None req-33c1b962-9710-4b72-a8de-b30d0a17bad1 tempest-ServerRescueTestJSONUnderV235-800430211 tempest-ServerRescueTestJSONUnderV235-800430211-project-member] Task: {'id': task-1111826, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.062271] env[63515]: DEBUG nova.network.neutron [None req-042daf1f-2fe3-4b1d-bb04-2f7a0479fcba tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: e925fc93-a731-498a-984c-9b1f3eabf353] Updating instance_info_cache with network_info: [{"id": "80493aae-217e-4846-84b0-b60116b4afda", "address": "fa:16:3e:dc:00:ff", "network": {"id": "581d448e-be1b-4d5b-8a6e-c27326aaffc5", "bridge": "br-int", "label": "tempest-ServersTestJSON-1769610073-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "299fdeff647f486390366d5bbf911518", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fb143ef7-8271-4a8a-a4aa-8eba9a89f6a1", "external-id": "nsx-vlan-transportzone-504", "segmentation_id": 504, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap80493aae-21", "ovs_interfaceid": "80493aae-217e-4846-84b0-b60116b4afda", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1020.185027] env[63515]: DEBUG nova.compute.manager [None req-a7fc067e-c3eb-4fcd-8441-645f050b9570 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: b76cf3dd-7b7e-4e5d-bf3c-e15bd37069ac] Stashing vm_state: active {{(pid=63515) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 1020.363048] env[63515]: DEBUG nova.network.neutron [-] [instance: 68766cd4-84be-475b-8494-d7ab43a9e969] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1020.495325] env[63515]: DEBUG oslo_vmware.api [None req-33c1b962-9710-4b72-a8de-b30d0a17bad1 tempest-ServerRescueTestJSONUnderV235-800430211 tempest-ServerRescueTestJSONUnderV235-800430211-project-member] Task: {'id': task-1111826, 'name': ReconfigVM_Task, 'duration_secs': 0.373784} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1020.495627] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-33c1b962-9710-4b72-a8de-b30d0a17bad1 tempest-ServerRescueTestJSONUnderV235-800430211 tempest-ServerRescueTestJSONUnderV235-800430211-project-member] [instance: fcfe3376-8dfa-4189-a267-ff6402713c1b] Reconfigured VM instance instance-00000063 to attach disk [datastore1] fcfe3376-8dfa-4189-a267-ff6402713c1b/8a120570-cb06-4099-b262-554ca0ad15c5-rescue.vmdk or device None with type thin {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1020.496471] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-922c30ea-84d8-4330-8d00-0fc65e1f5cf4 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.521518] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3340609b-9704-4f41-8930-ee5d646f5c1c {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.536093] env[63515]: DEBUG oslo_vmware.api [None req-33c1b962-9710-4b72-a8de-b30d0a17bad1 tempest-ServerRescueTestJSONUnderV235-800430211 tempest-ServerRescueTestJSONUnderV235-800430211-project-member] Waiting for the task: (returnval){ [ 1020.536093] env[63515]: value = "task-1111827" [ 1020.536093] env[63515]: _type = "Task" [ 1020.536093] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1020.543617] env[63515]: DEBUG oslo_vmware.api [None req-33c1b962-9710-4b72-a8de-b30d0a17bad1 tempest-ServerRescueTestJSONUnderV235-800430211 tempest-ServerRescueTestJSONUnderV235-800430211-project-member] Task: {'id': task-1111827, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.564811] env[63515]: DEBUG oslo_concurrency.lockutils [None req-042daf1f-2fe3-4b1d-bb04-2f7a0479fcba tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Releasing lock "refresh_cache-e925fc93-a731-498a-984c-9b1f3eabf353" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1020.566026] env[63515]: DEBUG nova.compute.manager [None req-042daf1f-2fe3-4b1d-bb04-2f7a0479fcba tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: e925fc93-a731-498a-984c-9b1f3eabf353] Instance network_info: |[{"id": "80493aae-217e-4846-84b0-b60116b4afda", "address": "fa:16:3e:dc:00:ff", "network": {"id": "581d448e-be1b-4d5b-8a6e-c27326aaffc5", "bridge": "br-int", "label": "tempest-ServersTestJSON-1769610073-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "299fdeff647f486390366d5bbf911518", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fb143ef7-8271-4a8a-a4aa-8eba9a89f6a1", "external-id": "nsx-vlan-transportzone-504", "segmentation_id": 504, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap80493aae-21", "ovs_interfaceid": "80493aae-217e-4846-84b0-b60116b4afda", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1020.566026] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-042daf1f-2fe3-4b1d-bb04-2f7a0479fcba tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: e925fc93-a731-498a-984c-9b1f3eabf353] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:dc:00:ff', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'fb143ef7-8271-4a8a-a4aa-8eba9a89f6a1', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '80493aae-217e-4846-84b0-b60116b4afda', 'vif_model': 'vmxnet3'}] {{(pid=63515) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1020.573172] env[63515]: DEBUG oslo.service.loopingcall [None req-042daf1f-2fe3-4b1d-bb04-2f7a0479fcba tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1020.573405] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e925fc93-a731-498a-984c-9b1f3eabf353] Creating VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1020.573635] env[63515]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-2e163969-67df-4291-8f36-b690a5bbf5a2 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.594046] env[63515]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1020.594046] env[63515]: value = "task-1111828" [ 1020.594046] env[63515]: _type = "Task" [ 1020.594046] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1020.602169] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1111828, 'name': CreateVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.640536] env[63515]: DEBUG nova.compute.manager [req-7389a178-4779-4258-8d29-0200e7e352e3 req-0e3ee008-b0e3-4d40-a9d1-21576f4f8153 service nova] [instance: e925fc93-a731-498a-984c-9b1f3eabf353] Received event network-changed-80493aae-217e-4846-84b0-b60116b4afda {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1020.640725] env[63515]: DEBUG nova.compute.manager [req-7389a178-4779-4258-8d29-0200e7e352e3 req-0e3ee008-b0e3-4d40-a9d1-21576f4f8153 service nova] [instance: e925fc93-a731-498a-984c-9b1f3eabf353] Refreshing instance network info cache due to event network-changed-80493aae-217e-4846-84b0-b60116b4afda. {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1020.640950] env[63515]: DEBUG oslo_concurrency.lockutils [req-7389a178-4779-4258-8d29-0200e7e352e3 req-0e3ee008-b0e3-4d40-a9d1-21576f4f8153 service nova] Acquiring lock "refresh_cache-e925fc93-a731-498a-984c-9b1f3eabf353" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1020.641123] env[63515]: DEBUG oslo_concurrency.lockutils [req-7389a178-4779-4258-8d29-0200e7e352e3 req-0e3ee008-b0e3-4d40-a9d1-21576f4f8153 service nova] Acquired lock "refresh_cache-e925fc93-a731-498a-984c-9b1f3eabf353" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1020.641290] env[63515]: DEBUG nova.network.neutron [req-7389a178-4779-4258-8d29-0200e7e352e3 req-0e3ee008-b0e3-4d40-a9d1-21576f4f8153 service nova] [instance: e925fc93-a731-498a-984c-9b1f3eabf353] Refreshing network info cache for port 80493aae-217e-4846-84b0-b60116b4afda {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1020.706511] env[63515]: DEBUG oslo_concurrency.lockutils [None req-a7fc067e-c3eb-4fcd-8441-645f050b9570 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1020.706825] env[63515]: DEBUG oslo_concurrency.lockutils [None req-a7fc067e-c3eb-4fcd-8441-645f050b9570 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1020.865839] env[63515]: INFO nova.compute.manager [-] [instance: 68766cd4-84be-475b-8494-d7ab43a9e969] Took 1.59 seconds to deallocate network for instance. [ 1020.934649] env[63515]: DEBUG oslo_concurrency.lockutils [None req-80fab3e3-466b-468e-b003-05f4a2e89df7 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Acquiring lock "interface-724b05bc-ce4b-4f99-ad14-f3abea5567f2-6a98da26-06f3-4eb5-aee8-02a448c34749" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1020.934917] env[63515]: DEBUG oslo_concurrency.lockutils [None req-80fab3e3-466b-468e-b003-05f4a2e89df7 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Lock "interface-724b05bc-ce4b-4f99-ad14-f3abea5567f2-6a98da26-06f3-4eb5-aee8-02a448c34749" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1020.935336] env[63515]: DEBUG nova.objects.instance [None req-80fab3e3-466b-468e-b003-05f4a2e89df7 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Lazy-loading 'flavor' on Instance uuid 724b05bc-ce4b-4f99-ad14-f3abea5567f2 {{(pid=63515) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1021.045994] env[63515]: DEBUG oslo_vmware.api [None req-33c1b962-9710-4b72-a8de-b30d0a17bad1 tempest-ServerRescueTestJSONUnderV235-800430211 tempest-ServerRescueTestJSONUnderV235-800430211-project-member] Task: {'id': task-1111827, 'name': ReconfigVM_Task, 'duration_secs': 0.152841} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1021.046307] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-33c1b962-9710-4b72-a8de-b30d0a17bad1 tempest-ServerRescueTestJSONUnderV235-800430211 tempest-ServerRescueTestJSONUnderV235-800430211-project-member] [instance: fcfe3376-8dfa-4189-a267-ff6402713c1b] Powering on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1021.046560] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-03d913fc-1f8d-4e9a-96e9-27cac5a5e508 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.052226] env[63515]: DEBUG oslo_vmware.api [None req-33c1b962-9710-4b72-a8de-b30d0a17bad1 tempest-ServerRescueTestJSONUnderV235-800430211 tempest-ServerRescueTestJSONUnderV235-800430211-project-member] Waiting for the task: (returnval){ [ 1021.052226] env[63515]: value = "task-1111829" [ 1021.052226] env[63515]: _type = "Task" [ 1021.052226] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1021.059456] env[63515]: DEBUG oslo_vmware.api [None req-33c1b962-9710-4b72-a8de-b30d0a17bad1 tempest-ServerRescueTestJSONUnderV235-800430211 tempest-ServerRescueTestJSONUnderV235-800430211-project-member] Task: {'id': task-1111829, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.104456] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1111828, 'name': CreateVM_Task} progress is 25%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.202130] env[63515]: DEBUG oslo_concurrency.lockutils [None req-878f8fb3-6690-40c9-9f6f-d6953e3612e1 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Acquiring lock "refresh_cache-7d572ac0-f6af-4622-96cc-e75983420222" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1021.202344] env[63515]: DEBUG oslo_concurrency.lockutils [None req-878f8fb3-6690-40c9-9f6f-d6953e3612e1 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Acquired lock "refresh_cache-7d572ac0-f6af-4622-96cc-e75983420222" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1021.202532] env[63515]: DEBUG nova.network.neutron [None req-878f8fb3-6690-40c9-9f6f-d6953e3612e1 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] [instance: 7d572ac0-f6af-4622-96cc-e75983420222] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1021.213407] env[63515]: INFO nova.compute.claims [None req-a7fc067e-c3eb-4fcd-8441-645f050b9570 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: b76cf3dd-7b7e-4e5d-bf3c-e15bd37069ac] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1021.372859] env[63515]: DEBUG oslo_concurrency.lockutils [None req-e8c24bdc-d1c1-4d98-8d1d-53bd436f7627 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1021.468560] env[63515]: DEBUG nova.network.neutron [req-7389a178-4779-4258-8d29-0200e7e352e3 req-0e3ee008-b0e3-4d40-a9d1-21576f4f8153 service nova] [instance: e925fc93-a731-498a-984c-9b1f3eabf353] Updated VIF entry in instance network info cache for port 80493aae-217e-4846-84b0-b60116b4afda. {{(pid=63515) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1021.468639] env[63515]: DEBUG nova.network.neutron [req-7389a178-4779-4258-8d29-0200e7e352e3 req-0e3ee008-b0e3-4d40-a9d1-21576f4f8153 service nova] [instance: e925fc93-a731-498a-984c-9b1f3eabf353] Updating instance_info_cache with network_info: [{"id": "80493aae-217e-4846-84b0-b60116b4afda", "address": "fa:16:3e:dc:00:ff", "network": {"id": "581d448e-be1b-4d5b-8a6e-c27326aaffc5", "bridge": "br-int", "label": "tempest-ServersTestJSON-1769610073-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "299fdeff647f486390366d5bbf911518", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fb143ef7-8271-4a8a-a4aa-8eba9a89f6a1", "external-id": "nsx-vlan-transportzone-504", "segmentation_id": 504, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap80493aae-21", "ovs_interfaceid": "80493aae-217e-4846-84b0-b60116b4afda", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1021.541482] env[63515]: DEBUG nova.objects.instance [None req-80fab3e3-466b-468e-b003-05f4a2e89df7 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Lazy-loading 'pci_requests' on Instance uuid 724b05bc-ce4b-4f99-ad14-f3abea5567f2 {{(pid=63515) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1021.561599] env[63515]: DEBUG oslo_vmware.api [None req-33c1b962-9710-4b72-a8de-b30d0a17bad1 tempest-ServerRescueTestJSONUnderV235-800430211 tempest-ServerRescueTestJSONUnderV235-800430211-project-member] Task: {'id': task-1111829, 'name': PowerOnVM_Task, 'duration_secs': 0.42479} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1021.561880] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-33c1b962-9710-4b72-a8de-b30d0a17bad1 tempest-ServerRescueTestJSONUnderV235-800430211 tempest-ServerRescueTestJSONUnderV235-800430211-project-member] [instance: fcfe3376-8dfa-4189-a267-ff6402713c1b] Powered on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1021.564480] env[63515]: DEBUG nova.compute.manager [None req-33c1b962-9710-4b72-a8de-b30d0a17bad1 tempest-ServerRescueTestJSONUnderV235-800430211 tempest-ServerRescueTestJSONUnderV235-800430211-project-member] [instance: fcfe3376-8dfa-4189-a267-ff6402713c1b] Checking state {{(pid=63515) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1021.565320] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9c014c9-50c5-4069-8f2a-8d3de52b6f88 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.604400] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1111828, 'name': CreateVM_Task, 'duration_secs': 0.919613} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1021.604544] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e925fc93-a731-498a-984c-9b1f3eabf353] Created VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1021.605212] env[63515]: DEBUG oslo_concurrency.lockutils [None req-042daf1f-2fe3-4b1d-bb04-2f7a0479fcba tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1021.605385] env[63515]: DEBUG oslo_concurrency.lockutils [None req-042daf1f-2fe3-4b1d-bb04-2f7a0479fcba tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1021.605711] env[63515]: DEBUG oslo_concurrency.lockutils [None req-042daf1f-2fe3-4b1d-bb04-2f7a0479fcba tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1021.605976] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-aa3e8742-64f8-4619-a893-ccfbb962aa6a {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.610230] env[63515]: DEBUG oslo_vmware.api [None req-042daf1f-2fe3-4b1d-bb04-2f7a0479fcba tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Waiting for the task: (returnval){ [ 1021.610230] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]52cf111e-6f5e-3357-186f-37a2531abe32" [ 1021.610230] env[63515]: _type = "Task" [ 1021.610230] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1021.618984] env[63515]: DEBUG oslo_vmware.api [None req-042daf1f-2fe3-4b1d-bb04-2f7a0479fcba tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52cf111e-6f5e-3357-186f-37a2531abe32, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.720971] env[63515]: INFO nova.compute.resource_tracker [None req-a7fc067e-c3eb-4fcd-8441-645f050b9570 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: b76cf3dd-7b7e-4e5d-bf3c-e15bd37069ac] Updating resource usage from migration ba725c49-c030-4294-a921-aba61789a9c4 [ 1021.839742] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-828163bb-625e-4b35-b49f-8f5dca3c9efa {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.849164] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5653a6d5-f7a9-4d68-b5f8-526b72ccdbf8 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.881647] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-800d7487-e222-4dfc-8870-c24b89fad3fb {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.888566] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4652e492-e162-441a-a841-04e6d78160ab {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.901496] env[63515]: DEBUG nova.compute.provider_tree [None req-a7fc067e-c3eb-4fcd-8441-645f050b9570 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1021.914114] env[63515]: DEBUG nova.network.neutron [None req-878f8fb3-6690-40c9-9f6f-d6953e3612e1 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] [instance: 7d572ac0-f6af-4622-96cc-e75983420222] Updating instance_info_cache with network_info: [{"id": "dbfd09c2-9ec4-4821-8613-43c6d6cf4724", "address": "fa:16:3e:6d:92:3d", "network": {"id": "09fb20ca-2b17-4d46-b915-6aff85b862d8", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1815529427-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.244", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "19ce3b05eec64475b685faad1fd801f6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ded18042-834c-4792-b3e8-b1c377446432", "external-id": "nsx-vlan-transportzone-293", "segmentation_id": 293, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdbfd09c2-9e", "ovs_interfaceid": "dbfd09c2-9ec4-4821-8613-43c6d6cf4724", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1021.973462] env[63515]: DEBUG oslo_concurrency.lockutils [req-7389a178-4779-4258-8d29-0200e7e352e3 req-0e3ee008-b0e3-4d40-a9d1-21576f4f8153 service nova] Releasing lock "refresh_cache-e925fc93-a731-498a-984c-9b1f3eabf353" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1021.973716] env[63515]: DEBUG nova.compute.manager [req-7389a178-4779-4258-8d29-0200e7e352e3 req-0e3ee008-b0e3-4d40-a9d1-21576f4f8153 service nova] [instance: 68766cd4-84be-475b-8494-d7ab43a9e969] Received event network-vif-deleted-28b59076-140f-46b5-b4a9-0bf530a1d43d {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1022.043021] env[63515]: DEBUG nova.objects.base [None req-80fab3e3-466b-468e-b003-05f4a2e89df7 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Object Instance<724b05bc-ce4b-4f99-ad14-f3abea5567f2> lazy-loaded attributes: flavor,pci_requests {{(pid=63515) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1022.043266] env[63515]: DEBUG nova.network.neutron [None req-80fab3e3-466b-468e-b003-05f4a2e89df7 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: 724b05bc-ce4b-4f99-ad14-f3abea5567f2] allocate_for_instance() {{(pid=63515) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1022.104177] env[63515]: DEBUG nova.policy [None req-80fab3e3-466b-468e-b003-05f4a2e89df7 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '16880abf1c5b4341800c94ada2c756c7', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7d4ada9ec35f42b19c6480a9101d21a4', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63515) authorize /opt/stack/nova/nova/policy.py:201}} [ 1022.119466] env[63515]: DEBUG oslo_vmware.api [None req-042daf1f-2fe3-4b1d-bb04-2f7a0479fcba tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52cf111e-6f5e-3357-186f-37a2531abe32, 'name': SearchDatastore_Task, 'duration_secs': 0.009932} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1022.119743] env[63515]: DEBUG oslo_concurrency.lockutils [None req-042daf1f-2fe3-4b1d-bb04-2f7a0479fcba tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1022.119976] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-042daf1f-2fe3-4b1d-bb04-2f7a0479fcba tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: e925fc93-a731-498a-984c-9b1f3eabf353] Processing image 8a120570-cb06-4099-b262-554ca0ad15c5 {{(pid=63515) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1022.120233] env[63515]: DEBUG oslo_concurrency.lockutils [None req-042daf1f-2fe3-4b1d-bb04-2f7a0479fcba tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1022.120717] env[63515]: DEBUG oslo_concurrency.lockutils [None req-042daf1f-2fe3-4b1d-bb04-2f7a0479fcba tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1022.120717] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-042daf1f-2fe3-4b1d-bb04-2f7a0479fcba tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1022.120798] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-8f2b6f4f-eac1-4453-9040-ac357854a396 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.129273] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-042daf1f-2fe3-4b1d-bb04-2f7a0479fcba tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1022.129518] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-042daf1f-2fe3-4b1d-bb04-2f7a0479fcba tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63515) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1022.130243] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-56e7e2c2-2291-4d0c-ae3f-d7304ed4224d {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.135471] env[63515]: DEBUG oslo_vmware.api [None req-042daf1f-2fe3-4b1d-bb04-2f7a0479fcba tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Waiting for the task: (returnval){ [ 1022.135471] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]525678eb-7cb9-c755-ce97-6ba4dd85da0b" [ 1022.135471] env[63515]: _type = "Task" [ 1022.135471] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1022.143541] env[63515]: DEBUG oslo_vmware.api [None req-042daf1f-2fe3-4b1d-bb04-2f7a0479fcba tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]525678eb-7cb9-c755-ce97-6ba4dd85da0b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.404628] env[63515]: DEBUG nova.scheduler.client.report [None req-a7fc067e-c3eb-4fcd-8441-645f050b9570 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1022.418028] env[63515]: DEBUG oslo_concurrency.lockutils [None req-878f8fb3-6690-40c9-9f6f-d6953e3612e1 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Releasing lock "refresh_cache-7d572ac0-f6af-4622-96cc-e75983420222" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1022.443601] env[63515]: DEBUG nova.virt.hardware [None req-878f8fb3-6690-40c9-9f6f-d6953e3612e1 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T02:52:14Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='a458fe56a2a283e7990c8b77baba1e4f',container_format='bare',created_at=2024-10-03T03:01:41Z,direct_url=,disk_format='vmdk',id=cb0243fe-58ca-4459-930c-548f47818dfc,min_disk=1,min_ram=0,name='tempest-AttachVolumeShelveTestJSON-server-1625896844-shelved',owner='19ce3b05eec64475b685faad1fd801f6',properties=ImageMetaProps,protected=,size=31663616,status='active',tags=,updated_at=2024-10-03T03:01:55Z,virtual_size=,visibility=), allow threads: False {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1022.443850] env[63515]: DEBUG nova.virt.hardware [None req-878f8fb3-6690-40c9-9f6f-d6953e3612e1 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Flavor limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1022.444022] env[63515]: DEBUG nova.virt.hardware [None req-878f8fb3-6690-40c9-9f6f-d6953e3612e1 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Image limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1022.444211] env[63515]: DEBUG nova.virt.hardware [None req-878f8fb3-6690-40c9-9f6f-d6953e3612e1 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Flavor pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1022.444359] env[63515]: DEBUG nova.virt.hardware [None req-878f8fb3-6690-40c9-9f6f-d6953e3612e1 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Image pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1022.444508] env[63515]: DEBUG nova.virt.hardware [None req-878f8fb3-6690-40c9-9f6f-d6953e3612e1 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1022.444715] env[63515]: DEBUG nova.virt.hardware [None req-878f8fb3-6690-40c9-9f6f-d6953e3612e1 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1022.444902] env[63515]: DEBUG nova.virt.hardware [None req-878f8fb3-6690-40c9-9f6f-d6953e3612e1 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1022.445095] env[63515]: DEBUG nova.virt.hardware [None req-878f8fb3-6690-40c9-9f6f-d6953e3612e1 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Got 1 possible topologies {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1022.445268] env[63515]: DEBUG nova.virt.hardware [None req-878f8fb3-6690-40c9-9f6f-d6953e3612e1 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1022.445445] env[63515]: DEBUG nova.virt.hardware [None req-878f8fb3-6690-40c9-9f6f-d6953e3612e1 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1022.446341] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b6aa17e-31cf-4422-b86b-8cb7697201b3 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.454245] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d77a1337-acb9-417e-bb28-d3aec60c2e3e {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.469051] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-878f8fb3-6690-40c9-9f6f-d6953e3612e1 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] [instance: 7d572ac0-f6af-4622-96cc-e75983420222] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:6d:92:3d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ded18042-834c-4792-b3e8-b1c377446432', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'dbfd09c2-9ec4-4821-8613-43c6d6cf4724', 'vif_model': 'vmxnet3'}] {{(pid=63515) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1022.476291] env[63515]: DEBUG oslo.service.loopingcall [None req-878f8fb3-6690-40c9-9f6f-d6953e3612e1 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1022.476521] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7d572ac0-f6af-4622-96cc-e75983420222] Creating VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1022.476722] env[63515]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f48442bf-f115-4557-9f5a-8bae863bf0e2 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.495843] env[63515]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1022.495843] env[63515]: value = "task-1111830" [ 1022.495843] env[63515]: _type = "Task" [ 1022.495843] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1022.503286] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1111830, 'name': CreateVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.648020] env[63515]: DEBUG oslo_vmware.api [None req-042daf1f-2fe3-4b1d-bb04-2f7a0479fcba tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]525678eb-7cb9-c755-ce97-6ba4dd85da0b, 'name': SearchDatastore_Task, 'duration_secs': 0.011592} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1022.648944] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-45b8b832-ede6-4f08-b7f5-dc7330351bee {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.654438] env[63515]: DEBUG oslo_vmware.api [None req-042daf1f-2fe3-4b1d-bb04-2f7a0479fcba tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Waiting for the task: (returnval){ [ 1022.654438] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]529823b6-3965-606e-b817-0fcab8d48622" [ 1022.654438] env[63515]: _type = "Task" [ 1022.654438] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1022.662669] env[63515]: DEBUG oslo_vmware.api [None req-042daf1f-2fe3-4b1d-bb04-2f7a0479fcba tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]529823b6-3965-606e-b817-0fcab8d48622, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.792164] env[63515]: DEBUG nova.compute.manager [req-866b450f-2e5a-4555-baa5-c61b0457bc11 req-53e22ed3-9adb-4347-8797-101b61c97048 service nova] [instance: 7d572ac0-f6af-4622-96cc-e75983420222] Received event network-vif-plugged-dbfd09c2-9ec4-4821-8613-43c6d6cf4724 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1022.792322] env[63515]: DEBUG oslo_concurrency.lockutils [req-866b450f-2e5a-4555-baa5-c61b0457bc11 req-53e22ed3-9adb-4347-8797-101b61c97048 service nova] Acquiring lock "7d572ac0-f6af-4622-96cc-e75983420222-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1022.792542] env[63515]: DEBUG oslo_concurrency.lockutils [req-866b450f-2e5a-4555-baa5-c61b0457bc11 req-53e22ed3-9adb-4347-8797-101b61c97048 service nova] Lock "7d572ac0-f6af-4622-96cc-e75983420222-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1022.792739] env[63515]: DEBUG oslo_concurrency.lockutils [req-866b450f-2e5a-4555-baa5-c61b0457bc11 req-53e22ed3-9adb-4347-8797-101b61c97048 service nova] Lock "7d572ac0-f6af-4622-96cc-e75983420222-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1022.792953] env[63515]: DEBUG nova.compute.manager [req-866b450f-2e5a-4555-baa5-c61b0457bc11 req-53e22ed3-9adb-4347-8797-101b61c97048 service nova] [instance: 7d572ac0-f6af-4622-96cc-e75983420222] No waiting events found dispatching network-vif-plugged-dbfd09c2-9ec4-4821-8613-43c6d6cf4724 {{(pid=63515) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1022.793341] env[63515]: WARNING nova.compute.manager [req-866b450f-2e5a-4555-baa5-c61b0457bc11 req-53e22ed3-9adb-4347-8797-101b61c97048 service nova] [instance: 7d572ac0-f6af-4622-96cc-e75983420222] Received unexpected event network-vif-plugged-dbfd09c2-9ec4-4821-8613-43c6d6cf4724 for instance with vm_state shelved_offloaded and task_state spawning. [ 1022.793494] env[63515]: DEBUG nova.compute.manager [req-866b450f-2e5a-4555-baa5-c61b0457bc11 req-53e22ed3-9adb-4347-8797-101b61c97048 service nova] [instance: 7d572ac0-f6af-4622-96cc-e75983420222] Received event network-changed-dbfd09c2-9ec4-4821-8613-43c6d6cf4724 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1022.793679] env[63515]: DEBUG nova.compute.manager [req-866b450f-2e5a-4555-baa5-c61b0457bc11 req-53e22ed3-9adb-4347-8797-101b61c97048 service nova] [instance: 7d572ac0-f6af-4622-96cc-e75983420222] Refreshing instance network info cache due to event network-changed-dbfd09c2-9ec4-4821-8613-43c6d6cf4724. {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1022.793877] env[63515]: DEBUG oslo_concurrency.lockutils [req-866b450f-2e5a-4555-baa5-c61b0457bc11 req-53e22ed3-9adb-4347-8797-101b61c97048 service nova] Acquiring lock "refresh_cache-7d572ac0-f6af-4622-96cc-e75983420222" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1022.794035] env[63515]: DEBUG oslo_concurrency.lockutils [req-866b450f-2e5a-4555-baa5-c61b0457bc11 req-53e22ed3-9adb-4347-8797-101b61c97048 service nova] Acquired lock "refresh_cache-7d572ac0-f6af-4622-96cc-e75983420222" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1022.794205] env[63515]: DEBUG nova.network.neutron [req-866b450f-2e5a-4555-baa5-c61b0457bc11 req-53e22ed3-9adb-4347-8797-101b61c97048 service nova] [instance: 7d572ac0-f6af-4622-96cc-e75983420222] Refreshing network info cache for port dbfd09c2-9ec4-4821-8613-43c6d6cf4724 {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1022.909967] env[63515]: DEBUG oslo_concurrency.lockutils [None req-a7fc067e-c3eb-4fcd-8441-645f050b9570 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.203s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1022.910274] env[63515]: INFO nova.compute.manager [None req-a7fc067e-c3eb-4fcd-8441-645f050b9570 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: b76cf3dd-7b7e-4e5d-bf3c-e15bd37069ac] Migrating [ 1022.916654] env[63515]: DEBUG oslo_concurrency.lockutils [None req-e8c24bdc-d1c1-4d98-8d1d-53bd436f7627 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.544s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1022.916910] env[63515]: DEBUG oslo_concurrency.lockutils [None req-e8c24bdc-d1c1-4d98-8d1d-53bd436f7627 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1022.939731] env[63515]: INFO nova.scheduler.client.report [None req-e8c24bdc-d1c1-4d98-8d1d-53bd436f7627 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Deleted allocations for instance 68766cd4-84be-475b-8494-d7ab43a9e969 [ 1023.006556] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1111830, 'name': CreateVM_Task, 'duration_secs': 0.3133} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1023.006727] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7d572ac0-f6af-4622-96cc-e75983420222] Created VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1023.007442] env[63515]: DEBUG oslo_concurrency.lockutils [None req-878f8fb3-6690-40c9-9f6f-d6953e3612e1 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/cb0243fe-58ca-4459-930c-548f47818dfc" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1023.007615] env[63515]: DEBUG oslo_concurrency.lockutils [None req-878f8fb3-6690-40c9-9f6f-d6953e3612e1 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Acquired lock "[datastore2] devstack-image-cache_base/cb0243fe-58ca-4459-930c-548f47818dfc" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1023.008138] env[63515]: DEBUG oslo_concurrency.lockutils [None req-878f8fb3-6690-40c9-9f6f-d6953e3612e1 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/cb0243fe-58ca-4459-930c-548f47818dfc" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1023.008254] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b370a87f-d01d-48d6-aeb1-fb0f8f008d54 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.012525] env[63515]: DEBUG oslo_vmware.api [None req-878f8fb3-6690-40c9-9f6f-d6953e3612e1 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Waiting for the task: (returnval){ [ 1023.012525] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]52625b8c-b399-811e-6eae-b21d3e8b9e8c" [ 1023.012525] env[63515]: _type = "Task" [ 1023.012525] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1023.020046] env[63515]: DEBUG oslo_vmware.api [None req-878f8fb3-6690-40c9-9f6f-d6953e3612e1 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52625b8c-b399-811e-6eae-b21d3e8b9e8c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.108491] env[63515]: DEBUG nova.compute.manager [req-e769c685-debb-4038-b6f9-199ed05daa21 req-5ed6a518-6b85-4c62-aa34-4e48a850b978 service nova] [instance: fcfe3376-8dfa-4189-a267-ff6402713c1b] Received event network-changed-4462b1aa-72ea-41cc-878a-68c38920c8cc {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1023.108491] env[63515]: DEBUG nova.compute.manager [req-e769c685-debb-4038-b6f9-199ed05daa21 req-5ed6a518-6b85-4c62-aa34-4e48a850b978 service nova] [instance: fcfe3376-8dfa-4189-a267-ff6402713c1b] Refreshing instance network info cache due to event network-changed-4462b1aa-72ea-41cc-878a-68c38920c8cc. {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1023.108572] env[63515]: DEBUG oslo_concurrency.lockutils [req-e769c685-debb-4038-b6f9-199ed05daa21 req-5ed6a518-6b85-4c62-aa34-4e48a850b978 service nova] Acquiring lock "refresh_cache-fcfe3376-8dfa-4189-a267-ff6402713c1b" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1023.108671] env[63515]: DEBUG oslo_concurrency.lockutils [req-e769c685-debb-4038-b6f9-199ed05daa21 req-5ed6a518-6b85-4c62-aa34-4e48a850b978 service nova] Acquired lock "refresh_cache-fcfe3376-8dfa-4189-a267-ff6402713c1b" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1023.108843] env[63515]: DEBUG nova.network.neutron [req-e769c685-debb-4038-b6f9-199ed05daa21 req-5ed6a518-6b85-4c62-aa34-4e48a850b978 service nova] [instance: fcfe3376-8dfa-4189-a267-ff6402713c1b] Refreshing network info cache for port 4462b1aa-72ea-41cc-878a-68c38920c8cc {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1023.164149] env[63515]: DEBUG oslo_vmware.api [None req-042daf1f-2fe3-4b1d-bb04-2f7a0479fcba tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]529823b6-3965-606e-b817-0fcab8d48622, 'name': SearchDatastore_Task, 'duration_secs': 0.009762} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1023.165255] env[63515]: DEBUG oslo_concurrency.lockutils [None req-042daf1f-2fe3-4b1d-bb04-2f7a0479fcba tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1023.165255] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-042daf1f-2fe3-4b1d-bb04-2f7a0479fcba tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk to [datastore1] e925fc93-a731-498a-984c-9b1f3eabf353/e925fc93-a731-498a-984c-9b1f3eabf353.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1023.165255] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-51b9e99d-4c82-4928-8895-e552d59c662d {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.171457] env[63515]: DEBUG oslo_vmware.api [None req-042daf1f-2fe3-4b1d-bb04-2f7a0479fcba tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Waiting for the task: (returnval){ [ 1023.171457] env[63515]: value = "task-1111831" [ 1023.171457] env[63515]: _type = "Task" [ 1023.171457] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1023.183125] env[63515]: DEBUG oslo_vmware.api [None req-042daf1f-2fe3-4b1d-bb04-2f7a0479fcba tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Task: {'id': task-1111831, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.428344] env[63515]: DEBUG oslo_concurrency.lockutils [None req-a7fc067e-c3eb-4fcd-8441-645f050b9570 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Acquiring lock "refresh_cache-b76cf3dd-7b7e-4e5d-bf3c-e15bd37069ac" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1023.428571] env[63515]: DEBUG oslo_concurrency.lockutils [None req-a7fc067e-c3eb-4fcd-8441-645f050b9570 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Acquired lock "refresh_cache-b76cf3dd-7b7e-4e5d-bf3c-e15bd37069ac" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1023.428760] env[63515]: DEBUG nova.network.neutron [None req-a7fc067e-c3eb-4fcd-8441-645f050b9570 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: b76cf3dd-7b7e-4e5d-bf3c-e15bd37069ac] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1023.446725] env[63515]: DEBUG oslo_concurrency.lockutils [None req-e8c24bdc-d1c1-4d98-8d1d-53bd436f7627 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Lock "68766cd4-84be-475b-8494-d7ab43a9e969" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.304s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1023.524988] env[63515]: DEBUG oslo_concurrency.lockutils [None req-878f8fb3-6690-40c9-9f6f-d6953e3612e1 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Releasing lock "[datastore2] devstack-image-cache_base/cb0243fe-58ca-4459-930c-548f47818dfc" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1023.525286] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-878f8fb3-6690-40c9-9f6f-d6953e3612e1 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] [instance: 7d572ac0-f6af-4622-96cc-e75983420222] Processing image cb0243fe-58ca-4459-930c-548f47818dfc {{(pid=63515) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1023.525536] env[63515]: DEBUG oslo_concurrency.lockutils [None req-878f8fb3-6690-40c9-9f6f-d6953e3612e1 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/cb0243fe-58ca-4459-930c-548f47818dfc/cb0243fe-58ca-4459-930c-548f47818dfc.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1023.525867] env[63515]: DEBUG oslo_concurrency.lockutils [None req-878f8fb3-6690-40c9-9f6f-d6953e3612e1 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Acquired lock "[datastore2] devstack-image-cache_base/cb0243fe-58ca-4459-930c-548f47818dfc/cb0243fe-58ca-4459-930c-548f47818dfc.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1023.525867] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-878f8fb3-6690-40c9-9f6f-d6953e3612e1 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1023.526623] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-060bfbfa-29ab-4722-aaaa-f4f02faf6b20 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.538302] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-878f8fb3-6690-40c9-9f6f-d6953e3612e1 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1023.538519] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-878f8fb3-6690-40c9-9f6f-d6953e3612e1 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63515) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1023.541851] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cd7daec8-2315-4639-8949-16fea52c430a {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.548770] env[63515]: DEBUG oslo_vmware.api [None req-878f8fb3-6690-40c9-9f6f-d6953e3612e1 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Waiting for the task: (returnval){ [ 1023.548770] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]52585d8e-69e1-f16b-5de4-67ec42745129" [ 1023.548770] env[63515]: _type = "Task" [ 1023.548770] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1023.559355] env[63515]: DEBUG oslo_vmware.api [None req-878f8fb3-6690-40c9-9f6f-d6953e3612e1 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52585d8e-69e1-f16b-5de4-67ec42745129, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.630313] env[63515]: DEBUG nova.network.neutron [None req-80fab3e3-466b-468e-b003-05f4a2e89df7 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: 724b05bc-ce4b-4f99-ad14-f3abea5567f2] Successfully updated port: 6a98da26-06f3-4eb5-aee8-02a448c34749 {{(pid=63515) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1023.672698] env[63515]: DEBUG nova.network.neutron [req-866b450f-2e5a-4555-baa5-c61b0457bc11 req-53e22ed3-9adb-4347-8797-101b61c97048 service nova] [instance: 7d572ac0-f6af-4622-96cc-e75983420222] Updated VIF entry in instance network info cache for port dbfd09c2-9ec4-4821-8613-43c6d6cf4724. {{(pid=63515) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1023.673092] env[63515]: DEBUG nova.network.neutron [req-866b450f-2e5a-4555-baa5-c61b0457bc11 req-53e22ed3-9adb-4347-8797-101b61c97048 service nova] [instance: 7d572ac0-f6af-4622-96cc-e75983420222] Updating instance_info_cache with network_info: [{"id": "dbfd09c2-9ec4-4821-8613-43c6d6cf4724", "address": "fa:16:3e:6d:92:3d", "network": {"id": "09fb20ca-2b17-4d46-b915-6aff85b862d8", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1815529427-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.244", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "19ce3b05eec64475b685faad1fd801f6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ded18042-834c-4792-b3e8-b1c377446432", "external-id": "nsx-vlan-transportzone-293", "segmentation_id": 293, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdbfd09c2-9e", "ovs_interfaceid": "dbfd09c2-9ec4-4821-8613-43c6d6cf4724", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1023.685355] env[63515]: DEBUG oslo_vmware.api [None req-042daf1f-2fe3-4b1d-bb04-2f7a0479fcba tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Task: {'id': task-1111831, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.494477} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1023.685716] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-042daf1f-2fe3-4b1d-bb04-2f7a0479fcba tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk to [datastore1] e925fc93-a731-498a-984c-9b1f3eabf353/e925fc93-a731-498a-984c-9b1f3eabf353.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1023.686013] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-042daf1f-2fe3-4b1d-bb04-2f7a0479fcba tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: e925fc93-a731-498a-984c-9b1f3eabf353] Extending root virtual disk to 1048576 {{(pid=63515) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1023.687218] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-76b8ae3b-be55-4e7c-9b56-af2cb1827d61 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.695436] env[63515]: DEBUG oslo_vmware.api [None req-042daf1f-2fe3-4b1d-bb04-2f7a0479fcba tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Waiting for the task: (returnval){ [ 1023.695436] env[63515]: value = "task-1111832" [ 1023.695436] env[63515]: _type = "Task" [ 1023.695436] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1023.707707] env[63515]: DEBUG oslo_vmware.api [None req-042daf1f-2fe3-4b1d-bb04-2f7a0479fcba tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Task: {'id': task-1111832, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.824138] env[63515]: DEBUG nova.network.neutron [req-e769c685-debb-4038-b6f9-199ed05daa21 req-5ed6a518-6b85-4c62-aa34-4e48a850b978 service nova] [instance: fcfe3376-8dfa-4189-a267-ff6402713c1b] Updated VIF entry in instance network info cache for port 4462b1aa-72ea-41cc-878a-68c38920c8cc. {{(pid=63515) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1023.824678] env[63515]: DEBUG nova.network.neutron [req-e769c685-debb-4038-b6f9-199ed05daa21 req-5ed6a518-6b85-4c62-aa34-4e48a850b978 service nova] [instance: fcfe3376-8dfa-4189-a267-ff6402713c1b] Updating instance_info_cache with network_info: [{"id": "4462b1aa-72ea-41cc-878a-68c38920c8cc", "address": "fa:16:3e:4f:bd:04", "network": {"id": "9fc2b9c9-6afc-4510-bf13-f87f1e1bd53b", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-513162532-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "0be5ebcf766b4da5b33be3a12387e737", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0248a27a-1d7f-4195-987b-06bfc8467347", "external-id": "nsx-vlan-transportzone-26", "segmentation_id": 26, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4462b1aa-72", "ovs_interfaceid": "4462b1aa-72ea-41cc-878a-68c38920c8cc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1024.059290] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-878f8fb3-6690-40c9-9f6f-d6953e3612e1 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] [instance: 7d572ac0-f6af-4622-96cc-e75983420222] Preparing fetch location {{(pid=63515) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 1024.059849] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-878f8fb3-6690-40c9-9f6f-d6953e3612e1 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] [instance: 7d572ac0-f6af-4622-96cc-e75983420222] Fetch image to [datastore2] OSTACK_IMG_516d6119-fb9a-490f-9429-53b230ce63a9/OSTACK_IMG_516d6119-fb9a-490f-9429-53b230ce63a9.vmdk {{(pid=63515) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 1024.060024] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-878f8fb3-6690-40c9-9f6f-d6953e3612e1 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] [instance: 7d572ac0-f6af-4622-96cc-e75983420222] Downloading stream optimized image cb0243fe-58ca-4459-930c-548f47818dfc to [datastore2] OSTACK_IMG_516d6119-fb9a-490f-9429-53b230ce63a9/OSTACK_IMG_516d6119-fb9a-490f-9429-53b230ce63a9.vmdk on the data store datastore2 as vApp {{(pid=63515) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 1024.060290] env[63515]: DEBUG nova.virt.vmwareapi.images [None req-878f8fb3-6690-40c9-9f6f-d6953e3612e1 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] [instance: 7d572ac0-f6af-4622-96cc-e75983420222] Downloading image file data cb0243fe-58ca-4459-930c-548f47818dfc to the ESX as VM named 'OSTACK_IMG_516d6119-fb9a-490f-9429-53b230ce63a9' {{(pid=63515) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 1024.135323] env[63515]: DEBUG oslo_concurrency.lockutils [None req-80fab3e3-466b-468e-b003-05f4a2e89df7 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Acquiring lock "refresh_cache-724b05bc-ce4b-4f99-ad14-f3abea5567f2" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1024.135542] env[63515]: DEBUG oslo_concurrency.lockutils [None req-80fab3e3-466b-468e-b003-05f4a2e89df7 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Acquired lock "refresh_cache-724b05bc-ce4b-4f99-ad14-f3abea5567f2" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1024.135743] env[63515]: DEBUG nova.network.neutron [None req-80fab3e3-466b-468e-b003-05f4a2e89df7 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: 724b05bc-ce4b-4f99-ad14-f3abea5567f2] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1024.142439] env[63515]: DEBUG oslo_vmware.rw_handles [None req-878f8fb3-6690-40c9-9f6f-d6953e3612e1 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 1024.142439] env[63515]: value = "resgroup-9" [ 1024.142439] env[63515]: _type = "ResourcePool" [ 1024.142439] env[63515]: }. {{(pid=63515) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 1024.143155] env[63515]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-8c230af1-0cea-4e5f-b715-c1b9a0e577ad {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.160919] env[63515]: DEBUG nova.network.neutron [None req-a7fc067e-c3eb-4fcd-8441-645f050b9570 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: b76cf3dd-7b7e-4e5d-bf3c-e15bd37069ac] Updating instance_info_cache with network_info: [{"id": "acea8927-7e43-427c-bed2-2420d4708f3b", "address": "fa:16:3e:21:d3:df", "network": {"id": "3e5914c8-0cd4-49ee-81b4-1f3e280440bb", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-641632277-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ea099ce07bfb4a8da014d9303cf552eb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "225b6979-9329-403b-91fa-138bd41f6e83", "external-id": "nsx-vlan-transportzone-38", "segmentation_id": 38, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapacea8927-7e", "ovs_interfaceid": "acea8927-7e43-427c-bed2-2420d4708f3b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1024.167922] env[63515]: DEBUG oslo_vmware.rw_handles [None req-878f8fb3-6690-40c9-9f6f-d6953e3612e1 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Lease: (returnval){ [ 1024.167922] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]52aec542-0629-5e3b-b217-0e701205c3c8" [ 1024.167922] env[63515]: _type = "HttpNfcLease" [ 1024.167922] env[63515]: } obtained for vApp import into resource pool (val){ [ 1024.167922] env[63515]: value = "resgroup-9" [ 1024.167922] env[63515]: _type = "ResourcePool" [ 1024.167922] env[63515]: }. {{(pid=63515) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 1024.168204] env[63515]: DEBUG oslo_vmware.api [None req-878f8fb3-6690-40c9-9f6f-d6953e3612e1 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Waiting for the lease: (returnval){ [ 1024.168204] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]52aec542-0629-5e3b-b217-0e701205c3c8" [ 1024.168204] env[63515]: _type = "HttpNfcLease" [ 1024.168204] env[63515]: } to be ready. {{(pid=63515) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1024.175313] env[63515]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1024.175313] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]52aec542-0629-5e3b-b217-0e701205c3c8" [ 1024.175313] env[63515]: _type = "HttpNfcLease" [ 1024.175313] env[63515]: } is initializing. {{(pid=63515) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1024.178904] env[63515]: DEBUG oslo_concurrency.lockutils [req-866b450f-2e5a-4555-baa5-c61b0457bc11 req-53e22ed3-9adb-4347-8797-101b61c97048 service nova] Releasing lock "refresh_cache-7d572ac0-f6af-4622-96cc-e75983420222" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1024.206923] env[63515]: DEBUG oslo_vmware.api [None req-042daf1f-2fe3-4b1d-bb04-2f7a0479fcba tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Task: {'id': task-1111832, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.105789} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1024.208146] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-042daf1f-2fe3-4b1d-bb04-2f7a0479fcba tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: e925fc93-a731-498a-984c-9b1f3eabf353] Extended root virtual disk {{(pid=63515) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1024.208390] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e5336e6-f15d-4d4d-8910-fbb6d9f10720 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.235022] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-042daf1f-2fe3-4b1d-bb04-2f7a0479fcba tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: e925fc93-a731-498a-984c-9b1f3eabf353] Reconfiguring VM instance instance-00000065 to attach disk [datastore1] e925fc93-a731-498a-984c-9b1f3eabf353/e925fc93-a731-498a-984c-9b1f3eabf353.vmdk or device None with type sparse {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1024.235022] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-31077541-59f1-4df6-bc10-edd9b069fcdd {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.252426] env[63515]: DEBUG oslo_vmware.api [None req-042daf1f-2fe3-4b1d-bb04-2f7a0479fcba tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Waiting for the task: (returnval){ [ 1024.252426] env[63515]: value = "task-1111834" [ 1024.252426] env[63515]: _type = "Task" [ 1024.252426] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1024.262240] env[63515]: DEBUG oslo_vmware.api [None req-042daf1f-2fe3-4b1d-bb04-2f7a0479fcba tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Task: {'id': task-1111834, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.327972] env[63515]: DEBUG oslo_concurrency.lockutils [req-e769c685-debb-4038-b6f9-199ed05daa21 req-5ed6a518-6b85-4c62-aa34-4e48a850b978 service nova] Releasing lock "refresh_cache-fcfe3376-8dfa-4189-a267-ff6402713c1b" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1024.663831] env[63515]: DEBUG oslo_concurrency.lockutils [None req-a7fc067e-c3eb-4fcd-8441-645f050b9570 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Releasing lock "refresh_cache-b76cf3dd-7b7e-4e5d-bf3c-e15bd37069ac" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1024.673889] env[63515]: WARNING nova.network.neutron [None req-80fab3e3-466b-468e-b003-05f4a2e89df7 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: 724b05bc-ce4b-4f99-ad14-f3abea5567f2] 88dfb29d-41e0-4df7-8f70-e52cda05620c already exists in list: networks containing: ['88dfb29d-41e0-4df7-8f70-e52cda05620c']. ignoring it [ 1024.680164] env[63515]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1024.680164] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]52aec542-0629-5e3b-b217-0e701205c3c8" [ 1024.680164] env[63515]: _type = "HttpNfcLease" [ 1024.680164] env[63515]: } is initializing. {{(pid=63515) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1024.765973] env[63515]: DEBUG oslo_vmware.api [None req-042daf1f-2fe3-4b1d-bb04-2f7a0479fcba tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Task: {'id': task-1111834, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.987691] env[63515]: DEBUG nova.network.neutron [None req-80fab3e3-466b-468e-b003-05f4a2e89df7 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: 724b05bc-ce4b-4f99-ad14-f3abea5567f2] Updating instance_info_cache with network_info: [{"id": "3dee43b6-2650-42e1-aa5a-4994a3ec7f05", "address": "fa:16:3e:89:06:d0", "network": {"id": "88dfb29d-41e0-4df7-8f70-e52cda05620c", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1355578400-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.146", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7d4ada9ec35f42b19c6480a9101d21a4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "357d2811-e990-4985-9f9e-b158d10d3699", "external-id": "nsx-vlan-transportzone-641", "segmentation_id": 641, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3dee43b6-26", "ovs_interfaceid": "3dee43b6-2650-42e1-aa5a-4994a3ec7f05", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "6a98da26-06f3-4eb5-aee8-02a448c34749", "address": "fa:16:3e:31:42:a5", "network": {"id": "88dfb29d-41e0-4df7-8f70-e52cda05620c", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1355578400-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7d4ada9ec35f42b19c6480a9101d21a4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "357d2811-e990-4985-9f9e-b158d10d3699", "external-id": "nsx-vlan-transportzone-641", "segmentation_id": 641, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6a98da26-06", "ovs_interfaceid": "6a98da26-06f3-4eb5-aee8-02a448c34749", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1025.183212] env[63515]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1025.183212] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]52aec542-0629-5e3b-b217-0e701205c3c8" [ 1025.183212] env[63515]: _type = "HttpNfcLease" [ 1025.183212] env[63515]: } is ready. {{(pid=63515) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1025.185079] env[63515]: DEBUG oslo_vmware.rw_handles [None req-878f8fb3-6690-40c9-9f6f-d6953e3612e1 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1025.185079] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]52aec542-0629-5e3b-b217-0e701205c3c8" [ 1025.185079] env[63515]: _type = "HttpNfcLease" [ 1025.185079] env[63515]: }. {{(pid=63515) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 1025.185079] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11799f8c-c703-48ef-984a-b94be74802a6 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.193040] env[63515]: DEBUG oslo_vmware.rw_handles [None req-878f8fb3-6690-40c9-9f6f-d6953e3612e1 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52e09a85-0324-a299-c245-46c30113acba/disk-0.vmdk from lease info. {{(pid=63515) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1025.193199] env[63515]: DEBUG oslo_vmware.rw_handles [None req-878f8fb3-6690-40c9-9f6f-d6953e3612e1 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Creating HTTP connection to write to file with size = 31663616 and URL = https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52e09a85-0324-a299-c245-46c30113acba/disk-0.vmdk. {{(pid=63515) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 1025.267305] env[63515]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-2c3b4938-1100-4993-a74a-8327d1b9cf55 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.269528] env[63515]: DEBUG oslo_vmware.api [None req-042daf1f-2fe3-4b1d-bb04-2f7a0479fcba tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Task: {'id': task-1111834, 'name': ReconfigVM_Task, 'duration_secs': 0.609748} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1025.271256] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-042daf1f-2fe3-4b1d-bb04-2f7a0479fcba tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: e925fc93-a731-498a-984c-9b1f3eabf353] Reconfigured VM instance instance-00000065 to attach disk [datastore1] e925fc93-a731-498a-984c-9b1f3eabf353/e925fc93-a731-498a-984c-9b1f3eabf353.vmdk or device None with type sparse {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1025.272307] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-211c59ac-3174-4310-8a14-7648a02c419e {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.278242] env[63515]: DEBUG oslo_vmware.api [None req-042daf1f-2fe3-4b1d-bb04-2f7a0479fcba tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Waiting for the task: (returnval){ [ 1025.278242] env[63515]: value = "task-1111835" [ 1025.278242] env[63515]: _type = "Task" [ 1025.278242] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1025.287033] env[63515]: DEBUG oslo_vmware.api [None req-042daf1f-2fe3-4b1d-bb04-2f7a0479fcba tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Task: {'id': task-1111835, 'name': Rename_Task} progress is 5%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.490735] env[63515]: DEBUG oslo_concurrency.lockutils [None req-80fab3e3-466b-468e-b003-05f4a2e89df7 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Releasing lock "refresh_cache-724b05bc-ce4b-4f99-ad14-f3abea5567f2" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1025.491524] env[63515]: DEBUG oslo_concurrency.lockutils [None req-80fab3e3-466b-468e-b003-05f4a2e89df7 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Acquiring lock "724b05bc-ce4b-4f99-ad14-f3abea5567f2" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1025.491709] env[63515]: DEBUG oslo_concurrency.lockutils [None req-80fab3e3-466b-468e-b003-05f4a2e89df7 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Acquired lock "724b05bc-ce4b-4f99-ad14-f3abea5567f2" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1025.492688] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7009624d-cab4-466e-a251-70b02f6058f1 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.510051] env[63515]: DEBUG nova.virt.hardware [None req-80fab3e3-466b-468e-b003-05f4a2e89df7 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T02:52:14Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1025.510259] env[63515]: DEBUG nova.virt.hardware [None req-80fab3e3-466b-468e-b003-05f4a2e89df7 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Flavor limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1025.510362] env[63515]: DEBUG nova.virt.hardware [None req-80fab3e3-466b-468e-b003-05f4a2e89df7 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Image limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1025.510553] env[63515]: DEBUG nova.virt.hardware [None req-80fab3e3-466b-468e-b003-05f4a2e89df7 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Flavor pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1025.510741] env[63515]: DEBUG nova.virt.hardware [None req-80fab3e3-466b-468e-b003-05f4a2e89df7 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Image pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1025.510875] env[63515]: DEBUG nova.virt.hardware [None req-80fab3e3-466b-468e-b003-05f4a2e89df7 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1025.511123] env[63515]: DEBUG nova.virt.hardware [None req-80fab3e3-466b-468e-b003-05f4a2e89df7 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1025.511339] env[63515]: DEBUG nova.virt.hardware [None req-80fab3e3-466b-468e-b003-05f4a2e89df7 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1025.511583] env[63515]: DEBUG nova.virt.hardware [None req-80fab3e3-466b-468e-b003-05f4a2e89df7 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Got 1 possible topologies {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1025.511777] env[63515]: DEBUG nova.virt.hardware [None req-80fab3e3-466b-468e-b003-05f4a2e89df7 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1025.512038] env[63515]: DEBUG nova.virt.hardware [None req-80fab3e3-466b-468e-b003-05f4a2e89df7 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1025.518744] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-80fab3e3-466b-468e-b003-05f4a2e89df7 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: 724b05bc-ce4b-4f99-ad14-f3abea5567f2] Reconfiguring VM to attach interface {{(pid=63515) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1923}} [ 1025.520788] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f63a21ac-f151-4542-a158-a1e8186564d0 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.539745] env[63515]: DEBUG oslo_vmware.api [None req-80fab3e3-466b-468e-b003-05f4a2e89df7 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Waiting for the task: (returnval){ [ 1025.539745] env[63515]: value = "task-1111836" [ 1025.539745] env[63515]: _type = "Task" [ 1025.539745] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1025.547829] env[63515]: DEBUG oslo_vmware.api [None req-80fab3e3-466b-468e-b003-05f4a2e89df7 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Task: {'id': task-1111836, 'name': ReconfigVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.790131] env[63515]: DEBUG oslo_vmware.api [None req-042daf1f-2fe3-4b1d-bb04-2f7a0479fcba tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Task: {'id': task-1111835, 'name': Rename_Task, 'duration_secs': 0.218136} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1025.792180] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-042daf1f-2fe3-4b1d-bb04-2f7a0479fcba tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: e925fc93-a731-498a-984c-9b1f3eabf353] Powering on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1025.792511] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ea725736-bc53-4818-a84a-fa99cee94bd4 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.799787] env[63515]: DEBUG oslo_vmware.api [None req-042daf1f-2fe3-4b1d-bb04-2f7a0479fcba tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Waiting for the task: (returnval){ [ 1025.799787] env[63515]: value = "task-1111837" [ 1025.799787] env[63515]: _type = "Task" [ 1025.799787] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1025.811518] env[63515]: DEBUG oslo_vmware.api [None req-042daf1f-2fe3-4b1d-bb04-2f7a0479fcba tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Task: {'id': task-1111837, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.940576] env[63515]: DEBUG nova.compute.manager [req-8d378247-cbf0-4efe-9bbb-fdd08040ab39 req-9e8f3e27-f4b5-4fee-84a1-b11f58cee9b6 service nova] [instance: 724b05bc-ce4b-4f99-ad14-f3abea5567f2] Received event network-vif-plugged-6a98da26-06f3-4eb5-aee8-02a448c34749 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1025.940794] env[63515]: DEBUG oslo_concurrency.lockutils [req-8d378247-cbf0-4efe-9bbb-fdd08040ab39 req-9e8f3e27-f4b5-4fee-84a1-b11f58cee9b6 service nova] Acquiring lock "724b05bc-ce4b-4f99-ad14-f3abea5567f2-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1025.941076] env[63515]: DEBUG oslo_concurrency.lockutils [req-8d378247-cbf0-4efe-9bbb-fdd08040ab39 req-9e8f3e27-f4b5-4fee-84a1-b11f58cee9b6 service nova] Lock "724b05bc-ce4b-4f99-ad14-f3abea5567f2-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1025.941464] env[63515]: DEBUG oslo_concurrency.lockutils [req-8d378247-cbf0-4efe-9bbb-fdd08040ab39 req-9e8f3e27-f4b5-4fee-84a1-b11f58cee9b6 service nova] Lock "724b05bc-ce4b-4f99-ad14-f3abea5567f2-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1025.941464] env[63515]: DEBUG nova.compute.manager [req-8d378247-cbf0-4efe-9bbb-fdd08040ab39 req-9e8f3e27-f4b5-4fee-84a1-b11f58cee9b6 service nova] [instance: 724b05bc-ce4b-4f99-ad14-f3abea5567f2] No waiting events found dispatching network-vif-plugged-6a98da26-06f3-4eb5-aee8-02a448c34749 {{(pid=63515) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1025.941639] env[63515]: WARNING nova.compute.manager [req-8d378247-cbf0-4efe-9bbb-fdd08040ab39 req-9e8f3e27-f4b5-4fee-84a1-b11f58cee9b6 service nova] [instance: 724b05bc-ce4b-4f99-ad14-f3abea5567f2] Received unexpected event network-vif-plugged-6a98da26-06f3-4eb5-aee8-02a448c34749 for instance with vm_state active and task_state None. [ 1025.941806] env[63515]: DEBUG nova.compute.manager [req-8d378247-cbf0-4efe-9bbb-fdd08040ab39 req-9e8f3e27-f4b5-4fee-84a1-b11f58cee9b6 service nova] [instance: 724b05bc-ce4b-4f99-ad14-f3abea5567f2] Received event network-changed-6a98da26-06f3-4eb5-aee8-02a448c34749 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1025.941964] env[63515]: DEBUG nova.compute.manager [req-8d378247-cbf0-4efe-9bbb-fdd08040ab39 req-9e8f3e27-f4b5-4fee-84a1-b11f58cee9b6 service nova] [instance: 724b05bc-ce4b-4f99-ad14-f3abea5567f2] Refreshing instance network info cache due to event network-changed-6a98da26-06f3-4eb5-aee8-02a448c34749. {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1025.942163] env[63515]: DEBUG oslo_concurrency.lockutils [req-8d378247-cbf0-4efe-9bbb-fdd08040ab39 req-9e8f3e27-f4b5-4fee-84a1-b11f58cee9b6 service nova] Acquiring lock "refresh_cache-724b05bc-ce4b-4f99-ad14-f3abea5567f2" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1025.942333] env[63515]: DEBUG oslo_concurrency.lockutils [req-8d378247-cbf0-4efe-9bbb-fdd08040ab39 req-9e8f3e27-f4b5-4fee-84a1-b11f58cee9b6 service nova] Acquired lock "refresh_cache-724b05bc-ce4b-4f99-ad14-f3abea5567f2" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1025.942490] env[63515]: DEBUG nova.network.neutron [req-8d378247-cbf0-4efe-9bbb-fdd08040ab39 req-9e8f3e27-f4b5-4fee-84a1-b11f58cee9b6 service nova] [instance: 724b05bc-ce4b-4f99-ad14-f3abea5567f2] Refreshing network info cache for port 6a98da26-06f3-4eb5-aee8-02a448c34749 {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1026.051245] env[63515]: DEBUG oslo_vmware.api [None req-80fab3e3-466b-468e-b003-05f4a2e89df7 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Task: {'id': task-1111836, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1026.072883] env[63515]: DEBUG nova.compute.manager [req-ab9100c9-a065-42e7-85ec-67775eafbd05 req-b177c9cc-3bc0-48ec-bd0a-0b13d6b691a2 service nova] [instance: fcfe3376-8dfa-4189-a267-ff6402713c1b] Received event network-changed-4462b1aa-72ea-41cc-878a-68c38920c8cc {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1026.073152] env[63515]: DEBUG nova.compute.manager [req-ab9100c9-a065-42e7-85ec-67775eafbd05 req-b177c9cc-3bc0-48ec-bd0a-0b13d6b691a2 service nova] [instance: fcfe3376-8dfa-4189-a267-ff6402713c1b] Refreshing instance network info cache due to event network-changed-4462b1aa-72ea-41cc-878a-68c38920c8cc. {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1026.073416] env[63515]: DEBUG oslo_concurrency.lockutils [req-ab9100c9-a065-42e7-85ec-67775eafbd05 req-b177c9cc-3bc0-48ec-bd0a-0b13d6b691a2 service nova] Acquiring lock "refresh_cache-fcfe3376-8dfa-4189-a267-ff6402713c1b" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1026.073593] env[63515]: DEBUG oslo_concurrency.lockutils [req-ab9100c9-a065-42e7-85ec-67775eafbd05 req-b177c9cc-3bc0-48ec-bd0a-0b13d6b691a2 service nova] Acquired lock "refresh_cache-fcfe3376-8dfa-4189-a267-ff6402713c1b" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1026.073821] env[63515]: DEBUG nova.network.neutron [req-ab9100c9-a065-42e7-85ec-67775eafbd05 req-b177c9cc-3bc0-48ec-bd0a-0b13d6b691a2 service nova] [instance: fcfe3376-8dfa-4189-a267-ff6402713c1b] Refreshing network info cache for port 4462b1aa-72ea-41cc-878a-68c38920c8cc {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1026.192036] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-373bd81c-5d91-493b-ae5a-16b2ff406d84 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.222488] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-a7fc067e-c3eb-4fcd-8441-645f050b9570 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: b76cf3dd-7b7e-4e5d-bf3c-e15bd37069ac] Updating instance 'b76cf3dd-7b7e-4e5d-bf3c-e15bd37069ac' progress to 0 {{(pid=63515) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1026.314069] env[63515]: DEBUG oslo_vmware.api [None req-042daf1f-2fe3-4b1d-bb04-2f7a0479fcba tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Task: {'id': task-1111837, 'name': PowerOnVM_Task} progress is 94%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1026.395876] env[63515]: DEBUG oslo_concurrency.lockutils [None req-123efa07-e143-4b76-a7d6-1e616f30576d tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Acquiring lock "aa83d035-9c47-434d-ab1a-140e6b3110b4" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1026.395876] env[63515]: DEBUG oslo_concurrency.lockutils [None req-123efa07-e143-4b76-a7d6-1e616f30576d tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Lock "aa83d035-9c47-434d-ab1a-140e6b3110b4" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1026.551441] env[63515]: DEBUG oslo_vmware.api [None req-80fab3e3-466b-468e-b003-05f4a2e89df7 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Task: {'id': task-1111836, 'name': ReconfigVM_Task, 'duration_secs': 0.870241} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1026.552084] env[63515]: DEBUG oslo_concurrency.lockutils [None req-80fab3e3-466b-468e-b003-05f4a2e89df7 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Releasing lock "724b05bc-ce4b-4f99-ad14-f3abea5567f2" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1026.552597] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-80fab3e3-466b-468e-b003-05f4a2e89df7 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: 724b05bc-ce4b-4f99-ad14-f3abea5567f2] Reconfigured VM to attach interface {{(pid=63515) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1937}} [ 1026.622822] env[63515]: DEBUG oslo_vmware.rw_handles [None req-878f8fb3-6690-40c9-9f6f-d6953e3612e1 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Completed reading data from the image iterator. {{(pid=63515) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 1026.623152] env[63515]: DEBUG oslo_vmware.rw_handles [None req-878f8fb3-6690-40c9-9f6f-d6953e3612e1 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52e09a85-0324-a299-c245-46c30113acba/disk-0.vmdk. {{(pid=63515) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1026.624385] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bdef330f-a82d-4e7d-8a79-e9fbb9643d0e {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.632348] env[63515]: DEBUG oslo_vmware.rw_handles [None req-878f8fb3-6690-40c9-9f6f-d6953e3612e1 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52e09a85-0324-a299-c245-46c30113acba/disk-0.vmdk is in state: ready. {{(pid=63515) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1026.632348] env[63515]: DEBUG oslo_vmware.rw_handles [None req-878f8fb3-6690-40c9-9f6f-d6953e3612e1 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Releasing lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52e09a85-0324-a299-c245-46c30113acba/disk-0.vmdk. {{(pid=63515) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 1026.632731] env[63515]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-f106f694-a205-4366-8130-89785eb71579 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.679090] env[63515]: DEBUG nova.network.neutron [req-8d378247-cbf0-4efe-9bbb-fdd08040ab39 req-9e8f3e27-f4b5-4fee-84a1-b11f58cee9b6 service nova] [instance: 724b05bc-ce4b-4f99-ad14-f3abea5567f2] Updated VIF entry in instance network info cache for port 6a98da26-06f3-4eb5-aee8-02a448c34749. {{(pid=63515) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1026.679687] env[63515]: DEBUG nova.network.neutron [req-8d378247-cbf0-4efe-9bbb-fdd08040ab39 req-9e8f3e27-f4b5-4fee-84a1-b11f58cee9b6 service nova] [instance: 724b05bc-ce4b-4f99-ad14-f3abea5567f2] Updating instance_info_cache with network_info: [{"id": "3dee43b6-2650-42e1-aa5a-4994a3ec7f05", "address": "fa:16:3e:89:06:d0", "network": {"id": "88dfb29d-41e0-4df7-8f70-e52cda05620c", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1355578400-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.146", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7d4ada9ec35f42b19c6480a9101d21a4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "357d2811-e990-4985-9f9e-b158d10d3699", "external-id": "nsx-vlan-transportzone-641", "segmentation_id": 641, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3dee43b6-26", "ovs_interfaceid": "3dee43b6-2650-42e1-aa5a-4994a3ec7f05", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "6a98da26-06f3-4eb5-aee8-02a448c34749", "address": "fa:16:3e:31:42:a5", "network": {"id": "88dfb29d-41e0-4df7-8f70-e52cda05620c", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1355578400-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7d4ada9ec35f42b19c6480a9101d21a4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "357d2811-e990-4985-9f9e-b158d10d3699", "external-id": "nsx-vlan-transportzone-641", "segmentation_id": 641, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6a98da26-06", "ovs_interfaceid": "6a98da26-06f3-4eb5-aee8-02a448c34749", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1026.736441] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-a7fc067e-c3eb-4fcd-8441-645f050b9570 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: b76cf3dd-7b7e-4e5d-bf3c-e15bd37069ac] Powering off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1026.736730] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2e9eec13-7fee-430f-8027-de2b0dcf2ac8 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.745113] env[63515]: DEBUG oslo_vmware.api [None req-a7fc067e-c3eb-4fcd-8441-645f050b9570 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Waiting for the task: (returnval){ [ 1026.745113] env[63515]: value = "task-1111838" [ 1026.745113] env[63515]: _type = "Task" [ 1026.745113] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1026.753540] env[63515]: DEBUG oslo_vmware.api [None req-a7fc067e-c3eb-4fcd-8441-645f050b9570 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Task: {'id': task-1111838, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1026.810574] env[63515]: DEBUG oslo_vmware.api [None req-042daf1f-2fe3-4b1d-bb04-2f7a0479fcba tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Task: {'id': task-1111837, 'name': PowerOnVM_Task, 'duration_secs': 0.569752} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1026.810574] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-042daf1f-2fe3-4b1d-bb04-2f7a0479fcba tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: e925fc93-a731-498a-984c-9b1f3eabf353] Powered on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1026.810886] env[63515]: INFO nova.compute.manager [None req-042daf1f-2fe3-4b1d-bb04-2f7a0479fcba tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: e925fc93-a731-498a-984c-9b1f3eabf353] Took 8.59 seconds to spawn the instance on the hypervisor. [ 1026.810886] env[63515]: DEBUG nova.compute.manager [None req-042daf1f-2fe3-4b1d-bb04-2f7a0479fcba tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: e925fc93-a731-498a-984c-9b1f3eabf353] Checking state {{(pid=63515) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1026.811884] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb099b36-9e9f-459c-ab90-926b4412e9d9 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.893221] env[63515]: DEBUG nova.network.neutron [req-ab9100c9-a065-42e7-85ec-67775eafbd05 req-b177c9cc-3bc0-48ec-bd0a-0b13d6b691a2 service nova] [instance: fcfe3376-8dfa-4189-a267-ff6402713c1b] Updated VIF entry in instance network info cache for port 4462b1aa-72ea-41cc-878a-68c38920c8cc. {{(pid=63515) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1026.893453] env[63515]: DEBUG nova.network.neutron [req-ab9100c9-a065-42e7-85ec-67775eafbd05 req-b177c9cc-3bc0-48ec-bd0a-0b13d6b691a2 service nova] [instance: fcfe3376-8dfa-4189-a267-ff6402713c1b] Updating instance_info_cache with network_info: [{"id": "4462b1aa-72ea-41cc-878a-68c38920c8cc", "address": "fa:16:3e:4f:bd:04", "network": {"id": "9fc2b9c9-6afc-4510-bf13-f87f1e1bd53b", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-513162532-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "0be5ebcf766b4da5b33be3a12387e737", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0248a27a-1d7f-4195-987b-06bfc8467347", "external-id": "nsx-vlan-transportzone-26", "segmentation_id": 26, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4462b1aa-72", "ovs_interfaceid": "4462b1aa-72ea-41cc-878a-68c38920c8cc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1026.897291] env[63515]: DEBUG nova.compute.manager [None req-123efa07-e143-4b76-a7d6-1e616f30576d tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: aa83d035-9c47-434d-ab1a-140e6b3110b4] Starting instance... {{(pid=63515) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1027.058577] env[63515]: DEBUG oslo_concurrency.lockutils [None req-80fab3e3-466b-468e-b003-05f4a2e89df7 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Lock "interface-724b05bc-ce4b-4f99-ad14-f3abea5567f2-6a98da26-06f3-4eb5-aee8-02a448c34749" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 6.123s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1027.183043] env[63515]: DEBUG oslo_concurrency.lockutils [req-8d378247-cbf0-4efe-9bbb-fdd08040ab39 req-9e8f3e27-f4b5-4fee-84a1-b11f58cee9b6 service nova] Releasing lock "refresh_cache-724b05bc-ce4b-4f99-ad14-f3abea5567f2" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1027.194887] env[63515]: DEBUG oslo_vmware.rw_handles [None req-878f8fb3-6690-40c9-9f6f-d6953e3612e1 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Closed VMDK write handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52e09a85-0324-a299-c245-46c30113acba/disk-0.vmdk. {{(pid=63515) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 1027.195244] env[63515]: INFO nova.virt.vmwareapi.images [None req-878f8fb3-6690-40c9-9f6f-d6953e3612e1 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] [instance: 7d572ac0-f6af-4622-96cc-e75983420222] Downloaded image file data cb0243fe-58ca-4459-930c-548f47818dfc [ 1027.196243] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05db8529-0ebe-4dcc-b230-ecd5962a56f3 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.214229] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-1adebddb-fd9a-46b2-a87f-736a0953eb57 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.254716] env[63515]: DEBUG oslo_vmware.api [None req-a7fc067e-c3eb-4fcd-8441-645f050b9570 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Task: {'id': task-1111838, 'name': PowerOffVM_Task} progress is 100%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1027.272956] env[63515]: INFO nova.virt.vmwareapi.images [None req-878f8fb3-6690-40c9-9f6f-d6953e3612e1 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] [instance: 7d572ac0-f6af-4622-96cc-e75983420222] The imported VM was unregistered [ 1027.275451] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-878f8fb3-6690-40c9-9f6f-d6953e3612e1 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] [instance: 7d572ac0-f6af-4622-96cc-e75983420222] Caching image {{(pid=63515) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 1027.275705] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-878f8fb3-6690-40c9-9f6f-d6953e3612e1 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Creating directory with path [datastore2] devstack-image-cache_base/cb0243fe-58ca-4459-930c-548f47818dfc {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1027.275993] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-361c8b7f-eda7-4a5e-bd07-3dc47598679d {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.302313] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-878f8fb3-6690-40c9-9f6f-d6953e3612e1 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Created directory with path [datastore2] devstack-image-cache_base/cb0243fe-58ca-4459-930c-548f47818dfc {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1027.302530] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-878f8fb3-6690-40c9-9f6f-d6953e3612e1 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Moving virtual disk from [datastore2] OSTACK_IMG_516d6119-fb9a-490f-9429-53b230ce63a9/OSTACK_IMG_516d6119-fb9a-490f-9429-53b230ce63a9.vmdk to [datastore2] devstack-image-cache_base/cb0243fe-58ca-4459-930c-548f47818dfc/cb0243fe-58ca-4459-930c-548f47818dfc.vmdk. {{(pid=63515) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 1027.302804] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-33886286-a5ce-4540-8220-2a5e09f34bc1 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.311669] env[63515]: DEBUG oslo_vmware.api [None req-878f8fb3-6690-40c9-9f6f-d6953e3612e1 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Waiting for the task: (returnval){ [ 1027.311669] env[63515]: value = "task-1111840" [ 1027.311669] env[63515]: _type = "Task" [ 1027.311669] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1027.325377] env[63515]: DEBUG oslo_vmware.api [None req-878f8fb3-6690-40c9-9f6f-d6953e3612e1 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Task: {'id': task-1111840, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1027.328178] env[63515]: INFO nova.compute.manager [None req-042daf1f-2fe3-4b1d-bb04-2f7a0479fcba tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: e925fc93-a731-498a-984c-9b1f3eabf353] Took 15.41 seconds to build instance. [ 1027.395938] env[63515]: DEBUG oslo_concurrency.lockutils [req-ab9100c9-a065-42e7-85ec-67775eafbd05 req-b177c9cc-3bc0-48ec-bd0a-0b13d6b691a2 service nova] Releasing lock "refresh_cache-fcfe3376-8dfa-4189-a267-ff6402713c1b" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1027.423753] env[63515]: DEBUG oslo_concurrency.lockutils [None req-123efa07-e143-4b76-a7d6-1e616f30576d tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1027.423819] env[63515]: DEBUG oslo_concurrency.lockutils [None req-123efa07-e143-4b76-a7d6-1e616f30576d tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1027.426072] env[63515]: INFO nova.compute.claims [None req-123efa07-e143-4b76-a7d6-1e616f30576d tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: aa83d035-9c47-434d-ab1a-140e6b3110b4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1027.755390] env[63515]: DEBUG oslo_vmware.api [None req-a7fc067e-c3eb-4fcd-8441-645f050b9570 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Task: {'id': task-1111838, 'name': PowerOffVM_Task, 'duration_secs': 0.522352} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1027.755663] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-a7fc067e-c3eb-4fcd-8441-645f050b9570 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: b76cf3dd-7b7e-4e5d-bf3c-e15bd37069ac] Powered off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1027.755850] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-a7fc067e-c3eb-4fcd-8441-645f050b9570 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: b76cf3dd-7b7e-4e5d-bf3c-e15bd37069ac] Updating instance 'b76cf3dd-7b7e-4e5d-bf3c-e15bd37069ac' progress to 17 {{(pid=63515) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1027.824270] env[63515]: DEBUG oslo_vmware.api [None req-878f8fb3-6690-40c9-9f6f-d6953e3612e1 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Task: {'id': task-1111840, 'name': MoveVirtualDisk_Task} progress is 9%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1027.831203] env[63515]: DEBUG oslo_concurrency.lockutils [None req-042daf1f-2fe3-4b1d-bb04-2f7a0479fcba tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Lock "e925fc93-a731-498a-984c-9b1f3eabf353" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 16.924s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1028.201223] env[63515]: DEBUG nova.compute.manager [req-a9550b49-292a-48c6-a575-f7bb823261e2 req-67219c49-85e2-422e-978f-d4011d15716f service nova] [instance: fcfe3376-8dfa-4189-a267-ff6402713c1b] Received event network-changed-4462b1aa-72ea-41cc-878a-68c38920c8cc {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1028.201490] env[63515]: DEBUG nova.compute.manager [req-a9550b49-292a-48c6-a575-f7bb823261e2 req-67219c49-85e2-422e-978f-d4011d15716f service nova] [instance: fcfe3376-8dfa-4189-a267-ff6402713c1b] Refreshing instance network info cache due to event network-changed-4462b1aa-72ea-41cc-878a-68c38920c8cc. {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1028.201674] env[63515]: DEBUG oslo_concurrency.lockutils [req-a9550b49-292a-48c6-a575-f7bb823261e2 req-67219c49-85e2-422e-978f-d4011d15716f service nova] Acquiring lock "refresh_cache-fcfe3376-8dfa-4189-a267-ff6402713c1b" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1028.201822] env[63515]: DEBUG oslo_concurrency.lockutils [req-a9550b49-292a-48c6-a575-f7bb823261e2 req-67219c49-85e2-422e-978f-d4011d15716f service nova] Acquired lock "refresh_cache-fcfe3376-8dfa-4189-a267-ff6402713c1b" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1028.201988] env[63515]: DEBUG nova.network.neutron [req-a9550b49-292a-48c6-a575-f7bb823261e2 req-67219c49-85e2-422e-978f-d4011d15716f service nova] [instance: fcfe3376-8dfa-4189-a267-ff6402713c1b] Refreshing network info cache for port 4462b1aa-72ea-41cc-878a-68c38920c8cc {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1028.264227] env[63515]: DEBUG nova.virt.hardware [None req-a7fc067e-c3eb-4fcd-8441-645f050b9570 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T02:52:16Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=8a120570-cb06-4099-b262-554ca0ad15c5,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1028.264575] env[63515]: DEBUG nova.virt.hardware [None req-a7fc067e-c3eb-4fcd-8441-645f050b9570 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Flavor limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1028.264746] env[63515]: DEBUG nova.virt.hardware [None req-a7fc067e-c3eb-4fcd-8441-645f050b9570 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Image limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1028.264936] env[63515]: DEBUG nova.virt.hardware [None req-a7fc067e-c3eb-4fcd-8441-645f050b9570 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Flavor pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1028.265102] env[63515]: DEBUG nova.virt.hardware [None req-a7fc067e-c3eb-4fcd-8441-645f050b9570 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Image pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1028.265257] env[63515]: DEBUG nova.virt.hardware [None req-a7fc067e-c3eb-4fcd-8441-645f050b9570 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1028.266054] env[63515]: DEBUG nova.virt.hardware [None req-a7fc067e-c3eb-4fcd-8441-645f050b9570 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1028.266054] env[63515]: DEBUG nova.virt.hardware [None req-a7fc067e-c3eb-4fcd-8441-645f050b9570 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1028.266054] env[63515]: DEBUG nova.virt.hardware [None req-a7fc067e-c3eb-4fcd-8441-645f050b9570 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Got 1 possible topologies {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1028.266054] env[63515]: DEBUG nova.virt.hardware [None req-a7fc067e-c3eb-4fcd-8441-645f050b9570 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1028.266251] env[63515]: DEBUG nova.virt.hardware [None req-a7fc067e-c3eb-4fcd-8441-645f050b9570 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1028.271996] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-538f7454-f1b9-4baa-9660-632b313af2a3 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.292593] env[63515]: DEBUG oslo_vmware.api [None req-a7fc067e-c3eb-4fcd-8441-645f050b9570 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Waiting for the task: (returnval){ [ 1028.292593] env[63515]: value = "task-1111841" [ 1028.292593] env[63515]: _type = "Task" [ 1028.292593] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1028.303042] env[63515]: DEBUG oslo_vmware.api [None req-a7fc067e-c3eb-4fcd-8441-645f050b9570 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Task: {'id': task-1111841, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.324582] env[63515]: DEBUG oslo_vmware.api [None req-878f8fb3-6690-40c9-9f6f-d6953e3612e1 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Task: {'id': task-1111840, 'name': MoveVirtualDisk_Task} progress is 24%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.406033] env[63515]: DEBUG oslo_concurrency.lockutils [None req-77b6ad21-5821-4c19-9cf4-41738b67053e tempest-ServerRescueTestJSONUnderV235-800430211 tempest-ServerRescueTestJSONUnderV235-800430211-project-member] Acquiring lock "fcfe3376-8dfa-4189-a267-ff6402713c1b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1028.406332] env[63515]: DEBUG oslo_concurrency.lockutils [None req-77b6ad21-5821-4c19-9cf4-41738b67053e tempest-ServerRescueTestJSONUnderV235-800430211 tempest-ServerRescueTestJSONUnderV235-800430211-project-member] Lock "fcfe3376-8dfa-4189-a267-ff6402713c1b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1028.406553] env[63515]: DEBUG oslo_concurrency.lockutils [None req-77b6ad21-5821-4c19-9cf4-41738b67053e tempest-ServerRescueTestJSONUnderV235-800430211 tempest-ServerRescueTestJSONUnderV235-800430211-project-member] Acquiring lock "fcfe3376-8dfa-4189-a267-ff6402713c1b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1028.406872] env[63515]: DEBUG oslo_concurrency.lockutils [None req-77b6ad21-5821-4c19-9cf4-41738b67053e tempest-ServerRescueTestJSONUnderV235-800430211 tempest-ServerRescueTestJSONUnderV235-800430211-project-member] Lock "fcfe3376-8dfa-4189-a267-ff6402713c1b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1028.407092] env[63515]: DEBUG oslo_concurrency.lockutils [None req-77b6ad21-5821-4c19-9cf4-41738b67053e tempest-ServerRescueTestJSONUnderV235-800430211 tempest-ServerRescueTestJSONUnderV235-800430211-project-member] Lock "fcfe3376-8dfa-4189-a267-ff6402713c1b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1028.410793] env[63515]: INFO nova.compute.manager [None req-77b6ad21-5821-4c19-9cf4-41738b67053e tempest-ServerRescueTestJSONUnderV235-800430211 tempest-ServerRescueTestJSONUnderV235-800430211-project-member] [instance: fcfe3376-8dfa-4189-a267-ff6402713c1b] Terminating instance [ 1028.413371] env[63515]: DEBUG nova.compute.manager [None req-77b6ad21-5821-4c19-9cf4-41738b67053e tempest-ServerRescueTestJSONUnderV235-800430211 tempest-ServerRescueTestJSONUnderV235-800430211-project-member] [instance: fcfe3376-8dfa-4189-a267-ff6402713c1b] Start destroying the instance on the hypervisor. {{(pid=63515) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1028.414457] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-77b6ad21-5821-4c19-9cf4-41738b67053e tempest-ServerRescueTestJSONUnderV235-800430211 tempest-ServerRescueTestJSONUnderV235-800430211-project-member] [instance: fcfe3376-8dfa-4189-a267-ff6402713c1b] Destroying instance {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1028.414576] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f9f3ed9-ff0e-4e09-ac74-8af547aa257e {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.423914] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-77b6ad21-5821-4c19-9cf4-41738b67053e tempest-ServerRescueTestJSONUnderV235-800430211 tempest-ServerRescueTestJSONUnderV235-800430211-project-member] [instance: fcfe3376-8dfa-4189-a267-ff6402713c1b] Powering off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1028.424480] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a678c455-0126-4380-8bbd-d38b1079ecce {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.430719] env[63515]: DEBUG oslo_vmware.api [None req-77b6ad21-5821-4c19-9cf4-41738b67053e tempest-ServerRescueTestJSONUnderV235-800430211 tempest-ServerRescueTestJSONUnderV235-800430211-project-member] Waiting for the task: (returnval){ [ 1028.430719] env[63515]: value = "task-1111842" [ 1028.430719] env[63515]: _type = "Task" [ 1028.430719] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1028.442725] env[63515]: DEBUG oslo_vmware.api [None req-77b6ad21-5821-4c19-9cf4-41738b67053e tempest-ServerRescueTestJSONUnderV235-800430211 tempest-ServerRescueTestJSONUnderV235-800430211-project-member] Task: {'id': task-1111842, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.597639] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d4d4d79-8536-456e-8c69-f416ec1d2980 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.605488] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-990032f3-47c2-41db-8f31-bddeee1efaa1 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.645673] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad7a228c-e687-4597-8b70-0e363c343cbb {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.654904] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23326c0b-72ec-4b16-8554-7c8f6953f462 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.673311] env[63515]: DEBUG nova.compute.provider_tree [None req-123efa07-e143-4b76-a7d6-1e616f30576d tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1028.788284] env[63515]: DEBUG oslo_concurrency.lockutils [None req-5a78568a-9d2c-4ab4-aaf0-b4782bd16af7 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Acquiring lock "interface-724b05bc-ce4b-4f99-ad14-f3abea5567f2-6a98da26-06f3-4eb5-aee8-02a448c34749" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1028.788581] env[63515]: DEBUG oslo_concurrency.lockutils [None req-5a78568a-9d2c-4ab4-aaf0-b4782bd16af7 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Lock "interface-724b05bc-ce4b-4f99-ad14-f3abea5567f2-6a98da26-06f3-4eb5-aee8-02a448c34749" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.001s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1028.806482] env[63515]: DEBUG oslo_vmware.api [None req-a7fc067e-c3eb-4fcd-8441-645f050b9570 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Task: {'id': task-1111841, 'name': ReconfigVM_Task, 'duration_secs': 0.212516} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1028.807137] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-a7fc067e-c3eb-4fcd-8441-645f050b9570 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: b76cf3dd-7b7e-4e5d-bf3c-e15bd37069ac] Updating instance 'b76cf3dd-7b7e-4e5d-bf3c-e15bd37069ac' progress to 33 {{(pid=63515) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1028.830696] env[63515]: DEBUG oslo_vmware.api [None req-878f8fb3-6690-40c9-9f6f-d6953e3612e1 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Task: {'id': task-1111840, 'name': MoveVirtualDisk_Task} progress is 46%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.853519] env[63515]: DEBUG oslo_concurrency.lockutils [None req-fd066576-98f6-4ad4-9fd9-bbb5e5a27c21 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Acquiring lock "e925fc93-a731-498a-984c-9b1f3eabf353" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1028.853942] env[63515]: DEBUG oslo_concurrency.lockutils [None req-fd066576-98f6-4ad4-9fd9-bbb5e5a27c21 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Lock "e925fc93-a731-498a-984c-9b1f3eabf353" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1028.854386] env[63515]: DEBUG oslo_concurrency.lockutils [None req-fd066576-98f6-4ad4-9fd9-bbb5e5a27c21 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Acquiring lock "e925fc93-a731-498a-984c-9b1f3eabf353-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1028.854759] env[63515]: DEBUG oslo_concurrency.lockutils [None req-fd066576-98f6-4ad4-9fd9-bbb5e5a27c21 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Lock "e925fc93-a731-498a-984c-9b1f3eabf353-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1028.855096] env[63515]: DEBUG oslo_concurrency.lockutils [None req-fd066576-98f6-4ad4-9fd9-bbb5e5a27c21 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Lock "e925fc93-a731-498a-984c-9b1f3eabf353-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1028.859681] env[63515]: INFO nova.compute.manager [None req-fd066576-98f6-4ad4-9fd9-bbb5e5a27c21 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: e925fc93-a731-498a-984c-9b1f3eabf353] Terminating instance [ 1028.862860] env[63515]: DEBUG nova.compute.manager [None req-fd066576-98f6-4ad4-9fd9-bbb5e5a27c21 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: e925fc93-a731-498a-984c-9b1f3eabf353] Start destroying the instance on the hypervisor. {{(pid=63515) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1028.863259] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-fd066576-98f6-4ad4-9fd9-bbb5e5a27c21 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: e925fc93-a731-498a-984c-9b1f3eabf353] Destroying instance {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1028.864930] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10db5a40-c0e8-4969-a9b8-c4c93464ce35 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.877811] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-fd066576-98f6-4ad4-9fd9-bbb5e5a27c21 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: e925fc93-a731-498a-984c-9b1f3eabf353] Powering off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1028.878283] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7ded82d8-680f-4b04-b4c7-f8fd9270fca2 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.887404] env[63515]: DEBUG oslo_vmware.api [None req-fd066576-98f6-4ad4-9fd9-bbb5e5a27c21 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Waiting for the task: (returnval){ [ 1028.887404] env[63515]: value = "task-1111843" [ 1028.887404] env[63515]: _type = "Task" [ 1028.887404] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1028.903022] env[63515]: DEBUG oslo_vmware.api [None req-fd066576-98f6-4ad4-9fd9-bbb5e5a27c21 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Task: {'id': task-1111843, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.942190] env[63515]: DEBUG oslo_vmware.api [None req-77b6ad21-5821-4c19-9cf4-41738b67053e tempest-ServerRescueTestJSONUnderV235-800430211 tempest-ServerRescueTestJSONUnderV235-800430211-project-member] Task: {'id': task-1111842, 'name': PowerOffVM_Task, 'duration_secs': 0.402334} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1028.943200] env[63515]: DEBUG nova.network.neutron [req-a9550b49-292a-48c6-a575-f7bb823261e2 req-67219c49-85e2-422e-978f-d4011d15716f service nova] [instance: fcfe3376-8dfa-4189-a267-ff6402713c1b] Updated VIF entry in instance network info cache for port 4462b1aa-72ea-41cc-878a-68c38920c8cc. {{(pid=63515) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1028.943562] env[63515]: DEBUG nova.network.neutron [req-a9550b49-292a-48c6-a575-f7bb823261e2 req-67219c49-85e2-422e-978f-d4011d15716f service nova] [instance: fcfe3376-8dfa-4189-a267-ff6402713c1b] Updating instance_info_cache with network_info: [{"id": "4462b1aa-72ea-41cc-878a-68c38920c8cc", "address": "fa:16:3e:4f:bd:04", "network": {"id": "9fc2b9c9-6afc-4510-bf13-f87f1e1bd53b", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-513162532-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "0be5ebcf766b4da5b33be3a12387e737", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0248a27a-1d7f-4195-987b-06bfc8467347", "external-id": "nsx-vlan-transportzone-26", "segmentation_id": 26, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4462b1aa-72", "ovs_interfaceid": "4462b1aa-72ea-41cc-878a-68c38920c8cc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1028.946144] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-77b6ad21-5821-4c19-9cf4-41738b67053e tempest-ServerRescueTestJSONUnderV235-800430211 tempest-ServerRescueTestJSONUnderV235-800430211-project-member] [instance: fcfe3376-8dfa-4189-a267-ff6402713c1b] Powered off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1028.946382] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-77b6ad21-5821-4c19-9cf4-41738b67053e tempest-ServerRescueTestJSONUnderV235-800430211 tempest-ServerRescueTestJSONUnderV235-800430211-project-member] [instance: fcfe3376-8dfa-4189-a267-ff6402713c1b] Unregistering the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1028.946971] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-093a5900-aba4-4b5a-96a6-47f038a560dd {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.012023] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-77b6ad21-5821-4c19-9cf4-41738b67053e tempest-ServerRescueTestJSONUnderV235-800430211 tempest-ServerRescueTestJSONUnderV235-800430211-project-member] [instance: fcfe3376-8dfa-4189-a267-ff6402713c1b] Unregistered the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1029.012297] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-77b6ad21-5821-4c19-9cf4-41738b67053e tempest-ServerRescueTestJSONUnderV235-800430211 tempest-ServerRescueTestJSONUnderV235-800430211-project-member] [instance: fcfe3376-8dfa-4189-a267-ff6402713c1b] Deleting contents of the VM from datastore datastore1 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1029.012504] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-77b6ad21-5821-4c19-9cf4-41738b67053e tempest-ServerRescueTestJSONUnderV235-800430211 tempest-ServerRescueTestJSONUnderV235-800430211-project-member] Deleting the datastore file [datastore1] fcfe3376-8dfa-4189-a267-ff6402713c1b {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1029.012780] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-93ff26d5-c95d-4e84-914c-7485ca33edf9 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.021872] env[63515]: DEBUG oslo_vmware.api [None req-77b6ad21-5821-4c19-9cf4-41738b67053e tempest-ServerRescueTestJSONUnderV235-800430211 tempest-ServerRescueTestJSONUnderV235-800430211-project-member] Waiting for the task: (returnval){ [ 1029.021872] env[63515]: value = "task-1111845" [ 1029.021872] env[63515]: _type = "Task" [ 1029.021872] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1029.030352] env[63515]: DEBUG oslo_vmware.api [None req-77b6ad21-5821-4c19-9cf4-41738b67053e tempest-ServerRescueTestJSONUnderV235-800430211 tempest-ServerRescueTestJSONUnderV235-800430211-project-member] Task: {'id': task-1111845, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1029.178085] env[63515]: DEBUG nova.scheduler.client.report [None req-123efa07-e143-4b76-a7d6-1e616f30576d tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1029.293997] env[63515]: DEBUG oslo_concurrency.lockutils [None req-5a78568a-9d2c-4ab4-aaf0-b4782bd16af7 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Acquiring lock "724b05bc-ce4b-4f99-ad14-f3abea5567f2" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1029.294364] env[63515]: DEBUG oslo_concurrency.lockutils [None req-5a78568a-9d2c-4ab4-aaf0-b4782bd16af7 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Acquired lock "724b05bc-ce4b-4f99-ad14-f3abea5567f2" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1029.295694] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72ecb12d-7799-4e5b-b8d5-5b11290a3fdc {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.316710] env[63515]: DEBUG nova.virt.hardware [None req-a7fc067e-c3eb-4fcd-8441-645f050b9570 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T02:52:14Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=8a120570-cb06-4099-b262-554ca0ad15c5,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1029.317048] env[63515]: DEBUG nova.virt.hardware [None req-a7fc067e-c3eb-4fcd-8441-645f050b9570 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Flavor limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1029.317263] env[63515]: DEBUG nova.virt.hardware [None req-a7fc067e-c3eb-4fcd-8441-645f050b9570 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Image limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1029.317466] env[63515]: DEBUG nova.virt.hardware [None req-a7fc067e-c3eb-4fcd-8441-645f050b9570 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Flavor pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1029.317622] env[63515]: DEBUG nova.virt.hardware [None req-a7fc067e-c3eb-4fcd-8441-645f050b9570 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Image pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1029.317775] env[63515]: DEBUG nova.virt.hardware [None req-a7fc067e-c3eb-4fcd-8441-645f050b9570 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1029.318088] env[63515]: DEBUG nova.virt.hardware [None req-a7fc067e-c3eb-4fcd-8441-645f050b9570 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1029.318296] env[63515]: DEBUG nova.virt.hardware [None req-a7fc067e-c3eb-4fcd-8441-645f050b9570 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1029.318547] env[63515]: DEBUG nova.virt.hardware [None req-a7fc067e-c3eb-4fcd-8441-645f050b9570 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Got 1 possible topologies {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1029.318747] env[63515]: DEBUG nova.virt.hardware [None req-a7fc067e-c3eb-4fcd-8441-645f050b9570 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1029.320160] env[63515]: DEBUG nova.virt.hardware [None req-a7fc067e-c3eb-4fcd-8441-645f050b9570 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1029.324399] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-a7fc067e-c3eb-4fcd-8441-645f050b9570 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: b76cf3dd-7b7e-4e5d-bf3c-e15bd37069ac] Reconfiguring VM instance instance-00000064 to detach disk 2000 {{(pid=63515) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1029.324956] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-346bfd6e-cc25-4af5-a25f-778fa0564fe2 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.342090] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca29a0ea-352a-4c26-8e11-3d0b2efd7e69 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.369636] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-5a78568a-9d2c-4ab4-aaf0-b4782bd16af7 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: 724b05bc-ce4b-4f99-ad14-f3abea5567f2] Reconfiguring VM to detach interface {{(pid=63515) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1968}} [ 1029.374477] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c024668d-0e8f-400e-85ae-b2c49ea0e179 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.388196] env[63515]: DEBUG oslo_vmware.api [None req-878f8fb3-6690-40c9-9f6f-d6953e3612e1 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Task: {'id': task-1111840, 'name': MoveVirtualDisk_Task} progress is 69%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1029.388563] env[63515]: DEBUG oslo_vmware.api [None req-a7fc067e-c3eb-4fcd-8441-645f050b9570 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Waiting for the task: (returnval){ [ 1029.388563] env[63515]: value = "task-1111846" [ 1029.388563] env[63515]: _type = "Task" [ 1029.388563] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1029.397291] env[63515]: DEBUG oslo_vmware.api [None req-5a78568a-9d2c-4ab4-aaf0-b4782bd16af7 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Waiting for the task: (returnval){ [ 1029.397291] env[63515]: value = "task-1111847" [ 1029.397291] env[63515]: _type = "Task" [ 1029.397291] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1029.400783] env[63515]: DEBUG oslo_vmware.api [None req-a7fc067e-c3eb-4fcd-8441-645f050b9570 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Task: {'id': task-1111846, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1029.406820] env[63515]: DEBUG oslo_vmware.api [None req-fd066576-98f6-4ad4-9fd9-bbb5e5a27c21 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Task: {'id': task-1111843, 'name': PowerOffVM_Task, 'duration_secs': 0.243521} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1029.407187] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-fd066576-98f6-4ad4-9fd9-bbb5e5a27c21 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: e925fc93-a731-498a-984c-9b1f3eabf353] Powered off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1029.407404] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-fd066576-98f6-4ad4-9fd9-bbb5e5a27c21 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: e925fc93-a731-498a-984c-9b1f3eabf353] Unregistering the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1029.408077] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-15593b8e-831a-43e2-b15a-c61527786f78 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.413831] env[63515]: DEBUG oslo_vmware.api [None req-5a78568a-9d2c-4ab4-aaf0-b4782bd16af7 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Task: {'id': task-1111847, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1029.448150] env[63515]: DEBUG oslo_concurrency.lockutils [req-a9550b49-292a-48c6-a575-f7bb823261e2 req-67219c49-85e2-422e-978f-d4011d15716f service nova] Releasing lock "refresh_cache-fcfe3376-8dfa-4189-a267-ff6402713c1b" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1029.448556] env[63515]: DEBUG nova.compute.manager [req-a9550b49-292a-48c6-a575-f7bb823261e2 req-67219c49-85e2-422e-978f-d4011d15716f service nova] [instance: fcfe3376-8dfa-4189-a267-ff6402713c1b] Received event network-changed-4462b1aa-72ea-41cc-878a-68c38920c8cc {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1029.448822] env[63515]: DEBUG nova.compute.manager [req-a9550b49-292a-48c6-a575-f7bb823261e2 req-67219c49-85e2-422e-978f-d4011d15716f service nova] [instance: fcfe3376-8dfa-4189-a267-ff6402713c1b] Refreshing instance network info cache due to event network-changed-4462b1aa-72ea-41cc-878a-68c38920c8cc. {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1029.449182] env[63515]: DEBUG oslo_concurrency.lockutils [req-a9550b49-292a-48c6-a575-f7bb823261e2 req-67219c49-85e2-422e-978f-d4011d15716f service nova] Acquiring lock "refresh_cache-fcfe3376-8dfa-4189-a267-ff6402713c1b" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1029.449364] env[63515]: DEBUG oslo_concurrency.lockutils [req-a9550b49-292a-48c6-a575-f7bb823261e2 req-67219c49-85e2-422e-978f-d4011d15716f service nova] Acquired lock "refresh_cache-fcfe3376-8dfa-4189-a267-ff6402713c1b" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1029.449596] env[63515]: DEBUG nova.network.neutron [req-a9550b49-292a-48c6-a575-f7bb823261e2 req-67219c49-85e2-422e-978f-d4011d15716f service nova] [instance: fcfe3376-8dfa-4189-a267-ff6402713c1b] Refreshing network info cache for port 4462b1aa-72ea-41cc-878a-68c38920c8cc {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1029.473223] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-fd066576-98f6-4ad4-9fd9-bbb5e5a27c21 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: e925fc93-a731-498a-984c-9b1f3eabf353] Unregistered the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1029.473532] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-fd066576-98f6-4ad4-9fd9-bbb5e5a27c21 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: e925fc93-a731-498a-984c-9b1f3eabf353] Deleting contents of the VM from datastore datastore1 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1029.473844] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-fd066576-98f6-4ad4-9fd9-bbb5e5a27c21 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Deleting the datastore file [datastore1] e925fc93-a731-498a-984c-9b1f3eabf353 {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1029.474185] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7ac78afe-26e2-4cf5-9921-0e4da1eb329a {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.482483] env[63515]: DEBUG oslo_vmware.api [None req-fd066576-98f6-4ad4-9fd9-bbb5e5a27c21 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Waiting for the task: (returnval){ [ 1029.482483] env[63515]: value = "task-1111849" [ 1029.482483] env[63515]: _type = "Task" [ 1029.482483] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1029.492192] env[63515]: DEBUG oslo_vmware.api [None req-fd066576-98f6-4ad4-9fd9-bbb5e5a27c21 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Task: {'id': task-1111849, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1029.533142] env[63515]: DEBUG oslo_vmware.api [None req-77b6ad21-5821-4c19-9cf4-41738b67053e tempest-ServerRescueTestJSONUnderV235-800430211 tempest-ServerRescueTestJSONUnderV235-800430211-project-member] Task: {'id': task-1111845, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.34228} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1029.533499] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-77b6ad21-5821-4c19-9cf4-41738b67053e tempest-ServerRescueTestJSONUnderV235-800430211 tempest-ServerRescueTestJSONUnderV235-800430211-project-member] Deleted the datastore file {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1029.533770] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-77b6ad21-5821-4c19-9cf4-41738b67053e tempest-ServerRescueTestJSONUnderV235-800430211 tempest-ServerRescueTestJSONUnderV235-800430211-project-member] [instance: fcfe3376-8dfa-4189-a267-ff6402713c1b] Deleted contents of the VM from datastore datastore1 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1029.534043] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-77b6ad21-5821-4c19-9cf4-41738b67053e tempest-ServerRescueTestJSONUnderV235-800430211 tempest-ServerRescueTestJSONUnderV235-800430211-project-member] [instance: fcfe3376-8dfa-4189-a267-ff6402713c1b] Instance destroyed {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1029.534263] env[63515]: INFO nova.compute.manager [None req-77b6ad21-5821-4c19-9cf4-41738b67053e tempest-ServerRescueTestJSONUnderV235-800430211 tempest-ServerRescueTestJSONUnderV235-800430211-project-member] [instance: fcfe3376-8dfa-4189-a267-ff6402713c1b] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1029.534590] env[63515]: DEBUG oslo.service.loopingcall [None req-77b6ad21-5821-4c19-9cf4-41738b67053e tempest-ServerRescueTestJSONUnderV235-800430211 tempest-ServerRescueTestJSONUnderV235-800430211-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1029.534806] env[63515]: DEBUG nova.compute.manager [-] [instance: fcfe3376-8dfa-4189-a267-ff6402713c1b] Deallocating network for instance {{(pid=63515) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1029.534959] env[63515]: DEBUG nova.network.neutron [-] [instance: fcfe3376-8dfa-4189-a267-ff6402713c1b] deallocate_for_instance() {{(pid=63515) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1029.686184] env[63515]: DEBUG oslo_concurrency.lockutils [None req-123efa07-e143-4b76-a7d6-1e616f30576d tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.262s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1029.687075] env[63515]: DEBUG nova.compute.manager [None req-123efa07-e143-4b76-a7d6-1e616f30576d tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: aa83d035-9c47-434d-ab1a-140e6b3110b4] Start building networks asynchronously for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1029.835436] env[63515]: DEBUG oslo_vmware.api [None req-878f8fb3-6690-40c9-9f6f-d6953e3612e1 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Task: {'id': task-1111840, 'name': MoveVirtualDisk_Task} progress is 88%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1029.898463] env[63515]: DEBUG oslo_vmware.api [None req-a7fc067e-c3eb-4fcd-8441-645f050b9570 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Task: {'id': task-1111846, 'name': ReconfigVM_Task, 'duration_secs': 0.243935} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1029.898771] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-a7fc067e-c3eb-4fcd-8441-645f050b9570 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: b76cf3dd-7b7e-4e5d-bf3c-e15bd37069ac] Reconfigured VM instance instance-00000064 to detach disk 2000 {{(pid=63515) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1029.899710] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99d4c5fc-c359-4650-99dd-61c09f7ba132 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.925964] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-a7fc067e-c3eb-4fcd-8441-645f050b9570 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: b76cf3dd-7b7e-4e5d-bf3c-e15bd37069ac] Reconfiguring VM instance instance-00000064 to attach disk [datastore1] b76cf3dd-7b7e-4e5d-bf3c-e15bd37069ac/b76cf3dd-7b7e-4e5d-bf3c-e15bd37069ac.vmdk or device None with type thin {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1029.929675] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-232c2b42-395f-4831-b696-cec4a5577246 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.943670] env[63515]: DEBUG oslo_vmware.api [None req-5a78568a-9d2c-4ab4-aaf0-b4782bd16af7 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Task: {'id': task-1111847, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1029.951574] env[63515]: DEBUG oslo_vmware.api [None req-a7fc067e-c3eb-4fcd-8441-645f050b9570 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Waiting for the task: (returnval){ [ 1029.951574] env[63515]: value = "task-1111850" [ 1029.951574] env[63515]: _type = "Task" [ 1029.951574] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1029.963592] env[63515]: DEBUG oslo_vmware.api [None req-a7fc067e-c3eb-4fcd-8441-645f050b9570 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Task: {'id': task-1111850, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1029.972238] env[63515]: INFO nova.network.neutron [req-a9550b49-292a-48c6-a575-f7bb823261e2 req-67219c49-85e2-422e-978f-d4011d15716f service nova] [instance: fcfe3376-8dfa-4189-a267-ff6402713c1b] Port 4462b1aa-72ea-41cc-878a-68c38920c8cc from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 1029.972507] env[63515]: DEBUG nova.network.neutron [req-a9550b49-292a-48c6-a575-f7bb823261e2 req-67219c49-85e2-422e-978f-d4011d15716f service nova] [instance: fcfe3376-8dfa-4189-a267-ff6402713c1b] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1029.995086] env[63515]: DEBUG oslo_vmware.api [None req-fd066576-98f6-4ad4-9fd9-bbb5e5a27c21 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Task: {'id': task-1111849, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.278822} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1029.995406] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-fd066576-98f6-4ad4-9fd9-bbb5e5a27c21 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Deleted the datastore file {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1029.995615] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-fd066576-98f6-4ad4-9fd9-bbb5e5a27c21 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: e925fc93-a731-498a-984c-9b1f3eabf353] Deleted contents of the VM from datastore datastore1 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1029.995968] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-fd066576-98f6-4ad4-9fd9-bbb5e5a27c21 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: e925fc93-a731-498a-984c-9b1f3eabf353] Instance destroyed {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1029.996099] env[63515]: INFO nova.compute.manager [None req-fd066576-98f6-4ad4-9fd9-bbb5e5a27c21 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: e925fc93-a731-498a-984c-9b1f3eabf353] Took 1.13 seconds to destroy the instance on the hypervisor. [ 1029.996293] env[63515]: DEBUG oslo.service.loopingcall [None req-fd066576-98f6-4ad4-9fd9-bbb5e5a27c21 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1029.996535] env[63515]: DEBUG nova.compute.manager [-] [instance: e925fc93-a731-498a-984c-9b1f3eabf353] Deallocating network for instance {{(pid=63515) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1029.996657] env[63515]: DEBUG nova.network.neutron [-] [instance: e925fc93-a731-498a-984c-9b1f3eabf353] deallocate_for_instance() {{(pid=63515) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1030.198036] env[63515]: DEBUG nova.compute.utils [None req-123efa07-e143-4b76-a7d6-1e616f30576d tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Using /dev/sd instead of None {{(pid=63515) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1030.198036] env[63515]: DEBUG nova.compute.manager [None req-123efa07-e143-4b76-a7d6-1e616f30576d tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: aa83d035-9c47-434d-ab1a-140e6b3110b4] Allocating IP information in the background. {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1030.198036] env[63515]: DEBUG nova.network.neutron [None req-123efa07-e143-4b76-a7d6-1e616f30576d tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: aa83d035-9c47-434d-ab1a-140e6b3110b4] allocate_for_instance() {{(pid=63515) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1030.232190] env[63515]: DEBUG nova.compute.manager [req-04cf6a48-8758-4562-b3d3-f2a2e2f12c40 req-63c938a9-b266-4e87-b85e-0937b828aea3 service nova] [instance: fcfe3376-8dfa-4189-a267-ff6402713c1b] Received event network-vif-deleted-4462b1aa-72ea-41cc-878a-68c38920c8cc {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1030.240674] env[63515]: DEBUG nova.policy [None req-123efa07-e143-4b76-a7d6-1e616f30576d tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '82fe830c309f41a4ab512a25d8481742', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b18d4daacbc84e758f9f9ae4f3f09d28', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63515) authorize /opt/stack/nova/nova/policy.py:201}} [ 1030.263967] env[63515]: DEBUG nova.network.neutron [-] [instance: fcfe3376-8dfa-4189-a267-ff6402713c1b] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1030.335796] env[63515]: DEBUG oslo_vmware.api [None req-878f8fb3-6690-40c9-9f6f-d6953e3612e1 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Task: {'id': task-1111840, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.803867} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1030.336035] env[63515]: INFO nova.virt.vmwareapi.ds_util [None req-878f8fb3-6690-40c9-9f6f-d6953e3612e1 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Moved virtual disk from [datastore2] OSTACK_IMG_516d6119-fb9a-490f-9429-53b230ce63a9/OSTACK_IMG_516d6119-fb9a-490f-9429-53b230ce63a9.vmdk to [datastore2] devstack-image-cache_base/cb0243fe-58ca-4459-930c-548f47818dfc/cb0243fe-58ca-4459-930c-548f47818dfc.vmdk. [ 1030.336229] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-878f8fb3-6690-40c9-9f6f-d6953e3612e1 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] [instance: 7d572ac0-f6af-4622-96cc-e75983420222] Cleaning up location [datastore2] OSTACK_IMG_516d6119-fb9a-490f-9429-53b230ce63a9 {{(pid=63515) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 1030.336386] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-878f8fb3-6690-40c9-9f6f-d6953e3612e1 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Deleting the datastore file [datastore2] OSTACK_IMG_516d6119-fb9a-490f-9429-53b230ce63a9 {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1030.336769] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-eec35e37-f4c8-42fa-98db-1d0c64c7c260 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.342811] env[63515]: DEBUG oslo_vmware.api [None req-878f8fb3-6690-40c9-9f6f-d6953e3612e1 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Waiting for the task: (returnval){ [ 1030.342811] env[63515]: value = "task-1111851" [ 1030.342811] env[63515]: _type = "Task" [ 1030.342811] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1030.350457] env[63515]: DEBUG oslo_vmware.api [None req-878f8fb3-6690-40c9-9f6f-d6953e3612e1 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Task: {'id': task-1111851, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.413435] env[63515]: DEBUG oslo_vmware.api [None req-5a78568a-9d2c-4ab4-aaf0-b4782bd16af7 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Task: {'id': task-1111847, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.463309] env[63515]: DEBUG oslo_vmware.api [None req-a7fc067e-c3eb-4fcd-8441-645f050b9570 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Task: {'id': task-1111850, 'name': ReconfigVM_Task, 'duration_secs': 0.316811} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1030.463695] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-a7fc067e-c3eb-4fcd-8441-645f050b9570 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: b76cf3dd-7b7e-4e5d-bf3c-e15bd37069ac] Reconfigured VM instance instance-00000064 to attach disk [datastore1] b76cf3dd-7b7e-4e5d-bf3c-e15bd37069ac/b76cf3dd-7b7e-4e5d-bf3c-e15bd37069ac.vmdk or device None with type thin {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1030.464100] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-a7fc067e-c3eb-4fcd-8441-645f050b9570 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: b76cf3dd-7b7e-4e5d-bf3c-e15bd37069ac] Updating instance 'b76cf3dd-7b7e-4e5d-bf3c-e15bd37069ac' progress to 50 {{(pid=63515) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1030.475323] env[63515]: DEBUG oslo_concurrency.lockutils [req-a9550b49-292a-48c6-a575-f7bb823261e2 req-67219c49-85e2-422e-978f-d4011d15716f service nova] Releasing lock "refresh_cache-fcfe3376-8dfa-4189-a267-ff6402713c1b" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1030.559418] env[63515]: DEBUG nova.network.neutron [None req-123efa07-e143-4b76-a7d6-1e616f30576d tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: aa83d035-9c47-434d-ab1a-140e6b3110b4] Successfully created port: 614aed8f-3ab8-4672-ac13-0ae8cb0b233d {{(pid=63515) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1030.699662] env[63515]: DEBUG nova.compute.manager [None req-123efa07-e143-4b76-a7d6-1e616f30576d tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: aa83d035-9c47-434d-ab1a-140e6b3110b4] Start building block device mappings for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1030.714410] env[63515]: DEBUG nova.network.neutron [-] [instance: e925fc93-a731-498a-984c-9b1f3eabf353] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1030.767525] env[63515]: INFO nova.compute.manager [-] [instance: fcfe3376-8dfa-4189-a267-ff6402713c1b] Took 1.23 seconds to deallocate network for instance. [ 1030.853346] env[63515]: DEBUG oslo_vmware.api [None req-878f8fb3-6690-40c9-9f6f-d6953e3612e1 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Task: {'id': task-1111851, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.034252} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1030.853560] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-878f8fb3-6690-40c9-9f6f-d6953e3612e1 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Deleted the datastore file {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1030.853736] env[63515]: DEBUG oslo_concurrency.lockutils [None req-878f8fb3-6690-40c9-9f6f-d6953e3612e1 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Releasing lock "[datastore2] devstack-image-cache_base/cb0243fe-58ca-4459-930c-548f47818dfc/cb0243fe-58ca-4459-930c-548f47818dfc.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1030.853994] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-878f8fb3-6690-40c9-9f6f-d6953e3612e1 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/cb0243fe-58ca-4459-930c-548f47818dfc/cb0243fe-58ca-4459-930c-548f47818dfc.vmdk to [datastore2] 7d572ac0-f6af-4622-96cc-e75983420222/7d572ac0-f6af-4622-96cc-e75983420222.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1030.854644] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3897ceff-d785-4b2c-84d7-ee82f0b4b18e {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.861050] env[63515]: DEBUG oslo_vmware.api [None req-878f8fb3-6690-40c9-9f6f-d6953e3612e1 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Waiting for the task: (returnval){ [ 1030.861050] env[63515]: value = "task-1111852" [ 1030.861050] env[63515]: _type = "Task" [ 1030.861050] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1030.869556] env[63515]: DEBUG oslo_vmware.api [None req-878f8fb3-6690-40c9-9f6f-d6953e3612e1 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Task: {'id': task-1111852, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.913691] env[63515]: DEBUG oslo_vmware.api [None req-5a78568a-9d2c-4ab4-aaf0-b4782bd16af7 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Task: {'id': task-1111847, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.972781] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b14841a7-1abf-4d5f-9a5e-1df37801a30d {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.991628] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0c6ab9a-50e8-4378-9c0d-0827fdd60813 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.009611] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-a7fc067e-c3eb-4fcd-8441-645f050b9570 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: b76cf3dd-7b7e-4e5d-bf3c-e15bd37069ac] Updating instance 'b76cf3dd-7b7e-4e5d-bf3c-e15bd37069ac' progress to 67 {{(pid=63515) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1031.217040] env[63515]: INFO nova.compute.manager [-] [instance: e925fc93-a731-498a-984c-9b1f3eabf353] Took 1.22 seconds to deallocate network for instance. [ 1031.273884] env[63515]: DEBUG oslo_concurrency.lockutils [None req-77b6ad21-5821-4c19-9cf4-41738b67053e tempest-ServerRescueTestJSONUnderV235-800430211 tempest-ServerRescueTestJSONUnderV235-800430211-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1031.274127] env[63515]: DEBUG oslo_concurrency.lockutils [None req-77b6ad21-5821-4c19-9cf4-41738b67053e tempest-ServerRescueTestJSONUnderV235-800430211 tempest-ServerRescueTestJSONUnderV235-800430211-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1031.274356] env[63515]: DEBUG nova.objects.instance [None req-77b6ad21-5821-4c19-9cf4-41738b67053e tempest-ServerRescueTestJSONUnderV235-800430211 tempest-ServerRescueTestJSONUnderV235-800430211-project-member] Lazy-loading 'resources' on Instance uuid fcfe3376-8dfa-4189-a267-ff6402713c1b {{(pid=63515) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1031.373496] env[63515]: DEBUG oslo_vmware.api [None req-878f8fb3-6690-40c9-9f6f-d6953e3612e1 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Task: {'id': task-1111852, 'name': CopyVirtualDisk_Task} progress is 15%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.417034] env[63515]: DEBUG oslo_vmware.api [None req-5a78568a-9d2c-4ab4-aaf0-b4782bd16af7 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Task: {'id': task-1111847, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.563105] env[63515]: DEBUG nova.network.neutron [None req-a7fc067e-c3eb-4fcd-8441-645f050b9570 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: b76cf3dd-7b7e-4e5d-bf3c-e15bd37069ac] Port acea8927-7e43-427c-bed2-2420d4708f3b binding to destination host cpu-1 is already ACTIVE {{(pid=63515) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 1031.711666] env[63515]: DEBUG nova.compute.manager [None req-123efa07-e143-4b76-a7d6-1e616f30576d tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: aa83d035-9c47-434d-ab1a-140e6b3110b4] Start spawning the instance on the hypervisor. {{(pid=63515) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1031.725264] env[63515]: DEBUG oslo_concurrency.lockutils [None req-fd066576-98f6-4ad4-9fd9-bbb5e5a27c21 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1031.740700] env[63515]: DEBUG nova.virt.hardware [None req-123efa07-e143-4b76-a7d6-1e616f30576d tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T02:52:14Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T02:51:56Z,direct_url=,disk_format='vmdk',id=8a120570-cb06-4099-b262-554ca0ad15c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b569255cc43e42e7ae3f2b4ad37c6ef4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T02:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1031.740966] env[63515]: DEBUG nova.virt.hardware [None req-123efa07-e143-4b76-a7d6-1e616f30576d tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Flavor limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1031.741148] env[63515]: DEBUG nova.virt.hardware [None req-123efa07-e143-4b76-a7d6-1e616f30576d tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Image limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1031.741336] env[63515]: DEBUG nova.virt.hardware [None req-123efa07-e143-4b76-a7d6-1e616f30576d tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Flavor pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1031.741546] env[63515]: DEBUG nova.virt.hardware [None req-123efa07-e143-4b76-a7d6-1e616f30576d tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Image pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1031.741704] env[63515]: DEBUG nova.virt.hardware [None req-123efa07-e143-4b76-a7d6-1e616f30576d tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1031.741917] env[63515]: DEBUG nova.virt.hardware [None req-123efa07-e143-4b76-a7d6-1e616f30576d tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1031.742094] env[63515]: DEBUG nova.virt.hardware [None req-123efa07-e143-4b76-a7d6-1e616f30576d tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1031.742267] env[63515]: DEBUG nova.virt.hardware [None req-123efa07-e143-4b76-a7d6-1e616f30576d tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Got 1 possible topologies {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1031.742437] env[63515]: DEBUG nova.virt.hardware [None req-123efa07-e143-4b76-a7d6-1e616f30576d tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1031.742677] env[63515]: DEBUG nova.virt.hardware [None req-123efa07-e143-4b76-a7d6-1e616f30576d tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1031.743614] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-566bb98b-574f-4148-bee4-24dab9b9f359 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.755032] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2bbcf706-965c-452d-a8cf-824f81c2903d {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.874131] env[63515]: DEBUG oslo_vmware.api [None req-878f8fb3-6690-40c9-9f6f-d6953e3612e1 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Task: {'id': task-1111852, 'name': CopyVirtualDisk_Task} progress is 35%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.918257] env[63515]: DEBUG oslo_vmware.api [None req-5a78568a-9d2c-4ab4-aaf0-b4782bd16af7 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Task: {'id': task-1111847, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.939015] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4216a7e-ca74-4746-8f74-d9b3bd6b83db {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.948337] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ce47dae-625f-419a-aea2-6e9ee9070fe9 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.981326] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4c4085f-2fe0-41fb-8071-ce512fdfccef {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.991078] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35d01f41-315b-495d-9730-7e666ff87d3e {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.998436] env[63515]: DEBUG nova.compute.manager [req-f23d0f17-aa3b-4f15-83cb-fcaa8a377d8e req-301a1fbb-f49c-4d38-a8c0-97187ff74bc6 service nova] [instance: aa83d035-9c47-434d-ab1a-140e6b3110b4] Received event network-vif-plugged-614aed8f-3ab8-4672-ac13-0ae8cb0b233d {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1031.998667] env[63515]: DEBUG oslo_concurrency.lockutils [req-f23d0f17-aa3b-4f15-83cb-fcaa8a377d8e req-301a1fbb-f49c-4d38-a8c0-97187ff74bc6 service nova] Acquiring lock "aa83d035-9c47-434d-ab1a-140e6b3110b4-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1031.998874] env[63515]: DEBUG oslo_concurrency.lockutils [req-f23d0f17-aa3b-4f15-83cb-fcaa8a377d8e req-301a1fbb-f49c-4d38-a8c0-97187ff74bc6 service nova] Lock "aa83d035-9c47-434d-ab1a-140e6b3110b4-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1031.999058] env[63515]: DEBUG oslo_concurrency.lockutils [req-f23d0f17-aa3b-4f15-83cb-fcaa8a377d8e req-301a1fbb-f49c-4d38-a8c0-97187ff74bc6 service nova] Lock "aa83d035-9c47-434d-ab1a-140e6b3110b4-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1031.999264] env[63515]: DEBUG nova.compute.manager [req-f23d0f17-aa3b-4f15-83cb-fcaa8a377d8e req-301a1fbb-f49c-4d38-a8c0-97187ff74bc6 service nova] [instance: aa83d035-9c47-434d-ab1a-140e6b3110b4] No waiting events found dispatching network-vif-plugged-614aed8f-3ab8-4672-ac13-0ae8cb0b233d {{(pid=63515) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1031.999507] env[63515]: WARNING nova.compute.manager [req-f23d0f17-aa3b-4f15-83cb-fcaa8a377d8e req-301a1fbb-f49c-4d38-a8c0-97187ff74bc6 service nova] [instance: aa83d035-9c47-434d-ab1a-140e6b3110b4] Received unexpected event network-vif-plugged-614aed8f-3ab8-4672-ac13-0ae8cb0b233d for instance with vm_state building and task_state spawning. [ 1032.012590] env[63515]: DEBUG nova.compute.provider_tree [None req-77b6ad21-5821-4c19-9cf4-41738b67053e tempest-ServerRescueTestJSONUnderV235-800430211 tempest-ServerRescueTestJSONUnderV235-800430211-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1032.094541] env[63515]: DEBUG nova.network.neutron [None req-123efa07-e143-4b76-a7d6-1e616f30576d tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: aa83d035-9c47-434d-ab1a-140e6b3110b4] Successfully updated port: 614aed8f-3ab8-4672-ac13-0ae8cb0b233d {{(pid=63515) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1032.259352] env[63515]: DEBUG nova.compute.manager [req-b4aebc25-25ad-45a2-b991-1aa01f22a379 req-b5334239-6c0f-4084-9970-acda69e6152f service nova] [instance: e925fc93-a731-498a-984c-9b1f3eabf353] Received event network-vif-deleted-80493aae-217e-4846-84b0-b60116b4afda {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1032.375296] env[63515]: DEBUG oslo_vmware.api [None req-878f8fb3-6690-40c9-9f6f-d6953e3612e1 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Task: {'id': task-1111852, 'name': CopyVirtualDisk_Task} progress is 57%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.418804] env[63515]: DEBUG oslo_vmware.api [None req-5a78568a-9d2c-4ab4-aaf0-b4782bd16af7 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Task: {'id': task-1111847, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.515491] env[63515]: DEBUG nova.scheduler.client.report [None req-77b6ad21-5821-4c19-9cf4-41738b67053e tempest-ServerRescueTestJSONUnderV235-800430211 tempest-ServerRescueTestJSONUnderV235-800430211-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1032.588597] env[63515]: DEBUG oslo_concurrency.lockutils [None req-a7fc067e-c3eb-4fcd-8441-645f050b9570 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Acquiring lock "b76cf3dd-7b7e-4e5d-bf3c-e15bd37069ac-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1032.588848] env[63515]: DEBUG oslo_concurrency.lockutils [None req-a7fc067e-c3eb-4fcd-8441-645f050b9570 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Lock "b76cf3dd-7b7e-4e5d-bf3c-e15bd37069ac-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1032.589038] env[63515]: DEBUG oslo_concurrency.lockutils [None req-a7fc067e-c3eb-4fcd-8441-645f050b9570 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Lock "b76cf3dd-7b7e-4e5d-bf3c-e15bd37069ac-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1032.597849] env[63515]: DEBUG oslo_concurrency.lockutils [None req-123efa07-e143-4b76-a7d6-1e616f30576d tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Acquiring lock "refresh_cache-aa83d035-9c47-434d-ab1a-140e6b3110b4" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1032.598113] env[63515]: DEBUG oslo_concurrency.lockutils [None req-123efa07-e143-4b76-a7d6-1e616f30576d tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Acquired lock "refresh_cache-aa83d035-9c47-434d-ab1a-140e6b3110b4" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1032.598181] env[63515]: DEBUG nova.network.neutron [None req-123efa07-e143-4b76-a7d6-1e616f30576d tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: aa83d035-9c47-434d-ab1a-140e6b3110b4] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1032.877462] env[63515]: DEBUG oslo_vmware.api [None req-878f8fb3-6690-40c9-9f6f-d6953e3612e1 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Task: {'id': task-1111852, 'name': CopyVirtualDisk_Task} progress is 80%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.918664] env[63515]: DEBUG oslo_vmware.api [None req-5a78568a-9d2c-4ab4-aaf0-b4782bd16af7 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Task: {'id': task-1111847, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.020867] env[63515]: DEBUG oslo_concurrency.lockutils [None req-77b6ad21-5821-4c19-9cf4-41738b67053e tempest-ServerRescueTestJSONUnderV235-800430211 tempest-ServerRescueTestJSONUnderV235-800430211-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.747s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1033.023724] env[63515]: DEBUG oslo_concurrency.lockutils [None req-fd066576-98f6-4ad4-9fd9-bbb5e5a27c21 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.298s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1033.024098] env[63515]: DEBUG nova.objects.instance [None req-fd066576-98f6-4ad4-9fd9-bbb5e5a27c21 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Lazy-loading 'resources' on Instance uuid e925fc93-a731-498a-984c-9b1f3eabf353 {{(pid=63515) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1033.046028] env[63515]: INFO nova.scheduler.client.report [None req-77b6ad21-5821-4c19-9cf4-41738b67053e tempest-ServerRescueTestJSONUnderV235-800430211 tempest-ServerRescueTestJSONUnderV235-800430211-project-member] Deleted allocations for instance fcfe3376-8dfa-4189-a267-ff6402713c1b [ 1033.130286] env[63515]: DEBUG nova.network.neutron [None req-123efa07-e143-4b76-a7d6-1e616f30576d tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: aa83d035-9c47-434d-ab1a-140e6b3110b4] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1033.319998] env[63515]: DEBUG nova.network.neutron [None req-123efa07-e143-4b76-a7d6-1e616f30576d tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: aa83d035-9c47-434d-ab1a-140e6b3110b4] Updating instance_info_cache with network_info: [{"id": "614aed8f-3ab8-4672-ac13-0ae8cb0b233d", "address": "fa:16:3e:98:68:38", "network": {"id": "5758c885-e1c9-45c8-86aa-27a802101f52", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-734458138-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b18d4daacbc84e758f9f9ae4f3f09d28", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69351262-8d39-441a-85ba-3a78df436d17", "external-id": "nsx-vlan-transportzone-205", "segmentation_id": 205, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap614aed8f-3a", "ovs_interfaceid": "614aed8f-3ab8-4672-ac13-0ae8cb0b233d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1033.376446] env[63515]: DEBUG oslo_vmware.api [None req-878f8fb3-6690-40c9-9f6f-d6953e3612e1 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Task: {'id': task-1111852, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.467283} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1033.376767] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-878f8fb3-6690-40c9-9f6f-d6953e3612e1 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/cb0243fe-58ca-4459-930c-548f47818dfc/cb0243fe-58ca-4459-930c-548f47818dfc.vmdk to [datastore2] 7d572ac0-f6af-4622-96cc-e75983420222/7d572ac0-f6af-4622-96cc-e75983420222.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1033.377512] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94fbd674-0dae-4a6a-8d77-ebe3e72303c9 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.398629] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-878f8fb3-6690-40c9-9f6f-d6953e3612e1 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] [instance: 7d572ac0-f6af-4622-96cc-e75983420222] Reconfiguring VM instance instance-00000058 to attach disk [datastore2] 7d572ac0-f6af-4622-96cc-e75983420222/7d572ac0-f6af-4622-96cc-e75983420222.vmdk or device None with type streamOptimized {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1033.399184] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b71508c8-0aca-4882-b708-7d7cc5cd087a {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.421773] env[63515]: DEBUG oslo_vmware.api [None req-5a78568a-9d2c-4ab4-aaf0-b4782bd16af7 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Task: {'id': task-1111847, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.423137] env[63515]: DEBUG oslo_vmware.api [None req-878f8fb3-6690-40c9-9f6f-d6953e3612e1 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Waiting for the task: (returnval){ [ 1033.423137] env[63515]: value = "task-1111853" [ 1033.423137] env[63515]: _type = "Task" [ 1033.423137] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1033.430527] env[63515]: DEBUG oslo_vmware.api [None req-878f8fb3-6690-40c9-9f6f-d6953e3612e1 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Task: {'id': task-1111853, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.557583] env[63515]: DEBUG oslo_concurrency.lockutils [None req-77b6ad21-5821-4c19-9cf4-41738b67053e tempest-ServerRescueTestJSONUnderV235-800430211 tempest-ServerRescueTestJSONUnderV235-800430211-project-member] Lock "fcfe3376-8dfa-4189-a267-ff6402713c1b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.151s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1033.624526] env[63515]: DEBUG oslo_concurrency.lockutils [None req-a7fc067e-c3eb-4fcd-8441-645f050b9570 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Acquiring lock "refresh_cache-b76cf3dd-7b7e-4e5d-bf3c-e15bd37069ac" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1033.624719] env[63515]: DEBUG oslo_concurrency.lockutils [None req-a7fc067e-c3eb-4fcd-8441-645f050b9570 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Acquired lock "refresh_cache-b76cf3dd-7b7e-4e5d-bf3c-e15bd37069ac" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1033.624901] env[63515]: DEBUG nova.network.neutron [None req-a7fc067e-c3eb-4fcd-8441-645f050b9570 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: b76cf3dd-7b7e-4e5d-bf3c-e15bd37069ac] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1033.664227] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1375578e-a09b-4c69-b04d-5fca6db1e640 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.671712] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6bd3fcbc-c2ea-4913-92dd-340392a8f614 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.701916] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15d1b3cd-57c3-4cb9-84f6-0fafc5eaf91f {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.709497] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e8ee0b3-c037-47cf-bd00-34976b772afc {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.724036] env[63515]: DEBUG nova.compute.provider_tree [None req-fd066576-98f6-4ad4-9fd9-bbb5e5a27c21 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1033.822520] env[63515]: DEBUG oslo_concurrency.lockutils [None req-123efa07-e143-4b76-a7d6-1e616f30576d tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Releasing lock "refresh_cache-aa83d035-9c47-434d-ab1a-140e6b3110b4" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1033.822846] env[63515]: DEBUG nova.compute.manager [None req-123efa07-e143-4b76-a7d6-1e616f30576d tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: aa83d035-9c47-434d-ab1a-140e6b3110b4] Instance network_info: |[{"id": "614aed8f-3ab8-4672-ac13-0ae8cb0b233d", "address": "fa:16:3e:98:68:38", "network": {"id": "5758c885-e1c9-45c8-86aa-27a802101f52", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-734458138-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b18d4daacbc84e758f9f9ae4f3f09d28", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69351262-8d39-441a-85ba-3a78df436d17", "external-id": "nsx-vlan-transportzone-205", "segmentation_id": 205, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap614aed8f-3a", "ovs_interfaceid": "614aed8f-3ab8-4672-ac13-0ae8cb0b233d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1033.823284] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-123efa07-e143-4b76-a7d6-1e616f30576d tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: aa83d035-9c47-434d-ab1a-140e6b3110b4] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:98:68:38', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '69351262-8d39-441a-85ba-3a78df436d17', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '614aed8f-3ab8-4672-ac13-0ae8cb0b233d', 'vif_model': 'vmxnet3'}] {{(pid=63515) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1033.830714] env[63515]: DEBUG oslo.service.loopingcall [None req-123efa07-e143-4b76-a7d6-1e616f30576d tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1033.831528] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: aa83d035-9c47-434d-ab1a-140e6b3110b4] Creating VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1033.831528] env[63515]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e049ec50-325b-4211-be0b-fd62814ce6c7 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.850706] env[63515]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1033.850706] env[63515]: value = "task-1111854" [ 1033.850706] env[63515]: _type = "Task" [ 1033.850706] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1033.858635] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1111854, 'name': CreateVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.922684] env[63515]: DEBUG oslo_vmware.api [None req-5a78568a-9d2c-4ab4-aaf0-b4782bd16af7 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Task: {'id': task-1111847, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.932289] env[63515]: DEBUG oslo_vmware.api [None req-878f8fb3-6690-40c9-9f6f-d6953e3612e1 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Task: {'id': task-1111853, 'name': ReconfigVM_Task, 'duration_secs': 0.269389} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1033.932576] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-878f8fb3-6690-40c9-9f6f-d6953e3612e1 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] [instance: 7d572ac0-f6af-4622-96cc-e75983420222] Reconfigured VM instance instance-00000058 to attach disk [datastore2] 7d572ac0-f6af-4622-96cc-e75983420222/7d572ac0-f6af-4622-96cc-e75983420222.vmdk or device None with type streamOptimized {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1033.933246] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-cc3683d7-48da-40f9-b4d6-683eee24b6c4 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.940076] env[63515]: DEBUG oslo_vmware.api [None req-878f8fb3-6690-40c9-9f6f-d6953e3612e1 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Waiting for the task: (returnval){ [ 1033.940076] env[63515]: value = "task-1111855" [ 1033.940076] env[63515]: _type = "Task" [ 1033.940076] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1033.948704] env[63515]: DEBUG oslo_vmware.api [None req-878f8fb3-6690-40c9-9f6f-d6953e3612e1 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Task: {'id': task-1111855, 'name': Rename_Task} progress is 5%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1034.034406] env[63515]: DEBUG nova.compute.manager [req-50a9694a-5ff0-4835-acc2-5b8c1ad3c0e7 req-c400b361-4cd2-4d2b-ad17-30547db6c02d service nova] [instance: aa83d035-9c47-434d-ab1a-140e6b3110b4] Received event network-changed-614aed8f-3ab8-4672-ac13-0ae8cb0b233d {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1034.034637] env[63515]: DEBUG nova.compute.manager [req-50a9694a-5ff0-4835-acc2-5b8c1ad3c0e7 req-c400b361-4cd2-4d2b-ad17-30547db6c02d service nova] [instance: aa83d035-9c47-434d-ab1a-140e6b3110b4] Refreshing instance network info cache due to event network-changed-614aed8f-3ab8-4672-ac13-0ae8cb0b233d. {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1034.034886] env[63515]: DEBUG oslo_concurrency.lockutils [req-50a9694a-5ff0-4835-acc2-5b8c1ad3c0e7 req-c400b361-4cd2-4d2b-ad17-30547db6c02d service nova] Acquiring lock "refresh_cache-aa83d035-9c47-434d-ab1a-140e6b3110b4" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1034.035169] env[63515]: DEBUG oslo_concurrency.lockutils [req-50a9694a-5ff0-4835-acc2-5b8c1ad3c0e7 req-c400b361-4cd2-4d2b-ad17-30547db6c02d service nova] Acquired lock "refresh_cache-aa83d035-9c47-434d-ab1a-140e6b3110b4" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1034.035313] env[63515]: DEBUG nova.network.neutron [req-50a9694a-5ff0-4835-acc2-5b8c1ad3c0e7 req-c400b361-4cd2-4d2b-ad17-30547db6c02d service nova] [instance: aa83d035-9c47-434d-ab1a-140e6b3110b4] Refreshing network info cache for port 614aed8f-3ab8-4672-ac13-0ae8cb0b233d {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1034.226620] env[63515]: DEBUG nova.scheduler.client.report [None req-fd066576-98f6-4ad4-9fd9-bbb5e5a27c21 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1034.326872] env[63515]: DEBUG nova.network.neutron [None req-a7fc067e-c3eb-4fcd-8441-645f050b9570 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: b76cf3dd-7b7e-4e5d-bf3c-e15bd37069ac] Updating instance_info_cache with network_info: [{"id": "acea8927-7e43-427c-bed2-2420d4708f3b", "address": "fa:16:3e:21:d3:df", "network": {"id": "3e5914c8-0cd4-49ee-81b4-1f3e280440bb", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-641632277-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ea099ce07bfb4a8da014d9303cf552eb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "225b6979-9329-403b-91fa-138bd41f6e83", "external-id": "nsx-vlan-transportzone-38", "segmentation_id": 38, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapacea8927-7e", "ovs_interfaceid": "acea8927-7e43-427c-bed2-2420d4708f3b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1034.360428] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1111854, 'name': CreateVM_Task, 'duration_secs': 0.399735} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1034.360613] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: aa83d035-9c47-434d-ab1a-140e6b3110b4] Created VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1034.361342] env[63515]: DEBUG oslo_concurrency.lockutils [None req-123efa07-e143-4b76-a7d6-1e616f30576d tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1034.361518] env[63515]: DEBUG oslo_concurrency.lockutils [None req-123efa07-e143-4b76-a7d6-1e616f30576d tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1034.361845] env[63515]: DEBUG oslo_concurrency.lockutils [None req-123efa07-e143-4b76-a7d6-1e616f30576d tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1034.362118] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8d08281a-9e47-467c-a315-ab33a835cc62 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.366361] env[63515]: DEBUG oslo_vmware.api [None req-123efa07-e143-4b76-a7d6-1e616f30576d tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Waiting for the task: (returnval){ [ 1034.366361] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]52879f38-9965-e945-14f0-e118c9d15822" [ 1034.366361] env[63515]: _type = "Task" [ 1034.366361] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1034.374189] env[63515]: DEBUG oslo_vmware.api [None req-123efa07-e143-4b76-a7d6-1e616f30576d tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52879f38-9965-e945-14f0-e118c9d15822, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1034.422419] env[63515]: DEBUG oslo_vmware.api [None req-5a78568a-9d2c-4ab4-aaf0-b4782bd16af7 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Task: {'id': task-1111847, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1034.449722] env[63515]: DEBUG oslo_vmware.api [None req-878f8fb3-6690-40c9-9f6f-d6953e3612e1 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Task: {'id': task-1111855, 'name': Rename_Task, 'duration_secs': 0.136449} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1034.449722] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-878f8fb3-6690-40c9-9f6f-d6953e3612e1 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] [instance: 7d572ac0-f6af-4622-96cc-e75983420222] Powering on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1034.449722] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8c453a0f-6265-440a-a54c-942ff08b9d6d {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.455259] env[63515]: DEBUG oslo_vmware.api [None req-878f8fb3-6690-40c9-9f6f-d6953e3612e1 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Waiting for the task: (returnval){ [ 1034.455259] env[63515]: value = "task-1111856" [ 1034.455259] env[63515]: _type = "Task" [ 1034.455259] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1034.464814] env[63515]: DEBUG oslo_vmware.api [None req-878f8fb3-6690-40c9-9f6f-d6953e3612e1 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Task: {'id': task-1111856, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1034.734347] env[63515]: DEBUG oslo_concurrency.lockutils [None req-fd066576-98f6-4ad4-9fd9-bbb5e5a27c21 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.711s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1034.760024] env[63515]: INFO nova.scheduler.client.report [None req-fd066576-98f6-4ad4-9fd9-bbb5e5a27c21 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Deleted allocations for instance e925fc93-a731-498a-984c-9b1f3eabf353 [ 1034.775173] env[63515]: DEBUG nova.network.neutron [req-50a9694a-5ff0-4835-acc2-5b8c1ad3c0e7 req-c400b361-4cd2-4d2b-ad17-30547db6c02d service nova] [instance: aa83d035-9c47-434d-ab1a-140e6b3110b4] Updated VIF entry in instance network info cache for port 614aed8f-3ab8-4672-ac13-0ae8cb0b233d. {{(pid=63515) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1034.775173] env[63515]: DEBUG nova.network.neutron [req-50a9694a-5ff0-4835-acc2-5b8c1ad3c0e7 req-c400b361-4cd2-4d2b-ad17-30547db6c02d service nova] [instance: aa83d035-9c47-434d-ab1a-140e6b3110b4] Updating instance_info_cache with network_info: [{"id": "614aed8f-3ab8-4672-ac13-0ae8cb0b233d", "address": "fa:16:3e:98:68:38", "network": {"id": "5758c885-e1c9-45c8-86aa-27a802101f52", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-734458138-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b18d4daacbc84e758f9f9ae4f3f09d28", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69351262-8d39-441a-85ba-3a78df436d17", "external-id": "nsx-vlan-transportzone-205", "segmentation_id": 205, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap614aed8f-3a", "ovs_interfaceid": "614aed8f-3ab8-4672-ac13-0ae8cb0b233d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1034.829624] env[63515]: DEBUG oslo_concurrency.lockutils [None req-a7fc067e-c3eb-4fcd-8441-645f050b9570 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Releasing lock "refresh_cache-b76cf3dd-7b7e-4e5d-bf3c-e15bd37069ac" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1034.880179] env[63515]: DEBUG oslo_vmware.api [None req-123efa07-e143-4b76-a7d6-1e616f30576d tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52879f38-9965-e945-14f0-e118c9d15822, 'name': SearchDatastore_Task, 'duration_secs': 0.008648} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1034.880719] env[63515]: DEBUG oslo_concurrency.lockutils [None req-123efa07-e143-4b76-a7d6-1e616f30576d tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1034.880824] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-123efa07-e143-4b76-a7d6-1e616f30576d tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: aa83d035-9c47-434d-ab1a-140e6b3110b4] Processing image 8a120570-cb06-4099-b262-554ca0ad15c5 {{(pid=63515) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1034.881061] env[63515]: DEBUG oslo_concurrency.lockutils [None req-123efa07-e143-4b76-a7d6-1e616f30576d tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1034.881440] env[63515]: DEBUG oslo_concurrency.lockutils [None req-123efa07-e143-4b76-a7d6-1e616f30576d tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1034.881654] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-123efa07-e143-4b76-a7d6-1e616f30576d tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1034.882640] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-64b3da82-139b-499b-bf78-64d111f20794 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.892264] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-123efa07-e143-4b76-a7d6-1e616f30576d tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1034.892496] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-123efa07-e143-4b76-a7d6-1e616f30576d tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63515) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1034.894261] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-763bbe77-cb76-4adf-a768-5c9d9d1a92f0 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.900282] env[63515]: DEBUG oslo_vmware.api [None req-123efa07-e143-4b76-a7d6-1e616f30576d tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Waiting for the task: (returnval){ [ 1034.900282] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]528308e9-59df-fdd1-cfb9-e1d93802d641" [ 1034.900282] env[63515]: _type = "Task" [ 1034.900282] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1034.908611] env[63515]: DEBUG oslo_vmware.api [None req-123efa07-e143-4b76-a7d6-1e616f30576d tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]528308e9-59df-fdd1-cfb9-e1d93802d641, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1034.924047] env[63515]: DEBUG oslo_vmware.api [None req-5a78568a-9d2c-4ab4-aaf0-b4782bd16af7 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Task: {'id': task-1111847, 'name': ReconfigVM_Task} progress is 18%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1034.965327] env[63515]: DEBUG oslo_vmware.api [None req-878f8fb3-6690-40c9-9f6f-d6953e3612e1 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Task: {'id': task-1111856, 'name': PowerOnVM_Task, 'duration_secs': 0.448499} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1034.965327] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-878f8fb3-6690-40c9-9f6f-d6953e3612e1 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] [instance: 7d572ac0-f6af-4622-96cc-e75983420222] Powered on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1035.075057] env[63515]: DEBUG nova.compute.manager [None req-878f8fb3-6690-40c9-9f6f-d6953e3612e1 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] [instance: 7d572ac0-f6af-4622-96cc-e75983420222] Checking state {{(pid=63515) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1035.076168] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa155bac-f459-4dfa-a3d0-d8a242603277 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.264838] env[63515]: DEBUG oslo_concurrency.lockutils [None req-fd066576-98f6-4ad4-9fd9-bbb5e5a27c21 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Lock "e925fc93-a731-498a-984c-9b1f3eabf353" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.411s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1035.277589] env[63515]: DEBUG oslo_concurrency.lockutils [req-50a9694a-5ff0-4835-acc2-5b8c1ad3c0e7 req-c400b361-4cd2-4d2b-ad17-30547db6c02d service nova] Releasing lock "refresh_cache-aa83d035-9c47-434d-ab1a-140e6b3110b4" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1035.354483] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46f112dc-1460-4c15-a440-f14c18ce20e4 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.374143] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-442133a8-1a7a-4f7d-bc0e-80e18c62ad76 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.381080] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-a7fc067e-c3eb-4fcd-8441-645f050b9570 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: b76cf3dd-7b7e-4e5d-bf3c-e15bd37069ac] Updating instance 'b76cf3dd-7b7e-4e5d-bf3c-e15bd37069ac' progress to 83 {{(pid=63515) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1035.411283] env[63515]: DEBUG oslo_vmware.api [None req-123efa07-e143-4b76-a7d6-1e616f30576d tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]528308e9-59df-fdd1-cfb9-e1d93802d641, 'name': SearchDatastore_Task, 'duration_secs': 0.008056} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1035.411837] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-caa9f6d1-2bfe-46d8-b898-96b7b926976e {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.420217] env[63515]: DEBUG oslo_vmware.api [None req-123efa07-e143-4b76-a7d6-1e616f30576d tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Waiting for the task: (returnval){ [ 1035.420217] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]52e122db-b376-33a1-6613-ca49433a78d0" [ 1035.420217] env[63515]: _type = "Task" [ 1035.420217] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1035.426232] env[63515]: DEBUG oslo_vmware.api [None req-5a78568a-9d2c-4ab4-aaf0-b4782bd16af7 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Task: {'id': task-1111847, 'name': ReconfigVM_Task, 'duration_secs': 5.76957} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1035.426783] env[63515]: DEBUG oslo_concurrency.lockutils [None req-5a78568a-9d2c-4ab4-aaf0-b4782bd16af7 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Releasing lock "724b05bc-ce4b-4f99-ad14-f3abea5567f2" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1035.427043] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-5a78568a-9d2c-4ab4-aaf0-b4782bd16af7 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: 724b05bc-ce4b-4f99-ad14-f3abea5567f2] Reconfigured VM to detach interface {{(pid=63515) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1978}} [ 1035.433868] env[63515]: DEBUG oslo_vmware.api [None req-123efa07-e143-4b76-a7d6-1e616f30576d tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52e122db-b376-33a1-6613-ca49433a78d0, 'name': SearchDatastore_Task, 'duration_secs': 0.009022} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1035.434986] env[63515]: DEBUG oslo_concurrency.lockutils [None req-123efa07-e143-4b76-a7d6-1e616f30576d tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1035.434986] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-123efa07-e143-4b76-a7d6-1e616f30576d tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk to [datastore2] aa83d035-9c47-434d-ab1a-140e6b3110b4/aa83d035-9c47-434d-ab1a-140e6b3110b4.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1035.434986] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5e9f9667-c43f-499a-8f66-a54eae0ccb28 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.443775] env[63515]: DEBUG oslo_vmware.api [None req-123efa07-e143-4b76-a7d6-1e616f30576d tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Waiting for the task: (returnval){ [ 1035.443775] env[63515]: value = "task-1111857" [ 1035.443775] env[63515]: _type = "Task" [ 1035.443775] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1035.452277] env[63515]: DEBUG oslo_vmware.api [None req-123efa07-e143-4b76-a7d6-1e616f30576d tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Task: {'id': task-1111857, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1035.600226] env[63515]: DEBUG oslo_concurrency.lockutils [None req-878f8fb3-6690-40c9-9f6f-d6953e3612e1 tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Lock "7d572ac0-f6af-4622-96cc-e75983420222" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 21.135s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1035.889311] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-a7fc067e-c3eb-4fcd-8441-645f050b9570 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: b76cf3dd-7b7e-4e5d-bf3c-e15bd37069ac] Powering on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1035.889311] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a7808d46-c6e8-43bb-8fd7-a16fac42ce50 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.900425] env[63515]: DEBUG oslo_vmware.api [None req-a7fc067e-c3eb-4fcd-8441-645f050b9570 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Waiting for the task: (returnval){ [ 1035.900425] env[63515]: value = "task-1111858" [ 1035.900425] env[63515]: _type = "Task" [ 1035.900425] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1035.909549] env[63515]: DEBUG oslo_vmware.api [None req-a7fc067e-c3eb-4fcd-8441-645f050b9570 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Task: {'id': task-1111858, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1035.934196] env[63515]: DEBUG oslo_concurrency.lockutils [None req-3a846554-7de2-4e42-88d0-6461fe3d7394 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Acquiring lock "a8d2cc01-3bc1-4792-85f5-b9095294b33c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1035.934196] env[63515]: DEBUG oslo_concurrency.lockutils [None req-3a846554-7de2-4e42-88d0-6461fe3d7394 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Lock "a8d2cc01-3bc1-4792-85f5-b9095294b33c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1035.955037] env[63515]: DEBUG oslo_vmware.api [None req-123efa07-e143-4b76-a7d6-1e616f30576d tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Task: {'id': task-1111857, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.486344} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1035.955592] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-123efa07-e143-4b76-a7d6-1e616f30576d tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk to [datastore2] aa83d035-9c47-434d-ab1a-140e6b3110b4/aa83d035-9c47-434d-ab1a-140e6b3110b4.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1035.955975] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-123efa07-e143-4b76-a7d6-1e616f30576d tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: aa83d035-9c47-434d-ab1a-140e6b3110b4] Extending root virtual disk to 1048576 {{(pid=63515) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1035.956364] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d8abbed5-f3d5-48ba-9513-34ef5cbcf8e8 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.963628] env[63515]: DEBUG oslo_vmware.api [None req-123efa07-e143-4b76-a7d6-1e616f30576d tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Waiting for the task: (returnval){ [ 1035.963628] env[63515]: value = "task-1111859" [ 1035.963628] env[63515]: _type = "Task" [ 1035.963628] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1035.973348] env[63515]: DEBUG oslo_vmware.api [None req-123efa07-e143-4b76-a7d6-1e616f30576d tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Task: {'id': task-1111859, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1036.410050] env[63515]: DEBUG oslo_vmware.api [None req-a7fc067e-c3eb-4fcd-8441-645f050b9570 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Task: {'id': task-1111858, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1036.436946] env[63515]: DEBUG nova.compute.manager [None req-3a846554-7de2-4e42-88d0-6461fe3d7394 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: a8d2cc01-3bc1-4792-85f5-b9095294b33c] Starting instance... {{(pid=63515) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1036.473693] env[63515]: DEBUG oslo_vmware.api [None req-123efa07-e143-4b76-a7d6-1e616f30576d tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Task: {'id': task-1111859, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.157549} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1036.473975] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-123efa07-e143-4b76-a7d6-1e616f30576d tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: aa83d035-9c47-434d-ab1a-140e6b3110b4] Extended root virtual disk {{(pid=63515) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1036.474854] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8fb48a1-3edf-45ec-bb75-45f3ce565b5b {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.500245] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-123efa07-e143-4b76-a7d6-1e616f30576d tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: aa83d035-9c47-434d-ab1a-140e6b3110b4] Reconfiguring VM instance instance-00000066 to attach disk [datastore2] aa83d035-9c47-434d-ab1a-140e6b3110b4/aa83d035-9c47-434d-ab1a-140e6b3110b4.vmdk or device None with type sparse {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1036.501592] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d1a24931-127e-437a-967f-a4468b5b2267 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.522775] env[63515]: DEBUG oslo_vmware.api [None req-123efa07-e143-4b76-a7d6-1e616f30576d tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Waiting for the task: (returnval){ [ 1036.522775] env[63515]: value = "task-1111860" [ 1036.522775] env[63515]: _type = "Task" [ 1036.522775] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1036.531511] env[63515]: DEBUG oslo_vmware.api [None req-123efa07-e143-4b76-a7d6-1e616f30576d tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Task: {'id': task-1111860, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1036.909653] env[63515]: DEBUG oslo_vmware.api [None req-a7fc067e-c3eb-4fcd-8441-645f050b9570 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Task: {'id': task-1111858, 'name': PowerOnVM_Task, 'duration_secs': 0.714536} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1036.909938] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-a7fc067e-c3eb-4fcd-8441-645f050b9570 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: b76cf3dd-7b7e-4e5d-bf3c-e15bd37069ac] Powered on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1036.910146] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-a7fc067e-c3eb-4fcd-8441-645f050b9570 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: b76cf3dd-7b7e-4e5d-bf3c-e15bd37069ac] Updating instance 'b76cf3dd-7b7e-4e5d-bf3c-e15bd37069ac' progress to 100 {{(pid=63515) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1036.962990] env[63515]: DEBUG oslo_concurrency.lockutils [None req-3a846554-7de2-4e42-88d0-6461fe3d7394 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1036.963278] env[63515]: DEBUG oslo_concurrency.lockutils [None req-3a846554-7de2-4e42-88d0-6461fe3d7394 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1036.964897] env[63515]: INFO nova.compute.claims [None req-3a846554-7de2-4e42-88d0-6461fe3d7394 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: a8d2cc01-3bc1-4792-85f5-b9095294b33c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1037.032722] env[63515]: DEBUG oslo_vmware.api [None req-123efa07-e143-4b76-a7d6-1e616f30576d tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Task: {'id': task-1111860, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1037.184314] env[63515]: DEBUG oslo_concurrency.lockutils [None req-5a78568a-9d2c-4ab4-aaf0-b4782bd16af7 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Acquiring lock "refresh_cache-724b05bc-ce4b-4f99-ad14-f3abea5567f2" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1037.184514] env[63515]: DEBUG oslo_concurrency.lockutils [None req-5a78568a-9d2c-4ab4-aaf0-b4782bd16af7 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Acquired lock "refresh_cache-724b05bc-ce4b-4f99-ad14-f3abea5567f2" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1037.184703] env[63515]: DEBUG nova.network.neutron [None req-5a78568a-9d2c-4ab4-aaf0-b4782bd16af7 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: 724b05bc-ce4b-4f99-ad14-f3abea5567f2] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1037.533064] env[63515]: DEBUG oslo_vmware.api [None req-123efa07-e143-4b76-a7d6-1e616f30576d tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Task: {'id': task-1111860, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1037.751397] env[63515]: DEBUG nova.compute.manager [req-94b82895-6c06-413d-bcb5-476f1f42cb9e req-9928d174-a8c8-4b67-9ad3-1766a130bb05 service nova] [instance: 724b05bc-ce4b-4f99-ad14-f3abea5567f2] Received event network-changed-3dee43b6-2650-42e1-aa5a-4994a3ec7f05 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1037.751567] env[63515]: DEBUG nova.compute.manager [req-94b82895-6c06-413d-bcb5-476f1f42cb9e req-9928d174-a8c8-4b67-9ad3-1766a130bb05 service nova] [instance: 724b05bc-ce4b-4f99-ad14-f3abea5567f2] Refreshing instance network info cache due to event network-changed-3dee43b6-2650-42e1-aa5a-4994a3ec7f05. {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1037.751757] env[63515]: DEBUG oslo_concurrency.lockutils [req-94b82895-6c06-413d-bcb5-476f1f42cb9e req-9928d174-a8c8-4b67-9ad3-1766a130bb05 service nova] Acquiring lock "refresh_cache-724b05bc-ce4b-4f99-ad14-f3abea5567f2" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1037.884591] env[63515]: INFO nova.network.neutron [None req-5a78568a-9d2c-4ab4-aaf0-b4782bd16af7 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: 724b05bc-ce4b-4f99-ad14-f3abea5567f2] Port 6a98da26-06f3-4eb5-aee8-02a448c34749 from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 1037.885074] env[63515]: DEBUG nova.network.neutron [None req-5a78568a-9d2c-4ab4-aaf0-b4782bd16af7 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: 724b05bc-ce4b-4f99-ad14-f3abea5567f2] Updating instance_info_cache with network_info: [{"id": "3dee43b6-2650-42e1-aa5a-4994a3ec7f05", "address": "fa:16:3e:89:06:d0", "network": {"id": "88dfb29d-41e0-4df7-8f70-e52cda05620c", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1355578400-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7d4ada9ec35f42b19c6480a9101d21a4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "357d2811-e990-4985-9f9e-b158d10d3699", "external-id": "nsx-vlan-transportzone-641", "segmentation_id": 641, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3dee43b6-26", "ovs_interfaceid": "3dee43b6-2650-42e1-aa5a-4994a3ec7f05", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1038.033592] env[63515]: DEBUG oslo_vmware.api [None req-123efa07-e143-4b76-a7d6-1e616f30576d tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Task: {'id': task-1111860, 'name': ReconfigVM_Task, 'duration_secs': 1.089882} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1038.036039] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-123efa07-e143-4b76-a7d6-1e616f30576d tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: aa83d035-9c47-434d-ab1a-140e6b3110b4] Reconfigured VM instance instance-00000066 to attach disk [datastore2] aa83d035-9c47-434d-ab1a-140e6b3110b4/aa83d035-9c47-434d-ab1a-140e6b3110b4.vmdk or device None with type sparse {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1038.036846] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-0cd6499e-0120-4cda-9026-0cf0c9bf48f9 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.043327] env[63515]: DEBUG oslo_vmware.api [None req-123efa07-e143-4b76-a7d6-1e616f30576d tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Waiting for the task: (returnval){ [ 1038.043327] env[63515]: value = "task-1111861" [ 1038.043327] env[63515]: _type = "Task" [ 1038.043327] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1038.053352] env[63515]: DEBUG oslo_vmware.api [None req-123efa07-e143-4b76-a7d6-1e616f30576d tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Task: {'id': task-1111861, 'name': Rename_Task} progress is 5%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1038.118351] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0269116f-a533-4b11-8924-f3798148af72 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.125938] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-899340d9-6525-4ac7-99ed-174afd01967c {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.850543] env[63515]: DEBUG oslo_concurrency.lockutils [None req-5a78568a-9d2c-4ab4-aaf0-b4782bd16af7 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Releasing lock "refresh_cache-724b05bc-ce4b-4f99-ad14-f3abea5567f2" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1038.861020] env[63515]: DEBUG oslo_concurrency.lockutils [None req-b7657298-0ac9-4492-ba07-d80007973987 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Acquiring lock "interface-c6ac42a9-a233-48b4-bcb2-fb791c7446bb-6a98da26-06f3-4eb5-aee8-02a448c34749" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1038.861020] env[63515]: DEBUG oslo_concurrency.lockutils [None req-b7657298-0ac9-4492-ba07-d80007973987 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Lock "interface-c6ac42a9-a233-48b4-bcb2-fb791c7446bb-6a98da26-06f3-4eb5-aee8-02a448c34749" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1038.861020] env[63515]: DEBUG nova.objects.instance [None req-b7657298-0ac9-4492-ba07-d80007973987 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Lazy-loading 'flavor' on Instance uuid c6ac42a9-a233-48b4-bcb2-fb791c7446bb {{(pid=63515) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1038.861943] env[63515]: DEBUG oslo_concurrency.lockutils [req-94b82895-6c06-413d-bcb5-476f1f42cb9e req-9928d174-a8c8-4b67-9ad3-1766a130bb05 service nova] Acquired lock "refresh_cache-724b05bc-ce4b-4f99-ad14-f3abea5567f2" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1038.862834] env[63515]: DEBUG nova.network.neutron [req-94b82895-6c06-413d-bcb5-476f1f42cb9e req-9928d174-a8c8-4b67-9ad3-1766a130bb05 service nova] [instance: 724b05bc-ce4b-4f99-ad14-f3abea5567f2] Refreshing network info cache for port 3dee43b6-2650-42e1-aa5a-4994a3ec7f05 {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1038.867230] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-556dc53a-e766-4996-92b5-7e53dd0a97e9 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.880558] env[63515]: DEBUG oslo_vmware.api [None req-123efa07-e143-4b76-a7d6-1e616f30576d tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Task: {'id': task-1111861, 'name': Rename_Task, 'duration_secs': 0.145412} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1038.883451] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-123efa07-e143-4b76-a7d6-1e616f30576d tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: aa83d035-9c47-434d-ab1a-140e6b3110b4] Powering on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1038.883916] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5ccfab8c-034e-4290-a1a8-f85a81060f1f {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.886422] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1266e958-a92a-4486-8cbd-4f5758129bcb {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.900285] env[63515]: DEBUG nova.compute.provider_tree [None req-3a846554-7de2-4e42-88d0-6461fe3d7394 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1038.903113] env[63515]: DEBUG oslo_vmware.api [None req-123efa07-e143-4b76-a7d6-1e616f30576d tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Waiting for the task: (returnval){ [ 1038.903113] env[63515]: value = "task-1111862" [ 1038.903113] env[63515]: _type = "Task" [ 1038.903113] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1038.911641] env[63515]: DEBUG oslo_vmware.api [None req-123efa07-e143-4b76-a7d6-1e616f30576d tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Task: {'id': task-1111862, 'name': PowerOnVM_Task} progress is 33%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1039.097691] env[63515]: DEBUG nova.network.neutron [req-94b82895-6c06-413d-bcb5-476f1f42cb9e req-9928d174-a8c8-4b67-9ad3-1766a130bb05 service nova] [instance: 724b05bc-ce4b-4f99-ad14-f3abea5567f2] Updated VIF entry in instance network info cache for port 3dee43b6-2650-42e1-aa5a-4994a3ec7f05. {{(pid=63515) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1039.098053] env[63515]: DEBUG nova.network.neutron [req-94b82895-6c06-413d-bcb5-476f1f42cb9e req-9928d174-a8c8-4b67-9ad3-1766a130bb05 service nova] [instance: 724b05bc-ce4b-4f99-ad14-f3abea5567f2] Updating instance_info_cache with network_info: [{"id": "3dee43b6-2650-42e1-aa5a-4994a3ec7f05", "address": "fa:16:3e:89:06:d0", "network": {"id": "88dfb29d-41e0-4df7-8f70-e52cda05620c", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1355578400-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7d4ada9ec35f42b19c6480a9101d21a4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "357d2811-e990-4985-9f9e-b158d10d3699", "external-id": "nsx-vlan-transportzone-641", "segmentation_id": 641, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3dee43b6-26", "ovs_interfaceid": "3dee43b6-2650-42e1-aa5a-4994a3ec7f05", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1039.361973] env[63515]: DEBUG oslo_concurrency.lockutils [None req-5a78568a-9d2c-4ab4-aaf0-b4782bd16af7 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Lock "interface-724b05bc-ce4b-4f99-ad14-f3abea5567f2-6a98da26-06f3-4eb5-aee8-02a448c34749" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 10.572s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1039.404739] env[63515]: DEBUG nova.scheduler.client.report [None req-3a846554-7de2-4e42-88d0-6461fe3d7394 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1039.419407] env[63515]: DEBUG oslo_vmware.api [None req-123efa07-e143-4b76-a7d6-1e616f30576d tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Task: {'id': task-1111862, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1039.467234] env[63515]: DEBUG nova.objects.instance [None req-b7657298-0ac9-4492-ba07-d80007973987 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Lazy-loading 'pci_requests' on Instance uuid c6ac42a9-a233-48b4-bcb2-fb791c7446bb {{(pid=63515) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1039.601288] env[63515]: DEBUG oslo_concurrency.lockutils [None req-833c0cde-a35b-468a-be6f-c5dbc0f2b161 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Acquiring lock "b76cf3dd-7b7e-4e5d-bf3c-e15bd37069ac" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1039.601563] env[63515]: DEBUG oslo_concurrency.lockutils [None req-833c0cde-a35b-468a-be6f-c5dbc0f2b161 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Lock "b76cf3dd-7b7e-4e5d-bf3c-e15bd37069ac" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.001s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1039.601786] env[63515]: DEBUG nova.compute.manager [None req-833c0cde-a35b-468a-be6f-c5dbc0f2b161 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: b76cf3dd-7b7e-4e5d-bf3c-e15bd37069ac] Going to confirm migration 5 {{(pid=63515) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 1039.603460] env[63515]: DEBUG oslo_concurrency.lockutils [req-94b82895-6c06-413d-bcb5-476f1f42cb9e req-9928d174-a8c8-4b67-9ad3-1766a130bb05 service nova] Releasing lock "refresh_cache-724b05bc-ce4b-4f99-ad14-f3abea5567f2" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1039.915125] env[63515]: DEBUG oslo_concurrency.lockutils [None req-3a846554-7de2-4e42-88d0-6461fe3d7394 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.952s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1039.915738] env[63515]: DEBUG nova.compute.manager [None req-3a846554-7de2-4e42-88d0-6461fe3d7394 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: a8d2cc01-3bc1-4792-85f5-b9095294b33c] Start building networks asynchronously for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1039.924174] env[63515]: DEBUG oslo_vmware.api [None req-123efa07-e143-4b76-a7d6-1e616f30576d tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Task: {'id': task-1111862, 'name': PowerOnVM_Task, 'duration_secs': 0.709311} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1039.924466] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-123efa07-e143-4b76-a7d6-1e616f30576d tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: aa83d035-9c47-434d-ab1a-140e6b3110b4] Powered on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1039.924673] env[63515]: INFO nova.compute.manager [None req-123efa07-e143-4b76-a7d6-1e616f30576d tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: aa83d035-9c47-434d-ab1a-140e6b3110b4] Took 8.21 seconds to spawn the instance on the hypervisor. [ 1039.924850] env[63515]: DEBUG nova.compute.manager [None req-123efa07-e143-4b76-a7d6-1e616f30576d tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: aa83d035-9c47-434d-ab1a-140e6b3110b4] Checking state {{(pid=63515) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1039.925696] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06683c4c-9675-48c2-b2c9-0d4658e95127 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.970197] env[63515]: DEBUG nova.objects.base [None req-b7657298-0ac9-4492-ba07-d80007973987 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Object Instance lazy-loaded attributes: flavor,pci_requests {{(pid=63515) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1039.970407] env[63515]: DEBUG nova.network.neutron [None req-b7657298-0ac9-4492-ba07-d80007973987 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: c6ac42a9-a233-48b4-bcb2-fb791c7446bb] allocate_for_instance() {{(pid=63515) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1040.041008] env[63515]: DEBUG nova.policy [None req-b7657298-0ac9-4492-ba07-d80007973987 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '16880abf1c5b4341800c94ada2c756c7', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7d4ada9ec35f42b19c6480a9101d21a4', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63515) authorize /opt/stack/nova/nova/policy.py:201}} [ 1040.140084] env[63515]: DEBUG oslo_concurrency.lockutils [None req-833c0cde-a35b-468a-be6f-c5dbc0f2b161 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Acquiring lock "refresh_cache-b76cf3dd-7b7e-4e5d-bf3c-e15bd37069ac" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1040.140294] env[63515]: DEBUG oslo_concurrency.lockutils [None req-833c0cde-a35b-468a-be6f-c5dbc0f2b161 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Acquired lock "refresh_cache-b76cf3dd-7b7e-4e5d-bf3c-e15bd37069ac" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1040.140475] env[63515]: DEBUG nova.network.neutron [None req-833c0cde-a35b-468a-be6f-c5dbc0f2b161 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: b76cf3dd-7b7e-4e5d-bf3c-e15bd37069ac] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1040.140659] env[63515]: DEBUG nova.objects.instance [None req-833c0cde-a35b-468a-be6f-c5dbc0f2b161 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Lazy-loading 'info_cache' on Instance uuid b76cf3dd-7b7e-4e5d-bf3c-e15bd37069ac {{(pid=63515) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1040.420468] env[63515]: DEBUG nova.compute.utils [None req-3a846554-7de2-4e42-88d0-6461fe3d7394 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Using /dev/sd instead of None {{(pid=63515) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1040.421866] env[63515]: DEBUG nova.compute.manager [None req-3a846554-7de2-4e42-88d0-6461fe3d7394 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: a8d2cc01-3bc1-4792-85f5-b9095294b33c] Allocating IP information in the background. {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1040.422046] env[63515]: DEBUG nova.network.neutron [None req-3a846554-7de2-4e42-88d0-6461fe3d7394 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: a8d2cc01-3bc1-4792-85f5-b9095294b33c] allocate_for_instance() {{(pid=63515) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1040.445368] env[63515]: INFO nova.compute.manager [None req-123efa07-e143-4b76-a7d6-1e616f30576d tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: aa83d035-9c47-434d-ab1a-140e6b3110b4] Took 13.04 seconds to build instance. [ 1040.462727] env[63515]: DEBUG nova.policy [None req-3a846554-7de2-4e42-88d0-6461fe3d7394 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9b19bb2d32d84c019541c3b2e711a202', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '299fdeff647f486390366d5bbf911518', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63515) authorize /opt/stack/nova/nova/policy.py:201}} [ 1040.739020] env[63515]: DEBUG nova.network.neutron [None req-3a846554-7de2-4e42-88d0-6461fe3d7394 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: a8d2cc01-3bc1-4792-85f5-b9095294b33c] Successfully created port: 2338962e-f2a1-4b5e-aa4e-7340cb8b76b5 {{(pid=63515) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1040.925035] env[63515]: DEBUG nova.compute.manager [None req-3a846554-7de2-4e42-88d0-6461fe3d7394 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: a8d2cc01-3bc1-4792-85f5-b9095294b33c] Start building block device mappings for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1040.947378] env[63515]: DEBUG oslo_concurrency.lockutils [None req-123efa07-e143-4b76-a7d6-1e616f30576d tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Lock "aa83d035-9c47-434d-ab1a-140e6b3110b4" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.553s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1041.407667] env[63515]: DEBUG nova.network.neutron [None req-833c0cde-a35b-468a-be6f-c5dbc0f2b161 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: b76cf3dd-7b7e-4e5d-bf3c-e15bd37069ac] Updating instance_info_cache with network_info: [{"id": "acea8927-7e43-427c-bed2-2420d4708f3b", "address": "fa:16:3e:21:d3:df", "network": {"id": "3e5914c8-0cd4-49ee-81b4-1f3e280440bb", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-641632277-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ea099ce07bfb4a8da014d9303cf552eb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "225b6979-9329-403b-91fa-138bd41f6e83", "external-id": "nsx-vlan-transportzone-38", "segmentation_id": 38, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapacea8927-7e", "ovs_interfaceid": "acea8927-7e43-427c-bed2-2420d4708f3b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1041.575150] env[63515]: DEBUG nova.network.neutron [None req-b7657298-0ac9-4492-ba07-d80007973987 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: c6ac42a9-a233-48b4-bcb2-fb791c7446bb] Successfully updated port: 6a98da26-06f3-4eb5-aee8-02a448c34749 {{(pid=63515) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1041.910488] env[63515]: DEBUG oslo_concurrency.lockutils [None req-833c0cde-a35b-468a-be6f-c5dbc0f2b161 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Releasing lock "refresh_cache-b76cf3dd-7b7e-4e5d-bf3c-e15bd37069ac" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1041.910777] env[63515]: DEBUG nova.objects.instance [None req-833c0cde-a35b-468a-be6f-c5dbc0f2b161 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Lazy-loading 'migration_context' on Instance uuid b76cf3dd-7b7e-4e5d-bf3c-e15bd37069ac {{(pid=63515) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1041.934108] env[63515]: DEBUG nova.compute.manager [None req-3a846554-7de2-4e42-88d0-6461fe3d7394 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: a8d2cc01-3bc1-4792-85f5-b9095294b33c] Start spawning the instance on the hypervisor. {{(pid=63515) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1041.959881] env[63515]: DEBUG nova.virt.hardware [None req-3a846554-7de2-4e42-88d0-6461fe3d7394 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T02:52:14Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T02:51:56Z,direct_url=,disk_format='vmdk',id=8a120570-cb06-4099-b262-554ca0ad15c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b569255cc43e42e7ae3f2b4ad37c6ef4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T02:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1041.960139] env[63515]: DEBUG nova.virt.hardware [None req-3a846554-7de2-4e42-88d0-6461fe3d7394 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Flavor limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1041.960305] env[63515]: DEBUG nova.virt.hardware [None req-3a846554-7de2-4e42-88d0-6461fe3d7394 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Image limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1041.960492] env[63515]: DEBUG nova.virt.hardware [None req-3a846554-7de2-4e42-88d0-6461fe3d7394 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Flavor pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1041.960640] env[63515]: DEBUG nova.virt.hardware [None req-3a846554-7de2-4e42-88d0-6461fe3d7394 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Image pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1041.960788] env[63515]: DEBUG nova.virt.hardware [None req-3a846554-7de2-4e42-88d0-6461fe3d7394 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1041.960989] env[63515]: DEBUG nova.virt.hardware [None req-3a846554-7de2-4e42-88d0-6461fe3d7394 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1041.961166] env[63515]: DEBUG nova.virt.hardware [None req-3a846554-7de2-4e42-88d0-6461fe3d7394 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1041.961335] env[63515]: DEBUG nova.virt.hardware [None req-3a846554-7de2-4e42-88d0-6461fe3d7394 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Got 1 possible topologies {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1041.961497] env[63515]: DEBUG nova.virt.hardware [None req-3a846554-7de2-4e42-88d0-6461fe3d7394 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1041.961666] env[63515]: DEBUG nova.virt.hardware [None req-3a846554-7de2-4e42-88d0-6461fe3d7394 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1041.962524] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78e11926-2710-4a60-b7be-cdd54fd6adfe {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.970084] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b6ed656-ff86-4997-908e-79aa7a992336 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.078155] env[63515]: DEBUG oslo_concurrency.lockutils [None req-b7657298-0ac9-4492-ba07-d80007973987 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Acquiring lock "refresh_cache-c6ac42a9-a233-48b4-bcb2-fb791c7446bb" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1042.078355] env[63515]: DEBUG oslo_concurrency.lockutils [None req-b7657298-0ac9-4492-ba07-d80007973987 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Acquired lock "refresh_cache-c6ac42a9-a233-48b4-bcb2-fb791c7446bb" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1042.078537] env[63515]: DEBUG nova.network.neutron [None req-b7657298-0ac9-4492-ba07-d80007973987 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: c6ac42a9-a233-48b4-bcb2-fb791c7446bb] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1042.222282] env[63515]: DEBUG nova.network.neutron [None req-3a846554-7de2-4e42-88d0-6461fe3d7394 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: a8d2cc01-3bc1-4792-85f5-b9095294b33c] Successfully updated port: 2338962e-f2a1-4b5e-aa4e-7340cb8b76b5 {{(pid=63515) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1042.413661] env[63515]: DEBUG nova.objects.base [None req-833c0cde-a35b-468a-be6f-c5dbc0f2b161 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Object Instance lazy-loaded attributes: info_cache,migration_context {{(pid=63515) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1042.414635] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4d2c579-710d-4328-811a-0099edea3e05 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.434077] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ef3e2042-952d-40e6-b915-6d81cfb6a3ce {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.439567] env[63515]: DEBUG oslo_vmware.api [None req-833c0cde-a35b-468a-be6f-c5dbc0f2b161 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Waiting for the task: (returnval){ [ 1042.439567] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]5294313c-1c52-6816-6f77-f1c936209d23" [ 1042.439567] env[63515]: _type = "Task" [ 1042.439567] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1042.447733] env[63515]: DEBUG oslo_vmware.api [None req-833c0cde-a35b-468a-be6f-c5dbc0f2b161 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]5294313c-1c52-6816-6f77-f1c936209d23, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1042.612410] env[63515]: WARNING nova.network.neutron [None req-b7657298-0ac9-4492-ba07-d80007973987 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: c6ac42a9-a233-48b4-bcb2-fb791c7446bb] 88dfb29d-41e0-4df7-8f70-e52cda05620c already exists in list: networks containing: ['88dfb29d-41e0-4df7-8f70-e52cda05620c']. ignoring it [ 1042.724476] env[63515]: DEBUG oslo_concurrency.lockutils [None req-3a846554-7de2-4e42-88d0-6461fe3d7394 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Acquiring lock "refresh_cache-a8d2cc01-3bc1-4792-85f5-b9095294b33c" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1042.724619] env[63515]: DEBUG oslo_concurrency.lockutils [None req-3a846554-7de2-4e42-88d0-6461fe3d7394 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Acquired lock "refresh_cache-a8d2cc01-3bc1-4792-85f5-b9095294b33c" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1042.725026] env[63515]: DEBUG nova.network.neutron [None req-3a846554-7de2-4e42-88d0-6461fe3d7394 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: a8d2cc01-3bc1-4792-85f5-b9095294b33c] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1042.867981] env[63515]: DEBUG nova.network.neutron [None req-b7657298-0ac9-4492-ba07-d80007973987 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: c6ac42a9-a233-48b4-bcb2-fb791c7446bb] Updating instance_info_cache with network_info: [{"id": "161e8482-76a0-4c82-8354-af132110794d", "address": "fa:16:3e:3e:77:7f", "network": {"id": "88dfb29d-41e0-4df7-8f70-e52cda05620c", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1355578400-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.146", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7d4ada9ec35f42b19c6480a9101d21a4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "357d2811-e990-4985-9f9e-b158d10d3699", "external-id": "nsx-vlan-transportzone-641", "segmentation_id": 641, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap161e8482-76", "ovs_interfaceid": "161e8482-76a0-4c82-8354-af132110794d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "6a98da26-06f3-4eb5-aee8-02a448c34749", "address": "fa:16:3e:31:42:a5", "network": {"id": "88dfb29d-41e0-4df7-8f70-e52cda05620c", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1355578400-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7d4ada9ec35f42b19c6480a9101d21a4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "357d2811-e990-4985-9f9e-b158d10d3699", "external-id": "nsx-vlan-transportzone-641", "segmentation_id": 641, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6a98da26-06", "ovs_interfaceid": "6a98da26-06f3-4eb5-aee8-02a448c34749", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1042.949538] env[63515]: DEBUG oslo_vmware.api [None req-833c0cde-a35b-468a-be6f-c5dbc0f2b161 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]5294313c-1c52-6816-6f77-f1c936209d23, 'name': SearchDatastore_Task, 'duration_secs': 0.008613} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1042.949974] env[63515]: DEBUG oslo_concurrency.lockutils [None req-833c0cde-a35b-468a-be6f-c5dbc0f2b161 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1042.950082] env[63515]: DEBUG oslo_concurrency.lockutils [None req-833c0cde-a35b-468a-be6f-c5dbc0f2b161 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1043.255542] env[63515]: DEBUG nova.network.neutron [None req-3a846554-7de2-4e42-88d0-6461fe3d7394 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: a8d2cc01-3bc1-4792-85f5-b9095294b33c] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1043.370582] env[63515]: DEBUG oslo_concurrency.lockutils [None req-b7657298-0ac9-4492-ba07-d80007973987 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Releasing lock "refresh_cache-c6ac42a9-a233-48b4-bcb2-fb791c7446bb" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1043.371273] env[63515]: DEBUG oslo_concurrency.lockutils [None req-b7657298-0ac9-4492-ba07-d80007973987 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Acquiring lock "c6ac42a9-a233-48b4-bcb2-fb791c7446bb" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1043.371436] env[63515]: DEBUG oslo_concurrency.lockutils [None req-b7657298-0ac9-4492-ba07-d80007973987 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Acquired lock "c6ac42a9-a233-48b4-bcb2-fb791c7446bb" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1043.372304] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e647532-b424-4f14-b63d-d5ebcab385d6 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.375677] env[63515]: DEBUG nova.network.neutron [None req-3a846554-7de2-4e42-88d0-6461fe3d7394 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: a8d2cc01-3bc1-4792-85f5-b9095294b33c] Updating instance_info_cache with network_info: [{"id": "2338962e-f2a1-4b5e-aa4e-7340cb8b76b5", "address": "fa:16:3e:89:21:12", "network": {"id": "581d448e-be1b-4d5b-8a6e-c27326aaffc5", "bridge": "br-int", "label": "tempest-ServersTestJSON-1769610073-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "299fdeff647f486390366d5bbf911518", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fb143ef7-8271-4a8a-a4aa-8eba9a89f6a1", "external-id": "nsx-vlan-transportzone-504", "segmentation_id": 504, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2338962e-f2", "ovs_interfaceid": "2338962e-f2a1-4b5e-aa4e-7340cb8b76b5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1043.391950] env[63515]: DEBUG nova.virt.hardware [None req-b7657298-0ac9-4492-ba07-d80007973987 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T02:52:14Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1043.392219] env[63515]: DEBUG nova.virt.hardware [None req-b7657298-0ac9-4492-ba07-d80007973987 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Flavor limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1043.392382] env[63515]: DEBUG nova.virt.hardware [None req-b7657298-0ac9-4492-ba07-d80007973987 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Image limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1043.392621] env[63515]: DEBUG nova.virt.hardware [None req-b7657298-0ac9-4492-ba07-d80007973987 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Flavor pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1043.392789] env[63515]: DEBUG nova.virt.hardware [None req-b7657298-0ac9-4492-ba07-d80007973987 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Image pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1043.392940] env[63515]: DEBUG nova.virt.hardware [None req-b7657298-0ac9-4492-ba07-d80007973987 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1043.393161] env[63515]: DEBUG nova.virt.hardware [None req-b7657298-0ac9-4492-ba07-d80007973987 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1043.393327] env[63515]: DEBUG nova.virt.hardware [None req-b7657298-0ac9-4492-ba07-d80007973987 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1043.393499] env[63515]: DEBUG nova.virt.hardware [None req-b7657298-0ac9-4492-ba07-d80007973987 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Got 1 possible topologies {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1043.393657] env[63515]: DEBUG nova.virt.hardware [None req-b7657298-0ac9-4492-ba07-d80007973987 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1043.393827] env[63515]: DEBUG nova.virt.hardware [None req-b7657298-0ac9-4492-ba07-d80007973987 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1043.400146] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-b7657298-0ac9-4492-ba07-d80007973987 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: c6ac42a9-a233-48b4-bcb2-fb791c7446bb] Reconfiguring VM to attach interface {{(pid=63515) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1923}} [ 1043.400988] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-36b58900-aa0e-4eef-b78a-99c6a471e67a {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.417368] env[63515]: DEBUG oslo_vmware.api [None req-b7657298-0ac9-4492-ba07-d80007973987 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Waiting for the task: (returnval){ [ 1043.417368] env[63515]: value = "task-1111863" [ 1043.417368] env[63515]: _type = "Task" [ 1043.417368] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1043.425020] env[63515]: DEBUG oslo_vmware.api [None req-b7657298-0ac9-4492-ba07-d80007973987 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Task: {'id': task-1111863, 'name': ReconfigVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1043.580139] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9b41844-2a41-4847-ac81-827bd1c195cd {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.587660] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f214565a-968a-4e7e-8d2f-3959faea308a {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.616083] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00488c1a-6c63-403a-a286-68c11355c561 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.623008] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5a04f78-a045-4b4f-a327-3c4b59562d14 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.635749] env[63515]: DEBUG nova.compute.provider_tree [None req-833c0cde-a35b-468a-be6f-c5dbc0f2b161 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1043.878783] env[63515]: DEBUG oslo_concurrency.lockutils [None req-3a846554-7de2-4e42-88d0-6461fe3d7394 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Releasing lock "refresh_cache-a8d2cc01-3bc1-4792-85f5-b9095294b33c" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1043.879145] env[63515]: DEBUG nova.compute.manager [None req-3a846554-7de2-4e42-88d0-6461fe3d7394 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: a8d2cc01-3bc1-4792-85f5-b9095294b33c] Instance network_info: |[{"id": "2338962e-f2a1-4b5e-aa4e-7340cb8b76b5", "address": "fa:16:3e:89:21:12", "network": {"id": "581d448e-be1b-4d5b-8a6e-c27326aaffc5", "bridge": "br-int", "label": "tempest-ServersTestJSON-1769610073-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "299fdeff647f486390366d5bbf911518", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fb143ef7-8271-4a8a-a4aa-8eba9a89f6a1", "external-id": "nsx-vlan-transportzone-504", "segmentation_id": 504, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2338962e-f2", "ovs_interfaceid": "2338962e-f2a1-4b5e-aa4e-7340cb8b76b5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1043.879669] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-3a846554-7de2-4e42-88d0-6461fe3d7394 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: a8d2cc01-3bc1-4792-85f5-b9095294b33c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:89:21:12', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'fb143ef7-8271-4a8a-a4aa-8eba9a89f6a1', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '2338962e-f2a1-4b5e-aa4e-7340cb8b76b5', 'vif_model': 'vmxnet3'}] {{(pid=63515) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1043.887288] env[63515]: DEBUG oslo.service.loopingcall [None req-3a846554-7de2-4e42-88d0-6461fe3d7394 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1043.887504] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a8d2cc01-3bc1-4792-85f5-b9095294b33c] Creating VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1043.887769] env[63515]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-0ae59301-710e-4903-8ee5-73b9c8027b34 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.909068] env[63515]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1043.909068] env[63515]: value = "task-1111864" [ 1043.909068] env[63515]: _type = "Task" [ 1043.909068] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1043.915969] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1111864, 'name': CreateVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1043.925595] env[63515]: DEBUG oslo_vmware.api [None req-b7657298-0ac9-4492-ba07-d80007973987 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Task: {'id': task-1111863, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1044.139394] env[63515]: DEBUG nova.scheduler.client.report [None req-833c0cde-a35b-468a-be6f-c5dbc0f2b161 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1044.418508] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1111864, 'name': CreateVM_Task, 'duration_secs': 0.318947} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1044.418722] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a8d2cc01-3bc1-4792-85f5-b9095294b33c] Created VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1044.421932] env[63515]: DEBUG oslo_concurrency.lockutils [None req-3a846554-7de2-4e42-88d0-6461fe3d7394 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1044.422128] env[63515]: DEBUG oslo_concurrency.lockutils [None req-3a846554-7de2-4e42-88d0-6461fe3d7394 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1044.422436] env[63515]: DEBUG oslo_concurrency.lockutils [None req-3a846554-7de2-4e42-88d0-6461fe3d7394 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1044.422681] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-238e9514-f5c8-4b79-ae62-45f8e83f03ec {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.428547] env[63515]: DEBUG oslo_vmware.api [None req-b7657298-0ac9-4492-ba07-d80007973987 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Task: {'id': task-1111863, 'name': ReconfigVM_Task, 'duration_secs': 0.691024} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1044.429798] env[63515]: DEBUG oslo_concurrency.lockutils [None req-b7657298-0ac9-4492-ba07-d80007973987 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Releasing lock "c6ac42a9-a233-48b4-bcb2-fb791c7446bb" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1044.430017] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-b7657298-0ac9-4492-ba07-d80007973987 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: c6ac42a9-a233-48b4-bcb2-fb791c7446bb] Reconfigured VM to attach interface {{(pid=63515) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1937}} [ 1044.432413] env[63515]: DEBUG oslo_vmware.api [None req-3a846554-7de2-4e42-88d0-6461fe3d7394 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Waiting for the task: (returnval){ [ 1044.432413] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]5229849a-e6dc-f33d-462c-e4a1fccaaaa8" [ 1044.432413] env[63515]: _type = "Task" [ 1044.432413] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1044.439835] env[63515]: DEBUG oslo_vmware.api [None req-3a846554-7de2-4e42-88d0-6461fe3d7394 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]5229849a-e6dc-f33d-462c-e4a1fccaaaa8, 'name': SearchDatastore_Task, 'duration_secs': 0.008998} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1044.440109] env[63515]: DEBUG oslo_concurrency.lockutils [None req-3a846554-7de2-4e42-88d0-6461fe3d7394 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1044.440339] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-3a846554-7de2-4e42-88d0-6461fe3d7394 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: a8d2cc01-3bc1-4792-85f5-b9095294b33c] Processing image 8a120570-cb06-4099-b262-554ca0ad15c5 {{(pid=63515) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1044.440565] env[63515]: DEBUG oslo_concurrency.lockutils [None req-3a846554-7de2-4e42-88d0-6461fe3d7394 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1044.440711] env[63515]: DEBUG oslo_concurrency.lockutils [None req-3a846554-7de2-4e42-88d0-6461fe3d7394 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1044.440889] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-3a846554-7de2-4e42-88d0-6461fe3d7394 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1044.441130] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1967771b-c757-4fea-a3e6-f260cd73ba98 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.448354] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-3a846554-7de2-4e42-88d0-6461fe3d7394 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1044.448524] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-3a846554-7de2-4e42-88d0-6461fe3d7394 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63515) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1044.449215] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-41a3f0fb-73e6-49ad-91ac-e20512c91c72 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.454516] env[63515]: DEBUG oslo_vmware.api [None req-3a846554-7de2-4e42-88d0-6461fe3d7394 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Waiting for the task: (returnval){ [ 1044.454516] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]52795d8b-b521-a92b-4ff9-6913453fafa7" [ 1044.454516] env[63515]: _type = "Task" [ 1044.454516] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1044.462171] env[63515]: DEBUG oslo_vmware.api [None req-3a846554-7de2-4e42-88d0-6461fe3d7394 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52795d8b-b521-a92b-4ff9-6913453fafa7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1044.934701] env[63515]: DEBUG oslo_concurrency.lockutils [None req-b7657298-0ac9-4492-ba07-d80007973987 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Lock "interface-c6ac42a9-a233-48b4-bcb2-fb791c7446bb-6a98da26-06f3-4eb5-aee8-02a448c34749" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 6.076s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1044.965559] env[63515]: DEBUG oslo_vmware.api [None req-3a846554-7de2-4e42-88d0-6461fe3d7394 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52795d8b-b521-a92b-4ff9-6913453fafa7, 'name': SearchDatastore_Task, 'duration_secs': 0.008021} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1044.966455] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f0cc49bd-ab29-4e60-ac9b-b34138a91a16 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.971871] env[63515]: DEBUG oslo_vmware.api [None req-3a846554-7de2-4e42-88d0-6461fe3d7394 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Waiting for the task: (returnval){ [ 1044.971871] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]52bec57d-9434-8d65-8b0e-030f38ef57dc" [ 1044.971871] env[63515]: _type = "Task" [ 1044.971871] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1044.982688] env[63515]: DEBUG oslo_vmware.api [None req-3a846554-7de2-4e42-88d0-6461fe3d7394 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52bec57d-9434-8d65-8b0e-030f38ef57dc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1045.150231] env[63515]: DEBUG oslo_concurrency.lockutils [None req-833c0cde-a35b-468a-be6f-c5dbc0f2b161 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.200s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1045.482132] env[63515]: DEBUG oslo_vmware.api [None req-3a846554-7de2-4e42-88d0-6461fe3d7394 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52bec57d-9434-8d65-8b0e-030f38ef57dc, 'name': SearchDatastore_Task, 'duration_secs': 0.016674} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1045.482356] env[63515]: DEBUG oslo_concurrency.lockutils [None req-3a846554-7de2-4e42-88d0-6461fe3d7394 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1045.482606] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-3a846554-7de2-4e42-88d0-6461fe3d7394 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk to [datastore1] a8d2cc01-3bc1-4792-85f5-b9095294b33c/a8d2cc01-3bc1-4792-85f5-b9095294b33c.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1045.482854] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7e2bba61-4910-4dfe-9e8c-378426d839cc {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.489588] env[63515]: DEBUG oslo_vmware.api [None req-3a846554-7de2-4e42-88d0-6461fe3d7394 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Waiting for the task: (returnval){ [ 1045.489588] env[63515]: value = "task-1111865" [ 1045.489588] env[63515]: _type = "Task" [ 1045.489588] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1045.496868] env[63515]: DEBUG oslo_vmware.api [None req-3a846554-7de2-4e42-88d0-6461fe3d7394 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Task: {'id': task-1111865, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1045.708511] env[63515]: INFO nova.scheduler.client.report [None req-833c0cde-a35b-468a-be6f-c5dbc0f2b161 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Deleted allocation for migration ba725c49-c030-4294-a921-aba61789a9c4 [ 1045.999639] env[63515]: DEBUG oslo_vmware.api [None req-3a846554-7de2-4e42-88d0-6461fe3d7394 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Task: {'id': task-1111865, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.502766} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1045.999900] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-3a846554-7de2-4e42-88d0-6461fe3d7394 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk to [datastore1] a8d2cc01-3bc1-4792-85f5-b9095294b33c/a8d2cc01-3bc1-4792-85f5-b9095294b33c.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1046.000133] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-3a846554-7de2-4e42-88d0-6461fe3d7394 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: a8d2cc01-3bc1-4792-85f5-b9095294b33c] Extending root virtual disk to 1048576 {{(pid=63515) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1046.000394] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-6b7034bb-f368-479d-8423-51843dc4b8d0 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.006841] env[63515]: DEBUG oslo_vmware.api [None req-3a846554-7de2-4e42-88d0-6461fe3d7394 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Waiting for the task: (returnval){ [ 1046.006841] env[63515]: value = "task-1111866" [ 1046.006841] env[63515]: _type = "Task" [ 1046.006841] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1046.014520] env[63515]: DEBUG oslo_vmware.api [None req-3a846554-7de2-4e42-88d0-6461fe3d7394 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Task: {'id': task-1111866, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1046.215217] env[63515]: DEBUG oslo_concurrency.lockutils [None req-833c0cde-a35b-468a-be6f-c5dbc0f2b161 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Lock "b76cf3dd-7b7e-4e5d-bf3c-e15bd37069ac" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 6.613s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1046.278675] env[63515]: DEBUG nova.compute.manager [req-c554d3a8-2e44-4715-b59a-ed6804259c87 req-6893c80c-ecbb-4cb0-bd1e-b4e59e26b2b6 service nova] [instance: c6ac42a9-a233-48b4-bcb2-fb791c7446bb] Received event network-changed-161e8482-76a0-4c82-8354-af132110794d {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1046.278881] env[63515]: DEBUG nova.compute.manager [req-c554d3a8-2e44-4715-b59a-ed6804259c87 req-6893c80c-ecbb-4cb0-bd1e-b4e59e26b2b6 service nova] [instance: c6ac42a9-a233-48b4-bcb2-fb791c7446bb] Refreshing instance network info cache due to event network-changed-161e8482-76a0-4c82-8354-af132110794d. {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1046.279120] env[63515]: DEBUG oslo_concurrency.lockutils [req-c554d3a8-2e44-4715-b59a-ed6804259c87 req-6893c80c-ecbb-4cb0-bd1e-b4e59e26b2b6 service nova] Acquiring lock "refresh_cache-c6ac42a9-a233-48b4-bcb2-fb791c7446bb" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1046.279270] env[63515]: DEBUG oslo_concurrency.lockutils [req-c554d3a8-2e44-4715-b59a-ed6804259c87 req-6893c80c-ecbb-4cb0-bd1e-b4e59e26b2b6 service nova] Acquired lock "refresh_cache-c6ac42a9-a233-48b4-bcb2-fb791c7446bb" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1046.279438] env[63515]: DEBUG nova.network.neutron [req-c554d3a8-2e44-4715-b59a-ed6804259c87 req-6893c80c-ecbb-4cb0-bd1e-b4e59e26b2b6 service nova] [instance: c6ac42a9-a233-48b4-bcb2-fb791c7446bb] Refreshing network info cache for port 161e8482-76a0-4c82-8354-af132110794d {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1046.293814] env[63515]: DEBUG nova.compute.manager [req-3c75ff60-ae43-4a2d-bb60-ceb8453e375f req-6aacef1a-3972-43be-b6bc-d93d8ee05df0 service nova] [instance: c6ac42a9-a233-48b4-bcb2-fb791c7446bb] Received event network-vif-plugged-6a98da26-06f3-4eb5-aee8-02a448c34749 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1046.294042] env[63515]: DEBUG oslo_concurrency.lockutils [req-3c75ff60-ae43-4a2d-bb60-ceb8453e375f req-6aacef1a-3972-43be-b6bc-d93d8ee05df0 service nova] Acquiring lock "c6ac42a9-a233-48b4-bcb2-fb791c7446bb-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1046.294272] env[63515]: DEBUG oslo_concurrency.lockutils [req-3c75ff60-ae43-4a2d-bb60-ceb8453e375f req-6aacef1a-3972-43be-b6bc-d93d8ee05df0 service nova] Lock "c6ac42a9-a233-48b4-bcb2-fb791c7446bb-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1046.294506] env[63515]: DEBUG oslo_concurrency.lockutils [req-3c75ff60-ae43-4a2d-bb60-ceb8453e375f req-6aacef1a-3972-43be-b6bc-d93d8ee05df0 service nova] Lock "c6ac42a9-a233-48b4-bcb2-fb791c7446bb-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1046.294685] env[63515]: DEBUG nova.compute.manager [req-3c75ff60-ae43-4a2d-bb60-ceb8453e375f req-6aacef1a-3972-43be-b6bc-d93d8ee05df0 service nova] [instance: c6ac42a9-a233-48b4-bcb2-fb791c7446bb] No waiting events found dispatching network-vif-plugged-6a98da26-06f3-4eb5-aee8-02a448c34749 {{(pid=63515) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1046.294856] env[63515]: WARNING nova.compute.manager [req-3c75ff60-ae43-4a2d-bb60-ceb8453e375f req-6aacef1a-3972-43be-b6bc-d93d8ee05df0 service nova] [instance: c6ac42a9-a233-48b4-bcb2-fb791c7446bb] Received unexpected event network-vif-plugged-6a98da26-06f3-4eb5-aee8-02a448c34749 for instance with vm_state active and task_state None. [ 1046.516315] env[63515]: DEBUG oslo_vmware.api [None req-3a846554-7de2-4e42-88d0-6461fe3d7394 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Task: {'id': task-1111866, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.069861} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1046.516587] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-3a846554-7de2-4e42-88d0-6461fe3d7394 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: a8d2cc01-3bc1-4792-85f5-b9095294b33c] Extended root virtual disk {{(pid=63515) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1046.517422] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c8edbbb-8c71-4ad2-bb2c-e41bb149cec3 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.539394] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-3a846554-7de2-4e42-88d0-6461fe3d7394 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: a8d2cc01-3bc1-4792-85f5-b9095294b33c] Reconfiguring VM instance instance-00000067 to attach disk [datastore1] a8d2cc01-3bc1-4792-85f5-b9095294b33c/a8d2cc01-3bc1-4792-85f5-b9095294b33c.vmdk or device None with type sparse {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1046.539716] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-687dc4d2-680c-4046-90c9-c0d361901204 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.559213] env[63515]: DEBUG oslo_vmware.api [None req-3a846554-7de2-4e42-88d0-6461fe3d7394 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Waiting for the task: (returnval){ [ 1046.559213] env[63515]: value = "task-1111867" [ 1046.559213] env[63515]: _type = "Task" [ 1046.559213] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1046.565128] env[63515]: DEBUG oslo_concurrency.lockutils [None req-a1919053-a12d-4d12-b905-db1a99127d7f tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Acquiring lock "d6c5a9f9-5e5c-4b14-9b9a-7825bbf526a5" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1046.565399] env[63515]: DEBUG oslo_concurrency.lockutils [None req-a1919053-a12d-4d12-b905-db1a99127d7f tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Lock "d6c5a9f9-5e5c-4b14-9b9a-7825bbf526a5" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1046.570051] env[63515]: DEBUG oslo_vmware.api [None req-3a846554-7de2-4e42-88d0-6461fe3d7394 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Task: {'id': task-1111867, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1046.994316] env[63515]: DEBUG nova.network.neutron [req-c554d3a8-2e44-4715-b59a-ed6804259c87 req-6893c80c-ecbb-4cb0-bd1e-b4e59e26b2b6 service nova] [instance: c6ac42a9-a233-48b4-bcb2-fb791c7446bb] Updated VIF entry in instance network info cache for port 161e8482-76a0-4c82-8354-af132110794d. {{(pid=63515) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1046.994839] env[63515]: DEBUG nova.network.neutron [req-c554d3a8-2e44-4715-b59a-ed6804259c87 req-6893c80c-ecbb-4cb0-bd1e-b4e59e26b2b6 service nova] [instance: c6ac42a9-a233-48b4-bcb2-fb791c7446bb] Updating instance_info_cache with network_info: [{"id": "161e8482-76a0-4c82-8354-af132110794d", "address": "fa:16:3e:3e:77:7f", "network": {"id": "88dfb29d-41e0-4df7-8f70-e52cda05620c", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1355578400-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.146", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7d4ada9ec35f42b19c6480a9101d21a4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "357d2811-e990-4985-9f9e-b158d10d3699", "external-id": "nsx-vlan-transportzone-641", "segmentation_id": 641, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap161e8482-76", "ovs_interfaceid": "161e8482-76a0-4c82-8354-af132110794d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "6a98da26-06f3-4eb5-aee8-02a448c34749", "address": "fa:16:3e:31:42:a5", "network": {"id": "88dfb29d-41e0-4df7-8f70-e52cda05620c", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1355578400-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7d4ada9ec35f42b19c6480a9101d21a4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "357d2811-e990-4985-9f9e-b158d10d3699", "external-id": "nsx-vlan-transportzone-641", "segmentation_id": 641, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6a98da26-06", "ovs_interfaceid": "6a98da26-06f3-4eb5-aee8-02a448c34749", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1047.068842] env[63515]: DEBUG oslo_vmware.api [None req-3a846554-7de2-4e42-88d0-6461fe3d7394 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Task: {'id': task-1111867, 'name': ReconfigVM_Task, 'duration_secs': 0.293259} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1047.068842] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-3a846554-7de2-4e42-88d0-6461fe3d7394 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: a8d2cc01-3bc1-4792-85f5-b9095294b33c] Reconfigured VM instance instance-00000067 to attach disk [datastore1] a8d2cc01-3bc1-4792-85f5-b9095294b33c/a8d2cc01-3bc1-4792-85f5-b9095294b33c.vmdk or device None with type sparse {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1047.069098] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6a85114f-f6c8-4337-8bf2-07f623ec6846 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.071396] env[63515]: DEBUG nova.compute.utils [None req-a1919053-a12d-4d12-b905-db1a99127d7f tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Using /dev/sd instead of None {{(pid=63515) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1047.077392] env[63515]: DEBUG oslo_vmware.api [None req-3a846554-7de2-4e42-88d0-6461fe3d7394 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Waiting for the task: (returnval){ [ 1047.077392] env[63515]: value = "task-1111868" [ 1047.077392] env[63515]: _type = "Task" [ 1047.077392] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1047.086098] env[63515]: DEBUG oslo_vmware.api [None req-3a846554-7de2-4e42-88d0-6461fe3d7394 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Task: {'id': task-1111868, 'name': Rename_Task} progress is 5%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1047.497810] env[63515]: DEBUG oslo_concurrency.lockutils [req-c554d3a8-2e44-4715-b59a-ed6804259c87 req-6893c80c-ecbb-4cb0-bd1e-b4e59e26b2b6 service nova] Releasing lock "refresh_cache-c6ac42a9-a233-48b4-bcb2-fb791c7446bb" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1047.574890] env[63515]: DEBUG oslo_concurrency.lockutils [None req-a1919053-a12d-4d12-b905-db1a99127d7f tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Lock "d6c5a9f9-5e5c-4b14-9b9a-7825bbf526a5" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.009s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1047.587110] env[63515]: DEBUG oslo_vmware.api [None req-3a846554-7de2-4e42-88d0-6461fe3d7394 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Task: {'id': task-1111868, 'name': Rename_Task, 'duration_secs': 0.321891} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1047.587483] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-3a846554-7de2-4e42-88d0-6461fe3d7394 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: a8d2cc01-3bc1-4792-85f5-b9095294b33c] Powering on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1047.587770] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c31102c0-3356-43e6-8839-f4a2aedbeadd {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.593787] env[63515]: DEBUG oslo_vmware.api [None req-3a846554-7de2-4e42-88d0-6461fe3d7394 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Waiting for the task: (returnval){ [ 1047.593787] env[63515]: value = "task-1111869" [ 1047.593787] env[63515]: _type = "Task" [ 1047.593787] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1047.601404] env[63515]: DEBUG oslo_vmware.api [None req-3a846554-7de2-4e42-88d0-6461fe3d7394 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Task: {'id': task-1111869, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1047.627188] env[63515]: DEBUG oslo_concurrency.lockutils [None req-833c0cde-a35b-468a-be6f-c5dbc0f2b161 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Acquiring lock "b76cf3dd-7b7e-4e5d-bf3c-e15bd37069ac" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1047.627518] env[63515]: DEBUG oslo_concurrency.lockutils [None req-833c0cde-a35b-468a-be6f-c5dbc0f2b161 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Lock "b76cf3dd-7b7e-4e5d-bf3c-e15bd37069ac" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1047.627738] env[63515]: DEBUG oslo_concurrency.lockutils [None req-833c0cde-a35b-468a-be6f-c5dbc0f2b161 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Acquiring lock "b76cf3dd-7b7e-4e5d-bf3c-e15bd37069ac-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1047.627924] env[63515]: DEBUG oslo_concurrency.lockutils [None req-833c0cde-a35b-468a-be6f-c5dbc0f2b161 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Lock "b76cf3dd-7b7e-4e5d-bf3c-e15bd37069ac-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1047.628120] env[63515]: DEBUG oslo_concurrency.lockutils [None req-833c0cde-a35b-468a-be6f-c5dbc0f2b161 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Lock "b76cf3dd-7b7e-4e5d-bf3c-e15bd37069ac-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1047.630348] env[63515]: INFO nova.compute.manager [None req-833c0cde-a35b-468a-be6f-c5dbc0f2b161 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: b76cf3dd-7b7e-4e5d-bf3c-e15bd37069ac] Terminating instance [ 1047.632212] env[63515]: DEBUG nova.compute.manager [None req-833c0cde-a35b-468a-be6f-c5dbc0f2b161 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: b76cf3dd-7b7e-4e5d-bf3c-e15bd37069ac] Start destroying the instance on the hypervisor. {{(pid=63515) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1047.632412] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-833c0cde-a35b-468a-be6f-c5dbc0f2b161 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: b76cf3dd-7b7e-4e5d-bf3c-e15bd37069ac] Destroying instance {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1047.633228] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-945c6905-cc6c-4f01-9316-914298264196 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.640221] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-833c0cde-a35b-468a-be6f-c5dbc0f2b161 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: b76cf3dd-7b7e-4e5d-bf3c-e15bd37069ac] Powering off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1047.640454] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-87e88c1f-976b-4edd-a053-6e359a70bee1 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.645979] env[63515]: DEBUG oslo_vmware.api [None req-833c0cde-a35b-468a-be6f-c5dbc0f2b161 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Waiting for the task: (returnval){ [ 1047.645979] env[63515]: value = "task-1111870" [ 1047.645979] env[63515]: _type = "Task" [ 1047.645979] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1047.653327] env[63515]: DEBUG oslo_vmware.api [None req-833c0cde-a35b-468a-be6f-c5dbc0f2b161 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Task: {'id': task-1111870, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1047.672980] env[63515]: DEBUG oslo_concurrency.lockutils [None req-2795ed8a-e4d1-4a5f-8e3c-baa30a1a2f05 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Acquiring lock "interface-c6ac42a9-a233-48b4-bcb2-fb791c7446bb-6a98da26-06f3-4eb5-aee8-02a448c34749" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1047.673315] env[63515]: DEBUG oslo_concurrency.lockutils [None req-2795ed8a-e4d1-4a5f-8e3c-baa30a1a2f05 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Lock "interface-c6ac42a9-a233-48b4-bcb2-fb791c7446bb-6a98da26-06f3-4eb5-aee8-02a448c34749" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1048.104765] env[63515]: DEBUG oslo_vmware.api [None req-3a846554-7de2-4e42-88d0-6461fe3d7394 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Task: {'id': task-1111869, 'name': PowerOnVM_Task} progress is 66%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1048.157386] env[63515]: DEBUG oslo_vmware.api [None req-833c0cde-a35b-468a-be6f-c5dbc0f2b161 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Task: {'id': task-1111870, 'name': PowerOffVM_Task, 'duration_secs': 0.23309} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1048.157386] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-833c0cde-a35b-468a-be6f-c5dbc0f2b161 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: b76cf3dd-7b7e-4e5d-bf3c-e15bd37069ac] Powered off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1048.157582] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-833c0cde-a35b-468a-be6f-c5dbc0f2b161 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: b76cf3dd-7b7e-4e5d-bf3c-e15bd37069ac] Unregistering the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1048.157840] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-339525fd-3e75-4b0d-84cb-9846bb17d8b0 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.176025] env[63515]: DEBUG oslo_concurrency.lockutils [None req-2795ed8a-e4d1-4a5f-8e3c-baa30a1a2f05 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Acquiring lock "c6ac42a9-a233-48b4-bcb2-fb791c7446bb" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1048.176225] env[63515]: DEBUG oslo_concurrency.lockutils [None req-2795ed8a-e4d1-4a5f-8e3c-baa30a1a2f05 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Acquired lock "c6ac42a9-a233-48b4-bcb2-fb791c7446bb" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1048.177077] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ffb3e5a2-a8e4-4b8c-b724-166cf754c7cf {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.195129] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c5b2d3a-1096-458d-a756-709cf3c9381f {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.221909] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-2795ed8a-e4d1-4a5f-8e3c-baa30a1a2f05 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: c6ac42a9-a233-48b4-bcb2-fb791c7446bb] Reconfiguring VM to detach interface {{(pid=63515) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1968}} [ 1048.222197] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9a831044-0294-402a-80d3-e4e35a3ce2cd {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.239523] env[63515]: DEBUG oslo_vmware.api [None req-2795ed8a-e4d1-4a5f-8e3c-baa30a1a2f05 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Waiting for the task: (returnval){ [ 1048.239523] env[63515]: value = "task-1111872" [ 1048.239523] env[63515]: _type = "Task" [ 1048.239523] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1048.247156] env[63515]: DEBUG oslo_vmware.api [None req-2795ed8a-e4d1-4a5f-8e3c-baa30a1a2f05 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Task: {'id': task-1111872, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1048.304322] env[63515]: DEBUG nova.compute.manager [req-e42462d1-697f-4a00-8d46-2007b7edcc22 req-18e70ef1-111e-4647-bdd0-3e94f5571383 service nova] [instance: a8d2cc01-3bc1-4792-85f5-b9095294b33c] Received event network-vif-plugged-2338962e-f2a1-4b5e-aa4e-7340cb8b76b5 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1048.304524] env[63515]: DEBUG oslo_concurrency.lockutils [req-e42462d1-697f-4a00-8d46-2007b7edcc22 req-18e70ef1-111e-4647-bdd0-3e94f5571383 service nova] Acquiring lock "a8d2cc01-3bc1-4792-85f5-b9095294b33c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1048.304745] env[63515]: DEBUG oslo_concurrency.lockutils [req-e42462d1-697f-4a00-8d46-2007b7edcc22 req-18e70ef1-111e-4647-bdd0-3e94f5571383 service nova] Lock "a8d2cc01-3bc1-4792-85f5-b9095294b33c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1048.304917] env[63515]: DEBUG oslo_concurrency.lockutils [req-e42462d1-697f-4a00-8d46-2007b7edcc22 req-18e70ef1-111e-4647-bdd0-3e94f5571383 service nova] Lock "a8d2cc01-3bc1-4792-85f5-b9095294b33c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1048.305314] env[63515]: DEBUG nova.compute.manager [req-e42462d1-697f-4a00-8d46-2007b7edcc22 req-18e70ef1-111e-4647-bdd0-3e94f5571383 service nova] [instance: a8d2cc01-3bc1-4792-85f5-b9095294b33c] No waiting events found dispatching network-vif-plugged-2338962e-f2a1-4b5e-aa4e-7340cb8b76b5 {{(pid=63515) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1048.305529] env[63515]: WARNING nova.compute.manager [req-e42462d1-697f-4a00-8d46-2007b7edcc22 req-18e70ef1-111e-4647-bdd0-3e94f5571383 service nova] [instance: a8d2cc01-3bc1-4792-85f5-b9095294b33c] Received unexpected event network-vif-plugged-2338962e-f2a1-4b5e-aa4e-7340cb8b76b5 for instance with vm_state building and task_state spawning. [ 1048.305700] env[63515]: DEBUG nova.compute.manager [req-e42462d1-697f-4a00-8d46-2007b7edcc22 req-18e70ef1-111e-4647-bdd0-3e94f5571383 service nova] [instance: a8d2cc01-3bc1-4792-85f5-b9095294b33c] Received event network-changed-2338962e-f2a1-4b5e-aa4e-7340cb8b76b5 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1048.305862] env[63515]: DEBUG nova.compute.manager [req-e42462d1-697f-4a00-8d46-2007b7edcc22 req-18e70ef1-111e-4647-bdd0-3e94f5571383 service nova] [instance: a8d2cc01-3bc1-4792-85f5-b9095294b33c] Refreshing instance network info cache due to event network-changed-2338962e-f2a1-4b5e-aa4e-7340cb8b76b5. {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1048.306855] env[63515]: DEBUG oslo_concurrency.lockutils [req-e42462d1-697f-4a00-8d46-2007b7edcc22 req-18e70ef1-111e-4647-bdd0-3e94f5571383 service nova] Acquiring lock "refresh_cache-a8d2cc01-3bc1-4792-85f5-b9095294b33c" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1048.306855] env[63515]: DEBUG oslo_concurrency.lockutils [req-e42462d1-697f-4a00-8d46-2007b7edcc22 req-18e70ef1-111e-4647-bdd0-3e94f5571383 service nova] Acquired lock "refresh_cache-a8d2cc01-3bc1-4792-85f5-b9095294b33c" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1048.306855] env[63515]: DEBUG nova.network.neutron [req-e42462d1-697f-4a00-8d46-2007b7edcc22 req-18e70ef1-111e-4647-bdd0-3e94f5571383 service nova] [instance: a8d2cc01-3bc1-4792-85f5-b9095294b33c] Refreshing network info cache for port 2338962e-f2a1-4b5e-aa4e-7340cb8b76b5 {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1048.323051] env[63515]: DEBUG nova.compute.manager [req-2a791d47-d678-4b51-ad20-a3cabe590076 req-8e6bfb34-8b80-4fb8-be8c-7d1b816fbbc6 service nova] [instance: c6ac42a9-a233-48b4-bcb2-fb791c7446bb] Received event network-changed-6a98da26-06f3-4eb5-aee8-02a448c34749 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1048.323051] env[63515]: DEBUG nova.compute.manager [req-2a791d47-d678-4b51-ad20-a3cabe590076 req-8e6bfb34-8b80-4fb8-be8c-7d1b816fbbc6 service nova] [instance: c6ac42a9-a233-48b4-bcb2-fb791c7446bb] Refreshing instance network info cache due to event network-changed-6a98da26-06f3-4eb5-aee8-02a448c34749. {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1048.323051] env[63515]: DEBUG oslo_concurrency.lockutils [req-2a791d47-d678-4b51-ad20-a3cabe590076 req-8e6bfb34-8b80-4fb8-be8c-7d1b816fbbc6 service nova] Acquiring lock "refresh_cache-c6ac42a9-a233-48b4-bcb2-fb791c7446bb" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1048.323159] env[63515]: DEBUG oslo_concurrency.lockutils [req-2a791d47-d678-4b51-ad20-a3cabe590076 req-8e6bfb34-8b80-4fb8-be8c-7d1b816fbbc6 service nova] Acquired lock "refresh_cache-c6ac42a9-a233-48b4-bcb2-fb791c7446bb" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1048.323320] env[63515]: DEBUG nova.network.neutron [req-2a791d47-d678-4b51-ad20-a3cabe590076 req-8e6bfb34-8b80-4fb8-be8c-7d1b816fbbc6 service nova] [instance: c6ac42a9-a233-48b4-bcb2-fb791c7446bb] Refreshing network info cache for port 6a98da26-06f3-4eb5-aee8-02a448c34749 {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1048.606677] env[63515]: DEBUG oslo_vmware.api [None req-3a846554-7de2-4e42-88d0-6461fe3d7394 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Task: {'id': task-1111869, 'name': PowerOnVM_Task, 'duration_secs': 0.57475} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1048.607037] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-3a846554-7de2-4e42-88d0-6461fe3d7394 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: a8d2cc01-3bc1-4792-85f5-b9095294b33c] Powered on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1048.607145] env[63515]: INFO nova.compute.manager [None req-3a846554-7de2-4e42-88d0-6461fe3d7394 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: a8d2cc01-3bc1-4792-85f5-b9095294b33c] Took 6.67 seconds to spawn the instance on the hypervisor. [ 1048.607356] env[63515]: DEBUG nova.compute.manager [None req-3a846554-7de2-4e42-88d0-6461fe3d7394 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: a8d2cc01-3bc1-4792-85f5-b9095294b33c] Checking state {{(pid=63515) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1048.608156] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d4bb06c-a5bc-40d9-9195-92002b8d5e61 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.634246] env[63515]: DEBUG oslo_concurrency.lockutils [None req-a1919053-a12d-4d12-b905-db1a99127d7f tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Acquiring lock "d6c5a9f9-5e5c-4b14-9b9a-7825bbf526a5" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1048.634516] env[63515]: DEBUG oslo_concurrency.lockutils [None req-a1919053-a12d-4d12-b905-db1a99127d7f tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Lock "d6c5a9f9-5e5c-4b14-9b9a-7825bbf526a5" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1048.634753] env[63515]: INFO nova.compute.manager [None req-a1919053-a12d-4d12-b905-db1a99127d7f tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: d6c5a9f9-5e5c-4b14-9b9a-7825bbf526a5] Attaching volume 71cdbc05-35d6-43f4-9224-cab57f2d0c9c to /dev/sdb [ 1048.666279] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3b47e7c-8524-4732-ba81-f11cd8e007c4 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.673562] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-607ef4ea-d156-442b-ab21-4f183f45c555 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.686745] env[63515]: DEBUG nova.virt.block_device [None req-a1919053-a12d-4d12-b905-db1a99127d7f tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: d6c5a9f9-5e5c-4b14-9b9a-7825bbf526a5] Updating existing volume attachment record: 534486cf-c3d3-42b8-90ce-88ac7b1f7baa {{(pid=63515) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1048.749235] env[63515]: DEBUG oslo_vmware.api [None req-2795ed8a-e4d1-4a5f-8e3c-baa30a1a2f05 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Task: {'id': task-1111872, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1049.033608] env[63515]: DEBUG nova.network.neutron [req-e42462d1-697f-4a00-8d46-2007b7edcc22 req-18e70ef1-111e-4647-bdd0-3e94f5571383 service nova] [instance: a8d2cc01-3bc1-4792-85f5-b9095294b33c] Updated VIF entry in instance network info cache for port 2338962e-f2a1-4b5e-aa4e-7340cb8b76b5. {{(pid=63515) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1049.033883] env[63515]: DEBUG nova.network.neutron [req-e42462d1-697f-4a00-8d46-2007b7edcc22 req-18e70ef1-111e-4647-bdd0-3e94f5571383 service nova] [instance: a8d2cc01-3bc1-4792-85f5-b9095294b33c] Updating instance_info_cache with network_info: [{"id": "2338962e-f2a1-4b5e-aa4e-7340cb8b76b5", "address": "fa:16:3e:89:21:12", "network": {"id": "581d448e-be1b-4d5b-8a6e-c27326aaffc5", "bridge": "br-int", "label": "tempest-ServersTestJSON-1769610073-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "299fdeff647f486390366d5bbf911518", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fb143ef7-8271-4a8a-a4aa-8eba9a89f6a1", "external-id": "nsx-vlan-transportzone-504", "segmentation_id": 504, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2338962e-f2", "ovs_interfaceid": "2338962e-f2a1-4b5e-aa4e-7340cb8b76b5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1049.035923] env[63515]: DEBUG nova.network.neutron [req-2a791d47-d678-4b51-ad20-a3cabe590076 req-8e6bfb34-8b80-4fb8-be8c-7d1b816fbbc6 service nova] [instance: c6ac42a9-a233-48b4-bcb2-fb791c7446bb] Updated VIF entry in instance network info cache for port 6a98da26-06f3-4eb5-aee8-02a448c34749. {{(pid=63515) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1049.036314] env[63515]: DEBUG nova.network.neutron [req-2a791d47-d678-4b51-ad20-a3cabe590076 req-8e6bfb34-8b80-4fb8-be8c-7d1b816fbbc6 service nova] [instance: c6ac42a9-a233-48b4-bcb2-fb791c7446bb] Updating instance_info_cache with network_info: [{"id": "161e8482-76a0-4c82-8354-af132110794d", "address": "fa:16:3e:3e:77:7f", "network": {"id": "88dfb29d-41e0-4df7-8f70-e52cda05620c", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1355578400-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.146", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7d4ada9ec35f42b19c6480a9101d21a4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "357d2811-e990-4985-9f9e-b158d10d3699", "external-id": "nsx-vlan-transportzone-641", "segmentation_id": 641, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap161e8482-76", "ovs_interfaceid": "161e8482-76a0-4c82-8354-af132110794d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "6a98da26-06f3-4eb5-aee8-02a448c34749", "address": "fa:16:3e:31:42:a5", "network": {"id": "88dfb29d-41e0-4df7-8f70-e52cda05620c", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1355578400-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7d4ada9ec35f42b19c6480a9101d21a4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "357d2811-e990-4985-9f9e-b158d10d3699", "external-id": "nsx-vlan-transportzone-641", "segmentation_id": 641, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6a98da26-06", "ovs_interfaceid": "6a98da26-06f3-4eb5-aee8-02a448c34749", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1049.081916] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-833c0cde-a35b-468a-be6f-c5dbc0f2b161 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: b76cf3dd-7b7e-4e5d-bf3c-e15bd37069ac] Unregistered the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1049.082175] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-833c0cde-a35b-468a-be6f-c5dbc0f2b161 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: b76cf3dd-7b7e-4e5d-bf3c-e15bd37069ac] Deleting contents of the VM from datastore datastore1 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1049.082396] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-833c0cde-a35b-468a-be6f-c5dbc0f2b161 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Deleting the datastore file [datastore1] b76cf3dd-7b7e-4e5d-bf3c-e15bd37069ac {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1049.082693] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3f878326-2022-498c-bf5f-890e7ebd049f {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.090726] env[63515]: DEBUG oslo_vmware.api [None req-833c0cde-a35b-468a-be6f-c5dbc0f2b161 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Waiting for the task: (returnval){ [ 1049.090726] env[63515]: value = "task-1111876" [ 1049.090726] env[63515]: _type = "Task" [ 1049.090726] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1049.099266] env[63515]: DEBUG oslo_vmware.api [None req-833c0cde-a35b-468a-be6f-c5dbc0f2b161 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Task: {'id': task-1111876, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1049.125873] env[63515]: INFO nova.compute.manager [None req-3a846554-7de2-4e42-88d0-6461fe3d7394 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: a8d2cc01-3bc1-4792-85f5-b9095294b33c] Took 12.18 seconds to build instance. [ 1049.252146] env[63515]: DEBUG oslo_vmware.api [None req-2795ed8a-e4d1-4a5f-8e3c-baa30a1a2f05 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Task: {'id': task-1111872, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1049.538923] env[63515]: DEBUG oslo_concurrency.lockutils [req-e42462d1-697f-4a00-8d46-2007b7edcc22 req-18e70ef1-111e-4647-bdd0-3e94f5571383 service nova] Releasing lock "refresh_cache-a8d2cc01-3bc1-4792-85f5-b9095294b33c" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1049.539546] env[63515]: DEBUG oslo_concurrency.lockutils [req-2a791d47-d678-4b51-ad20-a3cabe590076 req-8e6bfb34-8b80-4fb8-be8c-7d1b816fbbc6 service nova] Releasing lock "refresh_cache-c6ac42a9-a233-48b4-bcb2-fb791c7446bb" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1049.539794] env[63515]: DEBUG nova.compute.manager [req-2a791d47-d678-4b51-ad20-a3cabe590076 req-8e6bfb34-8b80-4fb8-be8c-7d1b816fbbc6 service nova] [instance: aa83d035-9c47-434d-ab1a-140e6b3110b4] Received event network-changed-614aed8f-3ab8-4672-ac13-0ae8cb0b233d {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1049.539991] env[63515]: DEBUG nova.compute.manager [req-2a791d47-d678-4b51-ad20-a3cabe590076 req-8e6bfb34-8b80-4fb8-be8c-7d1b816fbbc6 service nova] [instance: aa83d035-9c47-434d-ab1a-140e6b3110b4] Refreshing instance network info cache due to event network-changed-614aed8f-3ab8-4672-ac13-0ae8cb0b233d. {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1049.540202] env[63515]: DEBUG oslo_concurrency.lockutils [req-2a791d47-d678-4b51-ad20-a3cabe590076 req-8e6bfb34-8b80-4fb8-be8c-7d1b816fbbc6 service nova] Acquiring lock "refresh_cache-aa83d035-9c47-434d-ab1a-140e6b3110b4" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1049.540396] env[63515]: DEBUG oslo_concurrency.lockutils [req-2a791d47-d678-4b51-ad20-a3cabe590076 req-8e6bfb34-8b80-4fb8-be8c-7d1b816fbbc6 service nova] Acquired lock "refresh_cache-aa83d035-9c47-434d-ab1a-140e6b3110b4" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1049.540568] env[63515]: DEBUG nova.network.neutron [req-2a791d47-d678-4b51-ad20-a3cabe590076 req-8e6bfb34-8b80-4fb8-be8c-7d1b816fbbc6 service nova] [instance: aa83d035-9c47-434d-ab1a-140e6b3110b4] Refreshing network info cache for port 614aed8f-3ab8-4672-ac13-0ae8cb0b233d {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1049.600128] env[63515]: DEBUG oslo_vmware.api [None req-833c0cde-a35b-468a-be6f-c5dbc0f2b161 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Task: {'id': task-1111876, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.143406} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1049.600392] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-833c0cde-a35b-468a-be6f-c5dbc0f2b161 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Deleted the datastore file {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1049.600585] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-833c0cde-a35b-468a-be6f-c5dbc0f2b161 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: b76cf3dd-7b7e-4e5d-bf3c-e15bd37069ac] Deleted contents of the VM from datastore datastore1 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1049.600767] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-833c0cde-a35b-468a-be6f-c5dbc0f2b161 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: b76cf3dd-7b7e-4e5d-bf3c-e15bd37069ac] Instance destroyed {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1049.600946] env[63515]: INFO nova.compute.manager [None req-833c0cde-a35b-468a-be6f-c5dbc0f2b161 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] [instance: b76cf3dd-7b7e-4e5d-bf3c-e15bd37069ac] Took 1.97 seconds to destroy the instance on the hypervisor. [ 1049.601212] env[63515]: DEBUG oslo.service.loopingcall [None req-833c0cde-a35b-468a-be6f-c5dbc0f2b161 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1049.601412] env[63515]: DEBUG nova.compute.manager [-] [instance: b76cf3dd-7b7e-4e5d-bf3c-e15bd37069ac] Deallocating network for instance {{(pid=63515) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1049.601505] env[63515]: DEBUG nova.network.neutron [-] [instance: b76cf3dd-7b7e-4e5d-bf3c-e15bd37069ac] deallocate_for_instance() {{(pid=63515) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1049.628245] env[63515]: DEBUG oslo_concurrency.lockutils [None req-3a846554-7de2-4e42-88d0-6461fe3d7394 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Lock "a8d2cc01-3bc1-4792-85f5-b9095294b33c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.695s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1049.752083] env[63515]: DEBUG oslo_vmware.api [None req-2795ed8a-e4d1-4a5f-8e3c-baa30a1a2f05 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Task: {'id': task-1111872, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1050.247969] env[63515]: DEBUG nova.network.neutron [req-2a791d47-d678-4b51-ad20-a3cabe590076 req-8e6bfb34-8b80-4fb8-be8c-7d1b816fbbc6 service nova] [instance: aa83d035-9c47-434d-ab1a-140e6b3110b4] Updated VIF entry in instance network info cache for port 614aed8f-3ab8-4672-ac13-0ae8cb0b233d. {{(pid=63515) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1050.248403] env[63515]: DEBUG nova.network.neutron [req-2a791d47-d678-4b51-ad20-a3cabe590076 req-8e6bfb34-8b80-4fb8-be8c-7d1b816fbbc6 service nova] [instance: aa83d035-9c47-434d-ab1a-140e6b3110b4] Updating instance_info_cache with network_info: [{"id": "614aed8f-3ab8-4672-ac13-0ae8cb0b233d", "address": "fa:16:3e:98:68:38", "network": {"id": "5758c885-e1c9-45c8-86aa-27a802101f52", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-734458138-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.204", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b18d4daacbc84e758f9f9ae4f3f09d28", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69351262-8d39-441a-85ba-3a78df436d17", "external-id": "nsx-vlan-transportzone-205", "segmentation_id": 205, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap614aed8f-3a", "ovs_interfaceid": "614aed8f-3ab8-4672-ac13-0ae8cb0b233d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1050.253140] env[63515]: DEBUG oslo_vmware.api [None req-2795ed8a-e4d1-4a5f-8e3c-baa30a1a2f05 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Task: {'id': task-1111872, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1050.346957] env[63515]: DEBUG nova.compute.manager [req-29418fab-6f55-46cb-bc10-5769dc261563 req-75084859-5ac7-4d38-8b0e-a32ba825b18d service nova] [instance: b76cf3dd-7b7e-4e5d-bf3c-e15bd37069ac] Received event network-vif-deleted-acea8927-7e43-427c-bed2-2420d4708f3b {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1050.347192] env[63515]: INFO nova.compute.manager [req-29418fab-6f55-46cb-bc10-5769dc261563 req-75084859-5ac7-4d38-8b0e-a32ba825b18d service nova] [instance: b76cf3dd-7b7e-4e5d-bf3c-e15bd37069ac] Neutron deleted interface acea8927-7e43-427c-bed2-2420d4708f3b; detaching it from the instance and deleting it from the info cache [ 1050.347427] env[63515]: DEBUG nova.network.neutron [req-29418fab-6f55-46cb-bc10-5769dc261563 req-75084859-5ac7-4d38-8b0e-a32ba825b18d service nova] [instance: b76cf3dd-7b7e-4e5d-bf3c-e15bd37069ac] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1050.391140] env[63515]: DEBUG nova.network.neutron [-] [instance: b76cf3dd-7b7e-4e5d-bf3c-e15bd37069ac] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1050.500423] env[63515]: DEBUG oslo_concurrency.lockutils [None req-6bdaa058-8bd5-46fa-bd8d-458796ec2acc tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Acquiring lock "a8d2cc01-3bc1-4792-85f5-b9095294b33c" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1050.500669] env[63515]: DEBUG oslo_concurrency.lockutils [None req-6bdaa058-8bd5-46fa-bd8d-458796ec2acc tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Lock "a8d2cc01-3bc1-4792-85f5-b9095294b33c" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1050.500854] env[63515]: DEBUG nova.compute.manager [None req-6bdaa058-8bd5-46fa-bd8d-458796ec2acc tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: a8d2cc01-3bc1-4792-85f5-b9095294b33c] Checking state {{(pid=63515) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1050.501778] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9a70bb3-757d-45e2-9471-e7cb6b2576f7 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.508407] env[63515]: DEBUG nova.compute.manager [None req-6bdaa058-8bd5-46fa-bd8d-458796ec2acc tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: a8d2cc01-3bc1-4792-85f5-b9095294b33c] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=63515) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 1050.508980] env[63515]: DEBUG nova.objects.instance [None req-6bdaa058-8bd5-46fa-bd8d-458796ec2acc tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Lazy-loading 'flavor' on Instance uuid a8d2cc01-3bc1-4792-85f5-b9095294b33c {{(pid=63515) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1050.752159] env[63515]: DEBUG oslo_vmware.api [None req-2795ed8a-e4d1-4a5f-8e3c-baa30a1a2f05 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Task: {'id': task-1111872, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1050.754822] env[63515]: DEBUG oslo_concurrency.lockutils [req-2a791d47-d678-4b51-ad20-a3cabe590076 req-8e6bfb34-8b80-4fb8-be8c-7d1b816fbbc6 service nova] Releasing lock "refresh_cache-aa83d035-9c47-434d-ab1a-140e6b3110b4" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1050.850647] env[63515]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b580998b-7979-4939-b575-b9c305ccd992 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.860157] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7013cf4f-2683-4b58-b3a1-72a9ab63a5f2 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.890278] env[63515]: DEBUG nova.compute.manager [req-29418fab-6f55-46cb-bc10-5769dc261563 req-75084859-5ac7-4d38-8b0e-a32ba825b18d service nova] [instance: b76cf3dd-7b7e-4e5d-bf3c-e15bd37069ac] Detach interface failed, port_id=acea8927-7e43-427c-bed2-2420d4708f3b, reason: Instance b76cf3dd-7b7e-4e5d-bf3c-e15bd37069ac could not be found. {{(pid=63515) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1050.893975] env[63515]: INFO nova.compute.manager [-] [instance: b76cf3dd-7b7e-4e5d-bf3c-e15bd37069ac] Took 1.29 seconds to deallocate network for instance. [ 1051.013695] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-6bdaa058-8bd5-46fa-bd8d-458796ec2acc tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: a8d2cc01-3bc1-4792-85f5-b9095294b33c] Powering off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1051.013962] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-df2c8e2c-c0d4-4d46-a1d6-02af981d59f0 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.021369] env[63515]: DEBUG oslo_vmware.api [None req-6bdaa058-8bd5-46fa-bd8d-458796ec2acc tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Waiting for the task: (returnval){ [ 1051.021369] env[63515]: value = "task-1111877" [ 1051.021369] env[63515]: _type = "Task" [ 1051.021369] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1051.029377] env[63515]: DEBUG oslo_vmware.api [None req-6bdaa058-8bd5-46fa-bd8d-458796ec2acc tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Task: {'id': task-1111877, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1051.252365] env[63515]: DEBUG oslo_vmware.api [None req-2795ed8a-e4d1-4a5f-8e3c-baa30a1a2f05 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Task: {'id': task-1111872, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1051.401721] env[63515]: DEBUG oslo_concurrency.lockutils [None req-833c0cde-a35b-468a-be6f-c5dbc0f2b161 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1051.402012] env[63515]: DEBUG oslo_concurrency.lockutils [None req-833c0cde-a35b-468a-be6f-c5dbc0f2b161 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1051.402221] env[63515]: DEBUG oslo_concurrency.lockutils [None req-833c0cde-a35b-468a-be6f-c5dbc0f2b161 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1051.424470] env[63515]: INFO nova.scheduler.client.report [None req-833c0cde-a35b-468a-be6f-c5dbc0f2b161 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Deleted allocations for instance b76cf3dd-7b7e-4e5d-bf3c-e15bd37069ac [ 1051.530924] env[63515]: DEBUG oslo_vmware.api [None req-6bdaa058-8bd5-46fa-bd8d-458796ec2acc tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Task: {'id': task-1111877, 'name': PowerOffVM_Task, 'duration_secs': 0.192845} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1051.531639] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-6bdaa058-8bd5-46fa-bd8d-458796ec2acc tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: a8d2cc01-3bc1-4792-85f5-b9095294b33c] Powered off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1051.531959] env[63515]: DEBUG nova.compute.manager [None req-6bdaa058-8bd5-46fa-bd8d-458796ec2acc tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: a8d2cc01-3bc1-4792-85f5-b9095294b33c] Checking state {{(pid=63515) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1051.532982] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da1a9147-9373-4c8c-b63b-e7f507d53c30 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.753477] env[63515]: DEBUG oslo_vmware.api [None req-2795ed8a-e4d1-4a5f-8e3c-baa30a1a2f05 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Task: {'id': task-1111872, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1051.933151] env[63515]: DEBUG oslo_concurrency.lockutils [None req-833c0cde-a35b-468a-be6f-c5dbc0f2b161 tempest-DeleteServersTestJSON-1553408202 tempest-DeleteServersTestJSON-1553408202-project-member] Lock "b76cf3dd-7b7e-4e5d-bf3c-e15bd37069ac" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 4.305s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1052.044396] env[63515]: DEBUG oslo_concurrency.lockutils [None req-6bdaa058-8bd5-46fa-bd8d-458796ec2acc tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Lock "a8d2cc01-3bc1-4792-85f5-b9095294b33c" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.544s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1052.254331] env[63515]: DEBUG oslo_vmware.api [None req-2795ed8a-e4d1-4a5f-8e3c-baa30a1a2f05 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Task: {'id': task-1111872, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1052.754866] env[63515]: DEBUG oslo_vmware.api [None req-2795ed8a-e4d1-4a5f-8e3c-baa30a1a2f05 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Task: {'id': task-1111872, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1053.200548] env[63515]: DEBUG oslo_concurrency.lockutils [None req-dcdf2f7a-8723-4b23-a14d-c34b3389556f tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Acquiring lock "a8d2cc01-3bc1-4792-85f5-b9095294b33c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1053.200822] env[63515]: DEBUG oslo_concurrency.lockutils [None req-dcdf2f7a-8723-4b23-a14d-c34b3389556f tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Lock "a8d2cc01-3bc1-4792-85f5-b9095294b33c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1053.201061] env[63515]: DEBUG oslo_concurrency.lockutils [None req-dcdf2f7a-8723-4b23-a14d-c34b3389556f tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Acquiring lock "a8d2cc01-3bc1-4792-85f5-b9095294b33c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1053.201293] env[63515]: DEBUG oslo_concurrency.lockutils [None req-dcdf2f7a-8723-4b23-a14d-c34b3389556f tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Lock "a8d2cc01-3bc1-4792-85f5-b9095294b33c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1053.201496] env[63515]: DEBUG oslo_concurrency.lockutils [None req-dcdf2f7a-8723-4b23-a14d-c34b3389556f tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Lock "a8d2cc01-3bc1-4792-85f5-b9095294b33c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1053.203679] env[63515]: INFO nova.compute.manager [None req-dcdf2f7a-8723-4b23-a14d-c34b3389556f tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: a8d2cc01-3bc1-4792-85f5-b9095294b33c] Terminating instance [ 1053.205461] env[63515]: DEBUG nova.compute.manager [None req-dcdf2f7a-8723-4b23-a14d-c34b3389556f tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: a8d2cc01-3bc1-4792-85f5-b9095294b33c] Start destroying the instance on the hypervisor. {{(pid=63515) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1053.205665] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-dcdf2f7a-8723-4b23-a14d-c34b3389556f tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: a8d2cc01-3bc1-4792-85f5-b9095294b33c] Destroying instance {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1053.206491] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6649c11b-9d87-4255-b3f6-43f5a74f87eb {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.213859] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-dcdf2f7a-8723-4b23-a14d-c34b3389556f tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: a8d2cc01-3bc1-4792-85f5-b9095294b33c] Unregistering the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1053.214367] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-8926c444-5f6b-451c-8400-2e5f4a802283 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.231988] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-a1919053-a12d-4d12-b905-db1a99127d7f tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: d6c5a9f9-5e5c-4b14-9b9a-7825bbf526a5] Volume attach. Driver type: vmdk {{(pid=63515) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1053.232229] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-a1919053-a12d-4d12-b905-db1a99127d7f tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: d6c5a9f9-5e5c-4b14-9b9a-7825bbf526a5] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-243569', 'volume_id': '71cdbc05-35d6-43f4-9224-cab57f2d0c9c', 'name': 'volume-71cdbc05-35d6-43f4-9224-cab57f2d0c9c', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'd6c5a9f9-5e5c-4b14-9b9a-7825bbf526a5', 'attached_at': '', 'detached_at': '', 'volume_id': '71cdbc05-35d6-43f4-9224-cab57f2d0c9c', 'serial': '71cdbc05-35d6-43f4-9224-cab57f2d0c9c'} {{(pid=63515) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1053.232992] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a99661ae-e367-48f0-983e-b20787cd8d06 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.252562] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-208c720d-49c6-4a68-8d47-ce2438bd00a3 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.260071] env[63515]: DEBUG oslo_vmware.api [None req-2795ed8a-e4d1-4a5f-8e3c-baa30a1a2f05 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Task: {'id': task-1111872, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1053.279814] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-a1919053-a12d-4d12-b905-db1a99127d7f tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: d6c5a9f9-5e5c-4b14-9b9a-7825bbf526a5] Reconfiguring VM instance instance-00000062 to attach disk [datastore1] volume-71cdbc05-35d6-43f4-9224-cab57f2d0c9c/volume-71cdbc05-35d6-43f4-9224-cab57f2d0c9c.vmdk or device None with type thin {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1053.280067] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c9f1e412-f76a-4878-8a46-a860917b39c8 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.297739] env[63515]: DEBUG oslo_vmware.api [None req-a1919053-a12d-4d12-b905-db1a99127d7f tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Waiting for the task: (returnval){ [ 1053.297739] env[63515]: value = "task-1111881" [ 1053.297739] env[63515]: _type = "Task" [ 1053.297739] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1053.305495] env[63515]: DEBUG oslo_vmware.api [None req-a1919053-a12d-4d12-b905-db1a99127d7f tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Task: {'id': task-1111881, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1053.326775] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-dcdf2f7a-8723-4b23-a14d-c34b3389556f tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: a8d2cc01-3bc1-4792-85f5-b9095294b33c] Unregistered the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1053.327021] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-dcdf2f7a-8723-4b23-a14d-c34b3389556f tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: a8d2cc01-3bc1-4792-85f5-b9095294b33c] Deleting contents of the VM from datastore datastore1 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1053.327216] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-dcdf2f7a-8723-4b23-a14d-c34b3389556f tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Deleting the datastore file [datastore1] a8d2cc01-3bc1-4792-85f5-b9095294b33c {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1053.327538] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d76d7607-69d2-4f24-afdd-be36b210115f {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.333439] env[63515]: DEBUG oslo_vmware.api [None req-dcdf2f7a-8723-4b23-a14d-c34b3389556f tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Waiting for the task: (returnval){ [ 1053.333439] env[63515]: value = "task-1111882" [ 1053.333439] env[63515]: _type = "Task" [ 1053.333439] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1053.340959] env[63515]: DEBUG oslo_vmware.api [None req-dcdf2f7a-8723-4b23-a14d-c34b3389556f tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Task: {'id': task-1111882, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1053.758689] env[63515]: DEBUG oslo_vmware.api [None req-2795ed8a-e4d1-4a5f-8e3c-baa30a1a2f05 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Task: {'id': task-1111872, 'name': ReconfigVM_Task} progress is 18%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1053.805847] env[63515]: DEBUG oslo_vmware.api [None req-a1919053-a12d-4d12-b905-db1a99127d7f tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Task: {'id': task-1111881, 'name': ReconfigVM_Task, 'duration_secs': 0.34658} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1053.806125] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-a1919053-a12d-4d12-b905-db1a99127d7f tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: d6c5a9f9-5e5c-4b14-9b9a-7825bbf526a5] Reconfigured VM instance instance-00000062 to attach disk [datastore1] volume-71cdbc05-35d6-43f4-9224-cab57f2d0c9c/volume-71cdbc05-35d6-43f4-9224-cab57f2d0c9c.vmdk or device None with type thin {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1053.810736] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a9decec8-6045-4a0a-95f2-587c087297eb {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.825418] env[63515]: DEBUG oslo_vmware.api [None req-a1919053-a12d-4d12-b905-db1a99127d7f tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Waiting for the task: (returnval){ [ 1053.825418] env[63515]: value = "task-1111883" [ 1053.825418] env[63515]: _type = "Task" [ 1053.825418] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1053.833170] env[63515]: DEBUG oslo_vmware.api [None req-a1919053-a12d-4d12-b905-db1a99127d7f tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Task: {'id': task-1111883, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1053.841848] env[63515]: DEBUG oslo_vmware.api [None req-dcdf2f7a-8723-4b23-a14d-c34b3389556f tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Task: {'id': task-1111882, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.144858} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1053.842121] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-dcdf2f7a-8723-4b23-a14d-c34b3389556f tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Deleted the datastore file {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1053.842315] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-dcdf2f7a-8723-4b23-a14d-c34b3389556f tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: a8d2cc01-3bc1-4792-85f5-b9095294b33c] Deleted contents of the VM from datastore datastore1 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1053.842496] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-dcdf2f7a-8723-4b23-a14d-c34b3389556f tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: a8d2cc01-3bc1-4792-85f5-b9095294b33c] Instance destroyed {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1053.842672] env[63515]: INFO nova.compute.manager [None req-dcdf2f7a-8723-4b23-a14d-c34b3389556f tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: a8d2cc01-3bc1-4792-85f5-b9095294b33c] Took 0.64 seconds to destroy the instance on the hypervisor. [ 1053.842993] env[63515]: DEBUG oslo.service.loopingcall [None req-dcdf2f7a-8723-4b23-a14d-c34b3389556f tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1053.843292] env[63515]: DEBUG nova.compute.manager [-] [instance: a8d2cc01-3bc1-4792-85f5-b9095294b33c] Deallocating network for instance {{(pid=63515) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1053.843419] env[63515]: DEBUG nova.network.neutron [-] [instance: a8d2cc01-3bc1-4792-85f5-b9095294b33c] deallocate_for_instance() {{(pid=63515) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1054.121509] env[63515]: DEBUG nova.compute.manager [req-9426affc-8a96-45e6-b48a-7d0e1fb07c04 req-b0a7a856-7293-4c60-86c3-0e06cf6837de service nova] [instance: a8d2cc01-3bc1-4792-85f5-b9095294b33c] Received event network-vif-deleted-2338962e-f2a1-4b5e-aa4e-7340cb8b76b5 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1054.121722] env[63515]: INFO nova.compute.manager [req-9426affc-8a96-45e6-b48a-7d0e1fb07c04 req-b0a7a856-7293-4c60-86c3-0e06cf6837de service nova] [instance: a8d2cc01-3bc1-4792-85f5-b9095294b33c] Neutron deleted interface 2338962e-f2a1-4b5e-aa4e-7340cb8b76b5; detaching it from the instance and deleting it from the info cache [ 1054.121899] env[63515]: DEBUG nova.network.neutron [req-9426affc-8a96-45e6-b48a-7d0e1fb07c04 req-b0a7a856-7293-4c60-86c3-0e06cf6837de service nova] [instance: a8d2cc01-3bc1-4792-85f5-b9095294b33c] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1054.259687] env[63515]: DEBUG oslo_vmware.api [None req-2795ed8a-e4d1-4a5f-8e3c-baa30a1a2f05 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Task: {'id': task-1111872, 'name': ReconfigVM_Task, 'duration_secs': 5.761429} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1054.259926] env[63515]: DEBUG oslo_concurrency.lockutils [None req-2795ed8a-e4d1-4a5f-8e3c-baa30a1a2f05 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Releasing lock "c6ac42a9-a233-48b4-bcb2-fb791c7446bb" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1054.260155] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-2795ed8a-e4d1-4a5f-8e3c-baa30a1a2f05 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: c6ac42a9-a233-48b4-bcb2-fb791c7446bb] Reconfigured VM to detach interface {{(pid=63515) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1978}} [ 1054.335690] env[63515]: DEBUG oslo_vmware.api [None req-a1919053-a12d-4d12-b905-db1a99127d7f tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Task: {'id': task-1111883, 'name': ReconfigVM_Task, 'duration_secs': 0.137537} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1054.335979] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-a1919053-a12d-4d12-b905-db1a99127d7f tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: d6c5a9f9-5e5c-4b14-9b9a-7825bbf526a5] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-243569', 'volume_id': '71cdbc05-35d6-43f4-9224-cab57f2d0c9c', 'name': 'volume-71cdbc05-35d6-43f4-9224-cab57f2d0c9c', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'd6c5a9f9-5e5c-4b14-9b9a-7825bbf526a5', 'attached_at': '', 'detached_at': '', 'volume_id': '71cdbc05-35d6-43f4-9224-cab57f2d0c9c', 'serial': '71cdbc05-35d6-43f4-9224-cab57f2d0c9c'} {{(pid=63515) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1054.595792] env[63515]: DEBUG nova.network.neutron [-] [instance: a8d2cc01-3bc1-4792-85f5-b9095294b33c] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1054.625485] env[63515]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-65cb3f97-cf5a-468e-9922-b8f46db9ecfe {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.635903] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f61dae88-6ebc-4fb6-8ff0-8738bce753fa {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.663660] env[63515]: DEBUG nova.compute.manager [req-9426affc-8a96-45e6-b48a-7d0e1fb07c04 req-b0a7a856-7293-4c60-86c3-0e06cf6837de service nova] [instance: a8d2cc01-3bc1-4792-85f5-b9095294b33c] Detach interface failed, port_id=2338962e-f2a1-4b5e-aa4e-7340cb8b76b5, reason: Instance a8d2cc01-3bc1-4792-85f5-b9095294b33c could not be found. {{(pid=63515) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1055.099573] env[63515]: INFO nova.compute.manager [-] [instance: a8d2cc01-3bc1-4792-85f5-b9095294b33c] Took 1.26 seconds to deallocate network for instance. [ 1055.376127] env[63515]: DEBUG nova.objects.instance [None req-a1919053-a12d-4d12-b905-db1a99127d7f tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Lazy-loading 'flavor' on Instance uuid d6c5a9f9-5e5c-4b14-9b9a-7825bbf526a5 {{(pid=63515) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1055.550021] env[63515]: DEBUG oslo_concurrency.lockutils [None req-2795ed8a-e4d1-4a5f-8e3c-baa30a1a2f05 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Acquiring lock "refresh_cache-c6ac42a9-a233-48b4-bcb2-fb791c7446bb" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1055.550021] env[63515]: DEBUG oslo_concurrency.lockutils [None req-2795ed8a-e4d1-4a5f-8e3c-baa30a1a2f05 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Acquired lock "refresh_cache-c6ac42a9-a233-48b4-bcb2-fb791c7446bb" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1055.550021] env[63515]: DEBUG nova.network.neutron [None req-2795ed8a-e4d1-4a5f-8e3c-baa30a1a2f05 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: c6ac42a9-a233-48b4-bcb2-fb791c7446bb] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1055.607020] env[63515]: DEBUG oslo_concurrency.lockutils [None req-dcdf2f7a-8723-4b23-a14d-c34b3389556f tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1055.607020] env[63515]: DEBUG oslo_concurrency.lockutils [None req-dcdf2f7a-8723-4b23-a14d-c34b3389556f tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1055.607020] env[63515]: DEBUG nova.objects.instance [None req-dcdf2f7a-8723-4b23-a14d-c34b3389556f tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Lazy-loading 'resources' on Instance uuid a8d2cc01-3bc1-4792-85f5-b9095294b33c {{(pid=63515) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1055.880905] env[63515]: DEBUG oslo_concurrency.lockutils [None req-a1919053-a12d-4d12-b905-db1a99127d7f tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Lock "d6c5a9f9-5e5c-4b14-9b9a-7825bbf526a5" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.246s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1056.057018] env[63515]: DEBUG oslo_concurrency.lockutils [None req-bf817cf4-d89f-4edf-88cb-460ec0a43857 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Acquiring lock "c6ac42a9-a233-48b4-bcb2-fb791c7446bb" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1056.057368] env[63515]: DEBUG oslo_concurrency.lockutils [None req-bf817cf4-d89f-4edf-88cb-460ec0a43857 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Lock "c6ac42a9-a233-48b4-bcb2-fb791c7446bb" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1056.057616] env[63515]: DEBUG oslo_concurrency.lockutils [None req-bf817cf4-d89f-4edf-88cb-460ec0a43857 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Acquiring lock "c6ac42a9-a233-48b4-bcb2-fb791c7446bb-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1056.057848] env[63515]: DEBUG oslo_concurrency.lockutils [None req-bf817cf4-d89f-4edf-88cb-460ec0a43857 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Lock "c6ac42a9-a233-48b4-bcb2-fb791c7446bb-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1056.058419] env[63515]: DEBUG oslo_concurrency.lockutils [None req-bf817cf4-d89f-4edf-88cb-460ec0a43857 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Lock "c6ac42a9-a233-48b4-bcb2-fb791c7446bb-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1056.061268] env[63515]: INFO nova.compute.manager [None req-bf817cf4-d89f-4edf-88cb-460ec0a43857 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: c6ac42a9-a233-48b4-bcb2-fb791c7446bb] Terminating instance [ 1056.063429] env[63515]: DEBUG nova.compute.manager [None req-bf817cf4-d89f-4edf-88cb-460ec0a43857 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: c6ac42a9-a233-48b4-bcb2-fb791c7446bb] Start destroying the instance on the hypervisor. {{(pid=63515) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1056.063670] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-bf817cf4-d89f-4edf-88cb-460ec0a43857 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: c6ac42a9-a233-48b4-bcb2-fb791c7446bb] Destroying instance {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1056.064619] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cfe2706d-14dd-48e7-82d3-f69f18727843 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.073203] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-bf817cf4-d89f-4edf-88cb-460ec0a43857 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: c6ac42a9-a233-48b4-bcb2-fb791c7446bb] Powering off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1056.073475] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c91eaba0-abd6-4aef-b1a8-88ba645881cf {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.081482] env[63515]: DEBUG oslo_vmware.api [None req-bf817cf4-d89f-4edf-88cb-460ec0a43857 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Waiting for the task: (returnval){ [ 1056.081482] env[63515]: value = "task-1111884" [ 1056.081482] env[63515]: _type = "Task" [ 1056.081482] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1056.093853] env[63515]: DEBUG oslo_vmware.api [None req-bf817cf4-d89f-4edf-88cb-460ec0a43857 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Task: {'id': task-1111884, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1056.234566] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c92751f4-d733-4023-83e9-8098fa933847 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.242803] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89aa490e-4c0c-425d-9c84-cee4b3f3ed0a {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.274679] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0976473-4067-406c-9d3a-c9f6a123190c {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.283829] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ea0b660-d437-410a-9cf0-e4adc6a23cb9 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.302879] env[63515]: DEBUG nova.compute.provider_tree [None req-dcdf2f7a-8723-4b23-a14d-c34b3389556f tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1056.317479] env[63515]: INFO nova.network.neutron [None req-2795ed8a-e4d1-4a5f-8e3c-baa30a1a2f05 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: c6ac42a9-a233-48b4-bcb2-fb791c7446bb] Port 6a98da26-06f3-4eb5-aee8-02a448c34749 from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 1056.317479] env[63515]: DEBUG nova.network.neutron [None req-2795ed8a-e4d1-4a5f-8e3c-baa30a1a2f05 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: c6ac42a9-a233-48b4-bcb2-fb791c7446bb] Updating instance_info_cache with network_info: [{"id": "161e8482-76a0-4c82-8354-af132110794d", "address": "fa:16:3e:3e:77:7f", "network": {"id": "88dfb29d-41e0-4df7-8f70-e52cda05620c", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1355578400-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.146", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7d4ada9ec35f42b19c6480a9101d21a4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "357d2811-e990-4985-9f9e-b158d10d3699", "external-id": "nsx-vlan-transportzone-641", "segmentation_id": 641, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap161e8482-76", "ovs_interfaceid": "161e8482-76a0-4c82-8354-af132110794d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1056.357385] env[63515]: INFO nova.compute.manager [None req-f210ba7b-c0d6-4f5c-826f-ac5409d72c2c tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: d6c5a9f9-5e5c-4b14-9b9a-7825bbf526a5] Rebuilding instance [ 1056.410308] env[63515]: DEBUG nova.compute.manager [None req-f210ba7b-c0d6-4f5c-826f-ac5409d72c2c tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: d6c5a9f9-5e5c-4b14-9b9a-7825bbf526a5] Checking state {{(pid=63515) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1056.411200] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-190f4422-d001-4eab-a88d-f331e175c938 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.593027] env[63515]: DEBUG oslo_vmware.api [None req-bf817cf4-d89f-4edf-88cb-460ec0a43857 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Task: {'id': task-1111884, 'name': PowerOffVM_Task, 'duration_secs': 0.167616} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1056.593430] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-bf817cf4-d89f-4edf-88cb-460ec0a43857 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: c6ac42a9-a233-48b4-bcb2-fb791c7446bb] Powered off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1056.593674] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-bf817cf4-d89f-4edf-88cb-460ec0a43857 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: c6ac42a9-a233-48b4-bcb2-fb791c7446bb] Unregistering the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1056.593985] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ad836df2-0d83-480c-89e6-e97f2aa30ccc {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.660370] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-bf817cf4-d89f-4edf-88cb-460ec0a43857 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: c6ac42a9-a233-48b4-bcb2-fb791c7446bb] Unregistered the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1056.660720] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-bf817cf4-d89f-4edf-88cb-460ec0a43857 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: c6ac42a9-a233-48b4-bcb2-fb791c7446bb] Deleting contents of the VM from datastore datastore2 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1056.660985] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-bf817cf4-d89f-4edf-88cb-460ec0a43857 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Deleting the datastore file [datastore2] c6ac42a9-a233-48b4-bcb2-fb791c7446bb {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1056.661323] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-217c7147-7925-4cf4-8e22-af853df68713 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.670666] env[63515]: DEBUG oslo_vmware.api [None req-bf817cf4-d89f-4edf-88cb-460ec0a43857 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Waiting for the task: (returnval){ [ 1056.670666] env[63515]: value = "task-1111886" [ 1056.670666] env[63515]: _type = "Task" [ 1056.670666] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1056.679942] env[63515]: DEBUG oslo_vmware.api [None req-bf817cf4-d89f-4edf-88cb-460ec0a43857 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Task: {'id': task-1111886, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1056.806591] env[63515]: DEBUG nova.scheduler.client.report [None req-dcdf2f7a-8723-4b23-a14d-c34b3389556f tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1056.820218] env[63515]: DEBUG oslo_concurrency.lockutils [None req-2795ed8a-e4d1-4a5f-8e3c-baa30a1a2f05 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Releasing lock "refresh_cache-c6ac42a9-a233-48b4-bcb2-fb791c7446bb" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1056.922172] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-f210ba7b-c0d6-4f5c-826f-ac5409d72c2c tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: d6c5a9f9-5e5c-4b14-9b9a-7825bbf526a5] Powering off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1056.922499] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-413185d3-281f-40b7-beea-67a71d281493 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.929125] env[63515]: DEBUG oslo_vmware.api [None req-f210ba7b-c0d6-4f5c-826f-ac5409d72c2c tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Waiting for the task: (returnval){ [ 1056.929125] env[63515]: value = "task-1111887" [ 1056.929125] env[63515]: _type = "Task" [ 1056.929125] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1056.937468] env[63515]: DEBUG oslo_vmware.api [None req-f210ba7b-c0d6-4f5c-826f-ac5409d72c2c tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Task: {'id': task-1111887, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1057.180715] env[63515]: DEBUG oslo_vmware.api [None req-bf817cf4-d89f-4edf-88cb-460ec0a43857 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Task: {'id': task-1111886, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.15549} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1057.180984] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-bf817cf4-d89f-4edf-88cb-460ec0a43857 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Deleted the datastore file {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1057.181196] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-bf817cf4-d89f-4edf-88cb-460ec0a43857 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: c6ac42a9-a233-48b4-bcb2-fb791c7446bb] Deleted contents of the VM from datastore datastore2 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1057.181525] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-bf817cf4-d89f-4edf-88cb-460ec0a43857 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: c6ac42a9-a233-48b4-bcb2-fb791c7446bb] Instance destroyed {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1057.181737] env[63515]: INFO nova.compute.manager [None req-bf817cf4-d89f-4edf-88cb-460ec0a43857 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: c6ac42a9-a233-48b4-bcb2-fb791c7446bb] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1057.182014] env[63515]: DEBUG oslo.service.loopingcall [None req-bf817cf4-d89f-4edf-88cb-460ec0a43857 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1057.182241] env[63515]: DEBUG nova.compute.manager [-] [instance: c6ac42a9-a233-48b4-bcb2-fb791c7446bb] Deallocating network for instance {{(pid=63515) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1057.182339] env[63515]: DEBUG nova.network.neutron [-] [instance: c6ac42a9-a233-48b4-bcb2-fb791c7446bb] deallocate_for_instance() {{(pid=63515) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1057.314101] env[63515]: DEBUG oslo_concurrency.lockutils [None req-dcdf2f7a-8723-4b23-a14d-c34b3389556f tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.706s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1057.324616] env[63515]: DEBUG oslo_concurrency.lockutils [None req-2795ed8a-e4d1-4a5f-8e3c-baa30a1a2f05 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Lock "interface-c6ac42a9-a233-48b4-bcb2-fb791c7446bb-6a98da26-06f3-4eb5-aee8-02a448c34749" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 9.651s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1057.340393] env[63515]: INFO nova.scheduler.client.report [None req-dcdf2f7a-8723-4b23-a14d-c34b3389556f tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Deleted allocations for instance a8d2cc01-3bc1-4792-85f5-b9095294b33c [ 1057.444720] env[63515]: DEBUG oslo_vmware.api [None req-f210ba7b-c0d6-4f5c-826f-ac5409d72c2c tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Task: {'id': task-1111887, 'name': PowerOffVM_Task, 'duration_secs': 0.201748} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1057.445070] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-f210ba7b-c0d6-4f5c-826f-ac5409d72c2c tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: d6c5a9f9-5e5c-4b14-9b9a-7825bbf526a5] Powered off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1057.505452] env[63515]: INFO nova.compute.manager [None req-f210ba7b-c0d6-4f5c-826f-ac5409d72c2c tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: d6c5a9f9-5e5c-4b14-9b9a-7825bbf526a5] Detaching volume 71cdbc05-35d6-43f4-9224-cab57f2d0c9c [ 1057.543300] env[63515]: INFO nova.virt.block_device [None req-f210ba7b-c0d6-4f5c-826f-ac5409d72c2c tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: d6c5a9f9-5e5c-4b14-9b9a-7825bbf526a5] Attempting to driver detach volume 71cdbc05-35d6-43f4-9224-cab57f2d0c9c from mountpoint /dev/sdb [ 1057.543599] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-f210ba7b-c0d6-4f5c-826f-ac5409d72c2c tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: d6c5a9f9-5e5c-4b14-9b9a-7825bbf526a5] Volume detach. Driver type: vmdk {{(pid=63515) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1057.543731] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-f210ba7b-c0d6-4f5c-826f-ac5409d72c2c tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: d6c5a9f9-5e5c-4b14-9b9a-7825bbf526a5] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-243569', 'volume_id': '71cdbc05-35d6-43f4-9224-cab57f2d0c9c', 'name': 'volume-71cdbc05-35d6-43f4-9224-cab57f2d0c9c', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'd6c5a9f9-5e5c-4b14-9b9a-7825bbf526a5', 'attached_at': '', 'detached_at': '', 'volume_id': '71cdbc05-35d6-43f4-9224-cab57f2d0c9c', 'serial': '71cdbc05-35d6-43f4-9224-cab57f2d0c9c'} {{(pid=63515) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1057.544937] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-130e82f6-fd34-47cc-8a3f-d1a1a19ad119 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.567247] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ffd11981-da1c-4074-87f3-3065d6720acf {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.575808] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1be452a-0438-4ee3-b1fb-6ea5f643ba4a {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.602874] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0170f5cb-6f5b-4145-b1ff-3c0dbd704374 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.620769] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-f210ba7b-c0d6-4f5c-826f-ac5409d72c2c tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] The volume has not been displaced from its original location: [datastore1] volume-71cdbc05-35d6-43f4-9224-cab57f2d0c9c/volume-71cdbc05-35d6-43f4-9224-cab57f2d0c9c.vmdk. No consolidation needed. {{(pid=63515) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1057.626276] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-f210ba7b-c0d6-4f5c-826f-ac5409d72c2c tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: d6c5a9f9-5e5c-4b14-9b9a-7825bbf526a5] Reconfiguring VM instance instance-00000062 to detach disk 2001 {{(pid=63515) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1057.626797] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bc53b6a6-ce70-4f98-874e-9efb5648b467 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.646566] env[63515]: DEBUG oslo_vmware.api [None req-f210ba7b-c0d6-4f5c-826f-ac5409d72c2c tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Waiting for the task: (returnval){ [ 1057.646566] env[63515]: value = "task-1111888" [ 1057.646566] env[63515]: _type = "Task" [ 1057.646566] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1057.655024] env[63515]: DEBUG oslo_vmware.api [None req-f210ba7b-c0d6-4f5c-826f-ac5409d72c2c tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Task: {'id': task-1111888, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1057.852878] env[63515]: DEBUG oslo_concurrency.lockutils [None req-dcdf2f7a-8723-4b23-a14d-c34b3389556f tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Lock "a8d2cc01-3bc1-4792-85f5-b9095294b33c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 4.652s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1057.918213] env[63515]: DEBUG nova.compute.manager [req-a732c6b8-93b0-4a4e-a291-38f1e2e09f5a req-1f55e8d6-5c3f-43f6-abba-044dcb9599ce service nova] [instance: c6ac42a9-a233-48b4-bcb2-fb791c7446bb] Received event network-vif-deleted-161e8482-76a0-4c82-8354-af132110794d {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1057.918213] env[63515]: INFO nova.compute.manager [req-a732c6b8-93b0-4a4e-a291-38f1e2e09f5a req-1f55e8d6-5c3f-43f6-abba-044dcb9599ce service nova] [instance: c6ac42a9-a233-48b4-bcb2-fb791c7446bb] Neutron deleted interface 161e8482-76a0-4c82-8354-af132110794d; detaching it from the instance and deleting it from the info cache [ 1057.918213] env[63515]: DEBUG nova.network.neutron [req-a732c6b8-93b0-4a4e-a291-38f1e2e09f5a req-1f55e8d6-5c3f-43f6-abba-044dcb9599ce service nova] [instance: c6ac42a9-a233-48b4-bcb2-fb791c7446bb] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1058.160018] env[63515]: DEBUG oslo_vmware.api [None req-f210ba7b-c0d6-4f5c-826f-ac5409d72c2c tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Task: {'id': task-1111888, 'name': ReconfigVM_Task, 'duration_secs': 0.207497} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1058.160018] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-f210ba7b-c0d6-4f5c-826f-ac5409d72c2c tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: d6c5a9f9-5e5c-4b14-9b9a-7825bbf526a5] Reconfigured VM instance instance-00000062 to detach disk 2001 {{(pid=63515) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1058.162403] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a4beabf6-ee0e-49b1-8f60-10bd78fe641c {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.178591] env[63515]: DEBUG oslo_vmware.api [None req-f210ba7b-c0d6-4f5c-826f-ac5409d72c2c tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Waiting for the task: (returnval){ [ 1058.178591] env[63515]: value = "task-1111889" [ 1058.178591] env[63515]: _type = "Task" [ 1058.178591] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1058.189283] env[63515]: DEBUG oslo_vmware.api [None req-f210ba7b-c0d6-4f5c-826f-ac5409d72c2c tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Task: {'id': task-1111889, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1058.387553] env[63515]: DEBUG nova.network.neutron [-] [instance: c6ac42a9-a233-48b4-bcb2-fb791c7446bb] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1058.422440] env[63515]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-77d20f01-001b-414d-9bca-6a7b35fe0087 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.432340] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22158977-f217-4357-83d2-5a1a57b4de29 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.468236] env[63515]: DEBUG nova.compute.manager [req-a732c6b8-93b0-4a4e-a291-38f1e2e09f5a req-1f55e8d6-5c3f-43f6-abba-044dcb9599ce service nova] [instance: c6ac42a9-a233-48b4-bcb2-fb791c7446bb] Detach interface failed, port_id=161e8482-76a0-4c82-8354-af132110794d, reason: Instance c6ac42a9-a233-48b4-bcb2-fb791c7446bb could not be found. {{(pid=63515) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1058.690052] env[63515]: DEBUG oslo_vmware.api [None req-f210ba7b-c0d6-4f5c-826f-ac5409d72c2c tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Task: {'id': task-1111889, 'name': ReconfigVM_Task, 'duration_secs': 0.421898} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1058.690566] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-f210ba7b-c0d6-4f5c-826f-ac5409d72c2c tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: d6c5a9f9-5e5c-4b14-9b9a-7825bbf526a5] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-243569', 'volume_id': '71cdbc05-35d6-43f4-9224-cab57f2d0c9c', 'name': 'volume-71cdbc05-35d6-43f4-9224-cab57f2d0c9c', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'd6c5a9f9-5e5c-4b14-9b9a-7825bbf526a5', 'attached_at': '', 'detached_at': '', 'volume_id': '71cdbc05-35d6-43f4-9224-cab57f2d0c9c', 'serial': '71cdbc05-35d6-43f4-9224-cab57f2d0c9c'} {{(pid=63515) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1058.890075] env[63515]: INFO nova.compute.manager [-] [instance: c6ac42a9-a233-48b4-bcb2-fb791c7446bb] Took 1.71 seconds to deallocate network for instance. [ 1059.047027] env[63515]: DEBUG oslo_concurrency.lockutils [None req-5a3ece32-c98d-4ae4-a149-c214b90bef48 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Acquiring lock "19475235-dbec-40a5-bd8f-8070616b7ddb" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1059.047143] env[63515]: DEBUG oslo_concurrency.lockutils [None req-5a3ece32-c98d-4ae4-a149-c214b90bef48 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Lock "19475235-dbec-40a5-bd8f-8070616b7ddb" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1059.047365] env[63515]: DEBUG oslo_concurrency.lockutils [None req-5a3ece32-c98d-4ae4-a149-c214b90bef48 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Acquiring lock "19475235-dbec-40a5-bd8f-8070616b7ddb-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1059.047608] env[63515]: DEBUG oslo_concurrency.lockutils [None req-5a3ece32-c98d-4ae4-a149-c214b90bef48 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Lock "19475235-dbec-40a5-bd8f-8070616b7ddb-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1059.047886] env[63515]: DEBUG oslo_concurrency.lockutils [None req-5a3ece32-c98d-4ae4-a149-c214b90bef48 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Lock "19475235-dbec-40a5-bd8f-8070616b7ddb-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1059.050041] env[63515]: INFO nova.compute.manager [None req-5a3ece32-c98d-4ae4-a149-c214b90bef48 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: 19475235-dbec-40a5-bd8f-8070616b7ddb] Terminating instance [ 1059.051870] env[63515]: DEBUG nova.compute.manager [None req-5a3ece32-c98d-4ae4-a149-c214b90bef48 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: 19475235-dbec-40a5-bd8f-8070616b7ddb] Start destroying the instance on the hypervisor. {{(pid=63515) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1059.052086] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-5a3ece32-c98d-4ae4-a149-c214b90bef48 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: 19475235-dbec-40a5-bd8f-8070616b7ddb] Destroying instance {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1059.052927] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28536ffe-4f65-414f-9a31-cdc5e88f6a7b {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.060736] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-5a3ece32-c98d-4ae4-a149-c214b90bef48 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: 19475235-dbec-40a5-bd8f-8070616b7ddb] Powering off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1059.060974] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-27dd6043-139e-4f13-85b8-b925b4b14357 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.067227] env[63515]: DEBUG oslo_vmware.api [None req-5a3ece32-c98d-4ae4-a149-c214b90bef48 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Waiting for the task: (returnval){ [ 1059.067227] env[63515]: value = "task-1111890" [ 1059.067227] env[63515]: _type = "Task" [ 1059.067227] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1059.075549] env[63515]: DEBUG oslo_vmware.api [None req-5a3ece32-c98d-4ae4-a149-c214b90bef48 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Task: {'id': task-1111890, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1059.396550] env[63515]: DEBUG oslo_concurrency.lockutils [None req-bf817cf4-d89f-4edf-88cb-460ec0a43857 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1059.396849] env[63515]: DEBUG oslo_concurrency.lockutils [None req-bf817cf4-d89f-4edf-88cb-460ec0a43857 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1059.397059] env[63515]: DEBUG nova.objects.instance [None req-bf817cf4-d89f-4edf-88cb-460ec0a43857 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Lazy-loading 'resources' on Instance uuid c6ac42a9-a233-48b4-bcb2-fb791c7446bb {{(pid=63515) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1059.576507] env[63515]: DEBUG oslo_vmware.api [None req-5a3ece32-c98d-4ae4-a149-c214b90bef48 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Task: {'id': task-1111890, 'name': PowerOffVM_Task, 'duration_secs': 0.176998} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1059.576787] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-5a3ece32-c98d-4ae4-a149-c214b90bef48 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: 19475235-dbec-40a5-bd8f-8070616b7ddb] Powered off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1059.578419] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-5a3ece32-c98d-4ae4-a149-c214b90bef48 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: 19475235-dbec-40a5-bd8f-8070616b7ddb] Unregistering the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1059.578971] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c88e5206-b594-46a4-aafa-d66aa937f857 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.646020] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-5a3ece32-c98d-4ae4-a149-c214b90bef48 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: 19475235-dbec-40a5-bd8f-8070616b7ddb] Unregistered the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1059.650416] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-5a3ece32-c98d-4ae4-a149-c214b90bef48 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: 19475235-dbec-40a5-bd8f-8070616b7ddb] Deleting contents of the VM from datastore datastore2 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1059.650625] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-5a3ece32-c98d-4ae4-a149-c214b90bef48 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Deleting the datastore file [datastore2] 19475235-dbec-40a5-bd8f-8070616b7ddb {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1059.650909] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-8bdd6409-7b5a-4bd6-8cc3-b5e4e36f8a89 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.657753] env[63515]: DEBUG oslo_vmware.api [None req-5a3ece32-c98d-4ae4-a149-c214b90bef48 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Waiting for the task: (returnval){ [ 1059.657753] env[63515]: value = "task-1111892" [ 1059.657753] env[63515]: _type = "Task" [ 1059.657753] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1059.666537] env[63515]: DEBUG oslo_vmware.api [None req-5a3ece32-c98d-4ae4-a149-c214b90bef48 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Task: {'id': task-1111892, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1059.742055] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-f210ba7b-c0d6-4f5c-826f-ac5409d72c2c tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: d6c5a9f9-5e5c-4b14-9b9a-7825bbf526a5] Powering off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1059.742445] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-efae9b8f-223a-4c88-a466-cfa6323c8a2d {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.752244] env[63515]: DEBUG oslo_vmware.api [None req-f210ba7b-c0d6-4f5c-826f-ac5409d72c2c tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Waiting for the task: (returnval){ [ 1059.752244] env[63515]: value = "task-1111893" [ 1059.752244] env[63515]: _type = "Task" [ 1059.752244] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1059.761288] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-f210ba7b-c0d6-4f5c-826f-ac5409d72c2c tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: d6c5a9f9-5e5c-4b14-9b9a-7825bbf526a5] VM already powered off {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1059.761411] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-f210ba7b-c0d6-4f5c-826f-ac5409d72c2c tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: d6c5a9f9-5e5c-4b14-9b9a-7825bbf526a5] Volume detach. Driver type: vmdk {{(pid=63515) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1059.761714] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-f210ba7b-c0d6-4f5c-826f-ac5409d72c2c tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: d6c5a9f9-5e5c-4b14-9b9a-7825bbf526a5] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-243569', 'volume_id': '71cdbc05-35d6-43f4-9224-cab57f2d0c9c', 'name': 'volume-71cdbc05-35d6-43f4-9224-cab57f2d0c9c', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'd6c5a9f9-5e5c-4b14-9b9a-7825bbf526a5', 'attached_at': '', 'detached_at': '', 'volume_id': '71cdbc05-35d6-43f4-9224-cab57f2d0c9c', 'serial': '71cdbc05-35d6-43f4-9224-cab57f2d0c9c'} {{(pid=63515) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1059.762495] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a049bb2-6a83-41d7-b07d-dc10dc7ffc09 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.781040] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2124814-9a31-4ca4-ad8c-111549964411 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.789021] env[63515]: WARNING nova.virt.vmwareapi.driver [None req-f210ba7b-c0d6-4f5c-826f-ac5409d72c2c tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: d6c5a9f9-5e5c-4b14-9b9a-7825bbf526a5] The volume None does not exist!: nova.exception.DiskNotFound: Unable to find volume [ 1059.789021] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-f210ba7b-c0d6-4f5c-826f-ac5409d72c2c tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: d6c5a9f9-5e5c-4b14-9b9a-7825bbf526a5] Destroying instance {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1059.789021] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf3a418d-2c2b-4646-a17d-6205211ee25b {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.795900] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-f210ba7b-c0d6-4f5c-826f-ac5409d72c2c tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: d6c5a9f9-5e5c-4b14-9b9a-7825bbf526a5] Unregistering the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1059.796220] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-0deefb5b-a51f-463f-8571-16bf811548d2 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.855345] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-f210ba7b-c0d6-4f5c-826f-ac5409d72c2c tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: d6c5a9f9-5e5c-4b14-9b9a-7825bbf526a5] Unregistered the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1059.855747] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-f210ba7b-c0d6-4f5c-826f-ac5409d72c2c tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: d6c5a9f9-5e5c-4b14-9b9a-7825bbf526a5] Deleting contents of the VM from datastore datastore2 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1059.856055] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-f210ba7b-c0d6-4f5c-826f-ac5409d72c2c tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Deleting the datastore file [datastore2] d6c5a9f9-5e5c-4b14-9b9a-7825bbf526a5 {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1059.856445] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-79656220-1e13-4ed9-a595-3d939a3fdd78 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.863556] env[63515]: DEBUG oslo_vmware.api [None req-f210ba7b-c0d6-4f5c-826f-ac5409d72c2c tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Waiting for the task: (returnval){ [ 1059.863556] env[63515]: value = "task-1111895" [ 1059.863556] env[63515]: _type = "Task" [ 1059.863556] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1059.874994] env[63515]: DEBUG oslo_vmware.api [None req-f210ba7b-c0d6-4f5c-826f-ac5409d72c2c tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Task: {'id': task-1111895, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1060.020112] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c719702-5061-49bd-9a5f-69af258941f7 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.028069] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8530003e-0a3e-4c8d-be7b-523967b65312 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.059495] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4cc5b60d-695e-4604-8372-74e098f59564 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.066772] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3efbf1ec-6f1c-4b40-8d8b-a97d4bc010d0 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.080038] env[63515]: DEBUG nova.compute.provider_tree [None req-bf817cf4-d89f-4edf-88cb-460ec0a43857 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1060.172021] env[63515]: DEBUG oslo_vmware.api [None req-5a3ece32-c98d-4ae4-a149-c214b90bef48 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Task: {'id': task-1111892, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.140803} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1060.172021] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-5a3ece32-c98d-4ae4-a149-c214b90bef48 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Deleted the datastore file {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1060.173025] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-5a3ece32-c98d-4ae4-a149-c214b90bef48 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: 19475235-dbec-40a5-bd8f-8070616b7ddb] Deleted contents of the VM from datastore datastore2 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1060.173025] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-5a3ece32-c98d-4ae4-a149-c214b90bef48 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: 19475235-dbec-40a5-bd8f-8070616b7ddb] Instance destroyed {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1060.173025] env[63515]: INFO nova.compute.manager [None req-5a3ece32-c98d-4ae4-a149-c214b90bef48 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] [instance: 19475235-dbec-40a5-bd8f-8070616b7ddb] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1060.173289] env[63515]: DEBUG oslo.service.loopingcall [None req-5a3ece32-c98d-4ae4-a149-c214b90bef48 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1060.173579] env[63515]: DEBUG nova.compute.manager [-] [instance: 19475235-dbec-40a5-bd8f-8070616b7ddb] Deallocating network for instance {{(pid=63515) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1060.173732] env[63515]: DEBUG nova.network.neutron [-] [instance: 19475235-dbec-40a5-bd8f-8070616b7ddb] deallocate_for_instance() {{(pid=63515) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1060.376864] env[63515]: DEBUG oslo_vmware.api [None req-f210ba7b-c0d6-4f5c-826f-ac5409d72c2c tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Task: {'id': task-1111895, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.132412} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1060.377318] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-f210ba7b-c0d6-4f5c-826f-ac5409d72c2c tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Deleted the datastore file {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1060.377645] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-f210ba7b-c0d6-4f5c-826f-ac5409d72c2c tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: d6c5a9f9-5e5c-4b14-9b9a-7825bbf526a5] Deleted contents of the VM from datastore datastore2 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1060.377964] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-f210ba7b-c0d6-4f5c-826f-ac5409d72c2c tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: d6c5a9f9-5e5c-4b14-9b9a-7825bbf526a5] Instance destroyed {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1060.446085] env[63515]: DEBUG nova.compute.manager [req-5d191851-615a-458e-94b3-4125cef4c30e req-3aadc522-c119-4ad5-9882-5672a447d92e service nova] [instance: 19475235-dbec-40a5-bd8f-8070616b7ddb] Received event network-vif-deleted-372ea923-ddc7-4b61-92db-c55f8f352b44 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1060.446321] env[63515]: INFO nova.compute.manager [req-5d191851-615a-458e-94b3-4125cef4c30e req-3aadc522-c119-4ad5-9882-5672a447d92e service nova] [instance: 19475235-dbec-40a5-bd8f-8070616b7ddb] Neutron deleted interface 372ea923-ddc7-4b61-92db-c55f8f352b44; detaching it from the instance and deleting it from the info cache [ 1060.446515] env[63515]: DEBUG nova.network.neutron [req-5d191851-615a-458e-94b3-4125cef4c30e req-3aadc522-c119-4ad5-9882-5672a447d92e service nova] [instance: 19475235-dbec-40a5-bd8f-8070616b7ddb] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1060.583140] env[63515]: DEBUG nova.scheduler.client.report [None req-bf817cf4-d89f-4edf-88cb-460ec0a43857 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1060.885920] env[63515]: INFO nova.virt.block_device [None req-f210ba7b-c0d6-4f5c-826f-ac5409d72c2c tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: d6c5a9f9-5e5c-4b14-9b9a-7825bbf526a5] Booting with volume 71cdbc05-35d6-43f4-9224-cab57f2d0c9c at /dev/sdb [ 1060.928045] env[63515]: DEBUG nova.network.neutron [-] [instance: 19475235-dbec-40a5-bd8f-8070616b7ddb] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1060.934820] env[63515]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e00d419b-d37f-4c6b-8ed7-fb5846ed46bb {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.947148] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6e83b8c-adc1-4dbf-962c-bb62252525d3 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.960350] env[63515]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9f827a6d-3f6a-4dbe-a1ad-499a0385868f {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.970786] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c406913-88a5-43f3-98cc-7024f8addff2 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.988749] env[63515]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-fb603302-4fd4-467d-94ae-9a0702a53543 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.997853] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e306b3eb-578c-4623-89f2-2f700cb9e237 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.017434] env[63515]: DEBUG nova.compute.manager [req-5d191851-615a-458e-94b3-4125cef4c30e req-3aadc522-c119-4ad5-9882-5672a447d92e service nova] [instance: 19475235-dbec-40a5-bd8f-8070616b7ddb] Detach interface failed, port_id=372ea923-ddc7-4b61-92db-c55f8f352b44, reason: Instance 19475235-dbec-40a5-bd8f-8070616b7ddb could not be found. {{(pid=63515) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1061.031862] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0100b66c-072a-48e5-9d9d-516e82a13e34 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.038408] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f1e8fec-d198-4272-bae6-6fbb75afcf36 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.053397] env[63515]: DEBUG nova.virt.block_device [None req-f210ba7b-c0d6-4f5c-826f-ac5409d72c2c tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: d6c5a9f9-5e5c-4b14-9b9a-7825bbf526a5] Updating existing volume attachment record: fb983a95-4328-453e-bfa1-93226df6dd09 {{(pid=63515) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1061.088827] env[63515]: DEBUG oslo_concurrency.lockutils [None req-bf817cf4-d89f-4edf-88cb-460ec0a43857 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.692s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1061.126835] env[63515]: INFO nova.scheduler.client.report [None req-bf817cf4-d89f-4edf-88cb-460ec0a43857 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Deleted allocations for instance c6ac42a9-a233-48b4-bcb2-fb791c7446bb [ 1061.430261] env[63515]: INFO nova.compute.manager [-] [instance: 19475235-dbec-40a5-bd8f-8070616b7ddb] Took 1.26 seconds to deallocate network for instance. [ 1061.639825] env[63515]: DEBUG oslo_concurrency.lockutils [None req-bf817cf4-d89f-4edf-88cb-460ec0a43857 tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Lock "c6ac42a9-a233-48b4-bcb2-fb791c7446bb" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.582s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1061.940261] env[63515]: DEBUG oslo_concurrency.lockutils [None req-5a3ece32-c98d-4ae4-a149-c214b90bef48 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1061.940565] env[63515]: DEBUG oslo_concurrency.lockutils [None req-5a3ece32-c98d-4ae4-a149-c214b90bef48 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1061.940797] env[63515]: DEBUG nova.objects.instance [None req-5a3ece32-c98d-4ae4-a149-c214b90bef48 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Lazy-loading 'resources' on Instance uuid 19475235-dbec-40a5-bd8f-8070616b7ddb {{(pid=63515) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1062.336179] env[63515]: DEBUG oslo_concurrency.lockutils [None req-9dbc5332-6cd4-49ed-b269-132a6511b7a6 tempest-ServersNegativeTestMultiTenantJSON-1925445199 tempest-ServersNegativeTestMultiTenantJSON-1925445199-project-member] Acquiring lock "7fdc6303-2d53-4d87-9aea-bf40e476f790" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1062.336179] env[63515]: DEBUG oslo_concurrency.lockutils [None req-9dbc5332-6cd4-49ed-b269-132a6511b7a6 tempest-ServersNegativeTestMultiTenantJSON-1925445199 tempest-ServersNegativeTestMultiTenantJSON-1925445199-project-member] Lock "7fdc6303-2d53-4d87-9aea-bf40e476f790" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1062.548051] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3bad182b-649d-4019-a9ae-4e5f6d15bf5a {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.555819] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75577dd4-e961-44a6-9c9a-7f6c115888a6 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.595232] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb4f01e5-d08e-4601-abcb-e3a2cd779a3e {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.604679] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c23d674-bcda-4f00-81ac-57d5a714bc9b {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.619235] env[63515]: DEBUG nova.compute.provider_tree [None req-5a3ece32-c98d-4ae4-a149-c214b90bef48 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1062.747843] env[63515]: DEBUG oslo_concurrency.lockutils [None req-30134ddc-44b3-4354-8366-42d76e3947db tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Acquiring lock "724b05bc-ce4b-4f99-ad14-f3abea5567f2" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1062.748122] env[63515]: DEBUG oslo_concurrency.lockutils [None req-30134ddc-44b3-4354-8366-42d76e3947db tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Lock "724b05bc-ce4b-4f99-ad14-f3abea5567f2" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1062.748325] env[63515]: DEBUG oslo_concurrency.lockutils [None req-30134ddc-44b3-4354-8366-42d76e3947db tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Acquiring lock "724b05bc-ce4b-4f99-ad14-f3abea5567f2-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1062.749042] env[63515]: DEBUG oslo_concurrency.lockutils [None req-30134ddc-44b3-4354-8366-42d76e3947db tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Lock "724b05bc-ce4b-4f99-ad14-f3abea5567f2-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1062.749042] env[63515]: DEBUG oslo_concurrency.lockutils [None req-30134ddc-44b3-4354-8366-42d76e3947db tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Lock "724b05bc-ce4b-4f99-ad14-f3abea5567f2-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1062.751161] env[63515]: INFO nova.compute.manager [None req-30134ddc-44b3-4354-8366-42d76e3947db tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: 724b05bc-ce4b-4f99-ad14-f3abea5567f2] Terminating instance [ 1062.753030] env[63515]: DEBUG nova.compute.manager [None req-30134ddc-44b3-4354-8366-42d76e3947db tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: 724b05bc-ce4b-4f99-ad14-f3abea5567f2] Start destroying the instance on the hypervisor. {{(pid=63515) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1062.753176] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-30134ddc-44b3-4354-8366-42d76e3947db tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: 724b05bc-ce4b-4f99-ad14-f3abea5567f2] Destroying instance {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1062.753994] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e84fa11-e1e0-484b-a5bd-32a291a6dd11 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.762447] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-30134ddc-44b3-4354-8366-42d76e3947db tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: 724b05bc-ce4b-4f99-ad14-f3abea5567f2] Powering off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1062.762682] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a5804886-9364-4338-beba-a47ef82b1802 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.769436] env[63515]: DEBUG oslo_vmware.api [None req-30134ddc-44b3-4354-8366-42d76e3947db tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Waiting for the task: (returnval){ [ 1062.769436] env[63515]: value = "task-1111896" [ 1062.769436] env[63515]: _type = "Task" [ 1062.769436] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1062.776907] env[63515]: DEBUG oslo_vmware.api [None req-30134ddc-44b3-4354-8366-42d76e3947db tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Task: {'id': task-1111896, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1062.838140] env[63515]: DEBUG nova.compute.manager [None req-9dbc5332-6cd4-49ed-b269-132a6511b7a6 tempest-ServersNegativeTestMultiTenantJSON-1925445199 tempest-ServersNegativeTestMultiTenantJSON-1925445199-project-member] [instance: 7fdc6303-2d53-4d87-9aea-bf40e476f790] Starting instance... {{(pid=63515) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1063.127847] env[63515]: DEBUG nova.scheduler.client.report [None req-5a3ece32-c98d-4ae4-a149-c214b90bef48 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1063.183668] env[63515]: DEBUG nova.virt.hardware [None req-f210ba7b-c0d6-4f5c-826f-ac5409d72c2c tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T02:52:14Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T02:51:56Z,direct_url=,disk_format='vmdk',id=8a120570-cb06-4099-b262-554ca0ad15c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b569255cc43e42e7ae3f2b4ad37c6ef4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T02:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1063.183998] env[63515]: DEBUG nova.virt.hardware [None req-f210ba7b-c0d6-4f5c-826f-ac5409d72c2c tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Flavor limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1063.184120] env[63515]: DEBUG nova.virt.hardware [None req-f210ba7b-c0d6-4f5c-826f-ac5409d72c2c tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Image limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1063.184310] env[63515]: DEBUG nova.virt.hardware [None req-f210ba7b-c0d6-4f5c-826f-ac5409d72c2c tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Flavor pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1063.184490] env[63515]: DEBUG nova.virt.hardware [None req-f210ba7b-c0d6-4f5c-826f-ac5409d72c2c tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Image pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1063.184603] env[63515]: DEBUG nova.virt.hardware [None req-f210ba7b-c0d6-4f5c-826f-ac5409d72c2c tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1063.184813] env[63515]: DEBUG nova.virt.hardware [None req-f210ba7b-c0d6-4f5c-826f-ac5409d72c2c tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1063.184975] env[63515]: DEBUG nova.virt.hardware [None req-f210ba7b-c0d6-4f5c-826f-ac5409d72c2c tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1063.185417] env[63515]: DEBUG nova.virt.hardware [None req-f210ba7b-c0d6-4f5c-826f-ac5409d72c2c tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Got 1 possible topologies {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1063.185694] env[63515]: DEBUG nova.virt.hardware [None req-f210ba7b-c0d6-4f5c-826f-ac5409d72c2c tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1063.185872] env[63515]: DEBUG nova.virt.hardware [None req-f210ba7b-c0d6-4f5c-826f-ac5409d72c2c tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1063.186780] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-155dd1db-f689-4935-931f-c72f5c419422 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.196435] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af6c5a04-a06e-4268-be5c-d3d4bdfdc867 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.209956] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-f210ba7b-c0d6-4f5c-826f-ac5409d72c2c tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: d6c5a9f9-5e5c-4b14-9b9a-7825bbf526a5] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:21:f3:e4', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '3a0ddd7d-c321-4187-bdd8-b19044ea2c4a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ee82711d-2d77-4127-b804-7db65e838617', 'vif_model': 'vmxnet3'}] {{(pid=63515) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1063.217965] env[63515]: DEBUG oslo.service.loopingcall [None req-f210ba7b-c0d6-4f5c-826f-ac5409d72c2c tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1063.218254] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d6c5a9f9-5e5c-4b14-9b9a-7825bbf526a5] Creating VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1063.218502] env[63515]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-36a2ef8e-2e68-4b6d-a87d-e228678b4e73 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.237315] env[63515]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1063.237315] env[63515]: value = "task-1111897" [ 1063.237315] env[63515]: _type = "Task" [ 1063.237315] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1063.245819] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1111897, 'name': CreateVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1063.278343] env[63515]: DEBUG oslo_vmware.api [None req-30134ddc-44b3-4354-8366-42d76e3947db tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Task: {'id': task-1111896, 'name': PowerOffVM_Task, 'duration_secs': 0.216855} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1063.278676] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-30134ddc-44b3-4354-8366-42d76e3947db tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: 724b05bc-ce4b-4f99-ad14-f3abea5567f2] Powered off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1063.278891] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-30134ddc-44b3-4354-8366-42d76e3947db tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: 724b05bc-ce4b-4f99-ad14-f3abea5567f2] Unregistering the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1063.279193] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3acf30ec-cadc-4a59-85d8-793b063783db {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.338915] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-30134ddc-44b3-4354-8366-42d76e3947db tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: 724b05bc-ce4b-4f99-ad14-f3abea5567f2] Unregistered the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1063.339421] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-30134ddc-44b3-4354-8366-42d76e3947db tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: 724b05bc-ce4b-4f99-ad14-f3abea5567f2] Deleting contents of the VM from datastore datastore1 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1063.339740] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-30134ddc-44b3-4354-8366-42d76e3947db tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Deleting the datastore file [datastore1] 724b05bc-ce4b-4f99-ad14-f3abea5567f2 {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1063.340067] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-37d48f8d-410a-416e-8f77-115d26942f84 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.350009] env[63515]: DEBUG oslo_vmware.api [None req-30134ddc-44b3-4354-8366-42d76e3947db tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Waiting for the task: (returnval){ [ 1063.350009] env[63515]: value = "task-1111899" [ 1063.350009] env[63515]: _type = "Task" [ 1063.350009] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1063.358660] env[63515]: DEBUG oslo_vmware.api [None req-30134ddc-44b3-4354-8366-42d76e3947db tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Task: {'id': task-1111899, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1063.364635] env[63515]: DEBUG oslo_concurrency.lockutils [None req-9dbc5332-6cd4-49ed-b269-132a6511b7a6 tempest-ServersNegativeTestMultiTenantJSON-1925445199 tempest-ServersNegativeTestMultiTenantJSON-1925445199-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1063.643822] env[63515]: DEBUG oslo_concurrency.lockutils [None req-5a3ece32-c98d-4ae4-a149-c214b90bef48 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.702s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1063.645876] env[63515]: DEBUG oslo_concurrency.lockutils [None req-9dbc5332-6cd4-49ed-b269-132a6511b7a6 tempest-ServersNegativeTestMultiTenantJSON-1925445199 tempest-ServersNegativeTestMultiTenantJSON-1925445199-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.280s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1063.648551] env[63515]: INFO nova.compute.claims [None req-9dbc5332-6cd4-49ed-b269-132a6511b7a6 tempest-ServersNegativeTestMultiTenantJSON-1925445199 tempest-ServersNegativeTestMultiTenantJSON-1925445199-project-member] [instance: 7fdc6303-2d53-4d87-9aea-bf40e476f790] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1063.672569] env[63515]: INFO nova.scheduler.client.report [None req-5a3ece32-c98d-4ae4-a149-c214b90bef48 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Deleted allocations for instance 19475235-dbec-40a5-bd8f-8070616b7ddb [ 1063.746778] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1111897, 'name': CreateVM_Task, 'duration_secs': 0.309959} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1063.746990] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d6c5a9f9-5e5c-4b14-9b9a-7825bbf526a5] Created VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1063.747636] env[63515]: DEBUG oslo_concurrency.lockutils [None req-f210ba7b-c0d6-4f5c-826f-ac5409d72c2c tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1063.747805] env[63515]: DEBUG oslo_concurrency.lockutils [None req-f210ba7b-c0d6-4f5c-826f-ac5409d72c2c tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1063.748134] env[63515]: DEBUG oslo_concurrency.lockutils [None req-f210ba7b-c0d6-4f5c-826f-ac5409d72c2c tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1063.748388] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7f3cdd8f-1a92-43e5-b62f-cd9cf6aae66f {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.752743] env[63515]: DEBUG oslo_vmware.api [None req-f210ba7b-c0d6-4f5c-826f-ac5409d72c2c tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Waiting for the task: (returnval){ [ 1063.752743] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]52d1a568-408c-ef69-a106-88ac8e8b9c3e" [ 1063.752743] env[63515]: _type = "Task" [ 1063.752743] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1063.761152] env[63515]: DEBUG oslo_vmware.api [None req-f210ba7b-c0d6-4f5c-826f-ac5409d72c2c tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52d1a568-408c-ef69-a106-88ac8e8b9c3e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1063.859479] env[63515]: DEBUG oslo_vmware.api [None req-30134ddc-44b3-4354-8366-42d76e3947db tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Task: {'id': task-1111899, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.090321} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1063.859742] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-30134ddc-44b3-4354-8366-42d76e3947db tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Deleted the datastore file {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1063.859934] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-30134ddc-44b3-4354-8366-42d76e3947db tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: 724b05bc-ce4b-4f99-ad14-f3abea5567f2] Deleted contents of the VM from datastore datastore1 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1063.860127] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-30134ddc-44b3-4354-8366-42d76e3947db tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: 724b05bc-ce4b-4f99-ad14-f3abea5567f2] Instance destroyed {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1063.860306] env[63515]: INFO nova.compute.manager [None req-30134ddc-44b3-4354-8366-42d76e3947db tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] [instance: 724b05bc-ce4b-4f99-ad14-f3abea5567f2] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1063.860548] env[63515]: DEBUG oslo.service.loopingcall [None req-30134ddc-44b3-4354-8366-42d76e3947db tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1063.860738] env[63515]: DEBUG nova.compute.manager [-] [instance: 724b05bc-ce4b-4f99-ad14-f3abea5567f2] Deallocating network for instance {{(pid=63515) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1063.860831] env[63515]: DEBUG nova.network.neutron [-] [instance: 724b05bc-ce4b-4f99-ad14-f3abea5567f2] deallocate_for_instance() {{(pid=63515) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1064.183152] env[63515]: DEBUG oslo_concurrency.lockutils [None req-5a3ece32-c98d-4ae4-a149-c214b90bef48 tempest-ServersTestJSON-1324820170 tempest-ServersTestJSON-1324820170-project-member] Lock "19475235-dbec-40a5-bd8f-8070616b7ddb" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.136s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1064.263589] env[63515]: DEBUG oslo_vmware.api [None req-f210ba7b-c0d6-4f5c-826f-ac5409d72c2c tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52d1a568-408c-ef69-a106-88ac8e8b9c3e, 'name': SearchDatastore_Task, 'duration_secs': 0.009375} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1064.263906] env[63515]: DEBUG oslo_concurrency.lockutils [None req-f210ba7b-c0d6-4f5c-826f-ac5409d72c2c tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1064.264162] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-f210ba7b-c0d6-4f5c-826f-ac5409d72c2c tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: d6c5a9f9-5e5c-4b14-9b9a-7825bbf526a5] Processing image 8a120570-cb06-4099-b262-554ca0ad15c5 {{(pid=63515) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1064.264943] env[63515]: DEBUG oslo_concurrency.lockutils [None req-f210ba7b-c0d6-4f5c-826f-ac5409d72c2c tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1064.264943] env[63515]: DEBUG oslo_concurrency.lockutils [None req-f210ba7b-c0d6-4f5c-826f-ac5409d72c2c tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1064.264943] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-f210ba7b-c0d6-4f5c-826f-ac5409d72c2c tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1064.265364] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-7cc60664-1f0d-4e43-8d5e-7c7d86bf478b {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.279982] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-f210ba7b-c0d6-4f5c-826f-ac5409d72c2c tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1064.280194] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-f210ba7b-c0d6-4f5c-826f-ac5409d72c2c tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63515) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1064.280922] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-57d20c5c-d07c-4746-8277-e6e6dcaac690 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.286374] env[63515]: DEBUG oslo_vmware.api [None req-f210ba7b-c0d6-4f5c-826f-ac5409d72c2c tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Waiting for the task: (returnval){ [ 1064.286374] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]52aee1fb-7a46-7f37-04b6-cf985c06f89f" [ 1064.286374] env[63515]: _type = "Task" [ 1064.286374] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1064.293687] env[63515]: DEBUG oslo_vmware.api [None req-f210ba7b-c0d6-4f5c-826f-ac5409d72c2c tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52aee1fb-7a46-7f37-04b6-cf985c06f89f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1064.735484] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-abcff140-cdb1-4ee9-a752-56e25be50128 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.743000] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b854331a-a878-4c59-9b37-4fa3d37c90dc {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.773947] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a571d0a5-7cbb-4365-965e-da714bb8d52c {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.782019] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43b25595-3134-44c6-8438-30149f8703c0 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.794292] env[63515]: DEBUG nova.compute.provider_tree [None req-9dbc5332-6cd4-49ed-b269-132a6511b7a6 tempest-ServersNegativeTestMultiTenantJSON-1925445199 tempest-ServersNegativeTestMultiTenantJSON-1925445199-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1064.803619] env[63515]: DEBUG oslo_vmware.api [None req-f210ba7b-c0d6-4f5c-826f-ac5409d72c2c tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52aee1fb-7a46-7f37-04b6-cf985c06f89f, 'name': SearchDatastore_Task, 'duration_secs': 0.042531} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1064.804823] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4d6be810-9ff6-48b8-af5a-b6fe3ba97cae {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.810779] env[63515]: DEBUG oslo_vmware.api [None req-f210ba7b-c0d6-4f5c-826f-ac5409d72c2c tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Waiting for the task: (returnval){ [ 1064.810779] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]521ff3d7-b67c-644b-259b-f8f250f68787" [ 1064.810779] env[63515]: _type = "Task" [ 1064.810779] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1064.819914] env[63515]: DEBUG oslo_vmware.api [None req-f210ba7b-c0d6-4f5c-826f-ac5409d72c2c tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]521ff3d7-b67c-644b-259b-f8f250f68787, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1065.300187] env[63515]: DEBUG nova.scheduler.client.report [None req-9dbc5332-6cd4-49ed-b269-132a6511b7a6 tempest-ServersNegativeTestMultiTenantJSON-1925445199 tempest-ServersNegativeTestMultiTenantJSON-1925445199-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1065.321511] env[63515]: DEBUG oslo_vmware.api [None req-f210ba7b-c0d6-4f5c-826f-ac5409d72c2c tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]521ff3d7-b67c-644b-259b-f8f250f68787, 'name': SearchDatastore_Task, 'duration_secs': 0.011954} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1065.321511] env[63515]: DEBUG oslo_concurrency.lockutils [None req-f210ba7b-c0d6-4f5c-826f-ac5409d72c2c tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1065.321511] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-f210ba7b-c0d6-4f5c-826f-ac5409d72c2c tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk to [datastore2] d6c5a9f9-5e5c-4b14-9b9a-7825bbf526a5/d6c5a9f9-5e5c-4b14-9b9a-7825bbf526a5.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1065.321796] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-8cdaf426-a465-4bb8-bffc-e6ccd3549051 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.328157] env[63515]: DEBUG oslo_vmware.api [None req-f210ba7b-c0d6-4f5c-826f-ac5409d72c2c tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Waiting for the task: (returnval){ [ 1065.328157] env[63515]: value = "task-1111900" [ 1065.328157] env[63515]: _type = "Task" [ 1065.328157] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1065.335555] env[63515]: DEBUG oslo_vmware.api [None req-f210ba7b-c0d6-4f5c-826f-ac5409d72c2c tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Task: {'id': task-1111900, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1065.837906] env[63515]: DEBUG oslo_vmware.api [None req-f210ba7b-c0d6-4f5c-826f-ac5409d72c2c tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Task: {'id': task-1111900, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1066.338797] env[63515]: DEBUG oslo_vmware.api [None req-f210ba7b-c0d6-4f5c-826f-ac5409d72c2c tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Task: {'id': task-1111900, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1066.839124] env[63515]: DEBUG oslo_vmware.api [None req-f210ba7b-c0d6-4f5c-826f-ac5409d72c2c tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Task: {'id': task-1111900, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1067.307342] env[63515]: DEBUG oslo_concurrency.lockutils [None req-9dbc5332-6cd4-49ed-b269-132a6511b7a6 tempest-ServersNegativeTestMultiTenantJSON-1925445199 tempest-ServersNegativeTestMultiTenantJSON-1925445199-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.662s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1067.307992] env[63515]: DEBUG nova.compute.manager [None req-9dbc5332-6cd4-49ed-b269-132a6511b7a6 tempest-ServersNegativeTestMultiTenantJSON-1925445199 tempest-ServersNegativeTestMultiTenantJSON-1925445199-project-member] [instance: 7fdc6303-2d53-4d87-9aea-bf40e476f790] Start building networks asynchronously for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1067.318286] env[63515]: DEBUG nova.compute.manager [req-61970a46-743d-4b5c-abbe-a4806251bf81 req-c2f526b0-4af6-4dfc-96ea-1a05a9999f68 service nova] [instance: 724b05bc-ce4b-4f99-ad14-f3abea5567f2] Received event network-vif-deleted-3dee43b6-2650-42e1-aa5a-4994a3ec7f05 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1067.318536] env[63515]: INFO nova.compute.manager [req-61970a46-743d-4b5c-abbe-a4806251bf81 req-c2f526b0-4af6-4dfc-96ea-1a05a9999f68 service nova] [instance: 724b05bc-ce4b-4f99-ad14-f3abea5567f2] Neutron deleted interface 3dee43b6-2650-42e1-aa5a-4994a3ec7f05; detaching it from the instance and deleting it from the info cache [ 1067.318724] env[63515]: DEBUG nova.network.neutron [req-61970a46-743d-4b5c-abbe-a4806251bf81 req-c2f526b0-4af6-4dfc-96ea-1a05a9999f68 service nova] [instance: 724b05bc-ce4b-4f99-ad14-f3abea5567f2] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1067.342777] env[63515]: DEBUG oslo_vmware.api [None req-f210ba7b-c0d6-4f5c-826f-ac5409d72c2c tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Task: {'id': task-1111900, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1067.777989] env[63515]: DEBUG nova.network.neutron [-] [instance: 724b05bc-ce4b-4f99-ad14-f3abea5567f2] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1067.813156] env[63515]: DEBUG nova.compute.utils [None req-9dbc5332-6cd4-49ed-b269-132a6511b7a6 tempest-ServersNegativeTestMultiTenantJSON-1925445199 tempest-ServersNegativeTestMultiTenantJSON-1925445199-project-member] Using /dev/sd instead of None {{(pid=63515) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1067.814756] env[63515]: DEBUG nova.compute.manager [None req-9dbc5332-6cd4-49ed-b269-132a6511b7a6 tempest-ServersNegativeTestMultiTenantJSON-1925445199 tempest-ServersNegativeTestMultiTenantJSON-1925445199-project-member] [instance: 7fdc6303-2d53-4d87-9aea-bf40e476f790] Allocating IP information in the background. {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1067.814834] env[63515]: DEBUG nova.network.neutron [None req-9dbc5332-6cd4-49ed-b269-132a6511b7a6 tempest-ServersNegativeTestMultiTenantJSON-1925445199 tempest-ServersNegativeTestMultiTenantJSON-1925445199-project-member] [instance: 7fdc6303-2d53-4d87-9aea-bf40e476f790] allocate_for_instance() {{(pid=63515) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1067.821995] env[63515]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f303d5f9-3330-4165-8a4e-b4ce5ef25a02 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.831178] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7127f24e-ac85-4511-ad7b-8390cb262917 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.850897] env[63515]: DEBUG oslo_vmware.api [None req-f210ba7b-c0d6-4f5c-826f-ac5409d72c2c tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Task: {'id': task-1111900, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.271865} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1067.851214] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-f210ba7b-c0d6-4f5c-826f-ac5409d72c2c tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk to [datastore2] d6c5a9f9-5e5c-4b14-9b9a-7825bbf526a5/d6c5a9f9-5e5c-4b14-9b9a-7825bbf526a5.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1067.851415] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-f210ba7b-c0d6-4f5c-826f-ac5409d72c2c tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: d6c5a9f9-5e5c-4b14-9b9a-7825bbf526a5] Extending root virtual disk to 1048576 {{(pid=63515) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1067.851659] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-fb31c939-6b89-4a4b-9c85-de778f397081 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.859360] env[63515]: DEBUG nova.compute.manager [req-61970a46-743d-4b5c-abbe-a4806251bf81 req-c2f526b0-4af6-4dfc-96ea-1a05a9999f68 service nova] [instance: 724b05bc-ce4b-4f99-ad14-f3abea5567f2] Detach interface failed, port_id=3dee43b6-2650-42e1-aa5a-4994a3ec7f05, reason: Instance 724b05bc-ce4b-4f99-ad14-f3abea5567f2 could not be found. {{(pid=63515) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1067.864603] env[63515]: DEBUG oslo_vmware.api [None req-f210ba7b-c0d6-4f5c-826f-ac5409d72c2c tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Waiting for the task: (returnval){ [ 1067.864603] env[63515]: value = "task-1111901" [ 1067.864603] env[63515]: _type = "Task" [ 1067.864603] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1067.872726] env[63515]: DEBUG oslo_vmware.api [None req-f210ba7b-c0d6-4f5c-826f-ac5409d72c2c tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Task: {'id': task-1111901, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1067.888136] env[63515]: DEBUG nova.policy [None req-9dbc5332-6cd4-49ed-b269-132a6511b7a6 tempest-ServersNegativeTestMultiTenantJSON-1925445199 tempest-ServersNegativeTestMultiTenantJSON-1925445199-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '064f1ef36c694d55989fda32ee869867', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3cce2d6c191a4187932179eb598e8bad', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63515) authorize /opt/stack/nova/nova/policy.py:201}} [ 1068.174247] env[63515]: DEBUG nova.network.neutron [None req-9dbc5332-6cd4-49ed-b269-132a6511b7a6 tempest-ServersNegativeTestMultiTenantJSON-1925445199 tempest-ServersNegativeTestMultiTenantJSON-1925445199-project-member] [instance: 7fdc6303-2d53-4d87-9aea-bf40e476f790] Successfully created port: 0a17708c-c9b4-4715-bcfa-f2629d5b179f {{(pid=63515) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1068.282590] env[63515]: INFO nova.compute.manager [-] [instance: 724b05bc-ce4b-4f99-ad14-f3abea5567f2] Took 4.42 seconds to deallocate network for instance. [ 1068.319564] env[63515]: DEBUG nova.compute.manager [None req-9dbc5332-6cd4-49ed-b269-132a6511b7a6 tempest-ServersNegativeTestMultiTenantJSON-1925445199 tempest-ServersNegativeTestMultiTenantJSON-1925445199-project-member] [instance: 7fdc6303-2d53-4d87-9aea-bf40e476f790] Start building block device mappings for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1068.376045] env[63515]: DEBUG oslo_vmware.api [None req-f210ba7b-c0d6-4f5c-826f-ac5409d72c2c tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Task: {'id': task-1111901, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.067768} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1068.376695] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-f210ba7b-c0d6-4f5c-826f-ac5409d72c2c tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: d6c5a9f9-5e5c-4b14-9b9a-7825bbf526a5] Extended root virtual disk {{(pid=63515) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1068.378571] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1641c85d-5620-41d2-a060-cfc7e137837f {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.406529] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-f210ba7b-c0d6-4f5c-826f-ac5409d72c2c tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: d6c5a9f9-5e5c-4b14-9b9a-7825bbf526a5] Reconfiguring VM instance instance-00000062 to attach disk [datastore2] d6c5a9f9-5e5c-4b14-9b9a-7825bbf526a5/d6c5a9f9-5e5c-4b14-9b9a-7825bbf526a5.vmdk or device None with type sparse {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1068.406868] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-62baf06d-c25f-43fb-9a9f-20225a0aff0d {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.428406] env[63515]: DEBUG oslo_vmware.api [None req-f210ba7b-c0d6-4f5c-826f-ac5409d72c2c tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Waiting for the task: (returnval){ [ 1068.428406] env[63515]: value = "task-1111902" [ 1068.428406] env[63515]: _type = "Task" [ 1068.428406] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1068.439952] env[63515]: DEBUG oslo_vmware.api [None req-f210ba7b-c0d6-4f5c-826f-ac5409d72c2c tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Task: {'id': task-1111902, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1068.793074] env[63515]: DEBUG oslo_concurrency.lockutils [None req-30134ddc-44b3-4354-8366-42d76e3947db tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1068.793399] env[63515]: DEBUG oslo_concurrency.lockutils [None req-30134ddc-44b3-4354-8366-42d76e3947db tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1068.793705] env[63515]: DEBUG nova.objects.instance [None req-30134ddc-44b3-4354-8366-42d76e3947db tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Lazy-loading 'resources' on Instance uuid 724b05bc-ce4b-4f99-ad14-f3abea5567f2 {{(pid=63515) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1068.939488] env[63515]: DEBUG oslo_vmware.api [None req-f210ba7b-c0d6-4f5c-826f-ac5409d72c2c tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Task: {'id': task-1111902, 'name': ReconfigVM_Task, 'duration_secs': 0.320444} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1068.939776] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-f210ba7b-c0d6-4f5c-826f-ac5409d72c2c tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: d6c5a9f9-5e5c-4b14-9b9a-7825bbf526a5] Reconfigured VM instance instance-00000062 to attach disk [datastore2] d6c5a9f9-5e5c-4b14-9b9a-7825bbf526a5/d6c5a9f9-5e5c-4b14-9b9a-7825bbf526a5.vmdk or device None with type sparse {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1068.941028] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-f210ba7b-c0d6-4f5c-826f-ac5409d72c2c tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: d6c5a9f9-5e5c-4b14-9b9a-7825bbf526a5] Block device information present: {'root_device_name': '/dev/sda', 'image': [{'disk_bus': None, 'encrypted': False, 'encryption_secret_uuid': None, 'encryption_format': None, 'size': 0, 'guest_format': None, 'device_name': '/dev/sda', 'device_type': 'disk', 'boot_index': 0, 'encryption_options': None, 'image_id': '8a120570-cb06-4099-b262-554ca0ad15c5'}], 'ephemerals': [], 'block_device_mapping': [{'connection_info': {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-243569', 'volume_id': '71cdbc05-35d6-43f4-9224-cab57f2d0c9c', 'name': 'volume-71cdbc05-35d6-43f4-9224-cab57f2d0c9c', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'd6c5a9f9-5e5c-4b14-9b9a-7825bbf526a5', 'attached_at': '', 'detached_at': '', 'volume_id': '71cdbc05-35d6-43f4-9224-cab57f2d0c9c', 'serial': '71cdbc05-35d6-43f4-9224-cab57f2d0c9c'}, 'disk_bus': None, 'attachment_id': 'fb983a95-4328-453e-bfa1-93226df6dd09', 'guest_format': None, 'mount_device': '/dev/sdb', 'delete_on_termination': False, 'device_type': None, 'boot_index': None, 'volume_type': None}], 'swap': None} {{(pid=63515) spawn /opt/stack/nova/nova/virt/vmwareapi/vmops.py:799}} [ 1068.941186] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-f210ba7b-c0d6-4f5c-826f-ac5409d72c2c tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: d6c5a9f9-5e5c-4b14-9b9a-7825bbf526a5] Volume attach. Driver type: vmdk {{(pid=63515) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1068.941387] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-f210ba7b-c0d6-4f5c-826f-ac5409d72c2c tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: d6c5a9f9-5e5c-4b14-9b9a-7825bbf526a5] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-243569', 'volume_id': '71cdbc05-35d6-43f4-9224-cab57f2d0c9c', 'name': 'volume-71cdbc05-35d6-43f4-9224-cab57f2d0c9c', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'd6c5a9f9-5e5c-4b14-9b9a-7825bbf526a5', 'attached_at': '', 'detached_at': '', 'volume_id': '71cdbc05-35d6-43f4-9224-cab57f2d0c9c', 'serial': '71cdbc05-35d6-43f4-9224-cab57f2d0c9c'} {{(pid=63515) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1068.942193] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2cf4c11d-bf6b-4a60-bef3-e48b0cefb799 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.956940] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-deedf0e8-aeac-49e4-98bf-3902c8cc8f68 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.981711] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-f210ba7b-c0d6-4f5c-826f-ac5409d72c2c tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: d6c5a9f9-5e5c-4b14-9b9a-7825bbf526a5] Reconfiguring VM instance instance-00000062 to attach disk [datastore1] volume-71cdbc05-35d6-43f4-9224-cab57f2d0c9c/volume-71cdbc05-35d6-43f4-9224-cab57f2d0c9c.vmdk or device None with type thin {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1068.982046] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-68f3d541-ab77-4c42-af5b-90ca27fbbe87 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.000934] env[63515]: DEBUG oslo_vmware.api [None req-f210ba7b-c0d6-4f5c-826f-ac5409d72c2c tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Waiting for the task: (returnval){ [ 1069.000934] env[63515]: value = "task-1111903" [ 1069.000934] env[63515]: _type = "Task" [ 1069.000934] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1069.010973] env[63515]: DEBUG oslo_vmware.api [None req-f210ba7b-c0d6-4f5c-826f-ac5409d72c2c tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Task: {'id': task-1111903, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1069.327771] env[63515]: DEBUG nova.compute.manager [None req-9dbc5332-6cd4-49ed-b269-132a6511b7a6 tempest-ServersNegativeTestMultiTenantJSON-1925445199 tempest-ServersNegativeTestMultiTenantJSON-1925445199-project-member] [instance: 7fdc6303-2d53-4d87-9aea-bf40e476f790] Start spawning the instance on the hypervisor. {{(pid=63515) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1069.352955] env[63515]: DEBUG nova.virt.hardware [None req-9dbc5332-6cd4-49ed-b269-132a6511b7a6 tempest-ServersNegativeTestMultiTenantJSON-1925445199 tempest-ServersNegativeTestMultiTenantJSON-1925445199-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T02:52:14Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T02:51:56Z,direct_url=,disk_format='vmdk',id=8a120570-cb06-4099-b262-554ca0ad15c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b569255cc43e42e7ae3f2b4ad37c6ef4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T02:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1069.353289] env[63515]: DEBUG nova.virt.hardware [None req-9dbc5332-6cd4-49ed-b269-132a6511b7a6 tempest-ServersNegativeTestMultiTenantJSON-1925445199 tempest-ServersNegativeTestMultiTenantJSON-1925445199-project-member] Flavor limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1069.353527] env[63515]: DEBUG nova.virt.hardware [None req-9dbc5332-6cd4-49ed-b269-132a6511b7a6 tempest-ServersNegativeTestMultiTenantJSON-1925445199 tempest-ServersNegativeTestMultiTenantJSON-1925445199-project-member] Image limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1069.353738] env[63515]: DEBUG nova.virt.hardware [None req-9dbc5332-6cd4-49ed-b269-132a6511b7a6 tempest-ServersNegativeTestMultiTenantJSON-1925445199 tempest-ServersNegativeTestMultiTenantJSON-1925445199-project-member] Flavor pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1069.353915] env[63515]: DEBUG nova.virt.hardware [None req-9dbc5332-6cd4-49ed-b269-132a6511b7a6 tempest-ServersNegativeTestMultiTenantJSON-1925445199 tempest-ServersNegativeTestMultiTenantJSON-1925445199-project-member] Image pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1069.354192] env[63515]: DEBUG nova.virt.hardware [None req-9dbc5332-6cd4-49ed-b269-132a6511b7a6 tempest-ServersNegativeTestMultiTenantJSON-1925445199 tempest-ServersNegativeTestMultiTenantJSON-1925445199-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1069.354417] env[63515]: DEBUG nova.virt.hardware [None req-9dbc5332-6cd4-49ed-b269-132a6511b7a6 tempest-ServersNegativeTestMultiTenantJSON-1925445199 tempest-ServersNegativeTestMultiTenantJSON-1925445199-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1069.354677] env[63515]: DEBUG nova.virt.hardware [None req-9dbc5332-6cd4-49ed-b269-132a6511b7a6 tempest-ServersNegativeTestMultiTenantJSON-1925445199 tempest-ServersNegativeTestMultiTenantJSON-1925445199-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1069.354983] env[63515]: DEBUG nova.virt.hardware [None req-9dbc5332-6cd4-49ed-b269-132a6511b7a6 tempest-ServersNegativeTestMultiTenantJSON-1925445199 tempest-ServersNegativeTestMultiTenantJSON-1925445199-project-member] Got 1 possible topologies {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1069.355075] env[63515]: DEBUG nova.virt.hardware [None req-9dbc5332-6cd4-49ed-b269-132a6511b7a6 tempest-ServersNegativeTestMultiTenantJSON-1925445199 tempest-ServersNegativeTestMultiTenantJSON-1925445199-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1069.355310] env[63515]: DEBUG nova.virt.hardware [None req-9dbc5332-6cd4-49ed-b269-132a6511b7a6 tempest-ServersNegativeTestMultiTenantJSON-1925445199 tempest-ServersNegativeTestMultiTenantJSON-1925445199-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1069.356296] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4082f5f3-f564-4c10-98ae-e84209b9971d {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.367148] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-246b5b47-26c1-4372-8bb1-28f9579bb6b3 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.396232] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d048a3bb-e5e3-4f29-97f3-eb93fe9f8ec7 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.403321] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d43e9f6-aa73-4828-9463-ef276571e554 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.434482] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd621ad1-e4f5-4fd1-902d-fae475dbbbc5 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.442006] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-730e1be4-60fe-4c24-ad3d-3e728e016707 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.455309] env[63515]: DEBUG nova.compute.provider_tree [None req-30134ddc-44b3-4354-8366-42d76e3947db tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1069.458806] env[63515]: DEBUG oslo_service.periodic_task [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=63515) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1069.459026] env[63515]: DEBUG oslo_service.periodic_task [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=63515) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1069.459243] env[63515]: DEBUG nova.compute.manager [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Starting heal instance info cache {{(pid=63515) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9926}} [ 1069.510426] env[63515]: DEBUG oslo_vmware.api [None req-f210ba7b-c0d6-4f5c-826f-ac5409d72c2c tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Task: {'id': task-1111903, 'name': ReconfigVM_Task, 'duration_secs': 0.424424} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1069.510679] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-f210ba7b-c0d6-4f5c-826f-ac5409d72c2c tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: d6c5a9f9-5e5c-4b14-9b9a-7825bbf526a5] Reconfigured VM instance instance-00000062 to attach disk [datastore1] volume-71cdbc05-35d6-43f4-9224-cab57f2d0c9c/volume-71cdbc05-35d6-43f4-9224-cab57f2d0c9c.vmdk or device None with type thin {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1069.515285] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2979c5d8-fbc8-49d3-8298-587b15b0de3c {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.529671] env[63515]: DEBUG oslo_vmware.api [None req-f210ba7b-c0d6-4f5c-826f-ac5409d72c2c tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Waiting for the task: (returnval){ [ 1069.529671] env[63515]: value = "task-1111904" [ 1069.529671] env[63515]: _type = "Task" [ 1069.529671] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1069.538620] env[63515]: DEBUG oslo_vmware.api [None req-f210ba7b-c0d6-4f5c-826f-ac5409d72c2c tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Task: {'id': task-1111904, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1069.940765] env[63515]: DEBUG nova.compute.manager [req-ec772654-afcc-4b9b-8b92-b99a8dec383f req-d96df2e2-c51e-480d-93d6-30d9b1b4c0b6 service nova] [instance: 7fdc6303-2d53-4d87-9aea-bf40e476f790] Received event network-vif-plugged-0a17708c-c9b4-4715-bcfa-f2629d5b179f {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1069.940992] env[63515]: DEBUG oslo_concurrency.lockutils [req-ec772654-afcc-4b9b-8b92-b99a8dec383f req-d96df2e2-c51e-480d-93d6-30d9b1b4c0b6 service nova] Acquiring lock "7fdc6303-2d53-4d87-9aea-bf40e476f790-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1069.941221] env[63515]: DEBUG oslo_concurrency.lockutils [req-ec772654-afcc-4b9b-8b92-b99a8dec383f req-d96df2e2-c51e-480d-93d6-30d9b1b4c0b6 service nova] Lock "7fdc6303-2d53-4d87-9aea-bf40e476f790-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1069.941399] env[63515]: DEBUG oslo_concurrency.lockutils [req-ec772654-afcc-4b9b-8b92-b99a8dec383f req-d96df2e2-c51e-480d-93d6-30d9b1b4c0b6 service nova] Lock "7fdc6303-2d53-4d87-9aea-bf40e476f790-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1069.941586] env[63515]: DEBUG nova.compute.manager [req-ec772654-afcc-4b9b-8b92-b99a8dec383f req-d96df2e2-c51e-480d-93d6-30d9b1b4c0b6 service nova] [instance: 7fdc6303-2d53-4d87-9aea-bf40e476f790] No waiting events found dispatching network-vif-plugged-0a17708c-c9b4-4715-bcfa-f2629d5b179f {{(pid=63515) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1069.941771] env[63515]: WARNING nova.compute.manager [req-ec772654-afcc-4b9b-8b92-b99a8dec383f req-d96df2e2-c51e-480d-93d6-30d9b1b4c0b6 service nova] [instance: 7fdc6303-2d53-4d87-9aea-bf40e476f790] Received unexpected event network-vif-plugged-0a17708c-c9b4-4715-bcfa-f2629d5b179f for instance with vm_state building and task_state spawning. [ 1069.958397] env[63515]: DEBUG nova.scheduler.client.report [None req-30134ddc-44b3-4354-8366-42d76e3947db tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1070.042697] env[63515]: DEBUG oslo_vmware.api [None req-f210ba7b-c0d6-4f5c-826f-ac5409d72c2c tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Task: {'id': task-1111904, 'name': ReconfigVM_Task, 'duration_secs': 0.284918} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1070.043751] env[63515]: DEBUG nova.network.neutron [None req-9dbc5332-6cd4-49ed-b269-132a6511b7a6 tempest-ServersNegativeTestMultiTenantJSON-1925445199 tempest-ServersNegativeTestMultiTenantJSON-1925445199-project-member] [instance: 7fdc6303-2d53-4d87-9aea-bf40e476f790] Successfully updated port: 0a17708c-c9b4-4715-bcfa-f2629d5b179f {{(pid=63515) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1070.045346] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-f210ba7b-c0d6-4f5c-826f-ac5409d72c2c tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: d6c5a9f9-5e5c-4b14-9b9a-7825bbf526a5] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-243569', 'volume_id': '71cdbc05-35d6-43f4-9224-cab57f2d0c9c', 'name': 'volume-71cdbc05-35d6-43f4-9224-cab57f2d0c9c', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'd6c5a9f9-5e5c-4b14-9b9a-7825bbf526a5', 'attached_at': '', 'detached_at': '', 'volume_id': '71cdbc05-35d6-43f4-9224-cab57f2d0c9c', 'serial': '71cdbc05-35d6-43f4-9224-cab57f2d0c9c'} {{(pid=63515) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1070.046651] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d81685cc-a8cc-4b81-a3bb-2312b938fca5 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.054782] env[63515]: DEBUG oslo_vmware.api [None req-f210ba7b-c0d6-4f5c-826f-ac5409d72c2c tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Waiting for the task: (returnval){ [ 1070.054782] env[63515]: value = "task-1111905" [ 1070.054782] env[63515]: _type = "Task" [ 1070.054782] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1070.064636] env[63515]: DEBUG oslo_vmware.api [None req-f210ba7b-c0d6-4f5c-826f-ac5409d72c2c tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Task: {'id': task-1111905, 'name': Rename_Task} progress is 5%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1070.464523] env[63515]: DEBUG oslo_concurrency.lockutils [None req-30134ddc-44b3-4354-8366-42d76e3947db tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.671s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1070.489830] env[63515]: INFO nova.scheduler.client.report [None req-30134ddc-44b3-4354-8366-42d76e3947db tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Deleted allocations for instance 724b05bc-ce4b-4f99-ad14-f3abea5567f2 [ 1070.547205] env[63515]: DEBUG oslo_concurrency.lockutils [None req-9dbc5332-6cd4-49ed-b269-132a6511b7a6 tempest-ServersNegativeTestMultiTenantJSON-1925445199 tempest-ServersNegativeTestMultiTenantJSON-1925445199-project-member] Acquiring lock "refresh_cache-7fdc6303-2d53-4d87-9aea-bf40e476f790" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1070.547366] env[63515]: DEBUG oslo_concurrency.lockutils [None req-9dbc5332-6cd4-49ed-b269-132a6511b7a6 tempest-ServersNegativeTestMultiTenantJSON-1925445199 tempest-ServersNegativeTestMultiTenantJSON-1925445199-project-member] Acquired lock "refresh_cache-7fdc6303-2d53-4d87-9aea-bf40e476f790" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1070.547736] env[63515]: DEBUG nova.network.neutron [None req-9dbc5332-6cd4-49ed-b269-132a6511b7a6 tempest-ServersNegativeTestMultiTenantJSON-1925445199 tempest-ServersNegativeTestMultiTenantJSON-1925445199-project-member] [instance: 7fdc6303-2d53-4d87-9aea-bf40e476f790] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1070.566626] env[63515]: DEBUG oslo_vmware.api [None req-f210ba7b-c0d6-4f5c-826f-ac5409d72c2c tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Task: {'id': task-1111905, 'name': Rename_Task} progress is 14%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1070.996649] env[63515]: DEBUG oslo_concurrency.lockutils [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Acquiring lock "refresh_cache-e13da90c-28e6-43d2-99b7-19c5095954ca" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1070.996898] env[63515]: DEBUG oslo_concurrency.lockutils [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Acquired lock "refresh_cache-e13da90c-28e6-43d2-99b7-19c5095954ca" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1070.996936] env[63515]: DEBUG nova.network.neutron [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] [instance: e13da90c-28e6-43d2-99b7-19c5095954ca] Forcefully refreshing network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2004}} [ 1070.998205] env[63515]: DEBUG oslo_concurrency.lockutils [None req-30134ddc-44b3-4354-8366-42d76e3947db tempest-AttachInterfacesTestJSON-43695247 tempest-AttachInterfacesTestJSON-43695247-project-member] Lock "724b05bc-ce4b-4f99-ad14-f3abea5567f2" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.250s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1071.066081] env[63515]: DEBUG oslo_vmware.api [None req-f210ba7b-c0d6-4f5c-826f-ac5409d72c2c tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Task: {'id': task-1111905, 'name': Rename_Task, 'duration_secs': 0.879192} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1071.066372] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-f210ba7b-c0d6-4f5c-826f-ac5409d72c2c tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: d6c5a9f9-5e5c-4b14-9b9a-7825bbf526a5] Powering on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1071.066622] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-09c23048-30cc-4b8d-b38e-cdf27c2ed0e0 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.072724] env[63515]: DEBUG oslo_vmware.api [None req-f210ba7b-c0d6-4f5c-826f-ac5409d72c2c tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Waiting for the task: (returnval){ [ 1071.072724] env[63515]: value = "task-1111906" [ 1071.072724] env[63515]: _type = "Task" [ 1071.072724] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1071.078236] env[63515]: DEBUG nova.network.neutron [None req-9dbc5332-6cd4-49ed-b269-132a6511b7a6 tempest-ServersNegativeTestMultiTenantJSON-1925445199 tempest-ServersNegativeTestMultiTenantJSON-1925445199-project-member] [instance: 7fdc6303-2d53-4d87-9aea-bf40e476f790] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1071.089473] env[63515]: DEBUG oslo_vmware.api [None req-f210ba7b-c0d6-4f5c-826f-ac5409d72c2c tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Task: {'id': task-1111906, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1071.214646] env[63515]: DEBUG nova.network.neutron [None req-9dbc5332-6cd4-49ed-b269-132a6511b7a6 tempest-ServersNegativeTestMultiTenantJSON-1925445199 tempest-ServersNegativeTestMultiTenantJSON-1925445199-project-member] [instance: 7fdc6303-2d53-4d87-9aea-bf40e476f790] Updating instance_info_cache with network_info: [{"id": "0a17708c-c9b4-4715-bcfa-f2629d5b179f", "address": "fa:16:3e:8a:74:d4", "network": {"id": "c6edaab4-3f96-457c-8273-dd1571619804", "bridge": "br-int", "label": "tempest-ServersNegativeTestMultiTenantJSON-785361350-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3cce2d6c191a4187932179eb598e8bad", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "365ac5b1-6d83-4dfe-887f-60574d7f6124", "external-id": "nsx-vlan-transportzone-138", "segmentation_id": 138, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0a17708c-c9", "ovs_interfaceid": "0a17708c-c9b4-4715-bcfa-f2629d5b179f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1071.583079] env[63515]: DEBUG oslo_vmware.api [None req-f210ba7b-c0d6-4f5c-826f-ac5409d72c2c tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Task: {'id': task-1111906, 'name': PowerOnVM_Task, 'duration_secs': 0.453448} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1071.583369] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-f210ba7b-c0d6-4f5c-826f-ac5409d72c2c tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: d6c5a9f9-5e5c-4b14-9b9a-7825bbf526a5] Powered on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1071.583595] env[63515]: DEBUG nova.compute.manager [None req-f210ba7b-c0d6-4f5c-826f-ac5409d72c2c tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: d6c5a9f9-5e5c-4b14-9b9a-7825bbf526a5] Checking state {{(pid=63515) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1071.584374] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c81d4c3-17da-48a2-844a-c12822665e30 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.717739] env[63515]: DEBUG oslo_concurrency.lockutils [None req-9dbc5332-6cd4-49ed-b269-132a6511b7a6 tempest-ServersNegativeTestMultiTenantJSON-1925445199 tempest-ServersNegativeTestMultiTenantJSON-1925445199-project-member] Releasing lock "refresh_cache-7fdc6303-2d53-4d87-9aea-bf40e476f790" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1071.718086] env[63515]: DEBUG nova.compute.manager [None req-9dbc5332-6cd4-49ed-b269-132a6511b7a6 tempest-ServersNegativeTestMultiTenantJSON-1925445199 tempest-ServersNegativeTestMultiTenantJSON-1925445199-project-member] [instance: 7fdc6303-2d53-4d87-9aea-bf40e476f790] Instance network_info: |[{"id": "0a17708c-c9b4-4715-bcfa-f2629d5b179f", "address": "fa:16:3e:8a:74:d4", "network": {"id": "c6edaab4-3f96-457c-8273-dd1571619804", "bridge": "br-int", "label": "tempest-ServersNegativeTestMultiTenantJSON-785361350-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3cce2d6c191a4187932179eb598e8bad", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "365ac5b1-6d83-4dfe-887f-60574d7f6124", "external-id": "nsx-vlan-transportzone-138", "segmentation_id": 138, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0a17708c-c9", "ovs_interfaceid": "0a17708c-c9b4-4715-bcfa-f2629d5b179f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1071.718515] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-9dbc5332-6cd4-49ed-b269-132a6511b7a6 tempest-ServersNegativeTestMultiTenantJSON-1925445199 tempest-ServersNegativeTestMultiTenantJSON-1925445199-project-member] [instance: 7fdc6303-2d53-4d87-9aea-bf40e476f790] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:8a:74:d4', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '365ac5b1-6d83-4dfe-887f-60574d7f6124', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '0a17708c-c9b4-4715-bcfa-f2629d5b179f', 'vif_model': 'vmxnet3'}] {{(pid=63515) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1071.726503] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-9dbc5332-6cd4-49ed-b269-132a6511b7a6 tempest-ServersNegativeTestMultiTenantJSON-1925445199 tempest-ServersNegativeTestMultiTenantJSON-1925445199-project-member] Creating folder: Project (3cce2d6c191a4187932179eb598e8bad). Parent ref: group-v243370. {{(pid=63515) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1071.726843] env[63515]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-8dfe566b-7de4-4f54-b95e-61e455c0992e {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.738566] env[63515]: INFO nova.virt.vmwareapi.vm_util [None req-9dbc5332-6cd4-49ed-b269-132a6511b7a6 tempest-ServersNegativeTestMultiTenantJSON-1925445199 tempest-ServersNegativeTestMultiTenantJSON-1925445199-project-member] Created folder: Project (3cce2d6c191a4187932179eb598e8bad) in parent group-v243370. [ 1071.738803] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-9dbc5332-6cd4-49ed-b269-132a6511b7a6 tempest-ServersNegativeTestMultiTenantJSON-1925445199 tempest-ServersNegativeTestMultiTenantJSON-1925445199-project-member] Creating folder: Instances. Parent ref: group-v243571. {{(pid=63515) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1071.739059] env[63515]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-904cc3cd-eccf-498b-878e-156fc3aef16f {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.747570] env[63515]: INFO nova.virt.vmwareapi.vm_util [None req-9dbc5332-6cd4-49ed-b269-132a6511b7a6 tempest-ServersNegativeTestMultiTenantJSON-1925445199 tempest-ServersNegativeTestMultiTenantJSON-1925445199-project-member] Created folder: Instances in parent group-v243571. [ 1071.747720] env[63515]: DEBUG oslo.service.loopingcall [None req-9dbc5332-6cd4-49ed-b269-132a6511b7a6 tempest-ServersNegativeTestMultiTenantJSON-1925445199 tempest-ServersNegativeTestMultiTenantJSON-1925445199-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1071.747914] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7fdc6303-2d53-4d87-9aea-bf40e476f790] Creating VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1071.748899] env[63515]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-395a9e7e-bd60-4952-abc6-ed5f8a51c015 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.767850] env[63515]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1071.767850] env[63515]: value = "task-1111909" [ 1071.767850] env[63515]: _type = "Task" [ 1071.767850] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1071.775410] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1111909, 'name': CreateVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1071.965076] env[63515]: DEBUG nova.compute.manager [req-6a429081-1ea7-46c1-90aa-506ec045d2b1 req-5ead95dd-921a-4c62-94e9-a4a743ff9399 service nova] [instance: 7fdc6303-2d53-4d87-9aea-bf40e476f790] Received event network-changed-0a17708c-c9b4-4715-bcfa-f2629d5b179f {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1071.965299] env[63515]: DEBUG nova.compute.manager [req-6a429081-1ea7-46c1-90aa-506ec045d2b1 req-5ead95dd-921a-4c62-94e9-a4a743ff9399 service nova] [instance: 7fdc6303-2d53-4d87-9aea-bf40e476f790] Refreshing instance network info cache due to event network-changed-0a17708c-c9b4-4715-bcfa-f2629d5b179f. {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1071.965531] env[63515]: DEBUG oslo_concurrency.lockutils [req-6a429081-1ea7-46c1-90aa-506ec045d2b1 req-5ead95dd-921a-4c62-94e9-a4a743ff9399 service nova] Acquiring lock "refresh_cache-7fdc6303-2d53-4d87-9aea-bf40e476f790" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1071.965681] env[63515]: DEBUG oslo_concurrency.lockutils [req-6a429081-1ea7-46c1-90aa-506ec045d2b1 req-5ead95dd-921a-4c62-94e9-a4a743ff9399 service nova] Acquired lock "refresh_cache-7fdc6303-2d53-4d87-9aea-bf40e476f790" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1071.965846] env[63515]: DEBUG nova.network.neutron [req-6a429081-1ea7-46c1-90aa-506ec045d2b1 req-5ead95dd-921a-4c62-94e9-a4a743ff9399 service nova] [instance: 7fdc6303-2d53-4d87-9aea-bf40e476f790] Refreshing network info cache for port 0a17708c-c9b4-4715-bcfa-f2629d5b179f {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1072.102758] env[63515]: DEBUG oslo_concurrency.lockutils [None req-f210ba7b-c0d6-4f5c-826f-ac5409d72c2c tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1072.103206] env[63515]: DEBUG oslo_concurrency.lockutils [None req-f210ba7b-c0d6-4f5c-826f-ac5409d72c2c tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1072.103206] env[63515]: DEBUG nova.objects.instance [None req-f210ba7b-c0d6-4f5c-826f-ac5409d72c2c tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: d6c5a9f9-5e5c-4b14-9b9a-7825bbf526a5] Trying to apply a migration context that does not seem to be set for this instance {{(pid=63515) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 1072.278558] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1111909, 'name': CreateVM_Task, 'duration_secs': 0.375971} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1072.278821] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7fdc6303-2d53-4d87-9aea-bf40e476f790] Created VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1072.279508] env[63515]: DEBUG oslo_concurrency.lockutils [None req-9dbc5332-6cd4-49ed-b269-132a6511b7a6 tempest-ServersNegativeTestMultiTenantJSON-1925445199 tempest-ServersNegativeTestMultiTenantJSON-1925445199-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1072.279868] env[63515]: DEBUG oslo_concurrency.lockutils [None req-9dbc5332-6cd4-49ed-b269-132a6511b7a6 tempest-ServersNegativeTestMultiTenantJSON-1925445199 tempest-ServersNegativeTestMultiTenantJSON-1925445199-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1072.280246] env[63515]: DEBUG oslo_concurrency.lockutils [None req-9dbc5332-6cd4-49ed-b269-132a6511b7a6 tempest-ServersNegativeTestMultiTenantJSON-1925445199 tempest-ServersNegativeTestMultiTenantJSON-1925445199-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1072.280516] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-46e26210-3d94-4bd2-80a8-2ac151888e30 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.288453] env[63515]: DEBUG oslo_vmware.api [None req-9dbc5332-6cd4-49ed-b269-132a6511b7a6 tempest-ServersNegativeTestMultiTenantJSON-1925445199 tempest-ServersNegativeTestMultiTenantJSON-1925445199-project-member] Waiting for the task: (returnval){ [ 1072.288453] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]52f73aeb-9d57-3319-5c79-10ed243b03c7" [ 1072.288453] env[63515]: _type = "Task" [ 1072.288453] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1072.301569] env[63515]: DEBUG oslo_vmware.api [None req-9dbc5332-6cd4-49ed-b269-132a6511b7a6 tempest-ServersNegativeTestMultiTenantJSON-1925445199 tempest-ServersNegativeTestMultiTenantJSON-1925445199-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52f73aeb-9d57-3319-5c79-10ed243b03c7, 'name': SearchDatastore_Task, 'duration_secs': 0.009648} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1072.301867] env[63515]: DEBUG oslo_concurrency.lockutils [None req-9dbc5332-6cd4-49ed-b269-132a6511b7a6 tempest-ServersNegativeTestMultiTenantJSON-1925445199 tempest-ServersNegativeTestMultiTenantJSON-1925445199-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1072.302115] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-9dbc5332-6cd4-49ed-b269-132a6511b7a6 tempest-ServersNegativeTestMultiTenantJSON-1925445199 tempest-ServersNegativeTestMultiTenantJSON-1925445199-project-member] [instance: 7fdc6303-2d53-4d87-9aea-bf40e476f790] Processing image 8a120570-cb06-4099-b262-554ca0ad15c5 {{(pid=63515) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1072.302354] env[63515]: DEBUG oslo_concurrency.lockutils [None req-9dbc5332-6cd4-49ed-b269-132a6511b7a6 tempest-ServersNegativeTestMultiTenantJSON-1925445199 tempest-ServersNegativeTestMultiTenantJSON-1925445199-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1072.302502] env[63515]: DEBUG oslo_concurrency.lockutils [None req-9dbc5332-6cd4-49ed-b269-132a6511b7a6 tempest-ServersNegativeTestMultiTenantJSON-1925445199 tempest-ServersNegativeTestMultiTenantJSON-1925445199-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1072.302875] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-9dbc5332-6cd4-49ed-b269-132a6511b7a6 tempest-ServersNegativeTestMultiTenantJSON-1925445199 tempest-ServersNegativeTestMultiTenantJSON-1925445199-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1072.303167] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ea7d0788-69f3-4af2-acb9-fa6914485a55 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.311346] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-9dbc5332-6cd4-49ed-b269-132a6511b7a6 tempest-ServersNegativeTestMultiTenantJSON-1925445199 tempest-ServersNegativeTestMultiTenantJSON-1925445199-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1072.311346] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-9dbc5332-6cd4-49ed-b269-132a6511b7a6 tempest-ServersNegativeTestMultiTenantJSON-1925445199 tempest-ServersNegativeTestMultiTenantJSON-1925445199-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63515) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1072.312973] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-14f1c083-c573-4896-a1b8-3ca71d73c0ab {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.316857] env[63515]: DEBUG oslo_vmware.api [None req-9dbc5332-6cd4-49ed-b269-132a6511b7a6 tempest-ServersNegativeTestMultiTenantJSON-1925445199 tempest-ServersNegativeTestMultiTenantJSON-1925445199-project-member] Waiting for the task: (returnval){ [ 1072.316857] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]52c409e2-2427-efa6-c079-7444ec2a80cd" [ 1072.316857] env[63515]: _type = "Task" [ 1072.316857] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1072.325216] env[63515]: DEBUG oslo_vmware.api [None req-9dbc5332-6cd4-49ed-b269-132a6511b7a6 tempest-ServersNegativeTestMultiTenantJSON-1925445199 tempest-ServersNegativeTestMultiTenantJSON-1925445199-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52c409e2-2427-efa6-c079-7444ec2a80cd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1072.400241] env[63515]: DEBUG nova.network.neutron [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] [instance: e13da90c-28e6-43d2-99b7-19c5095954ca] Updating instance_info_cache with network_info: [{"id": "d9363b80-630e-4262-901f-c13e5c8b8c2a", "address": "fa:16:3e:dd:fe:76", "network": {"id": "cec614cb-7a5b-4b16-a83d-061ff62d6677", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1501468300-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.155", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "aef947bfaf7d4180b46b4be662e3d903", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3a0ddd7d-c321-4187-bdd8-b19044ea2c4a", "external-id": "nsx-vlan-transportzone-747", "segmentation_id": 747, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd9363b80-63", "ovs_interfaceid": "d9363b80-630e-4262-901f-c13e5c8b8c2a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1072.812112] env[63515]: DEBUG nova.network.neutron [req-6a429081-1ea7-46c1-90aa-506ec045d2b1 req-5ead95dd-921a-4c62-94e9-a4a743ff9399 service nova] [instance: 7fdc6303-2d53-4d87-9aea-bf40e476f790] Updated VIF entry in instance network info cache for port 0a17708c-c9b4-4715-bcfa-f2629d5b179f. {{(pid=63515) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1072.812535] env[63515]: DEBUG nova.network.neutron [req-6a429081-1ea7-46c1-90aa-506ec045d2b1 req-5ead95dd-921a-4c62-94e9-a4a743ff9399 service nova] [instance: 7fdc6303-2d53-4d87-9aea-bf40e476f790] Updating instance_info_cache with network_info: [{"id": "0a17708c-c9b4-4715-bcfa-f2629d5b179f", "address": "fa:16:3e:8a:74:d4", "network": {"id": "c6edaab4-3f96-457c-8273-dd1571619804", "bridge": "br-int", "label": "tempest-ServersNegativeTestMultiTenantJSON-785361350-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3cce2d6c191a4187932179eb598e8bad", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "365ac5b1-6d83-4dfe-887f-60574d7f6124", "external-id": "nsx-vlan-transportzone-138", "segmentation_id": 138, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0a17708c-c9", "ovs_interfaceid": "0a17708c-c9b4-4715-bcfa-f2629d5b179f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1072.826779] env[63515]: DEBUG oslo_vmware.api [None req-9dbc5332-6cd4-49ed-b269-132a6511b7a6 tempest-ServersNegativeTestMultiTenantJSON-1925445199 tempest-ServersNegativeTestMultiTenantJSON-1925445199-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52c409e2-2427-efa6-c079-7444ec2a80cd, 'name': SearchDatastore_Task, 'duration_secs': 0.009474} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1072.828908] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0001ab24-bc38-4958-ae5d-fd9f20cf8af3 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.833505] env[63515]: DEBUG oslo_vmware.api [None req-9dbc5332-6cd4-49ed-b269-132a6511b7a6 tempest-ServersNegativeTestMultiTenantJSON-1925445199 tempest-ServersNegativeTestMultiTenantJSON-1925445199-project-member] Waiting for the task: (returnval){ [ 1072.833505] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]5299458e-85b4-2598-624e-606305c16be6" [ 1072.833505] env[63515]: _type = "Task" [ 1072.833505] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1072.841557] env[63515]: DEBUG oslo_vmware.api [None req-9dbc5332-6cd4-49ed-b269-132a6511b7a6 tempest-ServersNegativeTestMultiTenantJSON-1925445199 tempest-ServersNegativeTestMultiTenantJSON-1925445199-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]5299458e-85b4-2598-624e-606305c16be6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1072.874473] env[63515]: DEBUG oslo_concurrency.lockutils [None req-fe382c79-9bcd-4ff6-a158-630c3ea4209f tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Acquiring lock "7d572ac0-f6af-4622-96cc-e75983420222" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1072.874754] env[63515]: DEBUG oslo_concurrency.lockutils [None req-fe382c79-9bcd-4ff6-a158-630c3ea4209f tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Lock "7d572ac0-f6af-4622-96cc-e75983420222" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1072.874966] env[63515]: DEBUG oslo_concurrency.lockutils [None req-fe382c79-9bcd-4ff6-a158-630c3ea4209f tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Acquiring lock "7d572ac0-f6af-4622-96cc-e75983420222-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1072.875174] env[63515]: DEBUG oslo_concurrency.lockutils [None req-fe382c79-9bcd-4ff6-a158-630c3ea4209f tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Lock "7d572ac0-f6af-4622-96cc-e75983420222-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1072.875351] env[63515]: DEBUG oslo_concurrency.lockutils [None req-fe382c79-9bcd-4ff6-a158-630c3ea4209f tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Lock "7d572ac0-f6af-4622-96cc-e75983420222-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1072.879634] env[63515]: INFO nova.compute.manager [None req-fe382c79-9bcd-4ff6-a158-630c3ea4209f tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] [instance: 7d572ac0-f6af-4622-96cc-e75983420222] Terminating instance [ 1072.882323] env[63515]: DEBUG nova.compute.manager [None req-fe382c79-9bcd-4ff6-a158-630c3ea4209f tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] [instance: 7d572ac0-f6af-4622-96cc-e75983420222] Start destroying the instance on the hypervisor. {{(pid=63515) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1072.882524] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-fe382c79-9bcd-4ff6-a158-630c3ea4209f tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] [instance: 7d572ac0-f6af-4622-96cc-e75983420222] Destroying instance {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1072.883467] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7405247-a637-4c69-91ac-2a865413b5da {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.892031] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-fe382c79-9bcd-4ff6-a158-630c3ea4209f tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] [instance: 7d572ac0-f6af-4622-96cc-e75983420222] Powering off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1072.892280] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-abf98ae2-5fa1-4159-8a11-d70e8752b5bc {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.898479] env[63515]: DEBUG oslo_vmware.api [None req-fe382c79-9bcd-4ff6-a158-630c3ea4209f tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Waiting for the task: (returnval){ [ 1072.898479] env[63515]: value = "task-1111910" [ 1072.898479] env[63515]: _type = "Task" [ 1072.898479] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1072.903261] env[63515]: DEBUG oslo_concurrency.lockutils [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Releasing lock "refresh_cache-e13da90c-28e6-43d2-99b7-19c5095954ca" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1072.903518] env[63515]: DEBUG nova.compute.manager [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] [instance: e13da90c-28e6-43d2-99b7-19c5095954ca] Updated the network info_cache for instance {{(pid=63515) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9997}} [ 1072.906926] env[63515]: DEBUG oslo_service.periodic_task [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=63515) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1072.907191] env[63515]: DEBUG oslo_vmware.api [None req-fe382c79-9bcd-4ff6-a158-630c3ea4209f tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Task: {'id': task-1111910, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1072.907495] env[63515]: DEBUG oslo_service.periodic_task [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=63515) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1072.907688] env[63515]: DEBUG oslo_service.periodic_task [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=63515) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1072.907843] env[63515]: DEBUG oslo_service.periodic_task [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=63515) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1072.907982] env[63515]: DEBUG oslo_service.periodic_task [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=63515) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1072.908141] env[63515]: DEBUG oslo_service.periodic_task [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=63515) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1072.908271] env[63515]: DEBUG nova.compute.manager [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=63515) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10545}} [ 1072.908412] env[63515]: DEBUG oslo_service.periodic_task [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Running periodic task ComputeManager.update_available_resource {{(pid=63515) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1073.113643] env[63515]: DEBUG oslo_concurrency.lockutils [None req-f210ba7b-c0d6-4f5c-826f-ac5409d72c2c tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.009s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1073.316720] env[63515]: DEBUG oslo_concurrency.lockutils [req-6a429081-1ea7-46c1-90aa-506ec045d2b1 req-5ead95dd-921a-4c62-94e9-a4a743ff9399 service nova] Releasing lock "refresh_cache-7fdc6303-2d53-4d87-9aea-bf40e476f790" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1073.345678] env[63515]: DEBUG oslo_vmware.api [None req-9dbc5332-6cd4-49ed-b269-132a6511b7a6 tempest-ServersNegativeTestMultiTenantJSON-1925445199 tempest-ServersNegativeTestMultiTenantJSON-1925445199-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]5299458e-85b4-2598-624e-606305c16be6, 'name': SearchDatastore_Task, 'duration_secs': 0.009542} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1073.345955] env[63515]: DEBUG oslo_concurrency.lockutils [None req-9dbc5332-6cd4-49ed-b269-132a6511b7a6 tempest-ServersNegativeTestMultiTenantJSON-1925445199 tempest-ServersNegativeTestMultiTenantJSON-1925445199-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1073.346231] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-9dbc5332-6cd4-49ed-b269-132a6511b7a6 tempest-ServersNegativeTestMultiTenantJSON-1925445199 tempest-ServersNegativeTestMultiTenantJSON-1925445199-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk to [datastore1] 7fdc6303-2d53-4d87-9aea-bf40e476f790/7fdc6303-2d53-4d87-9aea-bf40e476f790.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1073.346491] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-9ad79e10-f300-4d47-a79b-6e4346bc26d7 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.352675] env[63515]: DEBUG oslo_vmware.api [None req-9dbc5332-6cd4-49ed-b269-132a6511b7a6 tempest-ServersNegativeTestMultiTenantJSON-1925445199 tempest-ServersNegativeTestMultiTenantJSON-1925445199-project-member] Waiting for the task: (returnval){ [ 1073.352675] env[63515]: value = "task-1111911" [ 1073.352675] env[63515]: _type = "Task" [ 1073.352675] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1073.359876] env[63515]: DEBUG oslo_vmware.api [None req-9dbc5332-6cd4-49ed-b269-132a6511b7a6 tempest-ServersNegativeTestMultiTenantJSON-1925445199 tempest-ServersNegativeTestMultiTenantJSON-1925445199-project-member] Task: {'id': task-1111911, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1073.411022] env[63515]: DEBUG oslo_vmware.api [None req-fe382c79-9bcd-4ff6-a158-630c3ea4209f tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Task: {'id': task-1111910, 'name': PowerOffVM_Task, 'duration_secs': 0.188423} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1073.411022] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-fe382c79-9bcd-4ff6-a158-630c3ea4209f tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] [instance: 7d572ac0-f6af-4622-96cc-e75983420222] Powered off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1073.411022] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-fe382c79-9bcd-4ff6-a158-630c3ea4209f tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] [instance: 7d572ac0-f6af-4622-96cc-e75983420222] Unregistering the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1073.411022] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e63979ce-6f17-4db7-8a1a-db6bc9603d3a {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.411022] env[63515]: DEBUG oslo_concurrency.lockutils [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1073.411022] env[63515]: DEBUG oslo_concurrency.lockutils [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1073.411541] env[63515]: DEBUG oslo_concurrency.lockutils [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1073.416587] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=63515) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1073.416587] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf252078-0e68-4f08-8481-c2467871a475 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.420514] env[63515]: DEBUG oslo_concurrency.lockutils [None req-6ec44eb5-057e-4cc7-8109-11c3373d8aa3 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Acquiring lock "d6c5a9f9-5e5c-4b14-9b9a-7825bbf526a5" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1073.420925] env[63515]: DEBUG oslo_concurrency.lockutils [None req-6ec44eb5-057e-4cc7-8109-11c3373d8aa3 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Lock "d6c5a9f9-5e5c-4b14-9b9a-7825bbf526a5" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1073.423428] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0aee9dd-47a7-4af1-9cd6-02e05f0e36e0 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.439887] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e236eb77-5807-4740-aca1-2a5599855604 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.446894] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52907f98-4b5c-4486-b3af-bde0f910106a {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.476227] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180482MB free_disk=171GB free_vcpus=48 pci_devices=None {{(pid=63515) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1073.476474] env[63515]: DEBUG oslo_concurrency.lockutils [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1073.476724] env[63515]: DEBUG oslo_concurrency.lockutils [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1073.480098] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-fe382c79-9bcd-4ff6-a158-630c3ea4209f tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] [instance: 7d572ac0-f6af-4622-96cc-e75983420222] Unregistered the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1073.480098] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-fe382c79-9bcd-4ff6-a158-630c3ea4209f tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] [instance: 7d572ac0-f6af-4622-96cc-e75983420222] Deleting contents of the VM from datastore datastore2 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1073.480098] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-fe382c79-9bcd-4ff6-a158-630c3ea4209f tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Deleting the datastore file [datastore2] 7d572ac0-f6af-4622-96cc-e75983420222 {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1073.480454] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-467a1007-99e5-401a-89a7-29292bd2bf6b {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.486040] env[63515]: DEBUG oslo_vmware.api [None req-fe382c79-9bcd-4ff6-a158-630c3ea4209f tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Waiting for the task: (returnval){ [ 1073.486040] env[63515]: value = "task-1111913" [ 1073.486040] env[63515]: _type = "Task" [ 1073.486040] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1073.494173] env[63515]: DEBUG oslo_vmware.api [None req-fe382c79-9bcd-4ff6-a158-630c3ea4209f tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Task: {'id': task-1111913, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1073.863407] env[63515]: DEBUG oslo_vmware.api [None req-9dbc5332-6cd4-49ed-b269-132a6511b7a6 tempest-ServersNegativeTestMultiTenantJSON-1925445199 tempest-ServersNegativeTestMultiTenantJSON-1925445199-project-member] Task: {'id': task-1111911, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.464191} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1073.863802] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-9dbc5332-6cd4-49ed-b269-132a6511b7a6 tempest-ServersNegativeTestMultiTenantJSON-1925445199 tempest-ServersNegativeTestMultiTenantJSON-1925445199-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk to [datastore1] 7fdc6303-2d53-4d87-9aea-bf40e476f790/7fdc6303-2d53-4d87-9aea-bf40e476f790.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1073.864188] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-9dbc5332-6cd4-49ed-b269-132a6511b7a6 tempest-ServersNegativeTestMultiTenantJSON-1925445199 tempest-ServersNegativeTestMultiTenantJSON-1925445199-project-member] [instance: 7fdc6303-2d53-4d87-9aea-bf40e476f790] Extending root virtual disk to 1048576 {{(pid=63515) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1073.864358] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-fa5a8450-1cc5-42ae-8ead-b302cfdb88cf {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.869820] env[63515]: DEBUG oslo_vmware.api [None req-9dbc5332-6cd4-49ed-b269-132a6511b7a6 tempest-ServersNegativeTestMultiTenantJSON-1925445199 tempest-ServersNegativeTestMultiTenantJSON-1925445199-project-member] Waiting for the task: (returnval){ [ 1073.869820] env[63515]: value = "task-1111914" [ 1073.869820] env[63515]: _type = "Task" [ 1073.869820] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1073.878658] env[63515]: DEBUG oslo_vmware.api [None req-9dbc5332-6cd4-49ed-b269-132a6511b7a6 tempest-ServersNegativeTestMultiTenantJSON-1925445199 tempest-ServersNegativeTestMultiTenantJSON-1925445199-project-member] Task: {'id': task-1111914, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1073.930034] env[63515]: INFO nova.compute.manager [None req-6ec44eb5-057e-4cc7-8109-11c3373d8aa3 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: d6c5a9f9-5e5c-4b14-9b9a-7825bbf526a5] Detaching volume 71cdbc05-35d6-43f4-9224-cab57f2d0c9c [ 1073.965021] env[63515]: INFO nova.virt.block_device [None req-6ec44eb5-057e-4cc7-8109-11c3373d8aa3 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: d6c5a9f9-5e5c-4b14-9b9a-7825bbf526a5] Attempting to driver detach volume 71cdbc05-35d6-43f4-9224-cab57f2d0c9c from mountpoint /dev/sdb [ 1073.965021] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-6ec44eb5-057e-4cc7-8109-11c3373d8aa3 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: d6c5a9f9-5e5c-4b14-9b9a-7825bbf526a5] Volume detach. Driver type: vmdk {{(pid=63515) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1073.965771] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-6ec44eb5-057e-4cc7-8109-11c3373d8aa3 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: d6c5a9f9-5e5c-4b14-9b9a-7825bbf526a5] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-243569', 'volume_id': '71cdbc05-35d6-43f4-9224-cab57f2d0c9c', 'name': 'volume-71cdbc05-35d6-43f4-9224-cab57f2d0c9c', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'd6c5a9f9-5e5c-4b14-9b9a-7825bbf526a5', 'attached_at': '', 'detached_at': '', 'volume_id': '71cdbc05-35d6-43f4-9224-cab57f2d0c9c', 'serial': '71cdbc05-35d6-43f4-9224-cab57f2d0c9c'} {{(pid=63515) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1073.966995] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c180f6c6-4494-42d6-b3ae-5e0012095a76 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.994291] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e48c0561-e22e-4469-86e9-c8d4c59943ae {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.001990] env[63515]: DEBUG oslo_vmware.api [None req-fe382c79-9bcd-4ff6-a158-630c3ea4209f tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Task: {'id': task-1111913, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.345062} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1074.003884] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-fe382c79-9bcd-4ff6-a158-630c3ea4209f tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Deleted the datastore file {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1074.004025] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-fe382c79-9bcd-4ff6-a158-630c3ea4209f tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] [instance: 7d572ac0-f6af-4622-96cc-e75983420222] Deleted contents of the VM from datastore datastore2 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1074.004183] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-fe382c79-9bcd-4ff6-a158-630c3ea4209f tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] [instance: 7d572ac0-f6af-4622-96cc-e75983420222] Instance destroyed {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1074.004365] env[63515]: INFO nova.compute.manager [None req-fe382c79-9bcd-4ff6-a158-630c3ea4209f tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] [instance: 7d572ac0-f6af-4622-96cc-e75983420222] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1074.004616] env[63515]: DEBUG oslo.service.loopingcall [None req-fe382c79-9bcd-4ff6-a158-630c3ea4209f tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1074.005540] env[63515]: DEBUG nova.compute.manager [-] [instance: 7d572ac0-f6af-4622-96cc-e75983420222] Deallocating network for instance {{(pid=63515) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1074.005675] env[63515]: DEBUG nova.network.neutron [-] [instance: 7d572ac0-f6af-4622-96cc-e75983420222] deallocate_for_instance() {{(pid=63515) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1074.007752] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07f1b2a5-1101-411b-9518-96ce1ea4c7b5 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.030754] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40430e6d-b36b-44d3-89e5-b145f0f8f7e3 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.046609] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-6ec44eb5-057e-4cc7-8109-11c3373d8aa3 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] The volume has not been displaced from its original location: [datastore1] volume-71cdbc05-35d6-43f4-9224-cab57f2d0c9c/volume-71cdbc05-35d6-43f4-9224-cab57f2d0c9c.vmdk. No consolidation needed. {{(pid=63515) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1074.052119] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-6ec44eb5-057e-4cc7-8109-11c3373d8aa3 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: d6c5a9f9-5e5c-4b14-9b9a-7825bbf526a5] Reconfiguring VM instance instance-00000062 to detach disk 2001 {{(pid=63515) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1074.054497] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8507a1ca-922c-4656-ac15-12089f547c96 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.073220] env[63515]: DEBUG oslo_vmware.api [None req-6ec44eb5-057e-4cc7-8109-11c3373d8aa3 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Waiting for the task: (returnval){ [ 1074.073220] env[63515]: value = "task-1111915" [ 1074.073220] env[63515]: _type = "Task" [ 1074.073220] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1074.081933] env[63515]: DEBUG oslo_vmware.api [None req-6ec44eb5-057e-4cc7-8109-11c3373d8aa3 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Task: {'id': task-1111915, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1074.379919] env[63515]: DEBUG oslo_vmware.api [None req-9dbc5332-6cd4-49ed-b269-132a6511b7a6 tempest-ServersNegativeTestMultiTenantJSON-1925445199 tempest-ServersNegativeTestMultiTenantJSON-1925445199-project-member] Task: {'id': task-1111914, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.059544} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1074.380293] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-9dbc5332-6cd4-49ed-b269-132a6511b7a6 tempest-ServersNegativeTestMultiTenantJSON-1925445199 tempest-ServersNegativeTestMultiTenantJSON-1925445199-project-member] [instance: 7fdc6303-2d53-4d87-9aea-bf40e476f790] Extended root virtual disk {{(pid=63515) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1074.381071] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd33784b-fc6b-447c-bee8-3f33e1cfb4d9 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.405205] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-9dbc5332-6cd4-49ed-b269-132a6511b7a6 tempest-ServersNegativeTestMultiTenantJSON-1925445199 tempest-ServersNegativeTestMultiTenantJSON-1925445199-project-member] [instance: 7fdc6303-2d53-4d87-9aea-bf40e476f790] Reconfiguring VM instance instance-00000068 to attach disk [datastore1] 7fdc6303-2d53-4d87-9aea-bf40e476f790/7fdc6303-2d53-4d87-9aea-bf40e476f790.vmdk or device None with type sparse {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1074.405364] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-237cbf25-bf15-4c9d-9a9c-35d09f47b5bf {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.424851] env[63515]: DEBUG oslo_vmware.api [None req-9dbc5332-6cd4-49ed-b269-132a6511b7a6 tempest-ServersNegativeTestMultiTenantJSON-1925445199 tempest-ServersNegativeTestMultiTenantJSON-1925445199-project-member] Waiting for the task: (returnval){ [ 1074.424851] env[63515]: value = "task-1111916" [ 1074.424851] env[63515]: _type = "Task" [ 1074.424851] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1074.432607] env[63515]: DEBUG oslo_vmware.api [None req-9dbc5332-6cd4-49ed-b269-132a6511b7a6 tempest-ServersNegativeTestMultiTenantJSON-1925445199 tempest-ServersNegativeTestMultiTenantJSON-1925445199-project-member] Task: {'id': task-1111916, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1074.488924] env[63515]: DEBUG nova.compute.manager [req-1f3f6112-3b0a-433c-94d4-dde52f6a3d7b req-fd043def-caf9-4ef0-b337-9e5d687c522a service nova] [instance: 7d572ac0-f6af-4622-96cc-e75983420222] Received event network-vif-deleted-dbfd09c2-9ec4-4821-8613-43c6d6cf4724 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1074.488924] env[63515]: INFO nova.compute.manager [req-1f3f6112-3b0a-433c-94d4-dde52f6a3d7b req-fd043def-caf9-4ef0-b337-9e5d687c522a service nova] [instance: 7d572ac0-f6af-4622-96cc-e75983420222] Neutron deleted interface dbfd09c2-9ec4-4821-8613-43c6d6cf4724; detaching it from the instance and deleting it from the info cache [ 1074.489142] env[63515]: DEBUG nova.network.neutron [req-1f3f6112-3b0a-433c-94d4-dde52f6a3d7b req-fd043def-caf9-4ef0-b337-9e5d687c522a service nova] [instance: 7d572ac0-f6af-4622-96cc-e75983420222] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1074.518227] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Instance e13da90c-28e6-43d2-99b7-19c5095954ca actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63515) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1074.518393] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Instance d6c5a9f9-5e5c-4b14-9b9a-7825bbf526a5 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63515) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1074.518517] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Instance 7d572ac0-f6af-4622-96cc-e75983420222 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63515) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1074.518650] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Instance aa83d035-9c47-434d-ab1a-140e6b3110b4 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63515) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1074.518825] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Instance 7fdc6303-2d53-4d87-9aea-bf40e476f790 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63515) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1074.519459] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Total usable vcpus: 48, total allocated vcpus: 5 {{(pid=63515) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1074.519459] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1472MB phys_disk=200GB used_disk=5GB total_vcpus=48 used_vcpus=5 pci_stats=[] {{(pid=63515) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1074.586873] env[63515]: DEBUG oslo_vmware.api [None req-6ec44eb5-057e-4cc7-8109-11c3373d8aa3 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Task: {'id': task-1111915, 'name': ReconfigVM_Task, 'duration_secs': 0.280773} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1074.587018] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-6ec44eb5-057e-4cc7-8109-11c3373d8aa3 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: d6c5a9f9-5e5c-4b14-9b9a-7825bbf526a5] Reconfigured VM instance instance-00000062 to detach disk 2001 {{(pid=63515) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1074.592698] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-861add84-4d2c-46a2-890a-06679bf6d400 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.602971] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0646a33-c115-4182-b37f-79c664e2824d {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.611558] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8834304-3530-49b6-9d8d-775a6e24b580 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.614594] env[63515]: DEBUG oslo_vmware.api [None req-6ec44eb5-057e-4cc7-8109-11c3373d8aa3 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Waiting for the task: (returnval){ [ 1074.614594] env[63515]: value = "task-1111917" [ 1074.614594] env[63515]: _type = "Task" [ 1074.614594] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1074.643015] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67690b4c-7f03-4273-975b-d5d416772b53 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.648537] env[63515]: DEBUG oslo_vmware.api [None req-6ec44eb5-057e-4cc7-8109-11c3373d8aa3 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Task: {'id': task-1111917, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1074.653243] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ebef07d3-13da-4651-8196-595760c6d9d6 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.668104] env[63515]: DEBUG nova.compute.provider_tree [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1074.935024] env[63515]: DEBUG oslo_vmware.api [None req-9dbc5332-6cd4-49ed-b269-132a6511b7a6 tempest-ServersNegativeTestMultiTenantJSON-1925445199 tempest-ServersNegativeTestMultiTenantJSON-1925445199-project-member] Task: {'id': task-1111916, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1074.961935] env[63515]: DEBUG nova.network.neutron [-] [instance: 7d572ac0-f6af-4622-96cc-e75983420222] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1074.994292] env[63515]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a729f9fd-8cb2-477e-af1b-6430d9482bd4 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.001608] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62283731-8c90-4f7b-9498-c67c37811c72 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.031890] env[63515]: DEBUG nova.compute.manager [req-1f3f6112-3b0a-433c-94d4-dde52f6a3d7b req-fd043def-caf9-4ef0-b337-9e5d687c522a service nova] [instance: 7d572ac0-f6af-4622-96cc-e75983420222] Detach interface failed, port_id=dbfd09c2-9ec4-4821-8613-43c6d6cf4724, reason: Instance 7d572ac0-f6af-4622-96cc-e75983420222 could not be found. {{(pid=63515) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1075.127034] env[63515]: DEBUG oslo_vmware.api [None req-6ec44eb5-057e-4cc7-8109-11c3373d8aa3 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Task: {'id': task-1111917, 'name': ReconfigVM_Task, 'duration_secs': 0.265493} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1075.127034] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-6ec44eb5-057e-4cc7-8109-11c3373d8aa3 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: d6c5a9f9-5e5c-4b14-9b9a-7825bbf526a5] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-243569', 'volume_id': '71cdbc05-35d6-43f4-9224-cab57f2d0c9c', 'name': 'volume-71cdbc05-35d6-43f4-9224-cab57f2d0c9c', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'd6c5a9f9-5e5c-4b14-9b9a-7825bbf526a5', 'attached_at': '', 'detached_at': '', 'volume_id': '71cdbc05-35d6-43f4-9224-cab57f2d0c9c', 'serial': '71cdbc05-35d6-43f4-9224-cab57f2d0c9c'} {{(pid=63515) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1075.173268] env[63515]: DEBUG nova.scheduler.client.report [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1075.436975] env[63515]: DEBUG oslo_vmware.api [None req-9dbc5332-6cd4-49ed-b269-132a6511b7a6 tempest-ServersNegativeTestMultiTenantJSON-1925445199 tempest-ServersNegativeTestMultiTenantJSON-1925445199-project-member] Task: {'id': task-1111916, 'name': ReconfigVM_Task, 'duration_secs': 0.99718} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1075.438604] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-9dbc5332-6cd4-49ed-b269-132a6511b7a6 tempest-ServersNegativeTestMultiTenantJSON-1925445199 tempest-ServersNegativeTestMultiTenantJSON-1925445199-project-member] [instance: 7fdc6303-2d53-4d87-9aea-bf40e476f790] Reconfigured VM instance instance-00000068 to attach disk [datastore1] 7fdc6303-2d53-4d87-9aea-bf40e476f790/7fdc6303-2d53-4d87-9aea-bf40e476f790.vmdk or device None with type sparse {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1075.438895] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e9bd72ff-7f31-447d-aba2-8a09657bb00c {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.448018] env[63515]: DEBUG oslo_vmware.api [None req-9dbc5332-6cd4-49ed-b269-132a6511b7a6 tempest-ServersNegativeTestMultiTenantJSON-1925445199 tempest-ServersNegativeTestMultiTenantJSON-1925445199-project-member] Waiting for the task: (returnval){ [ 1075.448018] env[63515]: value = "task-1111918" [ 1075.448018] env[63515]: _type = "Task" [ 1075.448018] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1075.454880] env[63515]: DEBUG oslo_vmware.api [None req-9dbc5332-6cd4-49ed-b269-132a6511b7a6 tempest-ServersNegativeTestMultiTenantJSON-1925445199 tempest-ServersNegativeTestMultiTenantJSON-1925445199-project-member] Task: {'id': task-1111918, 'name': Rename_Task} progress is 5%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1075.465776] env[63515]: INFO nova.compute.manager [-] [instance: 7d572ac0-f6af-4622-96cc-e75983420222] Took 1.46 seconds to deallocate network for instance. [ 1075.676021] env[63515]: DEBUG nova.objects.instance [None req-6ec44eb5-057e-4cc7-8109-11c3373d8aa3 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Lazy-loading 'flavor' on Instance uuid d6c5a9f9-5e5c-4b14-9b9a-7825bbf526a5 {{(pid=63515) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1075.676978] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=63515) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1075.677804] env[63515]: DEBUG oslo_concurrency.lockutils [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.201s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1075.678422] env[63515]: DEBUG oslo_service.periodic_task [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Running periodic task ComputeManager._cleanup_expired_console_auth_tokens {{(pid=63515) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1075.957260] env[63515]: DEBUG oslo_vmware.api [None req-9dbc5332-6cd4-49ed-b269-132a6511b7a6 tempest-ServersNegativeTestMultiTenantJSON-1925445199 tempest-ServersNegativeTestMultiTenantJSON-1925445199-project-member] Task: {'id': task-1111918, 'name': Rename_Task, 'duration_secs': 0.330797} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1075.957260] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-9dbc5332-6cd4-49ed-b269-132a6511b7a6 tempest-ServersNegativeTestMultiTenantJSON-1925445199 tempest-ServersNegativeTestMultiTenantJSON-1925445199-project-member] [instance: 7fdc6303-2d53-4d87-9aea-bf40e476f790] Powering on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1075.957575] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-73837b1a-de82-4328-9ea5-11e3a7987a72 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.964230] env[63515]: DEBUG oslo_vmware.api [None req-9dbc5332-6cd4-49ed-b269-132a6511b7a6 tempest-ServersNegativeTestMultiTenantJSON-1925445199 tempest-ServersNegativeTestMultiTenantJSON-1925445199-project-member] Waiting for the task: (returnval){ [ 1075.964230] env[63515]: value = "task-1111919" [ 1075.964230] env[63515]: _type = "Task" [ 1075.964230] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1075.973905] env[63515]: DEBUG oslo_concurrency.lockutils [None req-fe382c79-9bcd-4ff6-a158-630c3ea4209f tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1075.974118] env[63515]: DEBUG oslo_concurrency.lockutils [None req-fe382c79-9bcd-4ff6-a158-630c3ea4209f tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1075.974206] env[63515]: DEBUG nova.objects.instance [None req-fe382c79-9bcd-4ff6-a158-630c3ea4209f tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Lazy-loading 'resources' on Instance uuid 7d572ac0-f6af-4622-96cc-e75983420222 {{(pid=63515) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1075.975454] env[63515]: DEBUG oslo_vmware.api [None req-9dbc5332-6cd4-49ed-b269-132a6511b7a6 tempest-ServersNegativeTestMultiTenantJSON-1925445199 tempest-ServersNegativeTestMultiTenantJSON-1925445199-project-member] Task: {'id': task-1111919, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1076.480797] env[63515]: DEBUG oslo_vmware.api [None req-9dbc5332-6cd4-49ed-b269-132a6511b7a6 tempest-ServersNegativeTestMultiTenantJSON-1925445199 tempest-ServersNegativeTestMultiTenantJSON-1925445199-project-member] Task: {'id': task-1111919, 'name': PowerOnVM_Task, 'duration_secs': 0.464556} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1076.481310] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-9dbc5332-6cd4-49ed-b269-132a6511b7a6 tempest-ServersNegativeTestMultiTenantJSON-1925445199 tempest-ServersNegativeTestMultiTenantJSON-1925445199-project-member] [instance: 7fdc6303-2d53-4d87-9aea-bf40e476f790] Powered on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1076.481529] env[63515]: INFO nova.compute.manager [None req-9dbc5332-6cd4-49ed-b269-132a6511b7a6 tempest-ServersNegativeTestMultiTenantJSON-1925445199 tempest-ServersNegativeTestMultiTenantJSON-1925445199-project-member] [instance: 7fdc6303-2d53-4d87-9aea-bf40e476f790] Took 7.15 seconds to spawn the instance on the hypervisor. [ 1076.481715] env[63515]: DEBUG nova.compute.manager [None req-9dbc5332-6cd4-49ed-b269-132a6511b7a6 tempest-ServersNegativeTestMultiTenantJSON-1925445199 tempest-ServersNegativeTestMultiTenantJSON-1925445199-project-member] [instance: 7fdc6303-2d53-4d87-9aea-bf40e476f790] Checking state {{(pid=63515) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1076.482545] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23cb4384-a6be-4121-b096-907394efa5d4 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.587115] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-289af096-0c11-4506-9d23-321398bb5174 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.593109] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e681779-731e-4c8f-b7ee-0e9c9fa227fd {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.639129] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7c98436-5c34-43bd-9d3a-1a03ca515dbe {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.649840] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b320fd22-a1df-4631-a2c6-89918689e7b7 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.670322] env[63515]: DEBUG nova.compute.provider_tree [None req-fe382c79-9bcd-4ff6-a158-630c3ea4209f tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1076.686417] env[63515]: DEBUG oslo_concurrency.lockutils [None req-6ec44eb5-057e-4cc7-8109-11c3373d8aa3 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Lock "d6c5a9f9-5e5c-4b14-9b9a-7825bbf526a5" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.265s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1076.999613] env[63515]: INFO nova.compute.manager [None req-9dbc5332-6cd4-49ed-b269-132a6511b7a6 tempest-ServersNegativeTestMultiTenantJSON-1925445199 tempest-ServersNegativeTestMultiTenantJSON-1925445199-project-member] [instance: 7fdc6303-2d53-4d87-9aea-bf40e476f790] Took 13.66 seconds to build instance. [ 1077.105380] env[63515]: DEBUG oslo_concurrency.lockutils [None req-680a9f3c-a655-45e5-b8db-642d274b2d94 tempest-InstanceActionsV221TestJSON-505260833 tempest-InstanceActionsV221TestJSON-505260833-project-member] Acquiring lock "6336f10d-97b3-4413-a828-823a27492215" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1077.105900] env[63515]: DEBUG oslo_concurrency.lockutils [None req-680a9f3c-a655-45e5-b8db-642d274b2d94 tempest-InstanceActionsV221TestJSON-505260833 tempest-InstanceActionsV221TestJSON-505260833-project-member] Lock "6336f10d-97b3-4413-a828-823a27492215" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1077.172873] env[63515]: DEBUG nova.scheduler.client.report [None req-fe382c79-9bcd-4ff6-a158-630c3ea4209f tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1077.502850] env[63515]: DEBUG oslo_concurrency.lockutils [None req-9dbc5332-6cd4-49ed-b269-132a6511b7a6 tempest-ServersNegativeTestMultiTenantJSON-1925445199 tempest-ServersNegativeTestMultiTenantJSON-1925445199-project-member] Lock "7fdc6303-2d53-4d87-9aea-bf40e476f790" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.167s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1077.610423] env[63515]: DEBUG nova.compute.manager [None req-680a9f3c-a655-45e5-b8db-642d274b2d94 tempest-InstanceActionsV221TestJSON-505260833 tempest-InstanceActionsV221TestJSON-505260833-project-member] [instance: 6336f10d-97b3-4413-a828-823a27492215] Starting instance... {{(pid=63515) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1077.685024] env[63515]: DEBUG oslo_concurrency.lockutils [None req-fe382c79-9bcd-4ff6-a158-630c3ea4209f tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.708s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1077.685024] env[63515]: DEBUG oslo_concurrency.lockutils [None req-3aff29fa-b372-46b8-bb63-99f0a302c6e3 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Acquiring lock "d6c5a9f9-5e5c-4b14-9b9a-7825bbf526a5" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1077.685024] env[63515]: DEBUG oslo_concurrency.lockutils [None req-3aff29fa-b372-46b8-bb63-99f0a302c6e3 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Lock "d6c5a9f9-5e5c-4b14-9b9a-7825bbf526a5" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1077.686881] env[63515]: DEBUG oslo_concurrency.lockutils [None req-3aff29fa-b372-46b8-bb63-99f0a302c6e3 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Acquiring lock "d6c5a9f9-5e5c-4b14-9b9a-7825bbf526a5-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1077.687307] env[63515]: DEBUG oslo_concurrency.lockutils [None req-3aff29fa-b372-46b8-bb63-99f0a302c6e3 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Lock "d6c5a9f9-5e5c-4b14-9b9a-7825bbf526a5-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1077.688689] env[63515]: DEBUG oslo_concurrency.lockutils [None req-3aff29fa-b372-46b8-bb63-99f0a302c6e3 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Lock "d6c5a9f9-5e5c-4b14-9b9a-7825bbf526a5-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.001s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1077.692456] env[63515]: INFO nova.compute.manager [None req-3aff29fa-b372-46b8-bb63-99f0a302c6e3 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: d6c5a9f9-5e5c-4b14-9b9a-7825bbf526a5] Terminating instance [ 1077.697089] env[63515]: DEBUG nova.compute.manager [None req-3aff29fa-b372-46b8-bb63-99f0a302c6e3 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: d6c5a9f9-5e5c-4b14-9b9a-7825bbf526a5] Start destroying the instance on the hypervisor. {{(pid=63515) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1077.700283] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-3aff29fa-b372-46b8-bb63-99f0a302c6e3 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: d6c5a9f9-5e5c-4b14-9b9a-7825bbf526a5] Destroying instance {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1077.700283] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-160f4a40-54f2-4ba1-9cce-37469ad85797 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.708033] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-3aff29fa-b372-46b8-bb63-99f0a302c6e3 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: d6c5a9f9-5e5c-4b14-9b9a-7825bbf526a5] Powering off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1077.708311] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e2a8735a-8704-4770-b5a6-fed0d68458d9 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.712018] env[63515]: INFO nova.scheduler.client.report [None req-fe382c79-9bcd-4ff6-a158-630c3ea4209f tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Deleted allocations for instance 7d572ac0-f6af-4622-96cc-e75983420222 [ 1077.719025] env[63515]: DEBUG oslo_vmware.api [None req-3aff29fa-b372-46b8-bb63-99f0a302c6e3 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Waiting for the task: (returnval){ [ 1077.719025] env[63515]: value = "task-1111920" [ 1077.719025] env[63515]: _type = "Task" [ 1077.719025] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1077.728523] env[63515]: DEBUG oslo_vmware.api [None req-3aff29fa-b372-46b8-bb63-99f0a302c6e3 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Task: {'id': task-1111920, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1078.120291] env[63515]: DEBUG oslo_concurrency.lockutils [None req-fb5a95f4-4d37-484f-9d77-185e6f446865 tempest-ServersNegativeTestMultiTenantJSON-1925445199 tempest-ServersNegativeTestMultiTenantJSON-1925445199-project-member] Acquiring lock "7fdc6303-2d53-4d87-9aea-bf40e476f790" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1078.120291] env[63515]: DEBUG oslo_concurrency.lockutils [None req-fb5a95f4-4d37-484f-9d77-185e6f446865 tempest-ServersNegativeTestMultiTenantJSON-1925445199 tempest-ServersNegativeTestMultiTenantJSON-1925445199-project-member] Lock "7fdc6303-2d53-4d87-9aea-bf40e476f790" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1078.120291] env[63515]: DEBUG oslo_concurrency.lockutils [None req-fb5a95f4-4d37-484f-9d77-185e6f446865 tempest-ServersNegativeTestMultiTenantJSON-1925445199 tempest-ServersNegativeTestMultiTenantJSON-1925445199-project-member] Acquiring lock "7fdc6303-2d53-4d87-9aea-bf40e476f790-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1078.120291] env[63515]: DEBUG oslo_concurrency.lockutils [None req-fb5a95f4-4d37-484f-9d77-185e6f446865 tempest-ServersNegativeTestMultiTenantJSON-1925445199 tempest-ServersNegativeTestMultiTenantJSON-1925445199-project-member] Lock "7fdc6303-2d53-4d87-9aea-bf40e476f790-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1078.120291] env[63515]: DEBUG oslo_concurrency.lockutils [None req-fb5a95f4-4d37-484f-9d77-185e6f446865 tempest-ServersNegativeTestMultiTenantJSON-1925445199 tempest-ServersNegativeTestMultiTenantJSON-1925445199-project-member] Lock "7fdc6303-2d53-4d87-9aea-bf40e476f790-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1078.121362] env[63515]: INFO nova.compute.manager [None req-fb5a95f4-4d37-484f-9d77-185e6f446865 tempest-ServersNegativeTestMultiTenantJSON-1925445199 tempest-ServersNegativeTestMultiTenantJSON-1925445199-project-member] [instance: 7fdc6303-2d53-4d87-9aea-bf40e476f790] Terminating instance [ 1078.126860] env[63515]: DEBUG nova.compute.manager [None req-fb5a95f4-4d37-484f-9d77-185e6f446865 tempest-ServersNegativeTestMultiTenantJSON-1925445199 tempest-ServersNegativeTestMultiTenantJSON-1925445199-project-member] [instance: 7fdc6303-2d53-4d87-9aea-bf40e476f790] Start destroying the instance on the hypervisor. {{(pid=63515) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1078.127100] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-fb5a95f4-4d37-484f-9d77-185e6f446865 tempest-ServersNegativeTestMultiTenantJSON-1925445199 tempest-ServersNegativeTestMultiTenantJSON-1925445199-project-member] [instance: 7fdc6303-2d53-4d87-9aea-bf40e476f790] Destroying instance {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1078.128016] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-680710de-aa73-4a42-bdae-e6c7e37d86a9 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.131698] env[63515]: DEBUG oslo_concurrency.lockutils [None req-680a9f3c-a655-45e5-b8db-642d274b2d94 tempest-InstanceActionsV221TestJSON-505260833 tempest-InstanceActionsV221TestJSON-505260833-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1078.131929] env[63515]: DEBUG oslo_concurrency.lockutils [None req-680a9f3c-a655-45e5-b8db-642d274b2d94 tempest-InstanceActionsV221TestJSON-505260833 tempest-InstanceActionsV221TestJSON-505260833-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1078.133594] env[63515]: INFO nova.compute.claims [None req-680a9f3c-a655-45e5-b8db-642d274b2d94 tempest-InstanceActionsV221TestJSON-505260833 tempest-InstanceActionsV221TestJSON-505260833-project-member] [instance: 6336f10d-97b3-4413-a828-823a27492215] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1078.145529] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-fb5a95f4-4d37-484f-9d77-185e6f446865 tempest-ServersNegativeTestMultiTenantJSON-1925445199 tempest-ServersNegativeTestMultiTenantJSON-1925445199-project-member] [instance: 7fdc6303-2d53-4d87-9aea-bf40e476f790] Powering off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1078.145915] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ec3eb0f0-3fd0-4882-81d3-af5c468ce246 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.154170] env[63515]: DEBUG oslo_vmware.api [None req-fb5a95f4-4d37-484f-9d77-185e6f446865 tempest-ServersNegativeTestMultiTenantJSON-1925445199 tempest-ServersNegativeTestMultiTenantJSON-1925445199-project-member] Waiting for the task: (returnval){ [ 1078.154170] env[63515]: value = "task-1111921" [ 1078.154170] env[63515]: _type = "Task" [ 1078.154170] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1078.165222] env[63515]: DEBUG oslo_vmware.api [None req-fb5a95f4-4d37-484f-9d77-185e6f446865 tempest-ServersNegativeTestMultiTenantJSON-1925445199 tempest-ServersNegativeTestMultiTenantJSON-1925445199-project-member] Task: {'id': task-1111921, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1078.221553] env[63515]: DEBUG oslo_concurrency.lockutils [None req-fe382c79-9bcd-4ff6-a158-630c3ea4209f tempest-AttachVolumeShelveTestJSON-1420457407 tempest-AttachVolumeShelveTestJSON-1420457407-project-member] Lock "7d572ac0-f6af-4622-96cc-e75983420222" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.347s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1078.229122] env[63515]: DEBUG oslo_vmware.api [None req-3aff29fa-b372-46b8-bb63-99f0a302c6e3 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Task: {'id': task-1111920, 'name': PowerOffVM_Task, 'duration_secs': 0.240441} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1078.229421] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-3aff29fa-b372-46b8-bb63-99f0a302c6e3 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: d6c5a9f9-5e5c-4b14-9b9a-7825bbf526a5] Powered off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1078.229614] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-3aff29fa-b372-46b8-bb63-99f0a302c6e3 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: d6c5a9f9-5e5c-4b14-9b9a-7825bbf526a5] Unregistering the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1078.230360] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-033078c4-297f-466e-b5e5-871457fe8cbb {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.303830] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-3aff29fa-b372-46b8-bb63-99f0a302c6e3 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: d6c5a9f9-5e5c-4b14-9b9a-7825bbf526a5] Unregistered the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1078.304152] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-3aff29fa-b372-46b8-bb63-99f0a302c6e3 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: d6c5a9f9-5e5c-4b14-9b9a-7825bbf526a5] Deleting contents of the VM from datastore datastore2 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1078.304405] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-3aff29fa-b372-46b8-bb63-99f0a302c6e3 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Deleting the datastore file [datastore2] d6c5a9f9-5e5c-4b14-9b9a-7825bbf526a5 {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1078.304709] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ef7b9346-cd02-4f3b-ae0c-838d9fdc7966 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.313871] env[63515]: DEBUG oslo_vmware.api [None req-3aff29fa-b372-46b8-bb63-99f0a302c6e3 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Waiting for the task: (returnval){ [ 1078.313871] env[63515]: value = "task-1111923" [ 1078.313871] env[63515]: _type = "Task" [ 1078.313871] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1078.321578] env[63515]: DEBUG oslo_vmware.api [None req-3aff29fa-b372-46b8-bb63-99f0a302c6e3 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Task: {'id': task-1111923, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1078.664882] env[63515]: DEBUG oslo_vmware.api [None req-fb5a95f4-4d37-484f-9d77-185e6f446865 tempest-ServersNegativeTestMultiTenantJSON-1925445199 tempest-ServersNegativeTestMultiTenantJSON-1925445199-project-member] Task: {'id': task-1111921, 'name': PowerOffVM_Task, 'duration_secs': 0.165434} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1078.665201] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-fb5a95f4-4d37-484f-9d77-185e6f446865 tempest-ServersNegativeTestMultiTenantJSON-1925445199 tempest-ServersNegativeTestMultiTenantJSON-1925445199-project-member] [instance: 7fdc6303-2d53-4d87-9aea-bf40e476f790] Powered off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1078.665353] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-fb5a95f4-4d37-484f-9d77-185e6f446865 tempest-ServersNegativeTestMultiTenantJSON-1925445199 tempest-ServersNegativeTestMultiTenantJSON-1925445199-project-member] [instance: 7fdc6303-2d53-4d87-9aea-bf40e476f790] Unregistering the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1078.665627] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d4478b6f-1b44-403a-a1c4-8732933d50fa {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.758277] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-fb5a95f4-4d37-484f-9d77-185e6f446865 tempest-ServersNegativeTestMultiTenantJSON-1925445199 tempest-ServersNegativeTestMultiTenantJSON-1925445199-project-member] [instance: 7fdc6303-2d53-4d87-9aea-bf40e476f790] Unregistered the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1078.758538] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-fb5a95f4-4d37-484f-9d77-185e6f446865 tempest-ServersNegativeTestMultiTenantJSON-1925445199 tempest-ServersNegativeTestMultiTenantJSON-1925445199-project-member] [instance: 7fdc6303-2d53-4d87-9aea-bf40e476f790] Deleting contents of the VM from datastore datastore1 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1078.758773] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-fb5a95f4-4d37-484f-9d77-185e6f446865 tempest-ServersNegativeTestMultiTenantJSON-1925445199 tempest-ServersNegativeTestMultiTenantJSON-1925445199-project-member] Deleting the datastore file [datastore1] 7fdc6303-2d53-4d87-9aea-bf40e476f790 {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1078.759098] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-96a41485-7c2a-45ac-9b9a-bfb2e199f53b {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.764868] env[63515]: DEBUG oslo_vmware.api [None req-fb5a95f4-4d37-484f-9d77-185e6f446865 tempest-ServersNegativeTestMultiTenantJSON-1925445199 tempest-ServersNegativeTestMultiTenantJSON-1925445199-project-member] Waiting for the task: (returnval){ [ 1078.764868] env[63515]: value = "task-1111925" [ 1078.764868] env[63515]: _type = "Task" [ 1078.764868] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1078.772999] env[63515]: DEBUG oslo_vmware.api [None req-fb5a95f4-4d37-484f-9d77-185e6f446865 tempest-ServersNegativeTestMultiTenantJSON-1925445199 tempest-ServersNegativeTestMultiTenantJSON-1925445199-project-member] Task: {'id': task-1111925, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1078.823504] env[63515]: DEBUG oslo_vmware.api [None req-3aff29fa-b372-46b8-bb63-99f0a302c6e3 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Task: {'id': task-1111923, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.16448} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1078.824292] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-3aff29fa-b372-46b8-bb63-99f0a302c6e3 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Deleted the datastore file {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1078.824292] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-3aff29fa-b372-46b8-bb63-99f0a302c6e3 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: d6c5a9f9-5e5c-4b14-9b9a-7825bbf526a5] Deleted contents of the VM from datastore datastore2 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1078.824433] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-3aff29fa-b372-46b8-bb63-99f0a302c6e3 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: d6c5a9f9-5e5c-4b14-9b9a-7825bbf526a5] Instance destroyed {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1078.824625] env[63515]: INFO nova.compute.manager [None req-3aff29fa-b372-46b8-bb63-99f0a302c6e3 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: d6c5a9f9-5e5c-4b14-9b9a-7825bbf526a5] Took 1.13 seconds to destroy the instance on the hypervisor. [ 1078.824852] env[63515]: DEBUG oslo.service.loopingcall [None req-3aff29fa-b372-46b8-bb63-99f0a302c6e3 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1078.825053] env[63515]: DEBUG nova.compute.manager [-] [instance: d6c5a9f9-5e5c-4b14-9b9a-7825bbf526a5] Deallocating network for instance {{(pid=63515) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1078.825154] env[63515]: DEBUG nova.network.neutron [-] [instance: d6c5a9f9-5e5c-4b14-9b9a-7825bbf526a5] deallocate_for_instance() {{(pid=63515) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1079.226113] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0519712-5908-45ec-af9c-c795b363d85c {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.233552] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-200575ae-09cf-4d9d-adfc-53be50e0b6b2 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.264329] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65a34e8e-423c-4198-b7e3-da8968e025e1 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.268892] env[63515]: DEBUG nova.compute.manager [req-0986f9ba-fad4-46c6-96a4-7f28b7bec2a5 req-36c92dbe-d32b-49d0-94e4-6d7fcaa17e94 service nova] [instance: d6c5a9f9-5e5c-4b14-9b9a-7825bbf526a5] Received event network-vif-deleted-ee82711d-2d77-4127-b804-7db65e838617 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1079.269093] env[63515]: INFO nova.compute.manager [req-0986f9ba-fad4-46c6-96a4-7f28b7bec2a5 req-36c92dbe-d32b-49d0-94e4-6d7fcaa17e94 service nova] [instance: d6c5a9f9-5e5c-4b14-9b9a-7825bbf526a5] Neutron deleted interface ee82711d-2d77-4127-b804-7db65e838617; detaching it from the instance and deleting it from the info cache [ 1079.269271] env[63515]: DEBUG nova.network.neutron [req-0986f9ba-fad4-46c6-96a4-7f28b7bec2a5 req-36c92dbe-d32b-49d0-94e4-6d7fcaa17e94 service nova] [instance: d6c5a9f9-5e5c-4b14-9b9a-7825bbf526a5] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1079.280566] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c93ea89-1820-4353-ac44-ae43dd4ab227 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.284658] env[63515]: DEBUG oslo_vmware.api [None req-fb5a95f4-4d37-484f-9d77-185e6f446865 tempest-ServersNegativeTestMultiTenantJSON-1925445199 tempest-ServersNegativeTestMultiTenantJSON-1925445199-project-member] Task: {'id': task-1111925, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.127467} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1079.285120] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-fb5a95f4-4d37-484f-9d77-185e6f446865 tempest-ServersNegativeTestMultiTenantJSON-1925445199 tempest-ServersNegativeTestMultiTenantJSON-1925445199-project-member] Deleted the datastore file {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1079.285317] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-fb5a95f4-4d37-484f-9d77-185e6f446865 tempest-ServersNegativeTestMultiTenantJSON-1925445199 tempest-ServersNegativeTestMultiTenantJSON-1925445199-project-member] [instance: 7fdc6303-2d53-4d87-9aea-bf40e476f790] Deleted contents of the VM from datastore datastore1 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1079.285500] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-fb5a95f4-4d37-484f-9d77-185e6f446865 tempest-ServersNegativeTestMultiTenantJSON-1925445199 tempest-ServersNegativeTestMultiTenantJSON-1925445199-project-member] [instance: 7fdc6303-2d53-4d87-9aea-bf40e476f790] Instance destroyed {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1079.285673] env[63515]: INFO nova.compute.manager [None req-fb5a95f4-4d37-484f-9d77-185e6f446865 tempest-ServersNegativeTestMultiTenantJSON-1925445199 tempest-ServersNegativeTestMultiTenantJSON-1925445199-project-member] [instance: 7fdc6303-2d53-4d87-9aea-bf40e476f790] Took 1.16 seconds to destroy the instance on the hypervisor. [ 1079.285940] env[63515]: DEBUG oslo.service.loopingcall [None req-fb5a95f4-4d37-484f-9d77-185e6f446865 tempest-ServersNegativeTestMultiTenantJSON-1925445199 tempest-ServersNegativeTestMultiTenantJSON-1925445199-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1079.286620] env[63515]: DEBUG nova.compute.manager [-] [instance: 7fdc6303-2d53-4d87-9aea-bf40e476f790] Deallocating network for instance {{(pid=63515) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1079.286732] env[63515]: DEBUG nova.network.neutron [-] [instance: 7fdc6303-2d53-4d87-9aea-bf40e476f790] deallocate_for_instance() {{(pid=63515) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1079.297509] env[63515]: DEBUG nova.compute.provider_tree [None req-680a9f3c-a655-45e5-b8db-642d274b2d94 tempest-InstanceActionsV221TestJSON-505260833 tempest-InstanceActionsV221TestJSON-505260833-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1079.763751] env[63515]: DEBUG nova.network.neutron [-] [instance: d6c5a9f9-5e5c-4b14-9b9a-7825bbf526a5] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1079.775342] env[63515]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-2b03f613-1e66-4326-976b-3c5c6ab3853d {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.784654] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c6b0eb1-a91d-4b16-bca4-0da6cf7e0137 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.800078] env[63515]: DEBUG nova.scheduler.client.report [None req-680a9f3c-a655-45e5-b8db-642d274b2d94 tempest-InstanceActionsV221TestJSON-505260833 tempest-InstanceActionsV221TestJSON-505260833-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1079.808583] env[63515]: DEBUG nova.compute.manager [req-0986f9ba-fad4-46c6-96a4-7f28b7bec2a5 req-36c92dbe-d32b-49d0-94e4-6d7fcaa17e94 service nova] [instance: d6c5a9f9-5e5c-4b14-9b9a-7825bbf526a5] Detach interface failed, port_id=ee82711d-2d77-4127-b804-7db65e838617, reason: Instance d6c5a9f9-5e5c-4b14-9b9a-7825bbf526a5 could not be found. {{(pid=63515) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1080.017636] env[63515]: DEBUG nova.network.neutron [-] [instance: 7fdc6303-2d53-4d87-9aea-bf40e476f790] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1080.267127] env[63515]: INFO nova.compute.manager [-] [instance: d6c5a9f9-5e5c-4b14-9b9a-7825bbf526a5] Took 1.44 seconds to deallocate network for instance. [ 1080.303776] env[63515]: DEBUG oslo_concurrency.lockutils [None req-680a9f3c-a655-45e5-b8db-642d274b2d94 tempest-InstanceActionsV221TestJSON-505260833 tempest-InstanceActionsV221TestJSON-505260833-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.172s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1080.304308] env[63515]: DEBUG nova.compute.manager [None req-680a9f3c-a655-45e5-b8db-642d274b2d94 tempest-InstanceActionsV221TestJSON-505260833 tempest-InstanceActionsV221TestJSON-505260833-project-member] [instance: 6336f10d-97b3-4413-a828-823a27492215] Start building networks asynchronously for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1080.520984] env[63515]: INFO nova.compute.manager [-] [instance: 7fdc6303-2d53-4d87-9aea-bf40e476f790] Took 1.23 seconds to deallocate network for instance. [ 1080.773452] env[63515]: DEBUG oslo_concurrency.lockutils [None req-3aff29fa-b372-46b8-bb63-99f0a302c6e3 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1080.773750] env[63515]: DEBUG oslo_concurrency.lockutils [None req-3aff29fa-b372-46b8-bb63-99f0a302c6e3 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1080.773956] env[63515]: DEBUG nova.objects.instance [None req-3aff29fa-b372-46b8-bb63-99f0a302c6e3 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Lazy-loading 'resources' on Instance uuid d6c5a9f9-5e5c-4b14-9b9a-7825bbf526a5 {{(pid=63515) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1080.808564] env[63515]: DEBUG nova.compute.utils [None req-680a9f3c-a655-45e5-b8db-642d274b2d94 tempest-InstanceActionsV221TestJSON-505260833 tempest-InstanceActionsV221TestJSON-505260833-project-member] Using /dev/sd instead of None {{(pid=63515) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1080.810010] env[63515]: DEBUG nova.compute.manager [None req-680a9f3c-a655-45e5-b8db-642d274b2d94 tempest-InstanceActionsV221TestJSON-505260833 tempest-InstanceActionsV221TestJSON-505260833-project-member] [instance: 6336f10d-97b3-4413-a828-823a27492215] Allocating IP information in the background. {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1080.811024] env[63515]: DEBUG nova.network.neutron [None req-680a9f3c-a655-45e5-b8db-642d274b2d94 tempest-InstanceActionsV221TestJSON-505260833 tempest-InstanceActionsV221TestJSON-505260833-project-member] [instance: 6336f10d-97b3-4413-a828-823a27492215] allocate_for_instance() {{(pid=63515) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1080.860364] env[63515]: DEBUG nova.policy [None req-680a9f3c-a655-45e5-b8db-642d274b2d94 tempest-InstanceActionsV221TestJSON-505260833 tempest-InstanceActionsV221TestJSON-505260833-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9b3b41b107da43a9ac5c47adb001beb8', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8350a1ba736243b39e35024c2de58b9b', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63515) authorize /opt/stack/nova/nova/policy.py:201}} [ 1081.027778] env[63515]: DEBUG oslo_concurrency.lockutils [None req-fb5a95f4-4d37-484f-9d77-185e6f446865 tempest-ServersNegativeTestMultiTenantJSON-1925445199 tempest-ServersNegativeTestMultiTenantJSON-1925445199-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1081.040344] env[63515]: DEBUG oslo_service.periodic_task [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=63515) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1081.145646] env[63515]: DEBUG nova.network.neutron [None req-680a9f3c-a655-45e5-b8db-642d274b2d94 tempest-InstanceActionsV221TestJSON-505260833 tempest-InstanceActionsV221TestJSON-505260833-project-member] [instance: 6336f10d-97b3-4413-a828-823a27492215] Successfully created port: 635a24b0-e5a3-43bf-989f-b7bdbc07f767 {{(pid=63515) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1081.295604] env[63515]: DEBUG nova.compute.manager [req-ce6b0094-c44e-4d40-86c6-b5f100399686 req-e253a40f-39f8-4149-8c5c-5c12767caffb service nova] [instance: 7fdc6303-2d53-4d87-9aea-bf40e476f790] Received event network-vif-deleted-0a17708c-c9b4-4715-bcfa-f2629d5b179f {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1081.314673] env[63515]: DEBUG nova.compute.manager [None req-680a9f3c-a655-45e5-b8db-642d274b2d94 tempest-InstanceActionsV221TestJSON-505260833 tempest-InstanceActionsV221TestJSON-505260833-project-member] [instance: 6336f10d-97b3-4413-a828-823a27492215] Start building block device mappings for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1081.364812] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5fd2e142-2d0e-451c-a161-2c4610675ebd {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.373281] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bcaf8310-cc1f-4d79-a431-a482082d4072 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.403722] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58855b3a-6692-43ce-b52e-cefbd476b38a {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.410973] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3050c9e4-6da1-4a4c-bc30-d5832ad876fa {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.424183] env[63515]: DEBUG nova.compute.provider_tree [None req-3aff29fa-b372-46b8-bb63-99f0a302c6e3 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1081.545047] env[63515]: DEBUG oslo_service.periodic_task [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=63515) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1081.545047] env[63515]: DEBUG nova.compute.manager [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Starting heal instance info cache {{(pid=63515) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9926}} [ 1081.926574] env[63515]: DEBUG nova.scheduler.client.report [None req-3aff29fa-b372-46b8-bb63-99f0a302c6e3 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1082.286077] env[63515]: DEBUG oslo_concurrency.lockutils [None req-c053fe74-7df3-47e0-afc9-192490b04524 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Acquiring lock "aa83d035-9c47-434d-ab1a-140e6b3110b4" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1082.286338] env[63515]: DEBUG oslo_concurrency.lockutils [None req-c053fe74-7df3-47e0-afc9-192490b04524 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Lock "aa83d035-9c47-434d-ab1a-140e6b3110b4" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1082.323144] env[63515]: DEBUG nova.compute.manager [None req-680a9f3c-a655-45e5-b8db-642d274b2d94 tempest-InstanceActionsV221TestJSON-505260833 tempest-InstanceActionsV221TestJSON-505260833-project-member] [instance: 6336f10d-97b3-4413-a828-823a27492215] Start spawning the instance on the hypervisor. {{(pid=63515) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1082.347944] env[63515]: DEBUG nova.virt.hardware [None req-680a9f3c-a655-45e5-b8db-642d274b2d94 tempest-InstanceActionsV221TestJSON-505260833 tempest-InstanceActionsV221TestJSON-505260833-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T02:52:14Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T02:51:56Z,direct_url=,disk_format='vmdk',id=8a120570-cb06-4099-b262-554ca0ad15c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b569255cc43e42e7ae3f2b4ad37c6ef4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T02:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1082.348239] env[63515]: DEBUG nova.virt.hardware [None req-680a9f3c-a655-45e5-b8db-642d274b2d94 tempest-InstanceActionsV221TestJSON-505260833 tempest-InstanceActionsV221TestJSON-505260833-project-member] Flavor limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1082.348401] env[63515]: DEBUG nova.virt.hardware [None req-680a9f3c-a655-45e5-b8db-642d274b2d94 tempest-InstanceActionsV221TestJSON-505260833 tempest-InstanceActionsV221TestJSON-505260833-project-member] Image limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1082.348590] env[63515]: DEBUG nova.virt.hardware [None req-680a9f3c-a655-45e5-b8db-642d274b2d94 tempest-InstanceActionsV221TestJSON-505260833 tempest-InstanceActionsV221TestJSON-505260833-project-member] Flavor pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1082.348779] env[63515]: DEBUG nova.virt.hardware [None req-680a9f3c-a655-45e5-b8db-642d274b2d94 tempest-InstanceActionsV221TestJSON-505260833 tempest-InstanceActionsV221TestJSON-505260833-project-member] Image pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1082.348917] env[63515]: DEBUG nova.virt.hardware [None req-680a9f3c-a655-45e5-b8db-642d274b2d94 tempest-InstanceActionsV221TestJSON-505260833 tempest-InstanceActionsV221TestJSON-505260833-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1082.349175] env[63515]: DEBUG nova.virt.hardware [None req-680a9f3c-a655-45e5-b8db-642d274b2d94 tempest-InstanceActionsV221TestJSON-505260833 tempest-InstanceActionsV221TestJSON-505260833-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1082.349348] env[63515]: DEBUG nova.virt.hardware [None req-680a9f3c-a655-45e5-b8db-642d274b2d94 tempest-InstanceActionsV221TestJSON-505260833 tempest-InstanceActionsV221TestJSON-505260833-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1082.349522] env[63515]: DEBUG nova.virt.hardware [None req-680a9f3c-a655-45e5-b8db-642d274b2d94 tempest-InstanceActionsV221TestJSON-505260833 tempest-InstanceActionsV221TestJSON-505260833-project-member] Got 1 possible topologies {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1082.349689] env[63515]: DEBUG nova.virt.hardware [None req-680a9f3c-a655-45e5-b8db-642d274b2d94 tempest-InstanceActionsV221TestJSON-505260833 tempest-InstanceActionsV221TestJSON-505260833-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1082.349866] env[63515]: DEBUG nova.virt.hardware [None req-680a9f3c-a655-45e5-b8db-642d274b2d94 tempest-InstanceActionsV221TestJSON-505260833 tempest-InstanceActionsV221TestJSON-505260833-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1082.350767] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d877394-2f91-4540-90af-90e503c89d54 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.359314] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-482f6cbd-ad37-4c83-a532-5949ac40cf89 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.431561] env[63515]: DEBUG oslo_concurrency.lockutils [None req-3aff29fa-b372-46b8-bb63-99f0a302c6e3 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.658s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1082.433752] env[63515]: DEBUG oslo_concurrency.lockutils [None req-fb5a95f4-4d37-484f-9d77-185e6f446865 tempest-ServersNegativeTestMultiTenantJSON-1925445199 tempest-ServersNegativeTestMultiTenantJSON-1925445199-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.406s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1082.433996] env[63515]: DEBUG nova.objects.instance [None req-fb5a95f4-4d37-484f-9d77-185e6f446865 tempest-ServersNegativeTestMultiTenantJSON-1925445199 tempest-ServersNegativeTestMultiTenantJSON-1925445199-project-member] Lazy-loading 'resources' on Instance uuid 7fdc6303-2d53-4d87-9aea-bf40e476f790 {{(pid=63515) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1082.453963] env[63515]: INFO nova.scheduler.client.report [None req-3aff29fa-b372-46b8-bb63-99f0a302c6e3 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Deleted allocations for instance d6c5a9f9-5e5c-4b14-9b9a-7825bbf526a5 [ 1082.639243] env[63515]: DEBUG nova.network.neutron [None req-680a9f3c-a655-45e5-b8db-642d274b2d94 tempest-InstanceActionsV221TestJSON-505260833 tempest-InstanceActionsV221TestJSON-505260833-project-member] [instance: 6336f10d-97b3-4413-a828-823a27492215] Successfully updated port: 635a24b0-e5a3-43bf-989f-b7bdbc07f767 {{(pid=63515) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1082.789456] env[63515]: DEBUG nova.compute.utils [None req-c053fe74-7df3-47e0-afc9-192490b04524 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Using /dev/sd instead of None {{(pid=63515) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1082.963620] env[63515]: DEBUG oslo_concurrency.lockutils [None req-3aff29fa-b372-46b8-bb63-99f0a302c6e3 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Lock "d6c5a9f9-5e5c-4b14-9b9a-7825bbf526a5" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.278s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1083.004071] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c4d76ea-ec1d-4c6b-8527-4134a88756ec {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.009737] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9dbc43e7-eccc-4686-8470-580983119777 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.043062] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f82c21f-563a-4e22-941d-56af9a0e523e {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.052179] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df6102de-97ea-4578-b323-a9978bb4fa2f {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.069823] env[63515]: DEBUG nova.compute.provider_tree [None req-fb5a95f4-4d37-484f-9d77-185e6f446865 tempest-ServersNegativeTestMultiTenantJSON-1925445199 tempest-ServersNegativeTestMultiTenantJSON-1925445199-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1083.142045] env[63515]: DEBUG oslo_concurrency.lockutils [None req-680a9f3c-a655-45e5-b8db-642d274b2d94 tempest-InstanceActionsV221TestJSON-505260833 tempest-InstanceActionsV221TestJSON-505260833-project-member] Acquiring lock "refresh_cache-6336f10d-97b3-4413-a828-823a27492215" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1083.142304] env[63515]: DEBUG oslo_concurrency.lockutils [None req-680a9f3c-a655-45e5-b8db-642d274b2d94 tempest-InstanceActionsV221TestJSON-505260833 tempest-InstanceActionsV221TestJSON-505260833-project-member] Acquired lock "refresh_cache-6336f10d-97b3-4413-a828-823a27492215" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1083.142468] env[63515]: DEBUG nova.network.neutron [None req-680a9f3c-a655-45e5-b8db-642d274b2d94 tempest-InstanceActionsV221TestJSON-505260833 tempest-InstanceActionsV221TestJSON-505260833-project-member] [instance: 6336f10d-97b3-4413-a828-823a27492215] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1083.292850] env[63515]: DEBUG oslo_concurrency.lockutils [None req-c053fe74-7df3-47e0-afc9-192490b04524 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Lock "aa83d035-9c47-434d-ab1a-140e6b3110b4" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.005s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1083.449998] env[63515]: DEBUG nova.compute.manager [req-ddb65374-6d93-4504-b682-e872a0390007 req-320bd8c3-9d16-490a-bfd3-ae7c6b88dee5 service nova] [instance: 6336f10d-97b3-4413-a828-823a27492215] Received event network-vif-plugged-635a24b0-e5a3-43bf-989f-b7bdbc07f767 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1083.450177] env[63515]: DEBUG oslo_concurrency.lockutils [req-ddb65374-6d93-4504-b682-e872a0390007 req-320bd8c3-9d16-490a-bfd3-ae7c6b88dee5 service nova] Acquiring lock "6336f10d-97b3-4413-a828-823a27492215-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1083.450358] env[63515]: DEBUG oslo_concurrency.lockutils [req-ddb65374-6d93-4504-b682-e872a0390007 req-320bd8c3-9d16-490a-bfd3-ae7c6b88dee5 service nova] Lock "6336f10d-97b3-4413-a828-823a27492215-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1083.451043] env[63515]: DEBUG oslo_concurrency.lockutils [req-ddb65374-6d93-4504-b682-e872a0390007 req-320bd8c3-9d16-490a-bfd3-ae7c6b88dee5 service nova] Lock "6336f10d-97b3-4413-a828-823a27492215-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1083.451043] env[63515]: DEBUG nova.compute.manager [req-ddb65374-6d93-4504-b682-e872a0390007 req-320bd8c3-9d16-490a-bfd3-ae7c6b88dee5 service nova] [instance: 6336f10d-97b3-4413-a828-823a27492215] No waiting events found dispatching network-vif-plugged-635a24b0-e5a3-43bf-989f-b7bdbc07f767 {{(pid=63515) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1083.451043] env[63515]: WARNING nova.compute.manager [req-ddb65374-6d93-4504-b682-e872a0390007 req-320bd8c3-9d16-490a-bfd3-ae7c6b88dee5 service nova] [instance: 6336f10d-97b3-4413-a828-823a27492215] Received unexpected event network-vif-plugged-635a24b0-e5a3-43bf-989f-b7bdbc07f767 for instance with vm_state building and task_state spawning. [ 1083.451043] env[63515]: DEBUG nova.compute.manager [req-ddb65374-6d93-4504-b682-e872a0390007 req-320bd8c3-9d16-490a-bfd3-ae7c6b88dee5 service nova] [instance: 6336f10d-97b3-4413-a828-823a27492215] Received event network-changed-635a24b0-e5a3-43bf-989f-b7bdbc07f767 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1083.451202] env[63515]: DEBUG nova.compute.manager [req-ddb65374-6d93-4504-b682-e872a0390007 req-320bd8c3-9d16-490a-bfd3-ae7c6b88dee5 service nova] [instance: 6336f10d-97b3-4413-a828-823a27492215] Refreshing instance network info cache due to event network-changed-635a24b0-e5a3-43bf-989f-b7bdbc07f767. {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1083.451353] env[63515]: DEBUG oslo_concurrency.lockutils [req-ddb65374-6d93-4504-b682-e872a0390007 req-320bd8c3-9d16-490a-bfd3-ae7c6b88dee5 service nova] Acquiring lock "refresh_cache-6336f10d-97b3-4413-a828-823a27492215" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1083.573280] env[63515]: DEBUG nova.scheduler.client.report [None req-fb5a95f4-4d37-484f-9d77-185e6f446865 tempest-ServersNegativeTestMultiTenantJSON-1925445199 tempest-ServersNegativeTestMultiTenantJSON-1925445199-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1083.699959] env[63515]: DEBUG nova.network.neutron [None req-680a9f3c-a655-45e5-b8db-642d274b2d94 tempest-InstanceActionsV221TestJSON-505260833 tempest-InstanceActionsV221TestJSON-505260833-project-member] [instance: 6336f10d-97b3-4413-a828-823a27492215] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1083.943497] env[63515]: DEBUG nova.network.neutron [None req-680a9f3c-a655-45e5-b8db-642d274b2d94 tempest-InstanceActionsV221TestJSON-505260833 tempest-InstanceActionsV221TestJSON-505260833-project-member] [instance: 6336f10d-97b3-4413-a828-823a27492215] Updating instance_info_cache with network_info: [{"id": "635a24b0-e5a3-43bf-989f-b7bdbc07f767", "address": "fa:16:3e:5e:ff:86", "network": {"id": "be106785-bb04-4c6c-93f6-1fafe9449aa0", "bridge": "br-int", "label": "tempest-InstanceActionsV221TestJSON-675645069-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8350a1ba736243b39e35024c2de58b9b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "39a4aca0-934b-4a91-8779-6a4360c3f967", "external-id": "nsx-vlan-transportzone-454", "segmentation_id": 454, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap635a24b0-e5", "ovs_interfaceid": "635a24b0-e5a3-43bf-989f-b7bdbc07f767", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1084.078915] env[63515]: DEBUG oslo_concurrency.lockutils [None req-fb5a95f4-4d37-484f-9d77-185e6f446865 tempest-ServersNegativeTestMultiTenantJSON-1925445199 tempest-ServersNegativeTestMultiTenantJSON-1925445199-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.645s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1084.098210] env[63515]: INFO nova.scheduler.client.report [None req-fb5a95f4-4d37-484f-9d77-185e6f446865 tempest-ServersNegativeTestMultiTenantJSON-1925445199 tempest-ServersNegativeTestMultiTenantJSON-1925445199-project-member] Deleted allocations for instance 7fdc6303-2d53-4d87-9aea-bf40e476f790 [ 1084.372735] env[63515]: DEBUG oslo_concurrency.lockutils [None req-c053fe74-7df3-47e0-afc9-192490b04524 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Acquiring lock "aa83d035-9c47-434d-ab1a-140e6b3110b4" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1084.373040] env[63515]: DEBUG oslo_concurrency.lockutils [None req-c053fe74-7df3-47e0-afc9-192490b04524 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Lock "aa83d035-9c47-434d-ab1a-140e6b3110b4" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1084.373296] env[63515]: INFO nova.compute.manager [None req-c053fe74-7df3-47e0-afc9-192490b04524 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: aa83d035-9c47-434d-ab1a-140e6b3110b4] Attaching volume 81e015c1-d604-49a7-9bad-778a82466e35 to /dev/sdb [ 1084.410064] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34c51131-a73a-443a-8ef0-043257542a73 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.418329] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11abaf80-af32-442f-964e-a472cf92e3f4 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.435099] env[63515]: DEBUG nova.virt.block_device [None req-c053fe74-7df3-47e0-afc9-192490b04524 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: aa83d035-9c47-434d-ab1a-140e6b3110b4] Updating existing volume attachment record: 3048a578-15c3-440d-8b5d-189a3f7d2aa1 {{(pid=63515) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1084.445906] env[63515]: DEBUG oslo_concurrency.lockutils [None req-680a9f3c-a655-45e5-b8db-642d274b2d94 tempest-InstanceActionsV221TestJSON-505260833 tempest-InstanceActionsV221TestJSON-505260833-project-member] Releasing lock "refresh_cache-6336f10d-97b3-4413-a828-823a27492215" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1084.446257] env[63515]: DEBUG nova.compute.manager [None req-680a9f3c-a655-45e5-b8db-642d274b2d94 tempest-InstanceActionsV221TestJSON-505260833 tempest-InstanceActionsV221TestJSON-505260833-project-member] [instance: 6336f10d-97b3-4413-a828-823a27492215] Instance network_info: |[{"id": "635a24b0-e5a3-43bf-989f-b7bdbc07f767", "address": "fa:16:3e:5e:ff:86", "network": {"id": "be106785-bb04-4c6c-93f6-1fafe9449aa0", "bridge": "br-int", "label": "tempest-InstanceActionsV221TestJSON-675645069-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8350a1ba736243b39e35024c2de58b9b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "39a4aca0-934b-4a91-8779-6a4360c3f967", "external-id": "nsx-vlan-transportzone-454", "segmentation_id": 454, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap635a24b0-e5", "ovs_interfaceid": "635a24b0-e5a3-43bf-989f-b7bdbc07f767", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1084.446582] env[63515]: DEBUG oslo_concurrency.lockutils [req-ddb65374-6d93-4504-b682-e872a0390007 req-320bd8c3-9d16-490a-bfd3-ae7c6b88dee5 service nova] Acquired lock "refresh_cache-6336f10d-97b3-4413-a828-823a27492215" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1084.446767] env[63515]: DEBUG nova.network.neutron [req-ddb65374-6d93-4504-b682-e872a0390007 req-320bd8c3-9d16-490a-bfd3-ae7c6b88dee5 service nova] [instance: 6336f10d-97b3-4413-a828-823a27492215] Refreshing network info cache for port 635a24b0-e5a3-43bf-989f-b7bdbc07f767 {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1084.448038] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-680a9f3c-a655-45e5-b8db-642d274b2d94 tempest-InstanceActionsV221TestJSON-505260833 tempest-InstanceActionsV221TestJSON-505260833-project-member] [instance: 6336f10d-97b3-4413-a828-823a27492215] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:5e:ff:86', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '39a4aca0-934b-4a91-8779-6a4360c3f967', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '635a24b0-e5a3-43bf-989f-b7bdbc07f767', 'vif_model': 'vmxnet3'}] {{(pid=63515) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1084.456083] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-680a9f3c-a655-45e5-b8db-642d274b2d94 tempest-InstanceActionsV221TestJSON-505260833 tempest-InstanceActionsV221TestJSON-505260833-project-member] Creating folder: Project (8350a1ba736243b39e35024c2de58b9b). Parent ref: group-v243370. {{(pid=63515) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1084.456718] env[63515]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e9324d83-3379-4ca1-8f12-e5adbe250e15 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.468010] env[63515]: INFO nova.virt.vmwareapi.vm_util [None req-680a9f3c-a655-45e5-b8db-642d274b2d94 tempest-InstanceActionsV221TestJSON-505260833 tempest-InstanceActionsV221TestJSON-505260833-project-member] Created folder: Project (8350a1ba736243b39e35024c2de58b9b) in parent group-v243370. [ 1084.468241] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-680a9f3c-a655-45e5-b8db-642d274b2d94 tempest-InstanceActionsV221TestJSON-505260833 tempest-InstanceActionsV221TestJSON-505260833-project-member] Creating folder: Instances. Parent ref: group-v243574. {{(pid=63515) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1084.468722] env[63515]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-6d28659f-9c33-48d0-9ff0-40b40b2ff581 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.482695] env[63515]: INFO nova.virt.vmwareapi.vm_util [None req-680a9f3c-a655-45e5-b8db-642d274b2d94 tempest-InstanceActionsV221TestJSON-505260833 tempest-InstanceActionsV221TestJSON-505260833-project-member] Created folder: Instances in parent group-v243574. [ 1084.482954] env[63515]: DEBUG oslo.service.loopingcall [None req-680a9f3c-a655-45e5-b8db-642d274b2d94 tempest-InstanceActionsV221TestJSON-505260833 tempest-InstanceActionsV221TestJSON-505260833-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1084.483173] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6336f10d-97b3-4413-a828-823a27492215] Creating VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1084.483381] env[63515]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-424506c4-2951-496a-bc8a-b8921af8096a {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.506009] env[63515]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1084.506009] env[63515]: value = "task-1111929" [ 1084.506009] env[63515]: _type = "Task" [ 1084.506009] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1084.513653] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1111929, 'name': CreateVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1084.572089] env[63515]: DEBUG nova.compute.manager [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Didn't find any instances for network info cache update. {{(pid=63515) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10012}} [ 1084.572089] env[63515]: DEBUG oslo_service.periodic_task [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=63515) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1084.572089] env[63515]: DEBUG oslo_service.periodic_task [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=63515) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1084.572089] env[63515]: DEBUG oslo_service.periodic_task [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=63515) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1084.572089] env[63515]: DEBUG oslo_service.periodic_task [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=63515) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1084.572089] env[63515]: DEBUG oslo_service.periodic_task [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=63515) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1084.572089] env[63515]: DEBUG oslo_service.periodic_task [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=63515) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1084.572089] env[63515]: DEBUG nova.compute.manager [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=63515) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10545}} [ 1084.572089] env[63515]: DEBUG oslo_service.periodic_task [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Running periodic task ComputeManager.update_available_resource {{(pid=63515) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1084.608680] env[63515]: DEBUG oslo_concurrency.lockutils [None req-fb5a95f4-4d37-484f-9d77-185e6f446865 tempest-ServersNegativeTestMultiTenantJSON-1925445199 tempest-ServersNegativeTestMultiTenantJSON-1925445199-project-member] Lock "7fdc6303-2d53-4d87-9aea-bf40e476f790" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.490s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1084.609939] env[63515]: DEBUG oslo_concurrency.lockutils [None req-f1f3c1ea-070f-4e78-a0c6-d376174a87f2 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Acquiring lock "ea2e0810-c9f1-49a6-8589-7db402847aea" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1084.610188] env[63515]: DEBUG oslo_concurrency.lockutils [None req-f1f3c1ea-070f-4e78-a0c6-d376174a87f2 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Lock "ea2e0810-c9f1-49a6-8589-7db402847aea" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1085.022055] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1111929, 'name': CreateVM_Task, 'duration_secs': 0.396862} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1085.022253] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6336f10d-97b3-4413-a828-823a27492215] Created VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1085.022929] env[63515]: DEBUG oslo_concurrency.lockutils [None req-680a9f3c-a655-45e5-b8db-642d274b2d94 tempest-InstanceActionsV221TestJSON-505260833 tempest-InstanceActionsV221TestJSON-505260833-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1085.023132] env[63515]: DEBUG oslo_concurrency.lockutils [None req-680a9f3c-a655-45e5-b8db-642d274b2d94 tempest-InstanceActionsV221TestJSON-505260833 tempest-InstanceActionsV221TestJSON-505260833-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1085.023460] env[63515]: DEBUG oslo_concurrency.lockutils [None req-680a9f3c-a655-45e5-b8db-642d274b2d94 tempest-InstanceActionsV221TestJSON-505260833 tempest-InstanceActionsV221TestJSON-505260833-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1085.023718] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e3d898fc-1203-4682-a911-1231da422d97 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.028440] env[63515]: DEBUG oslo_vmware.api [None req-680a9f3c-a655-45e5-b8db-642d274b2d94 tempest-InstanceActionsV221TestJSON-505260833 tempest-InstanceActionsV221TestJSON-505260833-project-member] Waiting for the task: (returnval){ [ 1085.028440] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]52a788eb-10a5-91b9-6cad-c246b7f947ac" [ 1085.028440] env[63515]: _type = "Task" [ 1085.028440] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1085.038654] env[63515]: DEBUG oslo_vmware.api [None req-680a9f3c-a655-45e5-b8db-642d274b2d94 tempest-InstanceActionsV221TestJSON-505260833 tempest-InstanceActionsV221TestJSON-505260833-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52a788eb-10a5-91b9-6cad-c246b7f947ac, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1085.074497] env[63515]: DEBUG oslo_concurrency.lockutils [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1085.074765] env[63515]: DEBUG oslo_concurrency.lockutils [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1085.074945] env[63515]: DEBUG oslo_concurrency.lockutils [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1085.075115] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=63515) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1085.076291] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50a5a472-7aff-415a-87e0-06d5e15e0d73 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.085553] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e542221-7c94-4b79-a05b-73435a3d18ad {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.102556] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0f6e49b-3f71-4266-8dc7-e7009fb3d620 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.113119] env[63515]: DEBUG nova.compute.manager [None req-f1f3c1ea-070f-4e78-a0c6-d376174a87f2 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: ea2e0810-c9f1-49a6-8589-7db402847aea] Starting instance... {{(pid=63515) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1085.116928] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-232adc42-db9a-4e72-959b-4fa5ff66f82c {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.149249] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180482MB free_disk=171GB free_vcpus=48 pci_devices=None {{(pid=63515) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1085.149428] env[63515]: DEBUG oslo_concurrency.lockutils [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1085.149657] env[63515]: DEBUG oslo_concurrency.lockutils [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1085.284301] env[63515]: DEBUG nova.network.neutron [req-ddb65374-6d93-4504-b682-e872a0390007 req-320bd8c3-9d16-490a-bfd3-ae7c6b88dee5 service nova] [instance: 6336f10d-97b3-4413-a828-823a27492215] Updated VIF entry in instance network info cache for port 635a24b0-e5a3-43bf-989f-b7bdbc07f767. {{(pid=63515) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1085.284719] env[63515]: DEBUG nova.network.neutron [req-ddb65374-6d93-4504-b682-e872a0390007 req-320bd8c3-9d16-490a-bfd3-ae7c6b88dee5 service nova] [instance: 6336f10d-97b3-4413-a828-823a27492215] Updating instance_info_cache with network_info: [{"id": "635a24b0-e5a3-43bf-989f-b7bdbc07f767", "address": "fa:16:3e:5e:ff:86", "network": {"id": "be106785-bb04-4c6c-93f6-1fafe9449aa0", "bridge": "br-int", "label": "tempest-InstanceActionsV221TestJSON-675645069-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8350a1ba736243b39e35024c2de58b9b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "39a4aca0-934b-4a91-8779-6a4360c3f967", "external-id": "nsx-vlan-transportzone-454", "segmentation_id": 454, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap635a24b0-e5", "ovs_interfaceid": "635a24b0-e5a3-43bf-989f-b7bdbc07f767", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1085.540222] env[63515]: DEBUG oslo_vmware.api [None req-680a9f3c-a655-45e5-b8db-642d274b2d94 tempest-InstanceActionsV221TestJSON-505260833 tempest-InstanceActionsV221TestJSON-505260833-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52a788eb-10a5-91b9-6cad-c246b7f947ac, 'name': SearchDatastore_Task, 'duration_secs': 0.011729} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1085.540550] env[63515]: DEBUG oslo_concurrency.lockutils [None req-680a9f3c-a655-45e5-b8db-642d274b2d94 tempest-InstanceActionsV221TestJSON-505260833 tempest-InstanceActionsV221TestJSON-505260833-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1085.540792] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-680a9f3c-a655-45e5-b8db-642d274b2d94 tempest-InstanceActionsV221TestJSON-505260833 tempest-InstanceActionsV221TestJSON-505260833-project-member] [instance: 6336f10d-97b3-4413-a828-823a27492215] Processing image 8a120570-cb06-4099-b262-554ca0ad15c5 {{(pid=63515) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1085.541076] env[63515]: DEBUG oslo_concurrency.lockutils [None req-680a9f3c-a655-45e5-b8db-642d274b2d94 tempest-InstanceActionsV221TestJSON-505260833 tempest-InstanceActionsV221TestJSON-505260833-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1085.541262] env[63515]: DEBUG oslo_concurrency.lockutils [None req-680a9f3c-a655-45e5-b8db-642d274b2d94 tempest-InstanceActionsV221TestJSON-505260833 tempest-InstanceActionsV221TestJSON-505260833-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1085.541481] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-680a9f3c-a655-45e5-b8db-642d274b2d94 tempest-InstanceActionsV221TestJSON-505260833 tempest-InstanceActionsV221TestJSON-505260833-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1085.541985] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e5f8bfcb-b40a-4adc-a0fb-9c2adb078a04 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.550076] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-680a9f3c-a655-45e5-b8db-642d274b2d94 tempest-InstanceActionsV221TestJSON-505260833 tempest-InstanceActionsV221TestJSON-505260833-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1085.550290] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-680a9f3c-a655-45e5-b8db-642d274b2d94 tempest-InstanceActionsV221TestJSON-505260833 tempest-InstanceActionsV221TestJSON-505260833-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63515) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1085.551049] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c4441698-5a85-4b9b-b384-e160d057153f {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.556387] env[63515]: DEBUG oslo_vmware.api [None req-680a9f3c-a655-45e5-b8db-642d274b2d94 tempest-InstanceActionsV221TestJSON-505260833 tempest-InstanceActionsV221TestJSON-505260833-project-member] Waiting for the task: (returnval){ [ 1085.556387] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]5275c97b-c314-6043-6e1f-00a3f9c79469" [ 1085.556387] env[63515]: _type = "Task" [ 1085.556387] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1085.565365] env[63515]: DEBUG oslo_vmware.api [None req-680a9f3c-a655-45e5-b8db-642d274b2d94 tempest-InstanceActionsV221TestJSON-505260833 tempest-InstanceActionsV221TestJSON-505260833-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]5275c97b-c314-6043-6e1f-00a3f9c79469, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1085.757923] env[63515]: DEBUG oslo_concurrency.lockutils [None req-f1f3c1ea-070f-4e78-a0c6-d376174a87f2 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1085.787759] env[63515]: DEBUG oslo_concurrency.lockutils [req-ddb65374-6d93-4504-b682-e872a0390007 req-320bd8c3-9d16-490a-bfd3-ae7c6b88dee5 service nova] Releasing lock "refresh_cache-6336f10d-97b3-4413-a828-823a27492215" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1086.067326] env[63515]: DEBUG oslo_vmware.api [None req-680a9f3c-a655-45e5-b8db-642d274b2d94 tempest-InstanceActionsV221TestJSON-505260833 tempest-InstanceActionsV221TestJSON-505260833-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]5275c97b-c314-6043-6e1f-00a3f9c79469, 'name': SearchDatastore_Task, 'duration_secs': 0.007765} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1086.068139] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d879849d-893a-4453-a37d-c88d3025af7f {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.073216] env[63515]: DEBUG oslo_vmware.api [None req-680a9f3c-a655-45e5-b8db-642d274b2d94 tempest-InstanceActionsV221TestJSON-505260833 tempest-InstanceActionsV221TestJSON-505260833-project-member] Waiting for the task: (returnval){ [ 1086.073216] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]52a85f11-bc51-b24a-8625-0ed521720572" [ 1086.073216] env[63515]: _type = "Task" [ 1086.073216] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1086.080918] env[63515]: DEBUG oslo_vmware.api [None req-680a9f3c-a655-45e5-b8db-642d274b2d94 tempest-InstanceActionsV221TestJSON-505260833 tempest-InstanceActionsV221TestJSON-505260833-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52a85f11-bc51-b24a-8625-0ed521720572, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1086.176331] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Instance e13da90c-28e6-43d2-99b7-19c5095954ca actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63515) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1086.176605] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Instance aa83d035-9c47-434d-ab1a-140e6b3110b4 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63515) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1086.176605] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Instance 6336f10d-97b3-4413-a828-823a27492215 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63515) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1086.583683] env[63515]: DEBUG oslo_vmware.api [None req-680a9f3c-a655-45e5-b8db-642d274b2d94 tempest-InstanceActionsV221TestJSON-505260833 tempest-InstanceActionsV221TestJSON-505260833-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52a85f11-bc51-b24a-8625-0ed521720572, 'name': SearchDatastore_Task, 'duration_secs': 0.009418} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1086.584068] env[63515]: DEBUG oslo_concurrency.lockutils [None req-680a9f3c-a655-45e5-b8db-642d274b2d94 tempest-InstanceActionsV221TestJSON-505260833 tempest-InstanceActionsV221TestJSON-505260833-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1086.584403] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-680a9f3c-a655-45e5-b8db-642d274b2d94 tempest-InstanceActionsV221TestJSON-505260833 tempest-InstanceActionsV221TestJSON-505260833-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk to [datastore1] 6336f10d-97b3-4413-a828-823a27492215/6336f10d-97b3-4413-a828-823a27492215.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1086.584722] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-bf176902-7976-4c09-9ea0-853192379b3b {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.592253] env[63515]: DEBUG oslo_vmware.api [None req-680a9f3c-a655-45e5-b8db-642d274b2d94 tempest-InstanceActionsV221TestJSON-505260833 tempest-InstanceActionsV221TestJSON-505260833-project-member] Waiting for the task: (returnval){ [ 1086.592253] env[63515]: value = "task-1111933" [ 1086.592253] env[63515]: _type = "Task" [ 1086.592253] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1086.600617] env[63515]: DEBUG oslo_vmware.api [None req-680a9f3c-a655-45e5-b8db-642d274b2d94 tempest-InstanceActionsV221TestJSON-505260833 tempest-InstanceActionsV221TestJSON-505260833-project-member] Task: {'id': task-1111933, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1086.679735] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Instance ea2e0810-c9f1-49a6-8589-7db402847aea has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63515) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 1086.680342] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Total usable vcpus: 48, total allocated vcpus: 3 {{(pid=63515) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1086.680655] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1088MB phys_disk=200GB used_disk=3GB total_vcpus=48 used_vcpus=3 pci_stats=[] {{(pid=63515) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1086.782930] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a79a8a0-4ab0-454a-9779-19d7b8abef65 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.793061] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f6994a0-a5a2-454c-ba9d-00adfcf29449 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.832620] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e811742-0881-4dba-a98b-2d0f2c20684c {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.840992] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27951ba6-3324-4057-9f4a-4e333fb27e44 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.857722] env[63515]: DEBUG nova.compute.provider_tree [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1087.103168] env[63515]: DEBUG oslo_vmware.api [None req-680a9f3c-a655-45e5-b8db-642d274b2d94 tempest-InstanceActionsV221TestJSON-505260833 tempest-InstanceActionsV221TestJSON-505260833-project-member] Task: {'id': task-1111933, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1087.362130] env[63515]: DEBUG nova.scheduler.client.report [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1087.602931] env[63515]: DEBUG oslo_vmware.api [None req-680a9f3c-a655-45e5-b8db-642d274b2d94 tempest-InstanceActionsV221TestJSON-505260833 tempest-InstanceActionsV221TestJSON-505260833-project-member] Task: {'id': task-1111933, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.513119} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1087.603996] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-680a9f3c-a655-45e5-b8db-642d274b2d94 tempest-InstanceActionsV221TestJSON-505260833 tempest-InstanceActionsV221TestJSON-505260833-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk to [datastore1] 6336f10d-97b3-4413-a828-823a27492215/6336f10d-97b3-4413-a828-823a27492215.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1087.603996] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-680a9f3c-a655-45e5-b8db-642d274b2d94 tempest-InstanceActionsV221TestJSON-505260833 tempest-InstanceActionsV221TestJSON-505260833-project-member] [instance: 6336f10d-97b3-4413-a828-823a27492215] Extending root virtual disk to 1048576 {{(pid=63515) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1087.603996] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-1fa7777b-f48c-4691-b3aa-23997d906a5e {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.610794] env[63515]: DEBUG oslo_vmware.api [None req-680a9f3c-a655-45e5-b8db-642d274b2d94 tempest-InstanceActionsV221TestJSON-505260833 tempest-InstanceActionsV221TestJSON-505260833-project-member] Waiting for the task: (returnval){ [ 1087.610794] env[63515]: value = "task-1111935" [ 1087.610794] env[63515]: _type = "Task" [ 1087.610794] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1087.619151] env[63515]: DEBUG oslo_vmware.api [None req-680a9f3c-a655-45e5-b8db-642d274b2d94 tempest-InstanceActionsV221TestJSON-505260833 tempest-InstanceActionsV221TestJSON-505260833-project-member] Task: {'id': task-1111935, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1087.866748] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=63515) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1087.867882] env[63515]: DEBUG oslo_concurrency.lockutils [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.718s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1087.868656] env[63515]: DEBUG oslo_concurrency.lockutils [None req-f1f3c1ea-070f-4e78-a0c6-d376174a87f2 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.111s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1087.870928] env[63515]: INFO nova.compute.claims [None req-f1f3c1ea-070f-4e78-a0c6-d376174a87f2 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: ea2e0810-c9f1-49a6-8589-7db402847aea] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1087.874610] env[63515]: DEBUG oslo_service.periodic_task [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Running periodic task ComputeManager._run_pending_deletes {{(pid=63515) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1087.874610] env[63515]: DEBUG nova.compute.manager [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Cleaning up deleted instances {{(pid=63515) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11226}} [ 1088.121855] env[63515]: DEBUG oslo_vmware.api [None req-680a9f3c-a655-45e5-b8db-642d274b2d94 tempest-InstanceActionsV221TestJSON-505260833 tempest-InstanceActionsV221TestJSON-505260833-project-member] Task: {'id': task-1111935, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.065621} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1088.122169] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-680a9f3c-a655-45e5-b8db-642d274b2d94 tempest-InstanceActionsV221TestJSON-505260833 tempest-InstanceActionsV221TestJSON-505260833-project-member] [instance: 6336f10d-97b3-4413-a828-823a27492215] Extended root virtual disk {{(pid=63515) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1088.122975] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52785f8d-f926-4e28-9eb4-3e767b61b2cf {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.144272] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-680a9f3c-a655-45e5-b8db-642d274b2d94 tempest-InstanceActionsV221TestJSON-505260833 tempest-InstanceActionsV221TestJSON-505260833-project-member] [instance: 6336f10d-97b3-4413-a828-823a27492215] Reconfiguring VM instance instance-00000069 to attach disk [datastore1] 6336f10d-97b3-4413-a828-823a27492215/6336f10d-97b3-4413-a828-823a27492215.vmdk or device None with type sparse {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1088.144870] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5f6d44e6-53fe-44e1-a104-56203390b4fd {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.165254] env[63515]: DEBUG oslo_vmware.api [None req-680a9f3c-a655-45e5-b8db-642d274b2d94 tempest-InstanceActionsV221TestJSON-505260833 tempest-InstanceActionsV221TestJSON-505260833-project-member] Waiting for the task: (returnval){ [ 1088.165254] env[63515]: value = "task-1111936" [ 1088.165254] env[63515]: _type = "Task" [ 1088.165254] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1088.174468] env[63515]: DEBUG oslo_vmware.api [None req-680a9f3c-a655-45e5-b8db-642d274b2d94 tempest-InstanceActionsV221TestJSON-505260833 tempest-InstanceActionsV221TestJSON-505260833-project-member] Task: {'id': task-1111936, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1088.399374] env[63515]: DEBUG nova.compute.manager [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] There are 67 instances to clean {{(pid=63515) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11235}} [ 1088.399695] env[63515]: DEBUG nova.compute.manager [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] [instance: 7fdc6303-2d53-4d87-9aea-bf40e476f790] Instance has had 0 of 5 cleanup attempts {{(pid=63515) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1088.473338] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bcfb93f4-b377-4975-8acc-8d69b32d6992 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.480863] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9ce0657-92a0-474c-b3f5-33d4a5546db9 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.512518] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-638edfc6-09a7-44eb-8073-c3cfebe87b1d {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.519571] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2915cf3c-a400-4f86-9127-b66011502ca3 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.534734] env[63515]: DEBUG nova.compute.provider_tree [None req-f1f3c1ea-070f-4e78-a0c6-d376174a87f2 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1088.678907] env[63515]: DEBUG oslo_vmware.api [None req-680a9f3c-a655-45e5-b8db-642d274b2d94 tempest-InstanceActionsV221TestJSON-505260833 tempest-InstanceActionsV221TestJSON-505260833-project-member] Task: {'id': task-1111936, 'name': ReconfigVM_Task, 'duration_secs': 0.283125} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1088.678907] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-680a9f3c-a655-45e5-b8db-642d274b2d94 tempest-InstanceActionsV221TestJSON-505260833 tempest-InstanceActionsV221TestJSON-505260833-project-member] [instance: 6336f10d-97b3-4413-a828-823a27492215] Reconfigured VM instance instance-00000069 to attach disk [datastore1] 6336f10d-97b3-4413-a828-823a27492215/6336f10d-97b3-4413-a828-823a27492215.vmdk or device None with type sparse {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1088.678907] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e7fca6d5-3a66-41db-a38c-a11a03280095 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.687023] env[63515]: DEBUG oslo_vmware.api [None req-680a9f3c-a655-45e5-b8db-642d274b2d94 tempest-InstanceActionsV221TestJSON-505260833 tempest-InstanceActionsV221TestJSON-505260833-project-member] Waiting for the task: (returnval){ [ 1088.687023] env[63515]: value = "task-1111937" [ 1088.687023] env[63515]: _type = "Task" [ 1088.687023] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1088.693572] env[63515]: DEBUG oslo_vmware.api [None req-680a9f3c-a655-45e5-b8db-642d274b2d94 tempest-InstanceActionsV221TestJSON-505260833 tempest-InstanceActionsV221TestJSON-505260833-project-member] Task: {'id': task-1111937, 'name': Rename_Task} progress is 5%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1088.903056] env[63515]: DEBUG nova.compute.manager [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] [instance: a8d2cc01-3bc1-4792-85f5-b9095294b33c] Instance has had 0 of 5 cleanup attempts {{(pid=63515) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1089.038491] env[63515]: DEBUG nova.scheduler.client.report [None req-f1f3c1ea-070f-4e78-a0c6-d376174a87f2 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1089.195661] env[63515]: DEBUG oslo_vmware.api [None req-680a9f3c-a655-45e5-b8db-642d274b2d94 tempest-InstanceActionsV221TestJSON-505260833 tempest-InstanceActionsV221TestJSON-505260833-project-member] Task: {'id': task-1111937, 'name': Rename_Task, 'duration_secs': 0.145567} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1089.196040] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-680a9f3c-a655-45e5-b8db-642d274b2d94 tempest-InstanceActionsV221TestJSON-505260833 tempest-InstanceActionsV221TestJSON-505260833-project-member] [instance: 6336f10d-97b3-4413-a828-823a27492215] Powering on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1089.196337] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d33e4d31-7f6d-42ce-b3fa-dbba72a8ef65 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.202909] env[63515]: DEBUG oslo_vmware.api [None req-680a9f3c-a655-45e5-b8db-642d274b2d94 tempest-InstanceActionsV221TestJSON-505260833 tempest-InstanceActionsV221TestJSON-505260833-project-member] Waiting for the task: (returnval){ [ 1089.202909] env[63515]: value = "task-1111938" [ 1089.202909] env[63515]: _type = "Task" [ 1089.202909] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1089.211689] env[63515]: DEBUG oslo_vmware.api [None req-680a9f3c-a655-45e5-b8db-642d274b2d94 tempest-InstanceActionsV221TestJSON-505260833 tempest-InstanceActionsV221TestJSON-505260833-project-member] Task: {'id': task-1111938, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1089.313830] env[63515]: DEBUG oslo_concurrency.lockutils [None req-129ac0c9-aca5-49a7-bd9c-0de3f76dcffa tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Acquiring lock "c8dbf542-70cb-45be-9308-5e12307c0be6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1089.314096] env[63515]: DEBUG oslo_concurrency.lockutils [None req-129ac0c9-aca5-49a7-bd9c-0de3f76dcffa tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Lock "c8dbf542-70cb-45be-9308-5e12307c0be6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1089.407398] env[63515]: DEBUG nova.compute.manager [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] [instance: e925fc93-a731-498a-984c-9b1f3eabf353] Instance has had 0 of 5 cleanup attempts {{(pid=63515) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1089.485486] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-c053fe74-7df3-47e0-afc9-192490b04524 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: aa83d035-9c47-434d-ab1a-140e6b3110b4] Volume attach. Driver type: vmdk {{(pid=63515) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1089.485741] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-c053fe74-7df3-47e0-afc9-192490b04524 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: aa83d035-9c47-434d-ab1a-140e6b3110b4] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-243578', 'volume_id': '81e015c1-d604-49a7-9bad-778a82466e35', 'name': 'volume-81e015c1-d604-49a7-9bad-778a82466e35', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'aa83d035-9c47-434d-ab1a-140e6b3110b4', 'attached_at': '', 'detached_at': '', 'volume_id': '81e015c1-d604-49a7-9bad-778a82466e35', 'serial': '81e015c1-d604-49a7-9bad-778a82466e35'} {{(pid=63515) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1089.486746] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e238b73f-58eb-4241-8e38-cc657d13ce90 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.502856] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9291183-cab4-4ca5-af39-c1aa39c7c9d9 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.525974] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-c053fe74-7df3-47e0-afc9-192490b04524 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: aa83d035-9c47-434d-ab1a-140e6b3110b4] Reconfiguring VM instance instance-00000066 to attach disk [datastore1] volume-81e015c1-d604-49a7-9bad-778a82466e35/volume-81e015c1-d604-49a7-9bad-778a82466e35.vmdk or device None with type thin {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1089.526231] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-008e70a1-8f14-4b5c-9e0e-68af2d72e43d {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.543612] env[63515]: DEBUG oslo_concurrency.lockutils [None req-f1f3c1ea-070f-4e78-a0c6-d376174a87f2 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 1.675s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1089.544099] env[63515]: DEBUG nova.compute.manager [None req-f1f3c1ea-070f-4e78-a0c6-d376174a87f2 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: ea2e0810-c9f1-49a6-8589-7db402847aea] Start building networks asynchronously for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1089.546957] env[63515]: DEBUG oslo_vmware.api [None req-c053fe74-7df3-47e0-afc9-192490b04524 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Waiting for the task: (returnval){ [ 1089.546957] env[63515]: value = "task-1111939" [ 1089.546957] env[63515]: _type = "Task" [ 1089.546957] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1089.557238] env[63515]: DEBUG oslo_vmware.api [None req-c053fe74-7df3-47e0-afc9-192490b04524 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Task: {'id': task-1111939, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1089.713607] env[63515]: DEBUG oslo_vmware.api [None req-680a9f3c-a655-45e5-b8db-642d274b2d94 tempest-InstanceActionsV221TestJSON-505260833 tempest-InstanceActionsV221TestJSON-505260833-project-member] Task: {'id': task-1111938, 'name': PowerOnVM_Task, 'duration_secs': 0.443818} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1089.713864] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-680a9f3c-a655-45e5-b8db-642d274b2d94 tempest-InstanceActionsV221TestJSON-505260833 tempest-InstanceActionsV221TestJSON-505260833-project-member] [instance: 6336f10d-97b3-4413-a828-823a27492215] Powered on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1089.714095] env[63515]: INFO nova.compute.manager [None req-680a9f3c-a655-45e5-b8db-642d274b2d94 tempest-InstanceActionsV221TestJSON-505260833 tempest-InstanceActionsV221TestJSON-505260833-project-member] [instance: 6336f10d-97b3-4413-a828-823a27492215] Took 7.39 seconds to spawn the instance on the hypervisor. [ 1089.714286] env[63515]: DEBUG nova.compute.manager [None req-680a9f3c-a655-45e5-b8db-642d274b2d94 tempest-InstanceActionsV221TestJSON-505260833 tempest-InstanceActionsV221TestJSON-505260833-project-member] [instance: 6336f10d-97b3-4413-a828-823a27492215] Checking state {{(pid=63515) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1089.715045] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a661fc1b-865d-4f87-8693-401ed24e5eb7 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.816535] env[63515]: DEBUG nova.compute.manager [None req-129ac0c9-aca5-49a7-bd9c-0de3f76dcffa tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] [instance: c8dbf542-70cb-45be-9308-5e12307c0be6] Starting instance... {{(pid=63515) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1089.910898] env[63515]: DEBUG nova.compute.manager [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] [instance: b76cf3dd-7b7e-4e5d-bf3c-e15bd37069ac] Instance has had 0 of 5 cleanup attempts {{(pid=63515) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1090.048805] env[63515]: DEBUG nova.compute.utils [None req-f1f3c1ea-070f-4e78-a0c6-d376174a87f2 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Using /dev/sd instead of None {{(pid=63515) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1090.050391] env[63515]: DEBUG nova.compute.manager [None req-f1f3c1ea-070f-4e78-a0c6-d376174a87f2 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: ea2e0810-c9f1-49a6-8589-7db402847aea] Allocating IP information in the background. {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1090.050581] env[63515]: DEBUG nova.network.neutron [None req-f1f3c1ea-070f-4e78-a0c6-d376174a87f2 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: ea2e0810-c9f1-49a6-8589-7db402847aea] allocate_for_instance() {{(pid=63515) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1090.061592] env[63515]: DEBUG oslo_vmware.api [None req-c053fe74-7df3-47e0-afc9-192490b04524 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Task: {'id': task-1111939, 'name': ReconfigVM_Task, 'duration_secs': 0.391747} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1090.062339] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-c053fe74-7df3-47e0-afc9-192490b04524 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: aa83d035-9c47-434d-ab1a-140e6b3110b4] Reconfigured VM instance instance-00000066 to attach disk [datastore1] volume-81e015c1-d604-49a7-9bad-778a82466e35/volume-81e015c1-d604-49a7-9bad-778a82466e35.vmdk or device None with type thin {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1090.067190] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c3731636-f37d-4ff1-93b0-92115d19fc55 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.081655] env[63515]: DEBUG oslo_vmware.api [None req-c053fe74-7df3-47e0-afc9-192490b04524 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Waiting for the task: (returnval){ [ 1090.081655] env[63515]: value = "task-1111940" [ 1090.081655] env[63515]: _type = "Task" [ 1090.081655] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1090.089545] env[63515]: DEBUG oslo_vmware.api [None req-c053fe74-7df3-47e0-afc9-192490b04524 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Task: {'id': task-1111940, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1090.094832] env[63515]: DEBUG nova.policy [None req-f1f3c1ea-070f-4e78-a0c6-d376174a87f2 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ed295ef8f0134bcaa668149ac9f4dd74', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'aef947bfaf7d4180b46b4be662e3d903', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63515) authorize /opt/stack/nova/nova/policy.py:201}} [ 1090.238683] env[63515]: INFO nova.compute.manager [None req-680a9f3c-a655-45e5-b8db-642d274b2d94 tempest-InstanceActionsV221TestJSON-505260833 tempest-InstanceActionsV221TestJSON-505260833-project-member] [instance: 6336f10d-97b3-4413-a828-823a27492215] Took 12.13 seconds to build instance. [ 1090.339910] env[63515]: DEBUG oslo_concurrency.lockutils [None req-129ac0c9-aca5-49a7-bd9c-0de3f76dcffa tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1090.340170] env[63515]: DEBUG oslo_concurrency.lockutils [None req-129ac0c9-aca5-49a7-bd9c-0de3f76dcffa tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1090.345939] env[63515]: INFO nova.compute.claims [None req-129ac0c9-aca5-49a7-bd9c-0de3f76dcffa tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] [instance: c8dbf542-70cb-45be-9308-5e12307c0be6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1090.354699] env[63515]: DEBUG nova.network.neutron [None req-f1f3c1ea-070f-4e78-a0c6-d376174a87f2 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: ea2e0810-c9f1-49a6-8589-7db402847aea] Successfully created port: 4e1d011e-2609-4c71-8317-2685198b6a8b {{(pid=63515) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1090.414070] env[63515]: DEBUG nova.compute.manager [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] [instance: fcfe3376-8dfa-4189-a267-ff6402713c1b] Instance has had 0 of 5 cleanup attempts {{(pid=63515) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1090.557338] env[63515]: DEBUG nova.compute.manager [None req-f1f3c1ea-070f-4e78-a0c6-d376174a87f2 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: ea2e0810-c9f1-49a6-8589-7db402847aea] Start building block device mappings for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1090.591152] env[63515]: DEBUG oslo_vmware.api [None req-c053fe74-7df3-47e0-afc9-192490b04524 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Task: {'id': task-1111940, 'name': ReconfigVM_Task, 'duration_secs': 0.134516} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1090.592062] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-c053fe74-7df3-47e0-afc9-192490b04524 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: aa83d035-9c47-434d-ab1a-140e6b3110b4] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-243578', 'volume_id': '81e015c1-d604-49a7-9bad-778a82466e35', 'name': 'volume-81e015c1-d604-49a7-9bad-778a82466e35', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'aa83d035-9c47-434d-ab1a-140e6b3110b4', 'attached_at': '', 'detached_at': '', 'volume_id': '81e015c1-d604-49a7-9bad-778a82466e35', 'serial': '81e015c1-d604-49a7-9bad-778a82466e35'} {{(pid=63515) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1090.734016] env[63515]: DEBUG oslo_concurrency.lockutils [None req-a445eac3-9e94-4d2f-a9a7-82f7ac36cc73 tempest-InstanceActionsV221TestJSON-505260833 tempest-InstanceActionsV221TestJSON-505260833-project-member] Acquiring lock "6336f10d-97b3-4413-a828-823a27492215" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1090.741142] env[63515]: DEBUG oslo_concurrency.lockutils [None req-680a9f3c-a655-45e5-b8db-642d274b2d94 tempest-InstanceActionsV221TestJSON-505260833 tempest-InstanceActionsV221TestJSON-505260833-project-member] Lock "6336f10d-97b3-4413-a828-823a27492215" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.635s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1090.741415] env[63515]: DEBUG oslo_concurrency.lockutils [None req-a445eac3-9e94-4d2f-a9a7-82f7ac36cc73 tempest-InstanceActionsV221TestJSON-505260833 tempest-InstanceActionsV221TestJSON-505260833-project-member] Lock "6336f10d-97b3-4413-a828-823a27492215" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.008s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1090.741638] env[63515]: DEBUG oslo_concurrency.lockutils [None req-a445eac3-9e94-4d2f-a9a7-82f7ac36cc73 tempest-InstanceActionsV221TestJSON-505260833 tempest-InstanceActionsV221TestJSON-505260833-project-member] Acquiring lock "6336f10d-97b3-4413-a828-823a27492215-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1090.741845] env[63515]: DEBUG oslo_concurrency.lockutils [None req-a445eac3-9e94-4d2f-a9a7-82f7ac36cc73 tempest-InstanceActionsV221TestJSON-505260833 tempest-InstanceActionsV221TestJSON-505260833-project-member] Lock "6336f10d-97b3-4413-a828-823a27492215-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1090.742023] env[63515]: DEBUG oslo_concurrency.lockutils [None req-a445eac3-9e94-4d2f-a9a7-82f7ac36cc73 tempest-InstanceActionsV221TestJSON-505260833 tempest-InstanceActionsV221TestJSON-505260833-project-member] Lock "6336f10d-97b3-4413-a828-823a27492215-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1090.744254] env[63515]: INFO nova.compute.manager [None req-a445eac3-9e94-4d2f-a9a7-82f7ac36cc73 tempest-InstanceActionsV221TestJSON-505260833 tempest-InstanceActionsV221TestJSON-505260833-project-member] [instance: 6336f10d-97b3-4413-a828-823a27492215] Terminating instance [ 1090.746107] env[63515]: DEBUG nova.compute.manager [None req-a445eac3-9e94-4d2f-a9a7-82f7ac36cc73 tempest-InstanceActionsV221TestJSON-505260833 tempest-InstanceActionsV221TestJSON-505260833-project-member] [instance: 6336f10d-97b3-4413-a828-823a27492215] Start destroying the instance on the hypervisor. {{(pid=63515) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1090.746374] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-a445eac3-9e94-4d2f-a9a7-82f7ac36cc73 tempest-InstanceActionsV221TestJSON-505260833 tempest-InstanceActionsV221TestJSON-505260833-project-member] [instance: 6336f10d-97b3-4413-a828-823a27492215] Destroying instance {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1090.747240] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83176d3a-ce5c-4fcd-8c09-55605743f04a {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.755242] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-a445eac3-9e94-4d2f-a9a7-82f7ac36cc73 tempest-InstanceActionsV221TestJSON-505260833 tempest-InstanceActionsV221TestJSON-505260833-project-member] [instance: 6336f10d-97b3-4413-a828-823a27492215] Powering off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1090.755815] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-eb1fc242-1231-4c26-8506-9b7c836384d4 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.761096] env[63515]: DEBUG oslo_vmware.api [None req-a445eac3-9e94-4d2f-a9a7-82f7ac36cc73 tempest-InstanceActionsV221TestJSON-505260833 tempest-InstanceActionsV221TestJSON-505260833-project-member] Waiting for the task: (returnval){ [ 1090.761096] env[63515]: value = "task-1111941" [ 1090.761096] env[63515]: _type = "Task" [ 1090.761096] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1090.768826] env[63515]: DEBUG oslo_vmware.api [None req-a445eac3-9e94-4d2f-a9a7-82f7ac36cc73 tempest-InstanceActionsV221TestJSON-505260833 tempest-InstanceActionsV221TestJSON-505260833-project-member] Task: {'id': task-1111941, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1090.917801] env[63515]: DEBUG nova.compute.manager [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] [instance: d6c5a9f9-5e5c-4b14-9b9a-7825bbf526a5] Instance has had 0 of 5 cleanup attempts {{(pid=63515) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1091.272648] env[63515]: DEBUG oslo_vmware.api [None req-a445eac3-9e94-4d2f-a9a7-82f7ac36cc73 tempest-InstanceActionsV221TestJSON-505260833 tempest-InstanceActionsV221TestJSON-505260833-project-member] Task: {'id': task-1111941, 'name': PowerOffVM_Task, 'duration_secs': 0.187913} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1091.272957] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-a445eac3-9e94-4d2f-a9a7-82f7ac36cc73 tempest-InstanceActionsV221TestJSON-505260833 tempest-InstanceActionsV221TestJSON-505260833-project-member] [instance: 6336f10d-97b3-4413-a828-823a27492215] Powered off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1091.273151] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-a445eac3-9e94-4d2f-a9a7-82f7ac36cc73 tempest-InstanceActionsV221TestJSON-505260833 tempest-InstanceActionsV221TestJSON-505260833-project-member] [instance: 6336f10d-97b3-4413-a828-823a27492215] Unregistering the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1091.273406] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9ca1485c-345d-4178-a767-1bd568283af1 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.340982] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-a445eac3-9e94-4d2f-a9a7-82f7ac36cc73 tempest-InstanceActionsV221TestJSON-505260833 tempest-InstanceActionsV221TestJSON-505260833-project-member] [instance: 6336f10d-97b3-4413-a828-823a27492215] Unregistered the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1091.340982] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-a445eac3-9e94-4d2f-a9a7-82f7ac36cc73 tempest-InstanceActionsV221TestJSON-505260833 tempest-InstanceActionsV221TestJSON-505260833-project-member] [instance: 6336f10d-97b3-4413-a828-823a27492215] Deleting contents of the VM from datastore datastore1 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1091.341179] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-a445eac3-9e94-4d2f-a9a7-82f7ac36cc73 tempest-InstanceActionsV221TestJSON-505260833 tempest-InstanceActionsV221TestJSON-505260833-project-member] Deleting the datastore file [datastore1] 6336f10d-97b3-4413-a828-823a27492215 {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1091.341472] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-96f50941-1802-4133-b1fa-1eb4893e6a36 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.348058] env[63515]: DEBUG oslo_vmware.api [None req-a445eac3-9e94-4d2f-a9a7-82f7ac36cc73 tempest-InstanceActionsV221TestJSON-505260833 tempest-InstanceActionsV221TestJSON-505260833-project-member] Waiting for the task: (returnval){ [ 1091.348058] env[63515]: value = "task-1111943" [ 1091.348058] env[63515]: _type = "Task" [ 1091.348058] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1091.359297] env[63515]: DEBUG oslo_vmware.api [None req-a445eac3-9e94-4d2f-a9a7-82f7ac36cc73 tempest-InstanceActionsV221TestJSON-505260833 tempest-InstanceActionsV221TestJSON-505260833-project-member] Task: {'id': task-1111943, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1091.420774] env[63515]: DEBUG nova.compute.manager [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] [instance: c6ac42a9-a233-48b4-bcb2-fb791c7446bb] Instance has had 0 of 5 cleanup attempts {{(pid=63515) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1091.450668] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5c8caa7-1e0b-457f-9c51-c9e02f38f223 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.458992] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-baead573-ddbe-40c5-917c-1bd331cf76e7 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.488953] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16e7d317-6f95-4fb9-8e34-afba318305aa {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.496472] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85e15d67-df41-4d87-af06-67630bd6b745 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.511318] env[63515]: DEBUG nova.compute.provider_tree [None req-129ac0c9-aca5-49a7-bd9c-0de3f76dcffa tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1091.568027] env[63515]: DEBUG nova.compute.manager [None req-f1f3c1ea-070f-4e78-a0c6-d376174a87f2 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: ea2e0810-c9f1-49a6-8589-7db402847aea] Start spawning the instance on the hypervisor. {{(pid=63515) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1091.596120] env[63515]: DEBUG nova.virt.hardware [None req-f1f3c1ea-070f-4e78-a0c6-d376174a87f2 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T02:52:14Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T02:51:56Z,direct_url=,disk_format='vmdk',id=8a120570-cb06-4099-b262-554ca0ad15c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b569255cc43e42e7ae3f2b4ad37c6ef4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T02:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1091.596120] env[63515]: DEBUG nova.virt.hardware [None req-f1f3c1ea-070f-4e78-a0c6-d376174a87f2 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Flavor limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1091.596120] env[63515]: DEBUG nova.virt.hardware [None req-f1f3c1ea-070f-4e78-a0c6-d376174a87f2 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Image limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1091.596120] env[63515]: DEBUG nova.virt.hardware [None req-f1f3c1ea-070f-4e78-a0c6-d376174a87f2 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Flavor pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1091.596120] env[63515]: DEBUG nova.virt.hardware [None req-f1f3c1ea-070f-4e78-a0c6-d376174a87f2 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Image pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1091.596120] env[63515]: DEBUG nova.virt.hardware [None req-f1f3c1ea-070f-4e78-a0c6-d376174a87f2 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1091.596120] env[63515]: DEBUG nova.virt.hardware [None req-f1f3c1ea-070f-4e78-a0c6-d376174a87f2 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1091.597037] env[63515]: DEBUG nova.virt.hardware [None req-f1f3c1ea-070f-4e78-a0c6-d376174a87f2 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1091.597402] env[63515]: DEBUG nova.virt.hardware [None req-f1f3c1ea-070f-4e78-a0c6-d376174a87f2 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Got 1 possible topologies {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1091.597745] env[63515]: DEBUG nova.virt.hardware [None req-f1f3c1ea-070f-4e78-a0c6-d376174a87f2 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1091.598118] env[63515]: DEBUG nova.virt.hardware [None req-f1f3c1ea-070f-4e78-a0c6-d376174a87f2 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1091.599132] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9430b2a5-473d-4885-964d-4bb0f35c3078 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.606964] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d016b945-3ff4-4d97-879f-a5adf222a9de {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.631714] env[63515]: DEBUG nova.objects.instance [None req-c053fe74-7df3-47e0-afc9-192490b04524 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Lazy-loading 'flavor' on Instance uuid aa83d035-9c47-434d-ab1a-140e6b3110b4 {{(pid=63515) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1091.839173] env[63515]: DEBUG nova.compute.manager [req-70832cc4-cc0d-4127-b0e7-4a8665bcd523 req-d5f9fd3c-d682-46ba-8785-475b11bf8b15 service nova] [instance: ea2e0810-c9f1-49a6-8589-7db402847aea] Received event network-vif-plugged-4e1d011e-2609-4c71-8317-2685198b6a8b {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1091.839173] env[63515]: DEBUG oslo_concurrency.lockutils [req-70832cc4-cc0d-4127-b0e7-4a8665bcd523 req-d5f9fd3c-d682-46ba-8785-475b11bf8b15 service nova] Acquiring lock "ea2e0810-c9f1-49a6-8589-7db402847aea-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1091.839173] env[63515]: DEBUG oslo_concurrency.lockutils [req-70832cc4-cc0d-4127-b0e7-4a8665bcd523 req-d5f9fd3c-d682-46ba-8785-475b11bf8b15 service nova] Lock "ea2e0810-c9f1-49a6-8589-7db402847aea-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1091.839173] env[63515]: DEBUG oslo_concurrency.lockutils [req-70832cc4-cc0d-4127-b0e7-4a8665bcd523 req-d5f9fd3c-d682-46ba-8785-475b11bf8b15 service nova] Lock "ea2e0810-c9f1-49a6-8589-7db402847aea-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1091.839870] env[63515]: DEBUG nova.compute.manager [req-70832cc4-cc0d-4127-b0e7-4a8665bcd523 req-d5f9fd3c-d682-46ba-8785-475b11bf8b15 service nova] [instance: ea2e0810-c9f1-49a6-8589-7db402847aea] No waiting events found dispatching network-vif-plugged-4e1d011e-2609-4c71-8317-2685198b6a8b {{(pid=63515) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1091.840210] env[63515]: WARNING nova.compute.manager [req-70832cc4-cc0d-4127-b0e7-4a8665bcd523 req-d5f9fd3c-d682-46ba-8785-475b11bf8b15 service nova] [instance: ea2e0810-c9f1-49a6-8589-7db402847aea] Received unexpected event network-vif-plugged-4e1d011e-2609-4c71-8317-2685198b6a8b for instance with vm_state building and task_state spawning. [ 1091.857981] env[63515]: DEBUG oslo_vmware.api [None req-a445eac3-9e94-4d2f-a9a7-82f7ac36cc73 tempest-InstanceActionsV221TestJSON-505260833 tempest-InstanceActionsV221TestJSON-505260833-project-member] Task: {'id': task-1111943, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.134391} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1091.858279] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-a445eac3-9e94-4d2f-a9a7-82f7ac36cc73 tempest-InstanceActionsV221TestJSON-505260833 tempest-InstanceActionsV221TestJSON-505260833-project-member] Deleted the datastore file {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1091.858472] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-a445eac3-9e94-4d2f-a9a7-82f7ac36cc73 tempest-InstanceActionsV221TestJSON-505260833 tempest-InstanceActionsV221TestJSON-505260833-project-member] [instance: 6336f10d-97b3-4413-a828-823a27492215] Deleted contents of the VM from datastore datastore1 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1091.858656] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-a445eac3-9e94-4d2f-a9a7-82f7ac36cc73 tempest-InstanceActionsV221TestJSON-505260833 tempest-InstanceActionsV221TestJSON-505260833-project-member] [instance: 6336f10d-97b3-4413-a828-823a27492215] Instance destroyed {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1091.858830] env[63515]: INFO nova.compute.manager [None req-a445eac3-9e94-4d2f-a9a7-82f7ac36cc73 tempest-InstanceActionsV221TestJSON-505260833 tempest-InstanceActionsV221TestJSON-505260833-project-member] [instance: 6336f10d-97b3-4413-a828-823a27492215] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1091.859095] env[63515]: DEBUG oslo.service.loopingcall [None req-a445eac3-9e94-4d2f-a9a7-82f7ac36cc73 tempest-InstanceActionsV221TestJSON-505260833 tempest-InstanceActionsV221TestJSON-505260833-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1091.859293] env[63515]: DEBUG nova.compute.manager [-] [instance: 6336f10d-97b3-4413-a828-823a27492215] Deallocating network for instance {{(pid=63515) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1091.859401] env[63515]: DEBUG nova.network.neutron [-] [instance: 6336f10d-97b3-4413-a828-823a27492215] deallocate_for_instance() {{(pid=63515) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1091.926558] env[63515]: DEBUG nova.compute.manager [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] [instance: 461a0c4c-dfdb-4669-ac72-8913c0277298] Instance has had 0 of 5 cleanup attempts {{(pid=63515) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1091.943464] env[63515]: DEBUG nova.network.neutron [None req-f1f3c1ea-070f-4e78-a0c6-d376174a87f2 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: ea2e0810-c9f1-49a6-8589-7db402847aea] Successfully updated port: 4e1d011e-2609-4c71-8317-2685198b6a8b {{(pid=63515) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1092.013847] env[63515]: DEBUG nova.scheduler.client.report [None req-129ac0c9-aca5-49a7-bd9c-0de3f76dcffa tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1092.097420] env[63515]: DEBUG nova.compute.manager [req-7faf5faf-b9df-4556-a650-a7991b2b1793 req-7996c990-5d0d-4e34-8560-7912e43f90bc service nova] [instance: 6336f10d-97b3-4413-a828-823a27492215] Received event network-vif-deleted-635a24b0-e5a3-43bf-989f-b7bdbc07f767 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1092.097636] env[63515]: INFO nova.compute.manager [req-7faf5faf-b9df-4556-a650-a7991b2b1793 req-7996c990-5d0d-4e34-8560-7912e43f90bc service nova] [instance: 6336f10d-97b3-4413-a828-823a27492215] Neutron deleted interface 635a24b0-e5a3-43bf-989f-b7bdbc07f767; detaching it from the instance and deleting it from the info cache [ 1092.097813] env[63515]: DEBUG nova.network.neutron [req-7faf5faf-b9df-4556-a650-a7991b2b1793 req-7996c990-5d0d-4e34-8560-7912e43f90bc service nova] [instance: 6336f10d-97b3-4413-a828-823a27492215] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1092.138862] env[63515]: DEBUG oslo_concurrency.lockutils [None req-c053fe74-7df3-47e0-afc9-192490b04524 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Lock "aa83d035-9c47-434d-ab1a-140e6b3110b4" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.766s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1092.428456] env[63515]: DEBUG nova.compute.manager [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] [instance: 724b05bc-ce4b-4f99-ad14-f3abea5567f2] Instance has had 0 of 5 cleanup attempts {{(pid=63515) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1092.446227] env[63515]: DEBUG oslo_concurrency.lockutils [None req-f1f3c1ea-070f-4e78-a0c6-d376174a87f2 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Acquiring lock "refresh_cache-ea2e0810-c9f1-49a6-8589-7db402847aea" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1092.446420] env[63515]: DEBUG oslo_concurrency.lockutils [None req-f1f3c1ea-070f-4e78-a0c6-d376174a87f2 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Acquired lock "refresh_cache-ea2e0810-c9f1-49a6-8589-7db402847aea" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1092.446621] env[63515]: DEBUG nova.network.neutron [None req-f1f3c1ea-070f-4e78-a0c6-d376174a87f2 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: ea2e0810-c9f1-49a6-8589-7db402847aea] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1092.518210] env[63515]: DEBUG oslo_concurrency.lockutils [None req-129ac0c9-aca5-49a7-bd9c-0de3f76dcffa tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.178s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1092.518745] env[63515]: DEBUG nova.compute.manager [None req-129ac0c9-aca5-49a7-bd9c-0de3f76dcffa tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] [instance: c8dbf542-70cb-45be-9308-5e12307c0be6] Start building networks asynchronously for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1092.575904] env[63515]: DEBUG nova.network.neutron [-] [instance: 6336f10d-97b3-4413-a828-823a27492215] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1092.601094] env[63515]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-482e81ef-1489-4c7f-a4c7-af9ee905d862 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.611369] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d47f6eb-7e69-4209-acde-123674fa997c {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.635312] env[63515]: DEBUG nova.compute.manager [req-7faf5faf-b9df-4556-a650-a7991b2b1793 req-7996c990-5d0d-4e34-8560-7912e43f90bc service nova] [instance: 6336f10d-97b3-4413-a828-823a27492215] Detach interface failed, port_id=635a24b0-e5a3-43bf-989f-b7bdbc07f767, reason: Instance 6336f10d-97b3-4413-a828-823a27492215 could not be found. {{(pid=63515) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1092.830612] env[63515]: DEBUG nova.compute.manager [None req-8f2dbdaa-d8cb-444f-a302-4aa071b6c146 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: aa83d035-9c47-434d-ab1a-140e6b3110b4] Stashing vm_state: active {{(pid=63515) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 1092.931233] env[63515]: DEBUG nova.compute.manager [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] [instance: 61cf2f9c-5d35-46c4-a477-d2cef9fcd5ae] Instance has had 0 of 5 cleanup attempts {{(pid=63515) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1092.979382] env[63515]: DEBUG nova.network.neutron [None req-f1f3c1ea-070f-4e78-a0c6-d376174a87f2 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: ea2e0810-c9f1-49a6-8589-7db402847aea] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1093.024468] env[63515]: DEBUG nova.compute.utils [None req-129ac0c9-aca5-49a7-bd9c-0de3f76dcffa tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Using /dev/sd instead of None {{(pid=63515) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1093.025822] env[63515]: DEBUG nova.compute.manager [None req-129ac0c9-aca5-49a7-bd9c-0de3f76dcffa tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] [instance: c8dbf542-70cb-45be-9308-5e12307c0be6] Allocating IP information in the background. {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1093.025991] env[63515]: DEBUG nova.network.neutron [None req-129ac0c9-aca5-49a7-bd9c-0de3f76dcffa tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] [instance: c8dbf542-70cb-45be-9308-5e12307c0be6] allocate_for_instance() {{(pid=63515) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1093.066931] env[63515]: DEBUG nova.policy [None req-129ac0c9-aca5-49a7-bd9c-0de3f76dcffa tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'cb4f41c6bfb94aa8ab60135e1d729a6a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '97f2e119708746038f38359572c99438', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63515) authorize /opt/stack/nova/nova/policy.py:201}} [ 1093.077848] env[63515]: INFO nova.compute.manager [-] [instance: 6336f10d-97b3-4413-a828-823a27492215] Took 1.22 seconds to deallocate network for instance. [ 1093.106019] env[63515]: DEBUG nova.network.neutron [None req-f1f3c1ea-070f-4e78-a0c6-d376174a87f2 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: ea2e0810-c9f1-49a6-8589-7db402847aea] Updating instance_info_cache with network_info: [{"id": "4e1d011e-2609-4c71-8317-2685198b6a8b", "address": "fa:16:3e:fb:cb:4f", "network": {"id": "cec614cb-7a5b-4b16-a83d-061ff62d6677", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1501468300-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "aef947bfaf7d4180b46b4be662e3d903", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3a0ddd7d-c321-4187-bdd8-b19044ea2c4a", "external-id": "nsx-vlan-transportzone-747", "segmentation_id": 747, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4e1d011e-26", "ovs_interfaceid": "4e1d011e-2609-4c71-8317-2685198b6a8b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1093.329818] env[63515]: DEBUG nova.network.neutron [None req-129ac0c9-aca5-49a7-bd9c-0de3f76dcffa tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] [instance: c8dbf542-70cb-45be-9308-5e12307c0be6] Successfully created port: 5569cb8c-d9c8-4cb5-abdc-5890ad7d7ede {{(pid=63515) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1093.350672] env[63515]: DEBUG oslo_concurrency.lockutils [None req-8f2dbdaa-d8cb-444f-a302-4aa071b6c146 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1093.350935] env[63515]: DEBUG oslo_concurrency.lockutils [None req-8f2dbdaa-d8cb-444f-a302-4aa071b6c146 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1093.435086] env[63515]: DEBUG nova.compute.manager [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] [instance: 6a66dd9b-7b24-4b70-b58f-0830cf39376c] Instance has had 0 of 5 cleanup attempts {{(pid=63515) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1093.528911] env[63515]: DEBUG nova.compute.manager [None req-129ac0c9-aca5-49a7-bd9c-0de3f76dcffa tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] [instance: c8dbf542-70cb-45be-9308-5e12307c0be6] Start building block device mappings for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1093.584582] env[63515]: DEBUG oslo_concurrency.lockutils [None req-a445eac3-9e94-4d2f-a9a7-82f7ac36cc73 tempest-InstanceActionsV221TestJSON-505260833 tempest-InstanceActionsV221TestJSON-505260833-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1093.608700] env[63515]: DEBUG oslo_concurrency.lockutils [None req-f1f3c1ea-070f-4e78-a0c6-d376174a87f2 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Releasing lock "refresh_cache-ea2e0810-c9f1-49a6-8589-7db402847aea" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1093.609048] env[63515]: DEBUG nova.compute.manager [None req-f1f3c1ea-070f-4e78-a0c6-d376174a87f2 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: ea2e0810-c9f1-49a6-8589-7db402847aea] Instance network_info: |[{"id": "4e1d011e-2609-4c71-8317-2685198b6a8b", "address": "fa:16:3e:fb:cb:4f", "network": {"id": "cec614cb-7a5b-4b16-a83d-061ff62d6677", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1501468300-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "aef947bfaf7d4180b46b4be662e3d903", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3a0ddd7d-c321-4187-bdd8-b19044ea2c4a", "external-id": "nsx-vlan-transportzone-747", "segmentation_id": 747, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4e1d011e-26", "ovs_interfaceid": "4e1d011e-2609-4c71-8317-2685198b6a8b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1093.609519] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-f1f3c1ea-070f-4e78-a0c6-d376174a87f2 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: ea2e0810-c9f1-49a6-8589-7db402847aea] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:fb:cb:4f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '3a0ddd7d-c321-4187-bdd8-b19044ea2c4a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '4e1d011e-2609-4c71-8317-2685198b6a8b', 'vif_model': 'vmxnet3'}] {{(pid=63515) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1093.617568] env[63515]: DEBUG oslo.service.loopingcall [None req-f1f3c1ea-070f-4e78-a0c6-d376174a87f2 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1093.617782] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ea2e0810-c9f1-49a6-8589-7db402847aea] Creating VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1093.618073] env[63515]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b6496e87-5ef4-4085-9179-2f6f4c5b6c6e {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.638793] env[63515]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1093.638793] env[63515]: value = "task-1111944" [ 1093.638793] env[63515]: _type = "Task" [ 1093.638793] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1093.646594] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1111944, 'name': CreateVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1093.856093] env[63515]: INFO nova.compute.claims [None req-8f2dbdaa-d8cb-444f-a302-4aa071b6c146 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: aa83d035-9c47-434d-ab1a-140e6b3110b4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1093.938791] env[63515]: DEBUG nova.compute.manager [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] [instance: d024b6b2-dd10-4112-89e8-ced57efe8208] Instance has had 0 of 5 cleanup attempts {{(pid=63515) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1093.959233] env[63515]: DEBUG nova.compute.manager [req-038a7cea-6a9f-4f92-ad35-34e63c6fb02a req-fa622182-0d6b-4185-a6e1-ffa47a2fe957 service nova] [instance: ea2e0810-c9f1-49a6-8589-7db402847aea] Received event network-changed-4e1d011e-2609-4c71-8317-2685198b6a8b {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1093.959507] env[63515]: DEBUG nova.compute.manager [req-038a7cea-6a9f-4f92-ad35-34e63c6fb02a req-fa622182-0d6b-4185-a6e1-ffa47a2fe957 service nova] [instance: ea2e0810-c9f1-49a6-8589-7db402847aea] Refreshing instance network info cache due to event network-changed-4e1d011e-2609-4c71-8317-2685198b6a8b. {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1093.959741] env[63515]: DEBUG oslo_concurrency.lockutils [req-038a7cea-6a9f-4f92-ad35-34e63c6fb02a req-fa622182-0d6b-4185-a6e1-ffa47a2fe957 service nova] Acquiring lock "refresh_cache-ea2e0810-c9f1-49a6-8589-7db402847aea" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1093.960244] env[63515]: DEBUG oslo_concurrency.lockutils [req-038a7cea-6a9f-4f92-ad35-34e63c6fb02a req-fa622182-0d6b-4185-a6e1-ffa47a2fe957 service nova] Acquired lock "refresh_cache-ea2e0810-c9f1-49a6-8589-7db402847aea" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1093.960244] env[63515]: DEBUG nova.network.neutron [req-038a7cea-6a9f-4f92-ad35-34e63c6fb02a req-fa622182-0d6b-4185-a6e1-ffa47a2fe957 service nova] [instance: ea2e0810-c9f1-49a6-8589-7db402847aea] Refreshing network info cache for port 4e1d011e-2609-4c71-8317-2685198b6a8b {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1094.149539] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1111944, 'name': CreateVM_Task, 'duration_secs': 0.306738} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1094.149695] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ea2e0810-c9f1-49a6-8589-7db402847aea] Created VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1094.150393] env[63515]: DEBUG oslo_concurrency.lockutils [None req-f1f3c1ea-070f-4e78-a0c6-d376174a87f2 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1094.150572] env[63515]: DEBUG oslo_concurrency.lockutils [None req-f1f3c1ea-070f-4e78-a0c6-d376174a87f2 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1094.150910] env[63515]: DEBUG oslo_concurrency.lockutils [None req-f1f3c1ea-070f-4e78-a0c6-d376174a87f2 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1094.151175] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8669efd2-897e-4f72-8455-a61ab12d1499 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.155398] env[63515]: DEBUG oslo_vmware.api [None req-f1f3c1ea-070f-4e78-a0c6-d376174a87f2 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Waiting for the task: (returnval){ [ 1094.155398] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]521aef25-08d9-e561-c1b8-e7692f51d1eb" [ 1094.155398] env[63515]: _type = "Task" [ 1094.155398] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1094.162450] env[63515]: DEBUG oslo_vmware.api [None req-f1f3c1ea-070f-4e78-a0c6-d376174a87f2 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]521aef25-08d9-e561-c1b8-e7692f51d1eb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1094.363980] env[63515]: INFO nova.compute.resource_tracker [None req-8f2dbdaa-d8cb-444f-a302-4aa071b6c146 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: aa83d035-9c47-434d-ab1a-140e6b3110b4] Updating resource usage from migration d9a7b867-b152-403e-8f62-d0daeb836c83 [ 1094.436418] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c07adae-126b-403e-8adc-ced8993ca81f {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.443371] env[63515]: DEBUG nova.compute.manager [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] [instance: d438dd30-b816-48a6-94c6-11e6e9bd5b71] Instance has had 0 of 5 cleanup attempts {{(pid=63515) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1094.445847] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4472488-56f5-44c6-a34c-6bf582a56dce {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.479084] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d36fa376-bcba-4514-b2c7-21dad1de993c {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.487449] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fbcf3394-31c6-4d32-89cc-fa53e0621fa6 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.500776] env[63515]: DEBUG nova.compute.provider_tree [None req-8f2dbdaa-d8cb-444f-a302-4aa071b6c146 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1094.539415] env[63515]: DEBUG nova.compute.manager [None req-129ac0c9-aca5-49a7-bd9c-0de3f76dcffa tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] [instance: c8dbf542-70cb-45be-9308-5e12307c0be6] Start spawning the instance on the hypervisor. {{(pid=63515) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1094.562896] env[63515]: DEBUG nova.virt.hardware [None req-129ac0c9-aca5-49a7-bd9c-0de3f76dcffa tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T02:52:14Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T02:51:56Z,direct_url=,disk_format='vmdk',id=8a120570-cb06-4099-b262-554ca0ad15c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b569255cc43e42e7ae3f2b4ad37c6ef4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T02:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1094.563174] env[63515]: DEBUG nova.virt.hardware [None req-129ac0c9-aca5-49a7-bd9c-0de3f76dcffa tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Flavor limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1094.563340] env[63515]: DEBUG nova.virt.hardware [None req-129ac0c9-aca5-49a7-bd9c-0de3f76dcffa tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Image limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1094.563523] env[63515]: DEBUG nova.virt.hardware [None req-129ac0c9-aca5-49a7-bd9c-0de3f76dcffa tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Flavor pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1094.563675] env[63515]: DEBUG nova.virt.hardware [None req-129ac0c9-aca5-49a7-bd9c-0de3f76dcffa tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Image pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1094.563828] env[63515]: DEBUG nova.virt.hardware [None req-129ac0c9-aca5-49a7-bd9c-0de3f76dcffa tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1094.564055] env[63515]: DEBUG nova.virt.hardware [None req-129ac0c9-aca5-49a7-bd9c-0de3f76dcffa tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1094.564235] env[63515]: DEBUG nova.virt.hardware [None req-129ac0c9-aca5-49a7-bd9c-0de3f76dcffa tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1094.564409] env[63515]: DEBUG nova.virt.hardware [None req-129ac0c9-aca5-49a7-bd9c-0de3f76dcffa tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Got 1 possible topologies {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1094.564574] env[63515]: DEBUG nova.virt.hardware [None req-129ac0c9-aca5-49a7-bd9c-0de3f76dcffa tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1094.564749] env[63515]: DEBUG nova.virt.hardware [None req-129ac0c9-aca5-49a7-bd9c-0de3f76dcffa tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1094.565862] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4c22bd7-0c60-4eec-ac1e-779290c3fb25 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.573318] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39cb4bdf-6874-47e5-8b3c-9facf64b75a4 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.664794] env[63515]: DEBUG oslo_vmware.api [None req-f1f3c1ea-070f-4e78-a0c6-d376174a87f2 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]521aef25-08d9-e561-c1b8-e7692f51d1eb, 'name': SearchDatastore_Task, 'duration_secs': 0.011415} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1094.665111] env[63515]: DEBUG oslo_concurrency.lockutils [None req-f1f3c1ea-070f-4e78-a0c6-d376174a87f2 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1094.665353] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-f1f3c1ea-070f-4e78-a0c6-d376174a87f2 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: ea2e0810-c9f1-49a6-8589-7db402847aea] Processing image 8a120570-cb06-4099-b262-554ca0ad15c5 {{(pid=63515) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1094.665588] env[63515]: DEBUG oslo_concurrency.lockutils [None req-f1f3c1ea-070f-4e78-a0c6-d376174a87f2 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1094.665738] env[63515]: DEBUG oslo_concurrency.lockutils [None req-f1f3c1ea-070f-4e78-a0c6-d376174a87f2 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1094.665924] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-f1f3c1ea-070f-4e78-a0c6-d376174a87f2 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1094.666219] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-7fdd8195-9603-4d86-8a3a-b2a3e3b4447d {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.674109] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-f1f3c1ea-070f-4e78-a0c6-d376174a87f2 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1094.674285] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-f1f3c1ea-070f-4e78-a0c6-d376174a87f2 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63515) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1094.675526] env[63515]: DEBUG nova.network.neutron [req-038a7cea-6a9f-4f92-ad35-34e63c6fb02a req-fa622182-0d6b-4185-a6e1-ffa47a2fe957 service nova] [instance: ea2e0810-c9f1-49a6-8589-7db402847aea] Updated VIF entry in instance network info cache for port 4e1d011e-2609-4c71-8317-2685198b6a8b. {{(pid=63515) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1094.675887] env[63515]: DEBUG nova.network.neutron [req-038a7cea-6a9f-4f92-ad35-34e63c6fb02a req-fa622182-0d6b-4185-a6e1-ffa47a2fe957 service nova] [instance: ea2e0810-c9f1-49a6-8589-7db402847aea] Updating instance_info_cache with network_info: [{"id": "4e1d011e-2609-4c71-8317-2685198b6a8b", "address": "fa:16:3e:fb:cb:4f", "network": {"id": "cec614cb-7a5b-4b16-a83d-061ff62d6677", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1501468300-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "aef947bfaf7d4180b46b4be662e3d903", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3a0ddd7d-c321-4187-bdd8-b19044ea2c4a", "external-id": "nsx-vlan-transportzone-747", "segmentation_id": 747, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4e1d011e-26", "ovs_interfaceid": "4e1d011e-2609-4c71-8317-2685198b6a8b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1094.677014] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ce221cee-e12f-482b-9b18-d694efdb9608 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.682603] env[63515]: DEBUG oslo_vmware.api [None req-f1f3c1ea-070f-4e78-a0c6-d376174a87f2 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Waiting for the task: (returnval){ [ 1094.682603] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]52de8ac4-2dc5-d230-2838-dc52be320d1b" [ 1094.682603] env[63515]: _type = "Task" [ 1094.682603] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1094.690462] env[63515]: DEBUG oslo_vmware.api [None req-f1f3c1ea-070f-4e78-a0c6-d376174a87f2 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52de8ac4-2dc5-d230-2838-dc52be320d1b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1094.950671] env[63515]: DEBUG nova.compute.manager [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] [instance: 75f4dc79-4b73-4ddc-be03-3653a5a0797e] Instance has had 0 of 5 cleanup attempts {{(pid=63515) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1095.005129] env[63515]: DEBUG nova.scheduler.client.report [None req-8f2dbdaa-d8cb-444f-a302-4aa071b6c146 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1095.032773] env[63515]: DEBUG nova.network.neutron [None req-129ac0c9-aca5-49a7-bd9c-0de3f76dcffa tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] [instance: c8dbf542-70cb-45be-9308-5e12307c0be6] Successfully updated port: 5569cb8c-d9c8-4cb5-abdc-5890ad7d7ede {{(pid=63515) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1095.181265] env[63515]: DEBUG oslo_concurrency.lockutils [req-038a7cea-6a9f-4f92-ad35-34e63c6fb02a req-fa622182-0d6b-4185-a6e1-ffa47a2fe957 service nova] Releasing lock "refresh_cache-ea2e0810-c9f1-49a6-8589-7db402847aea" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1095.194132] env[63515]: DEBUG oslo_vmware.api [None req-f1f3c1ea-070f-4e78-a0c6-d376174a87f2 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52de8ac4-2dc5-d230-2838-dc52be320d1b, 'name': SearchDatastore_Task, 'duration_secs': 0.008971} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1095.194939] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-10153923-7aa5-46f4-bd40-03d352f3fb64 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.199898] env[63515]: DEBUG oslo_vmware.api [None req-f1f3c1ea-070f-4e78-a0c6-d376174a87f2 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Waiting for the task: (returnval){ [ 1095.199898] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]5289563e-9554-ef4a-7434-f5ec733b6627" [ 1095.199898] env[63515]: _type = "Task" [ 1095.199898] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1095.208428] env[63515]: DEBUG oslo_vmware.api [None req-f1f3c1ea-070f-4e78-a0c6-d376174a87f2 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]5289563e-9554-ef4a-7434-f5ec733b6627, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1095.454191] env[63515]: DEBUG nova.compute.manager [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] [instance: 918d5b35-c46d-47c5-b59f-175831aed9d6] Instance has had 0 of 5 cleanup attempts {{(pid=63515) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1095.509757] env[63515]: DEBUG oslo_concurrency.lockutils [None req-8f2dbdaa-d8cb-444f-a302-4aa071b6c146 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.159s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1095.509971] env[63515]: INFO nova.compute.manager [None req-8f2dbdaa-d8cb-444f-a302-4aa071b6c146 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: aa83d035-9c47-434d-ab1a-140e6b3110b4] Migrating [ 1095.516526] env[63515]: DEBUG oslo_concurrency.lockutils [None req-a445eac3-9e94-4d2f-a9a7-82f7ac36cc73 tempest-InstanceActionsV221TestJSON-505260833 tempest-InstanceActionsV221TestJSON-505260833-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.932s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1095.516765] env[63515]: DEBUG nova.objects.instance [None req-a445eac3-9e94-4d2f-a9a7-82f7ac36cc73 tempest-InstanceActionsV221TestJSON-505260833 tempest-InstanceActionsV221TestJSON-505260833-project-member] Lazy-loading 'resources' on Instance uuid 6336f10d-97b3-4413-a828-823a27492215 {{(pid=63515) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1095.534806] env[63515]: DEBUG oslo_concurrency.lockutils [None req-129ac0c9-aca5-49a7-bd9c-0de3f76dcffa tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Acquiring lock "refresh_cache-c8dbf542-70cb-45be-9308-5e12307c0be6" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1095.534960] env[63515]: DEBUG oslo_concurrency.lockutils [None req-129ac0c9-aca5-49a7-bd9c-0de3f76dcffa tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Acquired lock "refresh_cache-c8dbf542-70cb-45be-9308-5e12307c0be6" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1095.535154] env[63515]: DEBUG nova.network.neutron [None req-129ac0c9-aca5-49a7-bd9c-0de3f76dcffa tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] [instance: c8dbf542-70cb-45be-9308-5e12307c0be6] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1095.710202] env[63515]: DEBUG oslo_vmware.api [None req-f1f3c1ea-070f-4e78-a0c6-d376174a87f2 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]5289563e-9554-ef4a-7434-f5ec733b6627, 'name': SearchDatastore_Task, 'duration_secs': 0.00875} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1095.710472] env[63515]: DEBUG oslo_concurrency.lockutils [None req-f1f3c1ea-070f-4e78-a0c6-d376174a87f2 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1095.710738] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-f1f3c1ea-070f-4e78-a0c6-d376174a87f2 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk to [datastore1] ea2e0810-c9f1-49a6-8589-7db402847aea/ea2e0810-c9f1-49a6-8589-7db402847aea.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1095.711012] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-fc84ee5e-7a87-421b-8cdc-e272fbe1e26d {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.717256] env[63515]: DEBUG oslo_vmware.api [None req-f1f3c1ea-070f-4e78-a0c6-d376174a87f2 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Waiting for the task: (returnval){ [ 1095.717256] env[63515]: value = "task-1111945" [ 1095.717256] env[63515]: _type = "Task" [ 1095.717256] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1095.726355] env[63515]: DEBUG oslo_vmware.api [None req-f1f3c1ea-070f-4e78-a0c6-d376174a87f2 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Task: {'id': task-1111945, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1095.959106] env[63515]: DEBUG nova.compute.manager [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] [instance: 7d572ac0-f6af-4622-96cc-e75983420222] Instance has had 0 of 5 cleanup attempts {{(pid=63515) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1095.987859] env[63515]: DEBUG nova.compute.manager [req-09ca46fa-3cec-4dbe-b17c-f33f461d8d9f req-bda210ca-6367-4b3b-82cb-d9a4e6cd0b74 service nova] [instance: c8dbf542-70cb-45be-9308-5e12307c0be6] Received event network-vif-plugged-5569cb8c-d9c8-4cb5-abdc-5890ad7d7ede {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1095.988335] env[63515]: DEBUG oslo_concurrency.lockutils [req-09ca46fa-3cec-4dbe-b17c-f33f461d8d9f req-bda210ca-6367-4b3b-82cb-d9a4e6cd0b74 service nova] Acquiring lock "c8dbf542-70cb-45be-9308-5e12307c0be6-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1095.988438] env[63515]: DEBUG oslo_concurrency.lockutils [req-09ca46fa-3cec-4dbe-b17c-f33f461d8d9f req-bda210ca-6367-4b3b-82cb-d9a4e6cd0b74 service nova] Lock "c8dbf542-70cb-45be-9308-5e12307c0be6-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1095.988704] env[63515]: DEBUG oslo_concurrency.lockutils [req-09ca46fa-3cec-4dbe-b17c-f33f461d8d9f req-bda210ca-6367-4b3b-82cb-d9a4e6cd0b74 service nova] Lock "c8dbf542-70cb-45be-9308-5e12307c0be6-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1095.988886] env[63515]: DEBUG nova.compute.manager [req-09ca46fa-3cec-4dbe-b17c-f33f461d8d9f req-bda210ca-6367-4b3b-82cb-d9a4e6cd0b74 service nova] [instance: c8dbf542-70cb-45be-9308-5e12307c0be6] No waiting events found dispatching network-vif-plugged-5569cb8c-d9c8-4cb5-abdc-5890ad7d7ede {{(pid=63515) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1095.989210] env[63515]: WARNING nova.compute.manager [req-09ca46fa-3cec-4dbe-b17c-f33f461d8d9f req-bda210ca-6367-4b3b-82cb-d9a4e6cd0b74 service nova] [instance: c8dbf542-70cb-45be-9308-5e12307c0be6] Received unexpected event network-vif-plugged-5569cb8c-d9c8-4cb5-abdc-5890ad7d7ede for instance with vm_state building and task_state spawning. [ 1095.989349] env[63515]: DEBUG nova.compute.manager [req-09ca46fa-3cec-4dbe-b17c-f33f461d8d9f req-bda210ca-6367-4b3b-82cb-d9a4e6cd0b74 service nova] [instance: c8dbf542-70cb-45be-9308-5e12307c0be6] Received event network-changed-5569cb8c-d9c8-4cb5-abdc-5890ad7d7ede {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1095.989597] env[63515]: DEBUG nova.compute.manager [req-09ca46fa-3cec-4dbe-b17c-f33f461d8d9f req-bda210ca-6367-4b3b-82cb-d9a4e6cd0b74 service nova] [instance: c8dbf542-70cb-45be-9308-5e12307c0be6] Refreshing instance network info cache due to event network-changed-5569cb8c-d9c8-4cb5-abdc-5890ad7d7ede. {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1095.989861] env[63515]: DEBUG oslo_concurrency.lockutils [req-09ca46fa-3cec-4dbe-b17c-f33f461d8d9f req-bda210ca-6367-4b3b-82cb-d9a4e6cd0b74 service nova] Acquiring lock "refresh_cache-c8dbf542-70cb-45be-9308-5e12307c0be6" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1096.025731] env[63515]: DEBUG oslo_concurrency.lockutils [None req-8f2dbdaa-d8cb-444f-a302-4aa071b6c146 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Acquiring lock "refresh_cache-aa83d035-9c47-434d-ab1a-140e6b3110b4" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1096.025975] env[63515]: DEBUG oslo_concurrency.lockutils [None req-8f2dbdaa-d8cb-444f-a302-4aa071b6c146 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Acquired lock "refresh_cache-aa83d035-9c47-434d-ab1a-140e6b3110b4" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1096.026233] env[63515]: DEBUG nova.network.neutron [None req-8f2dbdaa-d8cb-444f-a302-4aa071b6c146 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: aa83d035-9c47-434d-ab1a-140e6b3110b4] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1096.068479] env[63515]: DEBUG nova.network.neutron [None req-129ac0c9-aca5-49a7-bd9c-0de3f76dcffa tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] [instance: c8dbf542-70cb-45be-9308-5e12307c0be6] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1096.132367] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22d11880-30a3-4130-86d8-5271683b8730 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.141492] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b861a1b0-2277-43e8-9360-cc0997e14dd9 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.179317] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-865ad523-2bdb-438e-b7bb-dcd109ea2e7a {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.188287] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3dd69179-3507-4215-b6f8-b781fe1b8c5b {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.203126] env[63515]: DEBUG nova.compute.provider_tree [None req-a445eac3-9e94-4d2f-a9a7-82f7ac36cc73 tempest-InstanceActionsV221TestJSON-505260833 tempest-InstanceActionsV221TestJSON-505260833-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1096.229307] env[63515]: DEBUG oslo_vmware.api [None req-f1f3c1ea-070f-4e78-a0c6-d376174a87f2 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Task: {'id': task-1111945, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1096.302468] env[63515]: DEBUG nova.network.neutron [None req-129ac0c9-aca5-49a7-bd9c-0de3f76dcffa tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] [instance: c8dbf542-70cb-45be-9308-5e12307c0be6] Updating instance_info_cache with network_info: [{"id": "5569cb8c-d9c8-4cb5-abdc-5890ad7d7ede", "address": "fa:16:3e:8d:b8:b1", "network": {"id": "d6da2650-9ba9-4ef6-8035-e58b635a64b2", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-391564968-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "97f2e119708746038f38359572c99438", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "043ca97b-0fca-4b54-8be6-027123fa76d0", "external-id": "nsx-vlan-transportzone-702", "segmentation_id": 702, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5569cb8c-d9", "ovs_interfaceid": "5569cb8c-d9c8-4cb5-abdc-5890ad7d7ede", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1096.462105] env[63515]: DEBUG nova.compute.manager [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] [instance: b25bf2ba-cf08-44ac-a524-aa87cf785575] Instance has had 0 of 5 cleanup attempts {{(pid=63515) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1096.707875] env[63515]: DEBUG nova.scheduler.client.report [None req-a445eac3-9e94-4d2f-a9a7-82f7ac36cc73 tempest-InstanceActionsV221TestJSON-505260833 tempest-InstanceActionsV221TestJSON-505260833-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1096.728672] env[63515]: DEBUG oslo_vmware.api [None req-f1f3c1ea-070f-4e78-a0c6-d376174a87f2 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Task: {'id': task-1111945, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.533421} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1096.728932] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-f1f3c1ea-070f-4e78-a0c6-d376174a87f2 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk to [datastore1] ea2e0810-c9f1-49a6-8589-7db402847aea/ea2e0810-c9f1-49a6-8589-7db402847aea.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1096.729208] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-f1f3c1ea-070f-4e78-a0c6-d376174a87f2 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: ea2e0810-c9f1-49a6-8589-7db402847aea] Extending root virtual disk to 1048576 {{(pid=63515) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1096.729465] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-966aebde-869c-42c9-86e3-65a181ead60e {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.737176] env[63515]: DEBUG oslo_vmware.api [None req-f1f3c1ea-070f-4e78-a0c6-d376174a87f2 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Waiting for the task: (returnval){ [ 1096.737176] env[63515]: value = "task-1111946" [ 1096.737176] env[63515]: _type = "Task" [ 1096.737176] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1096.745516] env[63515]: DEBUG oslo_vmware.api [None req-f1f3c1ea-070f-4e78-a0c6-d376174a87f2 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Task: {'id': task-1111946, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1096.748071] env[63515]: DEBUG nova.network.neutron [None req-8f2dbdaa-d8cb-444f-a302-4aa071b6c146 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: aa83d035-9c47-434d-ab1a-140e6b3110b4] Updating instance_info_cache with network_info: [{"id": "614aed8f-3ab8-4672-ac13-0ae8cb0b233d", "address": "fa:16:3e:98:68:38", "network": {"id": "5758c885-e1c9-45c8-86aa-27a802101f52", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-734458138-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.204", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b18d4daacbc84e758f9f9ae4f3f09d28", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69351262-8d39-441a-85ba-3a78df436d17", "external-id": "nsx-vlan-transportzone-205", "segmentation_id": 205, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap614aed8f-3a", "ovs_interfaceid": "614aed8f-3ab8-4672-ac13-0ae8cb0b233d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1096.805366] env[63515]: DEBUG oslo_concurrency.lockutils [None req-129ac0c9-aca5-49a7-bd9c-0de3f76dcffa tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Releasing lock "refresh_cache-c8dbf542-70cb-45be-9308-5e12307c0be6" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1096.805734] env[63515]: DEBUG nova.compute.manager [None req-129ac0c9-aca5-49a7-bd9c-0de3f76dcffa tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] [instance: c8dbf542-70cb-45be-9308-5e12307c0be6] Instance network_info: |[{"id": "5569cb8c-d9c8-4cb5-abdc-5890ad7d7ede", "address": "fa:16:3e:8d:b8:b1", "network": {"id": "d6da2650-9ba9-4ef6-8035-e58b635a64b2", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-391564968-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "97f2e119708746038f38359572c99438", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "043ca97b-0fca-4b54-8be6-027123fa76d0", "external-id": "nsx-vlan-transportzone-702", "segmentation_id": 702, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5569cb8c-d9", "ovs_interfaceid": "5569cb8c-d9c8-4cb5-abdc-5890ad7d7ede", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1096.806095] env[63515]: DEBUG oslo_concurrency.lockutils [req-09ca46fa-3cec-4dbe-b17c-f33f461d8d9f req-bda210ca-6367-4b3b-82cb-d9a4e6cd0b74 service nova] Acquired lock "refresh_cache-c8dbf542-70cb-45be-9308-5e12307c0be6" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1096.806289] env[63515]: DEBUG nova.network.neutron [req-09ca46fa-3cec-4dbe-b17c-f33f461d8d9f req-bda210ca-6367-4b3b-82cb-d9a4e6cd0b74 service nova] [instance: c8dbf542-70cb-45be-9308-5e12307c0be6] Refreshing network info cache for port 5569cb8c-d9c8-4cb5-abdc-5890ad7d7ede {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1096.807492] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-129ac0c9-aca5-49a7-bd9c-0de3f76dcffa tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] [instance: c8dbf542-70cb-45be-9308-5e12307c0be6] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:8d:b8:b1', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '043ca97b-0fca-4b54-8be6-027123fa76d0', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '5569cb8c-d9c8-4cb5-abdc-5890ad7d7ede', 'vif_model': 'vmxnet3'}] {{(pid=63515) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1096.814922] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-129ac0c9-aca5-49a7-bd9c-0de3f76dcffa tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Creating folder: Project (97f2e119708746038f38359572c99438). Parent ref: group-v243370. {{(pid=63515) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1096.816022] env[63515]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-013256e6-ce28-494f-9cf1-efd5ac483973 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.828223] env[63515]: INFO nova.virt.vmwareapi.vm_util [None req-129ac0c9-aca5-49a7-bd9c-0de3f76dcffa tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Created folder: Project (97f2e119708746038f38359572c99438) in parent group-v243370. [ 1096.828456] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-129ac0c9-aca5-49a7-bd9c-0de3f76dcffa tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Creating folder: Instances. Parent ref: group-v243580. {{(pid=63515) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1096.828681] env[63515]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-98fe3a95-cb5a-4ca9-a903-fa51cc195632 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.839328] env[63515]: INFO nova.virt.vmwareapi.vm_util [None req-129ac0c9-aca5-49a7-bd9c-0de3f76dcffa tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Created folder: Instances in parent group-v243580. [ 1096.839572] env[63515]: DEBUG oslo.service.loopingcall [None req-129ac0c9-aca5-49a7-bd9c-0de3f76dcffa tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1096.839763] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c8dbf542-70cb-45be-9308-5e12307c0be6] Creating VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1096.839972] env[63515]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c407b199-271a-4519-9aac-166ff7a6c9a2 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.859945] env[63515]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1096.859945] env[63515]: value = "task-1111949" [ 1096.859945] env[63515]: _type = "Task" [ 1096.859945] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1096.867954] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1111949, 'name': CreateVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1096.966317] env[63515]: DEBUG nova.compute.manager [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] [instance: 014720fe-87b9-4e55-b9ba-295b009c7ff6] Instance has had 0 of 5 cleanup attempts {{(pid=63515) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1097.215050] env[63515]: DEBUG oslo_concurrency.lockutils [None req-a445eac3-9e94-4d2f-a9a7-82f7ac36cc73 tempest-InstanceActionsV221TestJSON-505260833 tempest-InstanceActionsV221TestJSON-505260833-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.698s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1097.237421] env[63515]: INFO nova.scheduler.client.report [None req-a445eac3-9e94-4d2f-a9a7-82f7ac36cc73 tempest-InstanceActionsV221TestJSON-505260833 tempest-InstanceActionsV221TestJSON-505260833-project-member] Deleted allocations for instance 6336f10d-97b3-4413-a828-823a27492215 [ 1097.251741] env[63515]: DEBUG oslo_concurrency.lockutils [None req-8f2dbdaa-d8cb-444f-a302-4aa071b6c146 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Releasing lock "refresh_cache-aa83d035-9c47-434d-ab1a-140e6b3110b4" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1097.252977] env[63515]: DEBUG oslo_vmware.api [None req-f1f3c1ea-070f-4e78-a0c6-d376174a87f2 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Task: {'id': task-1111946, 'name': ExtendVirtualDisk_Task} progress is 50%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1097.372140] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1111949, 'name': CreateVM_Task} progress is 25%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1097.469496] env[63515]: DEBUG nova.compute.manager [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] [instance: 2fa098b6-ecc0-491f-8cfc-5ee335bbf943] Instance has had 0 of 5 cleanup attempts {{(pid=63515) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1097.557553] env[63515]: DEBUG nova.network.neutron [req-09ca46fa-3cec-4dbe-b17c-f33f461d8d9f req-bda210ca-6367-4b3b-82cb-d9a4e6cd0b74 service nova] [instance: c8dbf542-70cb-45be-9308-5e12307c0be6] Updated VIF entry in instance network info cache for port 5569cb8c-d9c8-4cb5-abdc-5890ad7d7ede. {{(pid=63515) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1097.557913] env[63515]: DEBUG nova.network.neutron [req-09ca46fa-3cec-4dbe-b17c-f33f461d8d9f req-bda210ca-6367-4b3b-82cb-d9a4e6cd0b74 service nova] [instance: c8dbf542-70cb-45be-9308-5e12307c0be6] Updating instance_info_cache with network_info: [{"id": "5569cb8c-d9c8-4cb5-abdc-5890ad7d7ede", "address": "fa:16:3e:8d:b8:b1", "network": {"id": "d6da2650-9ba9-4ef6-8035-e58b635a64b2", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-391564968-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "97f2e119708746038f38359572c99438", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "043ca97b-0fca-4b54-8be6-027123fa76d0", "external-id": "nsx-vlan-transportzone-702", "segmentation_id": 702, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5569cb8c-d9", "ovs_interfaceid": "5569cb8c-d9c8-4cb5-abdc-5890ad7d7ede", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1097.750041] env[63515]: DEBUG oslo_vmware.api [None req-f1f3c1ea-070f-4e78-a0c6-d376174a87f2 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Task: {'id': task-1111946, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.73217} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1097.750041] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-f1f3c1ea-070f-4e78-a0c6-d376174a87f2 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: ea2e0810-c9f1-49a6-8589-7db402847aea] Extended root virtual disk {{(pid=63515) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1097.750789] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21a53769-3de8-4325-9a68-78916728f3f0 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.753699] env[63515]: DEBUG oslo_concurrency.lockutils [None req-a445eac3-9e94-4d2f-a9a7-82f7ac36cc73 tempest-InstanceActionsV221TestJSON-505260833 tempest-InstanceActionsV221TestJSON-505260833-project-member] Lock "6336f10d-97b3-4413-a828-823a27492215" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.012s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1097.777535] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-f1f3c1ea-070f-4e78-a0c6-d376174a87f2 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: ea2e0810-c9f1-49a6-8589-7db402847aea] Reconfiguring VM instance instance-0000006a to attach disk [datastore1] ea2e0810-c9f1-49a6-8589-7db402847aea/ea2e0810-c9f1-49a6-8589-7db402847aea.vmdk or device None with type sparse {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1097.777777] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-430f9cf9-4589-4d09-b8af-00b68cd09f96 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.797777] env[63515]: DEBUG oslo_vmware.api [None req-f1f3c1ea-070f-4e78-a0c6-d376174a87f2 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Waiting for the task: (returnval){ [ 1097.797777] env[63515]: value = "task-1111950" [ 1097.797777] env[63515]: _type = "Task" [ 1097.797777] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1097.805760] env[63515]: DEBUG oslo_vmware.api [None req-f1f3c1ea-070f-4e78-a0c6-d376174a87f2 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Task: {'id': task-1111950, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1097.871189] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1111949, 'name': CreateVM_Task} progress is 99%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1097.973041] env[63515]: DEBUG nova.compute.manager [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] [instance: a2203c5e-66b9-4c7e-8d76-e12a08b57b1d] Instance has had 0 of 5 cleanup attempts {{(pid=63515) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1098.060714] env[63515]: DEBUG oslo_concurrency.lockutils [req-09ca46fa-3cec-4dbe-b17c-f33f461d8d9f req-bda210ca-6367-4b3b-82cb-d9a4e6cd0b74 service nova] Releasing lock "refresh_cache-c8dbf542-70cb-45be-9308-5e12307c0be6" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1098.308348] env[63515]: DEBUG oslo_vmware.api [None req-f1f3c1ea-070f-4e78-a0c6-d376174a87f2 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Task: {'id': task-1111950, 'name': ReconfigVM_Task, 'duration_secs': 0.279509} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1098.308683] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-f1f3c1ea-070f-4e78-a0c6-d376174a87f2 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: ea2e0810-c9f1-49a6-8589-7db402847aea] Reconfigured VM instance instance-0000006a to attach disk [datastore1] ea2e0810-c9f1-49a6-8589-7db402847aea/ea2e0810-c9f1-49a6-8589-7db402847aea.vmdk or device None with type sparse {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1098.309469] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-11a819b4-5a7e-4a39-922a-8b433a4cec68 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.316209] env[63515]: DEBUG oslo_vmware.api [None req-f1f3c1ea-070f-4e78-a0c6-d376174a87f2 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Waiting for the task: (returnval){ [ 1098.316209] env[63515]: value = "task-1111951" [ 1098.316209] env[63515]: _type = "Task" [ 1098.316209] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1098.324362] env[63515]: DEBUG oslo_vmware.api [None req-f1f3c1ea-070f-4e78-a0c6-d376174a87f2 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Task: {'id': task-1111951, 'name': Rename_Task} progress is 5%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1098.372246] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1111949, 'name': CreateVM_Task, 'duration_secs': 1.042617} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1098.372513] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c8dbf542-70cb-45be-9308-5e12307c0be6] Created VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1098.373245] env[63515]: DEBUG oslo_concurrency.lockutils [None req-129ac0c9-aca5-49a7-bd9c-0de3f76dcffa tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1098.373480] env[63515]: DEBUG oslo_concurrency.lockutils [None req-129ac0c9-aca5-49a7-bd9c-0de3f76dcffa tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1098.373829] env[63515]: DEBUG oslo_concurrency.lockutils [None req-129ac0c9-aca5-49a7-bd9c-0de3f76dcffa tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1098.374351] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5ae7e772-4dde-494e-abb2-f5d1e34e7a2d {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.378860] env[63515]: DEBUG oslo_vmware.api [None req-129ac0c9-aca5-49a7-bd9c-0de3f76dcffa tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Waiting for the task: (returnval){ [ 1098.378860] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]5206772b-e74d-66b4-1f0a-fc305f1a0b5e" [ 1098.378860] env[63515]: _type = "Task" [ 1098.378860] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1098.391940] env[63515]: DEBUG oslo_vmware.api [None req-129ac0c9-aca5-49a7-bd9c-0de3f76dcffa tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]5206772b-e74d-66b4-1f0a-fc305f1a0b5e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1098.476764] env[63515]: DEBUG nova.compute.manager [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] [instance: ca6a8ac3-d880-4333-9816-2f076e423844] Instance has had 0 of 5 cleanup attempts {{(pid=63515) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1098.769026] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cbd9e297-4451-4d77-8126-901ac3a7c67b {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.787529] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-8f2dbdaa-d8cb-444f-a302-4aa071b6c146 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: aa83d035-9c47-434d-ab1a-140e6b3110b4] Updating instance 'aa83d035-9c47-434d-ab1a-140e6b3110b4' progress to 0 {{(pid=63515) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1098.827772] env[63515]: DEBUG oslo_vmware.api [None req-f1f3c1ea-070f-4e78-a0c6-d376174a87f2 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Task: {'id': task-1111951, 'name': Rename_Task, 'duration_secs': 0.132061} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1098.827772] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-f1f3c1ea-070f-4e78-a0c6-d376174a87f2 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: ea2e0810-c9f1-49a6-8589-7db402847aea] Powering on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1098.827772] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e4ef6422-fe38-43c8-8f7e-c653fdb7f620 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.837107] env[63515]: DEBUG oslo_vmware.api [None req-f1f3c1ea-070f-4e78-a0c6-d376174a87f2 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Waiting for the task: (returnval){ [ 1098.837107] env[63515]: value = "task-1111952" [ 1098.837107] env[63515]: _type = "Task" [ 1098.837107] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1098.843521] env[63515]: DEBUG oslo_vmware.api [None req-f1f3c1ea-070f-4e78-a0c6-d376174a87f2 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Task: {'id': task-1111952, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1098.888964] env[63515]: DEBUG oslo_vmware.api [None req-129ac0c9-aca5-49a7-bd9c-0de3f76dcffa tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]5206772b-e74d-66b4-1f0a-fc305f1a0b5e, 'name': SearchDatastore_Task, 'duration_secs': 0.010793} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1098.889259] env[63515]: DEBUG oslo_concurrency.lockutils [None req-129ac0c9-aca5-49a7-bd9c-0de3f76dcffa tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1098.890529] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-129ac0c9-aca5-49a7-bd9c-0de3f76dcffa tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] [instance: c8dbf542-70cb-45be-9308-5e12307c0be6] Processing image 8a120570-cb06-4099-b262-554ca0ad15c5 {{(pid=63515) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1098.890529] env[63515]: DEBUG oslo_concurrency.lockutils [None req-129ac0c9-aca5-49a7-bd9c-0de3f76dcffa tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1098.890529] env[63515]: DEBUG oslo_concurrency.lockutils [None req-129ac0c9-aca5-49a7-bd9c-0de3f76dcffa tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1098.890529] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-129ac0c9-aca5-49a7-bd9c-0de3f76dcffa tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1098.890529] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6ab31af0-9ef6-41a7-bd6a-2dee6c8af6f6 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.897914] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-129ac0c9-aca5-49a7-bd9c-0de3f76dcffa tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1098.898140] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-129ac0c9-aca5-49a7-bd9c-0de3f76dcffa tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63515) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1098.898822] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c5bb6eb4-0ac2-40f1-9052-1998eb76cca8 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.903784] env[63515]: DEBUG oslo_vmware.api [None req-129ac0c9-aca5-49a7-bd9c-0de3f76dcffa tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Waiting for the task: (returnval){ [ 1098.903784] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]52d5887c-6d66-83f6-7781-997fc4621e76" [ 1098.903784] env[63515]: _type = "Task" [ 1098.903784] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1098.912704] env[63515]: DEBUG oslo_vmware.api [None req-129ac0c9-aca5-49a7-bd9c-0de3f76dcffa tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52d5887c-6d66-83f6-7781-997fc4621e76, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1098.980355] env[63515]: DEBUG nova.compute.manager [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] [instance: 9236607d-47af-4272-93db-79826e918aa1] Instance has had 0 of 5 cleanup attempts {{(pid=63515) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1099.294595] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-8f2dbdaa-d8cb-444f-a302-4aa071b6c146 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: aa83d035-9c47-434d-ab1a-140e6b3110b4] Powering off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1099.294595] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b7c1fbbd-0986-4523-ba8b-8372273014c1 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.302858] env[63515]: DEBUG oslo_vmware.api [None req-8f2dbdaa-d8cb-444f-a302-4aa071b6c146 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Waiting for the task: (returnval){ [ 1099.302858] env[63515]: value = "task-1111953" [ 1099.302858] env[63515]: _type = "Task" [ 1099.302858] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1099.311866] env[63515]: DEBUG oslo_vmware.api [None req-8f2dbdaa-d8cb-444f-a302-4aa071b6c146 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Task: {'id': task-1111953, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1099.345503] env[63515]: DEBUG oslo_vmware.api [None req-f1f3c1ea-070f-4e78-a0c6-d376174a87f2 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Task: {'id': task-1111952, 'name': PowerOnVM_Task, 'duration_secs': 0.496164} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1099.345857] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-f1f3c1ea-070f-4e78-a0c6-d376174a87f2 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: ea2e0810-c9f1-49a6-8589-7db402847aea] Powered on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1099.346119] env[63515]: INFO nova.compute.manager [None req-f1f3c1ea-070f-4e78-a0c6-d376174a87f2 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: ea2e0810-c9f1-49a6-8589-7db402847aea] Took 7.78 seconds to spawn the instance on the hypervisor. [ 1099.346351] env[63515]: DEBUG nova.compute.manager [None req-f1f3c1ea-070f-4e78-a0c6-d376174a87f2 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: ea2e0810-c9f1-49a6-8589-7db402847aea] Checking state {{(pid=63515) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1099.347296] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3fc0e25-fab2-4e28-85e2-c1fba216a3f1 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.416192] env[63515]: DEBUG oslo_vmware.api [None req-129ac0c9-aca5-49a7-bd9c-0de3f76dcffa tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52d5887c-6d66-83f6-7781-997fc4621e76, 'name': SearchDatastore_Task, 'duration_secs': 0.008645} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1099.417185] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9f2b58f2-e940-4dec-b09e-65f7a5834789 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.422989] env[63515]: DEBUG oslo_vmware.api [None req-129ac0c9-aca5-49a7-bd9c-0de3f76dcffa tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Waiting for the task: (returnval){ [ 1099.422989] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]522be735-22a4-60c5-306a-66fb600fbbd5" [ 1099.422989] env[63515]: _type = "Task" [ 1099.422989] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1099.432683] env[63515]: DEBUG oslo_vmware.api [None req-129ac0c9-aca5-49a7-bd9c-0de3f76dcffa tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]522be735-22a4-60c5-306a-66fb600fbbd5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1099.484523] env[63515]: DEBUG nova.compute.manager [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] [instance: 86b1ddcf-585c-41d9-84fe-17b794f1abbc] Instance has had 0 of 5 cleanup attempts {{(pid=63515) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1099.812263] env[63515]: DEBUG oslo_vmware.api [None req-8f2dbdaa-d8cb-444f-a302-4aa071b6c146 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Task: {'id': task-1111953, 'name': PowerOffVM_Task, 'duration_secs': 0.213709} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1099.812617] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-8f2dbdaa-d8cb-444f-a302-4aa071b6c146 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: aa83d035-9c47-434d-ab1a-140e6b3110b4] Powered off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1099.812817] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-8f2dbdaa-d8cb-444f-a302-4aa071b6c146 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: aa83d035-9c47-434d-ab1a-140e6b3110b4] Updating instance 'aa83d035-9c47-434d-ab1a-140e6b3110b4' progress to 17 {{(pid=63515) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1099.863621] env[63515]: INFO nova.compute.manager [None req-f1f3c1ea-070f-4e78-a0c6-d376174a87f2 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: ea2e0810-c9f1-49a6-8589-7db402847aea] Took 14.24 seconds to build instance. [ 1099.933901] env[63515]: DEBUG oslo_vmware.api [None req-129ac0c9-aca5-49a7-bd9c-0de3f76dcffa tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]522be735-22a4-60c5-306a-66fb600fbbd5, 'name': SearchDatastore_Task, 'duration_secs': 0.010233} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1099.934538] env[63515]: DEBUG oslo_concurrency.lockutils [None req-129ac0c9-aca5-49a7-bd9c-0de3f76dcffa tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1099.934819] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-129ac0c9-aca5-49a7-bd9c-0de3f76dcffa tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk to [datastore1] c8dbf542-70cb-45be-9308-5e12307c0be6/c8dbf542-70cb-45be-9308-5e12307c0be6.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1099.935094] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-74387d28-e095-4be2-a82f-1afddd0d4346 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.944607] env[63515]: DEBUG oslo_vmware.api [None req-129ac0c9-aca5-49a7-bd9c-0de3f76dcffa tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Waiting for the task: (returnval){ [ 1099.944607] env[63515]: value = "task-1111954" [ 1099.944607] env[63515]: _type = "Task" [ 1099.944607] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1099.954420] env[63515]: DEBUG oslo_vmware.api [None req-129ac0c9-aca5-49a7-bd9c-0de3f76dcffa tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Task: {'id': task-1111954, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1099.980593] env[63515]: DEBUG nova.compute.manager [req-c841b169-d606-4014-857f-05deb8ddee34 req-3c74ce0b-891f-4e18-8886-6b732e505b06 service nova] [instance: ea2e0810-c9f1-49a6-8589-7db402847aea] Received event network-changed-4e1d011e-2609-4c71-8317-2685198b6a8b {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1099.980897] env[63515]: DEBUG nova.compute.manager [req-c841b169-d606-4014-857f-05deb8ddee34 req-3c74ce0b-891f-4e18-8886-6b732e505b06 service nova] [instance: ea2e0810-c9f1-49a6-8589-7db402847aea] Refreshing instance network info cache due to event network-changed-4e1d011e-2609-4c71-8317-2685198b6a8b. {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1099.981165] env[63515]: DEBUG oslo_concurrency.lockutils [req-c841b169-d606-4014-857f-05deb8ddee34 req-3c74ce0b-891f-4e18-8886-6b732e505b06 service nova] Acquiring lock "refresh_cache-ea2e0810-c9f1-49a6-8589-7db402847aea" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1099.981377] env[63515]: DEBUG oslo_concurrency.lockutils [req-c841b169-d606-4014-857f-05deb8ddee34 req-3c74ce0b-891f-4e18-8886-6b732e505b06 service nova] Acquired lock "refresh_cache-ea2e0810-c9f1-49a6-8589-7db402847aea" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1099.981554] env[63515]: DEBUG nova.network.neutron [req-c841b169-d606-4014-857f-05deb8ddee34 req-3c74ce0b-891f-4e18-8886-6b732e505b06 service nova] [instance: ea2e0810-c9f1-49a6-8589-7db402847aea] Refreshing network info cache for port 4e1d011e-2609-4c71-8317-2685198b6a8b {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1099.987567] env[63515]: DEBUG nova.compute.manager [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] [instance: 829e72d7-a56a-451c-b98a-94253e5900d5] Instance has had 0 of 5 cleanup attempts {{(pid=63515) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1100.319775] env[63515]: DEBUG nova.virt.hardware [None req-8f2dbdaa-d8cb-444f-a302-4aa071b6c146 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T02:52:16Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=8a120570-cb06-4099-b262-554ca0ad15c5,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1100.319969] env[63515]: DEBUG nova.virt.hardware [None req-8f2dbdaa-d8cb-444f-a302-4aa071b6c146 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Flavor limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1100.320530] env[63515]: DEBUG nova.virt.hardware [None req-8f2dbdaa-d8cb-444f-a302-4aa071b6c146 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Image limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1100.320979] env[63515]: DEBUG nova.virt.hardware [None req-8f2dbdaa-d8cb-444f-a302-4aa071b6c146 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Flavor pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1100.321162] env[63515]: DEBUG nova.virt.hardware [None req-8f2dbdaa-d8cb-444f-a302-4aa071b6c146 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Image pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1100.321372] env[63515]: DEBUG nova.virt.hardware [None req-8f2dbdaa-d8cb-444f-a302-4aa071b6c146 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1100.321703] env[63515]: DEBUG nova.virt.hardware [None req-8f2dbdaa-d8cb-444f-a302-4aa071b6c146 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1100.321984] env[63515]: DEBUG nova.virt.hardware [None req-8f2dbdaa-d8cb-444f-a302-4aa071b6c146 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1100.322242] env[63515]: DEBUG nova.virt.hardware [None req-8f2dbdaa-d8cb-444f-a302-4aa071b6c146 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Got 1 possible topologies {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1100.322509] env[63515]: DEBUG nova.virt.hardware [None req-8f2dbdaa-d8cb-444f-a302-4aa071b6c146 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1100.322723] env[63515]: DEBUG nova.virt.hardware [None req-8f2dbdaa-d8cb-444f-a302-4aa071b6c146 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1100.327965] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c5a90587-11d0-49dc-a1b6-50a10a9fd5a2 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.347082] env[63515]: DEBUG oslo_vmware.api [None req-8f2dbdaa-d8cb-444f-a302-4aa071b6c146 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Waiting for the task: (returnval){ [ 1100.347082] env[63515]: value = "task-1111955" [ 1100.347082] env[63515]: _type = "Task" [ 1100.347082] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1100.356555] env[63515]: DEBUG oslo_vmware.api [None req-8f2dbdaa-d8cb-444f-a302-4aa071b6c146 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Task: {'id': task-1111955, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1100.365374] env[63515]: DEBUG oslo_concurrency.lockutils [None req-f1f3c1ea-070f-4e78-a0c6-d376174a87f2 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Lock "ea2e0810-c9f1-49a6-8589-7db402847aea" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.755s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1100.456050] env[63515]: DEBUG oslo_vmware.api [None req-129ac0c9-aca5-49a7-bd9c-0de3f76dcffa tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Task: {'id': task-1111954, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.462009} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1100.456462] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-129ac0c9-aca5-49a7-bd9c-0de3f76dcffa tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk to [datastore1] c8dbf542-70cb-45be-9308-5e12307c0be6/c8dbf542-70cb-45be-9308-5e12307c0be6.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1100.456637] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-129ac0c9-aca5-49a7-bd9c-0de3f76dcffa tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] [instance: c8dbf542-70cb-45be-9308-5e12307c0be6] Extending root virtual disk to 1048576 {{(pid=63515) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1100.457013] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-0ff421d8-8b07-4c9c-9f71-c8a6b550138e {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.464264] env[63515]: DEBUG oslo_vmware.api [None req-129ac0c9-aca5-49a7-bd9c-0de3f76dcffa tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Waiting for the task: (returnval){ [ 1100.464264] env[63515]: value = "task-1111956" [ 1100.464264] env[63515]: _type = "Task" [ 1100.464264] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1100.473295] env[63515]: DEBUG oslo_vmware.api [None req-129ac0c9-aca5-49a7-bd9c-0de3f76dcffa tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Task: {'id': task-1111956, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1100.490890] env[63515]: DEBUG nova.compute.manager [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] [instance: 324f7e89-3c17-474d-af1b-6d7a6d042510] Instance has had 0 of 5 cleanup attempts {{(pid=63515) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1100.691868] env[63515]: DEBUG nova.network.neutron [req-c841b169-d606-4014-857f-05deb8ddee34 req-3c74ce0b-891f-4e18-8886-6b732e505b06 service nova] [instance: ea2e0810-c9f1-49a6-8589-7db402847aea] Updated VIF entry in instance network info cache for port 4e1d011e-2609-4c71-8317-2685198b6a8b. {{(pid=63515) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1100.692258] env[63515]: DEBUG nova.network.neutron [req-c841b169-d606-4014-857f-05deb8ddee34 req-3c74ce0b-891f-4e18-8886-6b732e505b06 service nova] [instance: ea2e0810-c9f1-49a6-8589-7db402847aea] Updating instance_info_cache with network_info: [{"id": "4e1d011e-2609-4c71-8317-2685198b6a8b", "address": "fa:16:3e:fb:cb:4f", "network": {"id": "cec614cb-7a5b-4b16-a83d-061ff62d6677", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1501468300-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "aef947bfaf7d4180b46b4be662e3d903", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3a0ddd7d-c321-4187-bdd8-b19044ea2c4a", "external-id": "nsx-vlan-transportzone-747", "segmentation_id": 747, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4e1d011e-26", "ovs_interfaceid": "4e1d011e-2609-4c71-8317-2685198b6a8b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1100.856614] env[63515]: DEBUG oslo_vmware.api [None req-8f2dbdaa-d8cb-444f-a302-4aa071b6c146 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Task: {'id': task-1111955, 'name': ReconfigVM_Task, 'duration_secs': 0.232462} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1100.857375] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-8f2dbdaa-d8cb-444f-a302-4aa071b6c146 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: aa83d035-9c47-434d-ab1a-140e6b3110b4] Updating instance 'aa83d035-9c47-434d-ab1a-140e6b3110b4' progress to 33 {{(pid=63515) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1100.978367] env[63515]: DEBUG oslo_vmware.api [None req-129ac0c9-aca5-49a7-bd9c-0de3f76dcffa tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Task: {'id': task-1111956, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.109998} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1100.983097] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-129ac0c9-aca5-49a7-bd9c-0de3f76dcffa tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] [instance: c8dbf542-70cb-45be-9308-5e12307c0be6] Extended root virtual disk {{(pid=63515) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1100.983963] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-232bdddb-d6d7-4672-9d63-6ec9dc1f2a4c {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.000848] env[63515]: DEBUG nova.compute.manager [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] [instance: b1bf5a0a-42e7-41a6-90fb-a0c311623fdd] Instance has had 0 of 5 cleanup attempts {{(pid=63515) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1101.013945] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-129ac0c9-aca5-49a7-bd9c-0de3f76dcffa tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] [instance: c8dbf542-70cb-45be-9308-5e12307c0be6] Reconfiguring VM instance instance-0000006b to attach disk [datastore1] c8dbf542-70cb-45be-9308-5e12307c0be6/c8dbf542-70cb-45be-9308-5e12307c0be6.vmdk or device None with type sparse {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1101.014779] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6bd96d6a-8c16-47e0-938c-ede14d606bb3 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.038191] env[63515]: DEBUG oslo_vmware.api [None req-129ac0c9-aca5-49a7-bd9c-0de3f76dcffa tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Waiting for the task: (returnval){ [ 1101.038191] env[63515]: value = "task-1111957" [ 1101.038191] env[63515]: _type = "Task" [ 1101.038191] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1101.196701] env[63515]: DEBUG oslo_concurrency.lockutils [req-c841b169-d606-4014-857f-05deb8ddee34 req-3c74ce0b-891f-4e18-8886-6b732e505b06 service nova] Releasing lock "refresh_cache-ea2e0810-c9f1-49a6-8589-7db402847aea" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1101.363778] env[63515]: DEBUG nova.virt.hardware [None req-8f2dbdaa-d8cb-444f-a302-4aa071b6c146 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T02:52:14Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=8a120570-cb06-4099-b262-554ca0ad15c5,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1101.364055] env[63515]: DEBUG nova.virt.hardware [None req-8f2dbdaa-d8cb-444f-a302-4aa071b6c146 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Flavor limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1101.364232] env[63515]: DEBUG nova.virt.hardware [None req-8f2dbdaa-d8cb-444f-a302-4aa071b6c146 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Image limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1101.364407] env[63515]: DEBUG nova.virt.hardware [None req-8f2dbdaa-d8cb-444f-a302-4aa071b6c146 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Flavor pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1101.364558] env[63515]: DEBUG nova.virt.hardware [None req-8f2dbdaa-d8cb-444f-a302-4aa071b6c146 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Image pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1101.364709] env[63515]: DEBUG nova.virt.hardware [None req-8f2dbdaa-d8cb-444f-a302-4aa071b6c146 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1101.364922] env[63515]: DEBUG nova.virt.hardware [None req-8f2dbdaa-d8cb-444f-a302-4aa071b6c146 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1101.365100] env[63515]: DEBUG nova.virt.hardware [None req-8f2dbdaa-d8cb-444f-a302-4aa071b6c146 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1101.365277] env[63515]: DEBUG nova.virt.hardware [None req-8f2dbdaa-d8cb-444f-a302-4aa071b6c146 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Got 1 possible topologies {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1101.365445] env[63515]: DEBUG nova.virt.hardware [None req-8f2dbdaa-d8cb-444f-a302-4aa071b6c146 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1101.365651] env[63515]: DEBUG nova.virt.hardware [None req-8f2dbdaa-d8cb-444f-a302-4aa071b6c146 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1101.370992] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-8f2dbdaa-d8cb-444f-a302-4aa071b6c146 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: aa83d035-9c47-434d-ab1a-140e6b3110b4] Reconfiguring VM instance instance-00000066 to detach disk 2000 {{(pid=63515) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1101.371323] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-eed933ac-6d24-4464-8f51-cd9f3f6cbfe5 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.391041] env[63515]: DEBUG oslo_vmware.api [None req-8f2dbdaa-d8cb-444f-a302-4aa071b6c146 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Waiting for the task: (returnval){ [ 1101.391041] env[63515]: value = "task-1111959" [ 1101.391041] env[63515]: _type = "Task" [ 1101.391041] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1101.406823] env[63515]: DEBUG oslo_vmware.api [None req-8f2dbdaa-d8cb-444f-a302-4aa071b6c146 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Task: {'id': task-1111959, 'name': ReconfigVM_Task} progress is 10%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1101.506537] env[63515]: DEBUG nova.compute.manager [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] [instance: a2e014a2-9e0f-4250-aa38-3a7154149903] Instance has had 0 of 5 cleanup attempts {{(pid=63515) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1101.550908] env[63515]: DEBUG oslo_vmware.api [None req-129ac0c9-aca5-49a7-bd9c-0de3f76dcffa tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Task: {'id': task-1111957, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1101.899876] env[63515]: DEBUG oslo_vmware.api [None req-8f2dbdaa-d8cb-444f-a302-4aa071b6c146 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Task: {'id': task-1111959, 'name': ReconfigVM_Task, 'duration_secs': 0.26423} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1101.900182] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-8f2dbdaa-d8cb-444f-a302-4aa071b6c146 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: aa83d035-9c47-434d-ab1a-140e6b3110b4] Reconfigured VM instance instance-00000066 to detach disk 2000 {{(pid=63515) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1101.901071] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16065e39-e868-454a-a78a-2e3de176427a {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.926836] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-8f2dbdaa-d8cb-444f-a302-4aa071b6c146 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: aa83d035-9c47-434d-ab1a-140e6b3110b4] Reconfiguring VM instance instance-00000066 to attach disk [datastore2] aa83d035-9c47-434d-ab1a-140e6b3110b4/aa83d035-9c47-434d-ab1a-140e6b3110b4.vmdk or device None with type thin {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1101.927205] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-449a0e26-8911-45dc-97d5-666dcb084c8d {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.945779] env[63515]: DEBUG oslo_vmware.api [None req-8f2dbdaa-d8cb-444f-a302-4aa071b6c146 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Waiting for the task: (returnval){ [ 1101.945779] env[63515]: value = "task-1111960" [ 1101.945779] env[63515]: _type = "Task" [ 1101.945779] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1101.954688] env[63515]: DEBUG oslo_vmware.api [None req-8f2dbdaa-d8cb-444f-a302-4aa071b6c146 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Task: {'id': task-1111960, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1102.007471] env[63515]: DEBUG nova.compute.manager [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] [instance: 0967283f-8dda-4692-b6b7-601545cabf2f] Instance has had 0 of 5 cleanup attempts {{(pid=63515) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1102.049364] env[63515]: DEBUG oslo_vmware.api [None req-129ac0c9-aca5-49a7-bd9c-0de3f76dcffa tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Task: {'id': task-1111957, 'name': ReconfigVM_Task, 'duration_secs': 0.997275} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1102.049668] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-129ac0c9-aca5-49a7-bd9c-0de3f76dcffa tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] [instance: c8dbf542-70cb-45be-9308-5e12307c0be6] Reconfigured VM instance instance-0000006b to attach disk [datastore1] c8dbf542-70cb-45be-9308-5e12307c0be6/c8dbf542-70cb-45be-9308-5e12307c0be6.vmdk or device None with type sparse {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1102.050325] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-9455e029-5f6f-43a2-a19d-71bf032d9e44 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.056573] env[63515]: DEBUG oslo_vmware.api [None req-129ac0c9-aca5-49a7-bd9c-0de3f76dcffa tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Waiting for the task: (returnval){ [ 1102.056573] env[63515]: value = "task-1111961" [ 1102.056573] env[63515]: _type = "Task" [ 1102.056573] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1102.065849] env[63515]: DEBUG oslo_vmware.api [None req-129ac0c9-aca5-49a7-bd9c-0de3f76dcffa tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Task: {'id': task-1111961, 'name': Rename_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1102.460041] env[63515]: DEBUG oslo_vmware.api [None req-8f2dbdaa-d8cb-444f-a302-4aa071b6c146 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Task: {'id': task-1111960, 'name': ReconfigVM_Task, 'duration_secs': 0.29775} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1102.460655] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-8f2dbdaa-d8cb-444f-a302-4aa071b6c146 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: aa83d035-9c47-434d-ab1a-140e6b3110b4] Reconfigured VM instance instance-00000066 to attach disk [datastore2] aa83d035-9c47-434d-ab1a-140e6b3110b4/aa83d035-9c47-434d-ab1a-140e6b3110b4.vmdk or device None with type thin {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1102.460655] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-8f2dbdaa-d8cb-444f-a302-4aa071b6c146 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: aa83d035-9c47-434d-ab1a-140e6b3110b4] Updating instance 'aa83d035-9c47-434d-ab1a-140e6b3110b4' progress to 50 {{(pid=63515) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1102.512093] env[63515]: DEBUG nova.compute.manager [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] [instance: 68766cd4-84be-475b-8494-d7ab43a9e969] Instance has had 0 of 5 cleanup attempts {{(pid=63515) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1102.566478] env[63515]: DEBUG oslo_vmware.api [None req-129ac0c9-aca5-49a7-bd9c-0de3f76dcffa tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Task: {'id': task-1111961, 'name': Rename_Task, 'duration_secs': 0.143773} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1102.566774] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-129ac0c9-aca5-49a7-bd9c-0de3f76dcffa tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] [instance: c8dbf542-70cb-45be-9308-5e12307c0be6] Powering on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1102.567036] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-275bdbf2-4067-46f3-baf0-f809c7a222f7 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.573420] env[63515]: DEBUG oslo_vmware.api [None req-129ac0c9-aca5-49a7-bd9c-0de3f76dcffa tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Waiting for the task: (returnval){ [ 1102.573420] env[63515]: value = "task-1111962" [ 1102.573420] env[63515]: _type = "Task" [ 1102.573420] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1102.581467] env[63515]: DEBUG oslo_vmware.api [None req-129ac0c9-aca5-49a7-bd9c-0de3f76dcffa tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Task: {'id': task-1111962, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1102.928724] env[63515]: DEBUG oslo_concurrency.lockutils [None req-3e2e3b54-760b-4bc4-9662-f0fa49db1c0e tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Acquiring lock "c9e07a6c-fd14-4071-8c69-6ece62e9109a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1102.929020] env[63515]: DEBUG oslo_concurrency.lockutils [None req-3e2e3b54-760b-4bc4-9662-f0fa49db1c0e tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Lock "c9e07a6c-fd14-4071-8c69-6ece62e9109a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1102.968593] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d61ea714-2e8f-4a80-a84b-b8e01b7c4313 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.990420] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17dc6bd9-707b-4298-8b75-9c02d442fde6 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.012130] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-8f2dbdaa-d8cb-444f-a302-4aa071b6c146 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: aa83d035-9c47-434d-ab1a-140e6b3110b4] Updating instance 'aa83d035-9c47-434d-ab1a-140e6b3110b4' progress to 67 {{(pid=63515) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1103.015565] env[63515]: DEBUG nova.compute.manager [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] [instance: 2ea99c7a-8b61-4718-bee0-f4ce4fc021af] Instance has had 0 of 5 cleanup attempts {{(pid=63515) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1103.083204] env[63515]: DEBUG oslo_vmware.api [None req-129ac0c9-aca5-49a7-bd9c-0de3f76dcffa tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Task: {'id': task-1111962, 'name': PowerOnVM_Task} progress is 87%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1103.434314] env[63515]: DEBUG nova.compute.manager [None req-3e2e3b54-760b-4bc4-9662-f0fa49db1c0e tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] [instance: c9e07a6c-fd14-4071-8c69-6ece62e9109a] Starting instance... {{(pid=63515) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1103.523331] env[63515]: DEBUG nova.compute.manager [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] [instance: eab16df9-7bb5-4576-bca0-769a561c5fe9] Instance has had 0 of 5 cleanup attempts {{(pid=63515) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1103.589333] env[63515]: DEBUG oslo_vmware.api [None req-129ac0c9-aca5-49a7-bd9c-0de3f76dcffa tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Task: {'id': task-1111962, 'name': PowerOnVM_Task, 'duration_secs': 0.954663} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1103.589737] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-129ac0c9-aca5-49a7-bd9c-0de3f76dcffa tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] [instance: c8dbf542-70cb-45be-9308-5e12307c0be6] Powered on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1103.590293] env[63515]: INFO nova.compute.manager [None req-129ac0c9-aca5-49a7-bd9c-0de3f76dcffa tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] [instance: c8dbf542-70cb-45be-9308-5e12307c0be6] Took 9.05 seconds to spawn the instance on the hypervisor. [ 1103.590455] env[63515]: DEBUG nova.compute.manager [None req-129ac0c9-aca5-49a7-bd9c-0de3f76dcffa tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] [instance: c8dbf542-70cb-45be-9308-5e12307c0be6] Checking state {{(pid=63515) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1103.591397] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86b5783c-c406-4b2b-bbdd-b388c0bde616 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.957642] env[63515]: DEBUG oslo_concurrency.lockutils [None req-3e2e3b54-760b-4bc4-9662-f0fa49db1c0e tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1103.957935] env[63515]: DEBUG oslo_concurrency.lockutils [None req-3e2e3b54-760b-4bc4-9662-f0fa49db1c0e tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1103.960067] env[63515]: INFO nova.compute.claims [None req-3e2e3b54-760b-4bc4-9662-f0fa49db1c0e tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] [instance: c9e07a6c-fd14-4071-8c69-6ece62e9109a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1104.028227] env[63515]: DEBUG nova.compute.manager [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] [instance: 0ee9cce2-94b9-4038-ab38-1b3f96a864bc] Instance has had 0 of 5 cleanup attempts {{(pid=63515) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1104.114638] env[63515]: INFO nova.compute.manager [None req-129ac0c9-aca5-49a7-bd9c-0de3f76dcffa tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] [instance: c8dbf542-70cb-45be-9308-5e12307c0be6] Took 13.79 seconds to build instance. [ 1104.532686] env[63515]: DEBUG nova.compute.manager [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] [instance: 2a29f37d-e786-48d3-b126-4854d6e7c67c] Instance has had 0 of 5 cleanup attempts {{(pid=63515) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1104.616278] env[63515]: DEBUG oslo_concurrency.lockutils [None req-129ac0c9-aca5-49a7-bd9c-0de3f76dcffa tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Lock "c8dbf542-70cb-45be-9308-5e12307c0be6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.302s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1104.728163] env[63515]: DEBUG nova.network.neutron [None req-8f2dbdaa-d8cb-444f-a302-4aa071b6c146 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: aa83d035-9c47-434d-ab1a-140e6b3110b4] Port 614aed8f-3ab8-4672-ac13-0ae8cb0b233d binding to destination host cpu-1 is already ACTIVE {{(pid=63515) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 1104.880692] env[63515]: DEBUG nova.compute.manager [req-8e6ce23b-2e5d-403d-b32e-e09b01791a81 req-ded1b186-fc0e-4675-983c-e85e48918e21 service nova] [instance: c8dbf542-70cb-45be-9308-5e12307c0be6] Received event network-changed-5569cb8c-d9c8-4cb5-abdc-5890ad7d7ede {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1104.881816] env[63515]: DEBUG nova.compute.manager [req-8e6ce23b-2e5d-403d-b32e-e09b01791a81 req-ded1b186-fc0e-4675-983c-e85e48918e21 service nova] [instance: c8dbf542-70cb-45be-9308-5e12307c0be6] Refreshing instance network info cache due to event network-changed-5569cb8c-d9c8-4cb5-abdc-5890ad7d7ede. {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1104.881816] env[63515]: DEBUG oslo_concurrency.lockutils [req-8e6ce23b-2e5d-403d-b32e-e09b01791a81 req-ded1b186-fc0e-4675-983c-e85e48918e21 service nova] Acquiring lock "refresh_cache-c8dbf542-70cb-45be-9308-5e12307c0be6" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1104.881816] env[63515]: DEBUG oslo_concurrency.lockutils [req-8e6ce23b-2e5d-403d-b32e-e09b01791a81 req-ded1b186-fc0e-4675-983c-e85e48918e21 service nova] Acquired lock "refresh_cache-c8dbf542-70cb-45be-9308-5e12307c0be6" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1104.881816] env[63515]: DEBUG nova.network.neutron [req-8e6ce23b-2e5d-403d-b32e-e09b01791a81 req-ded1b186-fc0e-4675-983c-e85e48918e21 service nova] [instance: c8dbf542-70cb-45be-9308-5e12307c0be6] Refreshing network info cache for port 5569cb8c-d9c8-4cb5-abdc-5890ad7d7ede {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1105.035906] env[63515]: DEBUG nova.compute.manager [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] [instance: 705b7e9b-3ab6-40ec-8cf6-d86bf0a1faf0] Instance has had 0 of 5 cleanup attempts {{(pid=63515) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1105.067930] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d964e112-3fd1-4647-86ca-23e770111779 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.076930] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-014b6de4-daa7-4c59-8163-3b71f443cb00 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.108933] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7bd8db2-77b2-46b5-a70c-9543ed555ba3 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.116467] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5cc5aaee-bd63-4496-b0d0-7066726cfdee {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.130014] env[63515]: DEBUG nova.compute.provider_tree [None req-3e2e3b54-760b-4bc4-9662-f0fa49db1c0e tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1105.541470] env[63515]: DEBUG nova.compute.manager [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] [instance: 19475235-dbec-40a5-bd8f-8070616b7ddb] Instance has had 0 of 5 cleanup attempts {{(pid=63515) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1105.606791] env[63515]: DEBUG nova.network.neutron [req-8e6ce23b-2e5d-403d-b32e-e09b01791a81 req-ded1b186-fc0e-4675-983c-e85e48918e21 service nova] [instance: c8dbf542-70cb-45be-9308-5e12307c0be6] Updated VIF entry in instance network info cache for port 5569cb8c-d9c8-4cb5-abdc-5890ad7d7ede. {{(pid=63515) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1105.607173] env[63515]: DEBUG nova.network.neutron [req-8e6ce23b-2e5d-403d-b32e-e09b01791a81 req-ded1b186-fc0e-4675-983c-e85e48918e21 service nova] [instance: c8dbf542-70cb-45be-9308-5e12307c0be6] Updating instance_info_cache with network_info: [{"id": "5569cb8c-d9c8-4cb5-abdc-5890ad7d7ede", "address": "fa:16:3e:8d:b8:b1", "network": {"id": "d6da2650-9ba9-4ef6-8035-e58b635a64b2", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-391564968-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.243", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "97f2e119708746038f38359572c99438", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "043ca97b-0fca-4b54-8be6-027123fa76d0", "external-id": "nsx-vlan-transportzone-702", "segmentation_id": 702, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5569cb8c-d9", "ovs_interfaceid": "5569cb8c-d9c8-4cb5-abdc-5890ad7d7ede", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1105.633067] env[63515]: DEBUG nova.scheduler.client.report [None req-3e2e3b54-760b-4bc4-9662-f0fa49db1c0e tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1105.747245] env[63515]: DEBUG oslo_concurrency.lockutils [None req-8f2dbdaa-d8cb-444f-a302-4aa071b6c146 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Acquiring lock "aa83d035-9c47-434d-ab1a-140e6b3110b4-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1105.747487] env[63515]: DEBUG oslo_concurrency.lockutils [None req-8f2dbdaa-d8cb-444f-a302-4aa071b6c146 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Lock "aa83d035-9c47-434d-ab1a-140e6b3110b4-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1105.747671] env[63515]: DEBUG oslo_concurrency.lockutils [None req-8f2dbdaa-d8cb-444f-a302-4aa071b6c146 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Lock "aa83d035-9c47-434d-ab1a-140e6b3110b4-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1106.044659] env[63515]: DEBUG nova.compute.manager [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] [instance: 17275472-2921-49c9-b4e0-5485649ebd1c] Instance has had 0 of 5 cleanup attempts {{(pid=63515) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1106.109764] env[63515]: DEBUG oslo_concurrency.lockutils [req-8e6ce23b-2e5d-403d-b32e-e09b01791a81 req-ded1b186-fc0e-4675-983c-e85e48918e21 service nova] Releasing lock "refresh_cache-c8dbf542-70cb-45be-9308-5e12307c0be6" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1106.137388] env[63515]: DEBUG oslo_concurrency.lockutils [None req-3e2e3b54-760b-4bc4-9662-f0fa49db1c0e tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.179s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1106.137915] env[63515]: DEBUG nova.compute.manager [None req-3e2e3b54-760b-4bc4-9662-f0fa49db1c0e tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] [instance: c9e07a6c-fd14-4071-8c69-6ece62e9109a] Start building networks asynchronously for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1106.548334] env[63515]: DEBUG nova.compute.manager [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] [instance: 70df69e5-687b-44fb-b6fc-cdb08e21dda0] Instance has had 0 of 5 cleanup attempts {{(pid=63515) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1106.643009] env[63515]: DEBUG nova.compute.utils [None req-3e2e3b54-760b-4bc4-9662-f0fa49db1c0e tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Using /dev/sd instead of None {{(pid=63515) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1106.644427] env[63515]: DEBUG nova.compute.manager [None req-3e2e3b54-760b-4bc4-9662-f0fa49db1c0e tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] [instance: c9e07a6c-fd14-4071-8c69-6ece62e9109a] Allocating IP information in the background. {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1106.644602] env[63515]: DEBUG nova.network.neutron [None req-3e2e3b54-760b-4bc4-9662-f0fa49db1c0e tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] [instance: c9e07a6c-fd14-4071-8c69-6ece62e9109a] allocate_for_instance() {{(pid=63515) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1106.690863] env[63515]: DEBUG nova.policy [None req-3e2e3b54-760b-4bc4-9662-f0fa49db1c0e tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e8da40be55174af2a0f4dd3bf2844cb4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'fe873600ed6d4a3eb1ee3d3a19fceaf7', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63515) authorize /opt/stack/nova/nova/policy.py:201}} [ 1106.804161] env[63515]: DEBUG oslo_concurrency.lockutils [None req-8f2dbdaa-d8cb-444f-a302-4aa071b6c146 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Acquiring lock "refresh_cache-aa83d035-9c47-434d-ab1a-140e6b3110b4" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1106.804366] env[63515]: DEBUG oslo_concurrency.lockutils [None req-8f2dbdaa-d8cb-444f-a302-4aa071b6c146 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Acquired lock "refresh_cache-aa83d035-9c47-434d-ab1a-140e6b3110b4" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1106.804544] env[63515]: DEBUG nova.network.neutron [None req-8f2dbdaa-d8cb-444f-a302-4aa071b6c146 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: aa83d035-9c47-434d-ab1a-140e6b3110b4] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1106.926697] env[63515]: DEBUG nova.network.neutron [None req-3e2e3b54-760b-4bc4-9662-f0fa49db1c0e tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] [instance: c9e07a6c-fd14-4071-8c69-6ece62e9109a] Successfully created port: bd316695-ec41-4f7d-b042-5cd5d319aa55 {{(pid=63515) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1107.052190] env[63515]: DEBUG nova.compute.manager [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] [instance: 9d18c7b6-ce8e-4042-9e70-696bb7f57cb3] Instance has had 0 of 5 cleanup attempts {{(pid=63515) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1107.147852] env[63515]: DEBUG nova.compute.manager [None req-3e2e3b54-760b-4bc4-9662-f0fa49db1c0e tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] [instance: c9e07a6c-fd14-4071-8c69-6ece62e9109a] Start building block device mappings for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1107.540701] env[63515]: DEBUG nova.network.neutron [None req-8f2dbdaa-d8cb-444f-a302-4aa071b6c146 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: aa83d035-9c47-434d-ab1a-140e6b3110b4] Updating instance_info_cache with network_info: [{"id": "614aed8f-3ab8-4672-ac13-0ae8cb0b233d", "address": "fa:16:3e:98:68:38", "network": {"id": "5758c885-e1c9-45c8-86aa-27a802101f52", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-734458138-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.204", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b18d4daacbc84e758f9f9ae4f3f09d28", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69351262-8d39-441a-85ba-3a78df436d17", "external-id": "nsx-vlan-transportzone-205", "segmentation_id": 205, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap614aed8f-3a", "ovs_interfaceid": "614aed8f-3ab8-4672-ac13-0ae8cb0b233d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1107.555153] env[63515]: DEBUG nova.compute.manager [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] [instance: d926c699-a64a-4942-9ef4-f0166414661d] Instance has had 0 of 5 cleanup attempts {{(pid=63515) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1108.043885] env[63515]: DEBUG oslo_concurrency.lockutils [None req-8f2dbdaa-d8cb-444f-a302-4aa071b6c146 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Releasing lock "refresh_cache-aa83d035-9c47-434d-ab1a-140e6b3110b4" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1108.058297] env[63515]: DEBUG nova.compute.manager [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] [instance: 4ef8066a-b68c-457b-9964-b1c34bab0fc3] Instance has had 0 of 5 cleanup attempts {{(pid=63515) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1108.156840] env[63515]: DEBUG nova.compute.manager [None req-3e2e3b54-760b-4bc4-9662-f0fa49db1c0e tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] [instance: c9e07a6c-fd14-4071-8c69-6ece62e9109a] Start spawning the instance on the hypervisor. {{(pid=63515) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1108.182014] env[63515]: DEBUG nova.virt.hardware [None req-3e2e3b54-760b-4bc4-9662-f0fa49db1c0e tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T02:52:14Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T02:51:56Z,direct_url=,disk_format='vmdk',id=8a120570-cb06-4099-b262-554ca0ad15c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b569255cc43e42e7ae3f2b4ad37c6ef4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T02:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1108.182295] env[63515]: DEBUG nova.virt.hardware [None req-3e2e3b54-760b-4bc4-9662-f0fa49db1c0e tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Flavor limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1108.182461] env[63515]: DEBUG nova.virt.hardware [None req-3e2e3b54-760b-4bc4-9662-f0fa49db1c0e tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Image limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1108.182647] env[63515]: DEBUG nova.virt.hardware [None req-3e2e3b54-760b-4bc4-9662-f0fa49db1c0e tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Flavor pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1108.182797] env[63515]: DEBUG nova.virt.hardware [None req-3e2e3b54-760b-4bc4-9662-f0fa49db1c0e tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Image pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1108.182947] env[63515]: DEBUG nova.virt.hardware [None req-3e2e3b54-760b-4bc4-9662-f0fa49db1c0e tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1108.183188] env[63515]: DEBUG nova.virt.hardware [None req-3e2e3b54-760b-4bc4-9662-f0fa49db1c0e tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1108.183359] env[63515]: DEBUG nova.virt.hardware [None req-3e2e3b54-760b-4bc4-9662-f0fa49db1c0e tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1108.183531] env[63515]: DEBUG nova.virt.hardware [None req-3e2e3b54-760b-4bc4-9662-f0fa49db1c0e tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Got 1 possible topologies {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1108.183733] env[63515]: DEBUG nova.virt.hardware [None req-3e2e3b54-760b-4bc4-9662-f0fa49db1c0e tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1108.183924] env[63515]: DEBUG nova.virt.hardware [None req-3e2e3b54-760b-4bc4-9662-f0fa49db1c0e tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1108.184817] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a28c209-d00b-4f18-aa98-ec268983d81c {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.193233] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c195884-8fd9-42b0-8322-1cc589cee97c {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.319795] env[63515]: DEBUG nova.compute.manager [req-c4d80776-d411-4137-9479-683bbf6afdbb req-f131b285-2aa7-4f84-95c4-98f9abe75543 service nova] [instance: c9e07a6c-fd14-4071-8c69-6ece62e9109a] Received event network-vif-plugged-bd316695-ec41-4f7d-b042-5cd5d319aa55 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1108.320105] env[63515]: DEBUG oslo_concurrency.lockutils [req-c4d80776-d411-4137-9479-683bbf6afdbb req-f131b285-2aa7-4f84-95c4-98f9abe75543 service nova] Acquiring lock "c9e07a6c-fd14-4071-8c69-6ece62e9109a-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1108.320325] env[63515]: DEBUG oslo_concurrency.lockutils [req-c4d80776-d411-4137-9479-683bbf6afdbb req-f131b285-2aa7-4f84-95c4-98f9abe75543 service nova] Lock "c9e07a6c-fd14-4071-8c69-6ece62e9109a-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1108.320509] env[63515]: DEBUG oslo_concurrency.lockutils [req-c4d80776-d411-4137-9479-683bbf6afdbb req-f131b285-2aa7-4f84-95c4-98f9abe75543 service nova] Lock "c9e07a6c-fd14-4071-8c69-6ece62e9109a-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1108.320682] env[63515]: DEBUG nova.compute.manager [req-c4d80776-d411-4137-9479-683bbf6afdbb req-f131b285-2aa7-4f84-95c4-98f9abe75543 service nova] [instance: c9e07a6c-fd14-4071-8c69-6ece62e9109a] No waiting events found dispatching network-vif-plugged-bd316695-ec41-4f7d-b042-5cd5d319aa55 {{(pid=63515) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1108.320853] env[63515]: WARNING nova.compute.manager [req-c4d80776-d411-4137-9479-683bbf6afdbb req-f131b285-2aa7-4f84-95c4-98f9abe75543 service nova] [instance: c9e07a6c-fd14-4071-8c69-6ece62e9109a] Received unexpected event network-vif-plugged-bd316695-ec41-4f7d-b042-5cd5d319aa55 for instance with vm_state building and task_state spawning. [ 1108.404467] env[63515]: DEBUG nova.network.neutron [None req-3e2e3b54-760b-4bc4-9662-f0fa49db1c0e tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] [instance: c9e07a6c-fd14-4071-8c69-6ece62e9109a] Successfully updated port: bd316695-ec41-4f7d-b042-5cd5d319aa55 {{(pid=63515) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1108.552511] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f4b3779-e0b3-4b59-a4e5-c0e9ff3c2b30 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.559715] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3640e6df-40a7-44b8-8801-0d2c3ded9f7f {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.562123] env[63515]: DEBUG nova.compute.manager [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] [instance: 4e453127-1f3e-40ea-819f-6678479826c8] Instance has had 0 of 5 cleanup attempts {{(pid=63515) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1108.907514] env[63515]: DEBUG oslo_concurrency.lockutils [None req-3e2e3b54-760b-4bc4-9662-f0fa49db1c0e tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Acquiring lock "refresh_cache-c9e07a6c-fd14-4071-8c69-6ece62e9109a" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1108.907657] env[63515]: DEBUG oslo_concurrency.lockutils [None req-3e2e3b54-760b-4bc4-9662-f0fa49db1c0e tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Acquired lock "refresh_cache-c9e07a6c-fd14-4071-8c69-6ece62e9109a" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1108.907810] env[63515]: DEBUG nova.network.neutron [None req-3e2e3b54-760b-4bc4-9662-f0fa49db1c0e tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] [instance: c9e07a6c-fd14-4071-8c69-6ece62e9109a] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1109.064537] env[63515]: DEBUG nova.compute.manager [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] [instance: e666825c-ff4e-4a0e-93c0-43c00f167bbb] Instance has had 0 of 5 cleanup attempts {{(pid=63515) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1109.448620] env[63515]: DEBUG nova.network.neutron [None req-3e2e3b54-760b-4bc4-9662-f0fa49db1c0e tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] [instance: c9e07a6c-fd14-4071-8c69-6ece62e9109a] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1109.568178] env[63515]: DEBUG nova.compute.manager [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] [instance: 63ce797d-7180-4209-ac2c-81978bf7607a] Instance has had 0 of 5 cleanup attempts {{(pid=63515) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1109.578437] env[63515]: DEBUG nova.network.neutron [None req-3e2e3b54-760b-4bc4-9662-f0fa49db1c0e tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] [instance: c9e07a6c-fd14-4071-8c69-6ece62e9109a] Updating instance_info_cache with network_info: [{"id": "bd316695-ec41-4f7d-b042-5cd5d319aa55", "address": "fa:16:3e:da:4b:8e", "network": {"id": "0e23b3a4-91ea-4bbe-9732-a3539d56b382", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1727906976-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fe873600ed6d4a3eb1ee3d3a19fceaf7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c7d2575f-b92f-44ec-a863-634cb76631a2", "external-id": "nsx-vlan-transportzone-794", "segmentation_id": 794, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbd316695-ec", "ovs_interfaceid": "bd316695-ec41-4f7d-b042-5cd5d319aa55", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1109.655678] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ebdbb43-86d4-4161-8233-39f028d9ac08 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.677376] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f54d64fe-3818-410b-8f16-107ae4ad89e4 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.684777] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-8f2dbdaa-d8cb-444f-a302-4aa071b6c146 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: aa83d035-9c47-434d-ab1a-140e6b3110b4] Updating instance 'aa83d035-9c47-434d-ab1a-140e6b3110b4' progress to 83 {{(pid=63515) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1110.071119] env[63515]: DEBUG nova.compute.manager [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] [instance: 5cbce760-0163-4b27-8ae3-e46c926c8916] Instance has had 0 of 5 cleanup attempts {{(pid=63515) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1110.080738] env[63515]: DEBUG oslo_concurrency.lockutils [None req-3e2e3b54-760b-4bc4-9662-f0fa49db1c0e tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Releasing lock "refresh_cache-c9e07a6c-fd14-4071-8c69-6ece62e9109a" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1110.081017] env[63515]: DEBUG nova.compute.manager [None req-3e2e3b54-760b-4bc4-9662-f0fa49db1c0e tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] [instance: c9e07a6c-fd14-4071-8c69-6ece62e9109a] Instance network_info: |[{"id": "bd316695-ec41-4f7d-b042-5cd5d319aa55", "address": "fa:16:3e:da:4b:8e", "network": {"id": "0e23b3a4-91ea-4bbe-9732-a3539d56b382", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1727906976-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fe873600ed6d4a3eb1ee3d3a19fceaf7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c7d2575f-b92f-44ec-a863-634cb76631a2", "external-id": "nsx-vlan-transportzone-794", "segmentation_id": 794, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbd316695-ec", "ovs_interfaceid": "bd316695-ec41-4f7d-b042-5cd5d319aa55", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1110.081410] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-3e2e3b54-760b-4bc4-9662-f0fa49db1c0e tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] [instance: c9e07a6c-fd14-4071-8c69-6ece62e9109a] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:da:4b:8e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c7d2575f-b92f-44ec-a863-634cb76631a2', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'bd316695-ec41-4f7d-b042-5cd5d319aa55', 'vif_model': 'vmxnet3'}] {{(pid=63515) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1110.088869] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-3e2e3b54-760b-4bc4-9662-f0fa49db1c0e tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Creating folder: Project (fe873600ed6d4a3eb1ee3d3a19fceaf7). Parent ref: group-v243370. {{(pid=63515) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1110.089142] env[63515]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a5cce9c6-5ae8-4e6d-a093-34840a428df6 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.099878] env[63515]: INFO nova.virt.vmwareapi.vm_util [None req-3e2e3b54-760b-4bc4-9662-f0fa49db1c0e tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Created folder: Project (fe873600ed6d4a3eb1ee3d3a19fceaf7) in parent group-v243370. [ 1110.100075] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-3e2e3b54-760b-4bc4-9662-f0fa49db1c0e tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Creating folder: Instances. Parent ref: group-v243584. {{(pid=63515) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1110.100506] env[63515]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-97eaf029-a164-4dd4-8fea-b2082a368878 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.109969] env[63515]: INFO nova.virt.vmwareapi.vm_util [None req-3e2e3b54-760b-4bc4-9662-f0fa49db1c0e tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Created folder: Instances in parent group-v243584. [ 1110.110207] env[63515]: DEBUG oslo.service.loopingcall [None req-3e2e3b54-760b-4bc4-9662-f0fa49db1c0e tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1110.110390] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c9e07a6c-fd14-4071-8c69-6ece62e9109a] Creating VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1110.110579] env[63515]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-10c05872-b79c-4f2c-872c-51f6d06c4e8c {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.130255] env[63515]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1110.130255] env[63515]: value = "task-1111968" [ 1110.130255] env[63515]: _type = "Task" [ 1110.130255] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1110.137289] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1111968, 'name': CreateVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1110.191167] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-8f2dbdaa-d8cb-444f-a302-4aa071b6c146 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: aa83d035-9c47-434d-ab1a-140e6b3110b4] Powering on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1110.191551] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b44e7471-07b4-4773-b70e-c77aa17fc2e0 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.198612] env[63515]: DEBUG oslo_vmware.api [None req-8f2dbdaa-d8cb-444f-a302-4aa071b6c146 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Waiting for the task: (returnval){ [ 1110.198612] env[63515]: value = "task-1111969" [ 1110.198612] env[63515]: _type = "Task" [ 1110.198612] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1110.207832] env[63515]: DEBUG oslo_vmware.api [None req-8f2dbdaa-d8cb-444f-a302-4aa071b6c146 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Task: {'id': task-1111969, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1110.346705] env[63515]: DEBUG nova.compute.manager [req-16466f57-9cf0-495d-a7da-885af84b9c18 req-1d56ff03-5696-4d29-b651-4794d6a1c0af service nova] [instance: c9e07a6c-fd14-4071-8c69-6ece62e9109a] Received event network-changed-bd316695-ec41-4f7d-b042-5cd5d319aa55 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1110.346705] env[63515]: DEBUG nova.compute.manager [req-16466f57-9cf0-495d-a7da-885af84b9c18 req-1d56ff03-5696-4d29-b651-4794d6a1c0af service nova] [instance: c9e07a6c-fd14-4071-8c69-6ece62e9109a] Refreshing instance network info cache due to event network-changed-bd316695-ec41-4f7d-b042-5cd5d319aa55. {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1110.346705] env[63515]: DEBUG oslo_concurrency.lockutils [req-16466f57-9cf0-495d-a7da-885af84b9c18 req-1d56ff03-5696-4d29-b651-4794d6a1c0af service nova] Acquiring lock "refresh_cache-c9e07a6c-fd14-4071-8c69-6ece62e9109a" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1110.347771] env[63515]: DEBUG oslo_concurrency.lockutils [req-16466f57-9cf0-495d-a7da-885af84b9c18 req-1d56ff03-5696-4d29-b651-4794d6a1c0af service nova] Acquired lock "refresh_cache-c9e07a6c-fd14-4071-8c69-6ece62e9109a" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1110.347771] env[63515]: DEBUG nova.network.neutron [req-16466f57-9cf0-495d-a7da-885af84b9c18 req-1d56ff03-5696-4d29-b651-4794d6a1c0af service nova] [instance: c9e07a6c-fd14-4071-8c69-6ece62e9109a] Refreshing network info cache for port bd316695-ec41-4f7d-b042-5cd5d319aa55 {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1110.574598] env[63515]: DEBUG nova.compute.manager [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] [instance: 192137b0-03e5-4bc4-b911-4b4f1a874f74] Instance has had 0 of 5 cleanup attempts {{(pid=63515) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1110.640728] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1111968, 'name': CreateVM_Task, 'duration_secs': 0.356782} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1110.640949] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c9e07a6c-fd14-4071-8c69-6ece62e9109a] Created VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1110.641710] env[63515]: DEBUG oslo_concurrency.lockutils [None req-3e2e3b54-760b-4bc4-9662-f0fa49db1c0e tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1110.641895] env[63515]: DEBUG oslo_concurrency.lockutils [None req-3e2e3b54-760b-4bc4-9662-f0fa49db1c0e tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1110.642259] env[63515]: DEBUG oslo_concurrency.lockutils [None req-3e2e3b54-760b-4bc4-9662-f0fa49db1c0e tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1110.642536] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-299907d0-76e8-445c-b092-5e1dcdf46b9a {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.647412] env[63515]: DEBUG oslo_vmware.api [None req-3e2e3b54-760b-4bc4-9662-f0fa49db1c0e tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Waiting for the task: (returnval){ [ 1110.647412] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]52715e29-83c6-cde2-28ec-291259660a0d" [ 1110.647412] env[63515]: _type = "Task" [ 1110.647412] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1110.655340] env[63515]: DEBUG oslo_vmware.api [None req-3e2e3b54-760b-4bc4-9662-f0fa49db1c0e tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52715e29-83c6-cde2-28ec-291259660a0d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1110.708170] env[63515]: DEBUG oslo_vmware.api [None req-8f2dbdaa-d8cb-444f-a302-4aa071b6c146 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Task: {'id': task-1111969, 'name': PowerOnVM_Task, 'duration_secs': 0.441553} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1110.708507] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-8f2dbdaa-d8cb-444f-a302-4aa071b6c146 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: aa83d035-9c47-434d-ab1a-140e6b3110b4] Powered on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1110.708709] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-8f2dbdaa-d8cb-444f-a302-4aa071b6c146 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: aa83d035-9c47-434d-ab1a-140e6b3110b4] Updating instance 'aa83d035-9c47-434d-ab1a-140e6b3110b4' progress to 100 {{(pid=63515) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1111.078191] env[63515]: DEBUG nova.compute.manager [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] [instance: 70ed982f-affd-4dd1-bc90-c64e7c6d49d2] Instance has had 0 of 5 cleanup attempts {{(pid=63515) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1111.086373] env[63515]: DEBUG nova.network.neutron [req-16466f57-9cf0-495d-a7da-885af84b9c18 req-1d56ff03-5696-4d29-b651-4794d6a1c0af service nova] [instance: c9e07a6c-fd14-4071-8c69-6ece62e9109a] Updated VIF entry in instance network info cache for port bd316695-ec41-4f7d-b042-5cd5d319aa55. {{(pid=63515) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1111.086373] env[63515]: DEBUG nova.network.neutron [req-16466f57-9cf0-495d-a7da-885af84b9c18 req-1d56ff03-5696-4d29-b651-4794d6a1c0af service nova] [instance: c9e07a6c-fd14-4071-8c69-6ece62e9109a] Updating instance_info_cache with network_info: [{"id": "bd316695-ec41-4f7d-b042-5cd5d319aa55", "address": "fa:16:3e:da:4b:8e", "network": {"id": "0e23b3a4-91ea-4bbe-9732-a3539d56b382", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1727906976-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fe873600ed6d4a3eb1ee3d3a19fceaf7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c7d2575f-b92f-44ec-a863-634cb76631a2", "external-id": "nsx-vlan-transportzone-794", "segmentation_id": 794, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbd316695-ec", "ovs_interfaceid": "bd316695-ec41-4f7d-b042-5cd5d319aa55", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1111.158541] env[63515]: DEBUG oslo_vmware.api [None req-3e2e3b54-760b-4bc4-9662-f0fa49db1c0e tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52715e29-83c6-cde2-28ec-291259660a0d, 'name': SearchDatastore_Task, 'duration_secs': 0.015029} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1111.158923] env[63515]: DEBUG oslo_concurrency.lockutils [None req-3e2e3b54-760b-4bc4-9662-f0fa49db1c0e tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1111.159251] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-3e2e3b54-760b-4bc4-9662-f0fa49db1c0e tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] [instance: c9e07a6c-fd14-4071-8c69-6ece62e9109a] Processing image 8a120570-cb06-4099-b262-554ca0ad15c5 {{(pid=63515) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1111.159604] env[63515]: DEBUG oslo_concurrency.lockutils [None req-3e2e3b54-760b-4bc4-9662-f0fa49db1c0e tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1111.159725] env[63515]: DEBUG oslo_concurrency.lockutils [None req-3e2e3b54-760b-4bc4-9662-f0fa49db1c0e tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1111.159952] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-3e2e3b54-760b-4bc4-9662-f0fa49db1c0e tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1111.160262] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-3c53bd70-9cae-4f17-88f4-ff329f671e5d {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.170593] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-3e2e3b54-760b-4bc4-9662-f0fa49db1c0e tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1111.170802] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-3e2e3b54-760b-4bc4-9662-f0fa49db1c0e tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63515) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1111.171860] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f2d6b6bf-7e06-4543-ba26-2f7b59f2e607 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.177154] env[63515]: DEBUG oslo_vmware.api [None req-3e2e3b54-760b-4bc4-9662-f0fa49db1c0e tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Waiting for the task: (returnval){ [ 1111.177154] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]529e8e17-11b6-0c3a-28cf-148a158b16af" [ 1111.177154] env[63515]: _type = "Task" [ 1111.177154] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1111.184608] env[63515]: DEBUG oslo_vmware.api [None req-3e2e3b54-760b-4bc4-9662-f0fa49db1c0e tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]529e8e17-11b6-0c3a-28cf-148a158b16af, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1111.581935] env[63515]: DEBUG nova.compute.manager [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] [instance: 84dbf321-2c70-4e08-b430-cb5a06fc6829] Instance has had 0 of 5 cleanup attempts {{(pid=63515) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1111.587721] env[63515]: DEBUG oslo_concurrency.lockutils [req-16466f57-9cf0-495d-a7da-885af84b9c18 req-1d56ff03-5696-4d29-b651-4794d6a1c0af service nova] Releasing lock "refresh_cache-c9e07a6c-fd14-4071-8c69-6ece62e9109a" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1111.688113] env[63515]: DEBUG oslo_vmware.api [None req-3e2e3b54-760b-4bc4-9662-f0fa49db1c0e tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]529e8e17-11b6-0c3a-28cf-148a158b16af, 'name': SearchDatastore_Task, 'duration_secs': 0.018677} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1111.688928] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-544e4073-5e66-4d04-b76d-04e3ae889109 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.694381] env[63515]: DEBUG oslo_vmware.api [None req-3e2e3b54-760b-4bc4-9662-f0fa49db1c0e tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Waiting for the task: (returnval){ [ 1111.694381] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]524917cf-db79-eb35-d9b4-8e92d2e04b3c" [ 1111.694381] env[63515]: _type = "Task" [ 1111.694381] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1111.701715] env[63515]: DEBUG oslo_vmware.api [None req-3e2e3b54-760b-4bc4-9662-f0fa49db1c0e tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]524917cf-db79-eb35-d9b4-8e92d2e04b3c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1112.085584] env[63515]: DEBUG nova.compute.manager [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] [instance: 5d50fbd6-a7fb-422c-9dd7-df7140fd0aae] Instance has had 0 of 5 cleanup attempts {{(pid=63515) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1112.205534] env[63515]: DEBUG oslo_vmware.api [None req-3e2e3b54-760b-4bc4-9662-f0fa49db1c0e tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]524917cf-db79-eb35-d9b4-8e92d2e04b3c, 'name': SearchDatastore_Task, 'duration_secs': 0.091706} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1112.205753] env[63515]: DEBUG oslo_concurrency.lockutils [None req-3e2e3b54-760b-4bc4-9662-f0fa49db1c0e tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1112.206045] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-3e2e3b54-760b-4bc4-9662-f0fa49db1c0e tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk to [datastore1] c9e07a6c-fd14-4071-8c69-6ece62e9109a/c9e07a6c-fd14-4071-8c69-6ece62e9109a.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1112.206356] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5ee7d65e-dd5d-476f-9853-23d25fa55eb2 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.216142] env[63515]: DEBUG oslo_vmware.api [None req-3e2e3b54-760b-4bc4-9662-f0fa49db1c0e tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Waiting for the task: (returnval){ [ 1112.216142] env[63515]: value = "task-1111971" [ 1112.216142] env[63515]: _type = "Task" [ 1112.216142] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1112.224286] env[63515]: DEBUG oslo_vmware.api [None req-3e2e3b54-760b-4bc4-9662-f0fa49db1c0e tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Task: {'id': task-1111971, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1112.589304] env[63515]: DEBUG nova.compute.manager [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] [instance: 30b8b872-40ca-4297-b98e-a64c83a75483] Instance has had 0 of 5 cleanup attempts {{(pid=63515) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1112.728066] env[63515]: DEBUG oslo_vmware.api [None req-3e2e3b54-760b-4bc4-9662-f0fa49db1c0e tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Task: {'id': task-1111971, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.505183} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1112.728226] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-3e2e3b54-760b-4bc4-9662-f0fa49db1c0e tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk to [datastore1] c9e07a6c-fd14-4071-8c69-6ece62e9109a/c9e07a6c-fd14-4071-8c69-6ece62e9109a.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1112.728434] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-3e2e3b54-760b-4bc4-9662-f0fa49db1c0e tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] [instance: c9e07a6c-fd14-4071-8c69-6ece62e9109a] Extending root virtual disk to 1048576 {{(pid=63515) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1112.728652] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-17ce2fc0-a0a5-4f58-8f9f-4f7b5f7a925d {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.734186] env[63515]: DEBUG oslo_vmware.api [None req-3e2e3b54-760b-4bc4-9662-f0fa49db1c0e tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Waiting for the task: (returnval){ [ 1112.734186] env[63515]: value = "task-1111972" [ 1112.734186] env[63515]: _type = "Task" [ 1112.734186] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1112.742641] env[63515]: DEBUG oslo_vmware.api [None req-3e2e3b54-760b-4bc4-9662-f0fa49db1c0e tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Task: {'id': task-1111972, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1112.756213] env[63515]: DEBUG nova.network.neutron [None req-121e8bf6-aa17-4f45-a156-987d42701ccb tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: aa83d035-9c47-434d-ab1a-140e6b3110b4] Port 614aed8f-3ab8-4672-ac13-0ae8cb0b233d binding to destination host cpu-1 is already ACTIVE {{(pid=63515) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 1112.756495] env[63515]: DEBUG oslo_concurrency.lockutils [None req-121e8bf6-aa17-4f45-a156-987d42701ccb tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Acquiring lock "refresh_cache-aa83d035-9c47-434d-ab1a-140e6b3110b4" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1112.756654] env[63515]: DEBUG oslo_concurrency.lockutils [None req-121e8bf6-aa17-4f45-a156-987d42701ccb tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Acquired lock "refresh_cache-aa83d035-9c47-434d-ab1a-140e6b3110b4" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1112.756822] env[63515]: DEBUG nova.network.neutron [None req-121e8bf6-aa17-4f45-a156-987d42701ccb tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: aa83d035-9c47-434d-ab1a-140e6b3110b4] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1113.093356] env[63515]: DEBUG nova.compute.manager [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] [instance: 8eb17506-ff93-4d25-b9af-ec5886569e65] Instance has had 0 of 5 cleanup attempts {{(pid=63515) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1113.244599] env[63515]: DEBUG oslo_vmware.api [None req-3e2e3b54-760b-4bc4-9662-f0fa49db1c0e tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Task: {'id': task-1111972, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.067265} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1113.244893] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-3e2e3b54-760b-4bc4-9662-f0fa49db1c0e tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] [instance: c9e07a6c-fd14-4071-8c69-6ece62e9109a] Extended root virtual disk {{(pid=63515) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1113.245714] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49564254-b2d7-4868-a46e-0b6bef57edfe {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.270342] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-3e2e3b54-760b-4bc4-9662-f0fa49db1c0e tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] [instance: c9e07a6c-fd14-4071-8c69-6ece62e9109a] Reconfiguring VM instance instance-0000006c to attach disk [datastore1] c9e07a6c-fd14-4071-8c69-6ece62e9109a/c9e07a6c-fd14-4071-8c69-6ece62e9109a.vmdk or device None with type sparse {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1113.270660] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5617670b-31a7-4370-a732-ce21869f8d75 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.292033] env[63515]: DEBUG oslo_vmware.api [None req-3e2e3b54-760b-4bc4-9662-f0fa49db1c0e tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Waiting for the task: (returnval){ [ 1113.292033] env[63515]: value = "task-1111973" [ 1113.292033] env[63515]: _type = "Task" [ 1113.292033] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1113.299741] env[63515]: DEBUG oslo_vmware.api [None req-3e2e3b54-760b-4bc4-9662-f0fa49db1c0e tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Task: {'id': task-1111973, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1113.596959] env[63515]: DEBUG nova.compute.manager [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] [instance: 38d6d030-06b0-4185-904d-44a038b3a752] Instance has had 0 of 5 cleanup attempts {{(pid=63515) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1113.742532] env[63515]: DEBUG nova.network.neutron [None req-121e8bf6-aa17-4f45-a156-987d42701ccb tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: aa83d035-9c47-434d-ab1a-140e6b3110b4] Updating instance_info_cache with network_info: [{"id": "614aed8f-3ab8-4672-ac13-0ae8cb0b233d", "address": "fa:16:3e:98:68:38", "network": {"id": "5758c885-e1c9-45c8-86aa-27a802101f52", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-734458138-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.204", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b18d4daacbc84e758f9f9ae4f3f09d28", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69351262-8d39-441a-85ba-3a78df436d17", "external-id": "nsx-vlan-transportzone-205", "segmentation_id": 205, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap614aed8f-3a", "ovs_interfaceid": "614aed8f-3ab8-4672-ac13-0ae8cb0b233d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1113.802545] env[63515]: DEBUG oslo_vmware.api [None req-3e2e3b54-760b-4bc4-9662-f0fa49db1c0e tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Task: {'id': task-1111973, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1114.101132] env[63515]: DEBUG nova.compute.manager [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] [instance: 6688cdaa-29ba-413a-8131-4f834cdb70e4] Instance has had 0 of 5 cleanup attempts {{(pid=63515) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1114.245172] env[63515]: DEBUG oslo_concurrency.lockutils [None req-121e8bf6-aa17-4f45-a156-987d42701ccb tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Releasing lock "refresh_cache-aa83d035-9c47-434d-ab1a-140e6b3110b4" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1114.301692] env[63515]: DEBUG oslo_vmware.api [None req-3e2e3b54-760b-4bc4-9662-f0fa49db1c0e tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Task: {'id': task-1111973, 'name': ReconfigVM_Task, 'duration_secs': 0.751435} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1114.301967] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-3e2e3b54-760b-4bc4-9662-f0fa49db1c0e tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] [instance: c9e07a6c-fd14-4071-8c69-6ece62e9109a] Reconfigured VM instance instance-0000006c to attach disk [datastore1] c9e07a6c-fd14-4071-8c69-6ece62e9109a/c9e07a6c-fd14-4071-8c69-6ece62e9109a.vmdk or device None with type sparse {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1114.302601] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ea08771f-9d6d-4c48-98c7-ebbc741f03fb {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.308658] env[63515]: DEBUG oslo_vmware.api [None req-3e2e3b54-760b-4bc4-9662-f0fa49db1c0e tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Waiting for the task: (returnval){ [ 1114.308658] env[63515]: value = "task-1111974" [ 1114.308658] env[63515]: _type = "Task" [ 1114.308658] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1114.317496] env[63515]: DEBUG oslo_vmware.api [None req-3e2e3b54-760b-4bc4-9662-f0fa49db1c0e tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Task: {'id': task-1111974, 'name': Rename_Task} progress is 5%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1114.604191] env[63515]: DEBUG nova.compute.manager [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] [instance: d2a4f66a-f1a2-4f0d-826d-0b7c4f04bd23] Instance has had 0 of 5 cleanup attempts {{(pid=63515) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1114.748812] env[63515]: DEBUG nova.compute.manager [None req-121e8bf6-aa17-4f45-a156-987d42701ccb tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: aa83d035-9c47-434d-ab1a-140e6b3110b4] Hypervisor driver does not support instance shared storage check, assuming it's not on shared storage {{(pid=63515) _is_instance_storage_shared /opt/stack/nova/nova/compute/manager.py:897}} [ 1114.819051] env[63515]: DEBUG oslo_vmware.api [None req-3e2e3b54-760b-4bc4-9662-f0fa49db1c0e tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Task: {'id': task-1111974, 'name': Rename_Task, 'duration_secs': 0.164259} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1114.819172] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-3e2e3b54-760b-4bc4-9662-f0fa49db1c0e tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] [instance: c9e07a6c-fd14-4071-8c69-6ece62e9109a] Powering on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1114.819386] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-bba4db1b-6f32-4457-8f9d-e062cf159976 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.825123] env[63515]: DEBUG oslo_vmware.api [None req-3e2e3b54-760b-4bc4-9662-f0fa49db1c0e tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Waiting for the task: (returnval){ [ 1114.825123] env[63515]: value = "task-1111976" [ 1114.825123] env[63515]: _type = "Task" [ 1114.825123] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1114.833441] env[63515]: DEBUG oslo_vmware.api [None req-3e2e3b54-760b-4bc4-9662-f0fa49db1c0e tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Task: {'id': task-1111976, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1115.107414] env[63515]: DEBUG nova.compute.manager [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] [instance: 001a48e1-b237-41f8-a6f4-2fe98d173512] Instance has had 0 of 5 cleanup attempts {{(pid=63515) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1115.335274] env[63515]: DEBUG oslo_vmware.api [None req-3e2e3b54-760b-4bc4-9662-f0fa49db1c0e tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Task: {'id': task-1111976, 'name': PowerOnVM_Task, 'duration_secs': 0.453505} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1115.335487] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-3e2e3b54-760b-4bc4-9662-f0fa49db1c0e tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] [instance: c9e07a6c-fd14-4071-8c69-6ece62e9109a] Powered on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1115.335697] env[63515]: INFO nova.compute.manager [None req-3e2e3b54-760b-4bc4-9662-f0fa49db1c0e tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] [instance: c9e07a6c-fd14-4071-8c69-6ece62e9109a] Took 7.18 seconds to spawn the instance on the hypervisor. [ 1115.335880] env[63515]: DEBUG nova.compute.manager [None req-3e2e3b54-760b-4bc4-9662-f0fa49db1c0e tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] [instance: c9e07a6c-fd14-4071-8c69-6ece62e9109a] Checking state {{(pid=63515) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1115.336724] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17865e5d-8eb4-4dec-9f4b-a65ae2d6e4d5 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.610568] env[63515]: DEBUG nova.compute.manager [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] [instance: de210780-5c0f-4fba-883c-13707566a2e1] Instance has had 0 of 5 cleanup attempts {{(pid=63515) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1115.852090] env[63515]: DEBUG oslo_concurrency.lockutils [None req-121e8bf6-aa17-4f45-a156-987d42701ccb tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1115.852330] env[63515]: DEBUG oslo_concurrency.lockutils [None req-121e8bf6-aa17-4f45-a156-987d42701ccb tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1115.855983] env[63515]: INFO nova.compute.manager [None req-3e2e3b54-760b-4bc4-9662-f0fa49db1c0e tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] [instance: c9e07a6c-fd14-4071-8c69-6ece62e9109a] Took 11.92 seconds to build instance. [ 1116.113714] env[63515]: DEBUG nova.compute.manager [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] [instance: ed550b10-d58f-45b8-b766-198f431c3788] Instance has had 0 of 5 cleanup attempts {{(pid=63515) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1116.356544] env[63515]: DEBUG nova.objects.instance [None req-121e8bf6-aa17-4f45-a156-987d42701ccb tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Lazy-loading 'migration_context' on Instance uuid aa83d035-9c47-434d-ab1a-140e6b3110b4 {{(pid=63515) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1116.357893] env[63515]: DEBUG oslo_concurrency.lockutils [None req-3e2e3b54-760b-4bc4-9662-f0fa49db1c0e tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Lock "c9e07a6c-fd14-4071-8c69-6ece62e9109a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.429s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1116.471290] env[63515]: DEBUG oslo_concurrency.lockutils [None req-66223c7a-f7fb-4712-8ddb-378277f7abf5 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Acquiring lock "7b0df8d3-1b14-4e00-90c0-3dab3efc153c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1116.471568] env[63515]: DEBUG oslo_concurrency.lockutils [None req-66223c7a-f7fb-4712-8ddb-378277f7abf5 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Lock "7b0df8d3-1b14-4e00-90c0-3dab3efc153c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1116.617339] env[63515]: DEBUG nova.compute.manager [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] [instance: e91aa479-1540-4950-851b-b2409e5f89f1] Instance has had 0 of 5 cleanup attempts {{(pid=63515) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1116.953826] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20dd33e3-be27-4b58-bd35-b86624630b96 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.961468] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e811f58-f334-44e9-82e4-a4ce49b45b3d {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.991415] env[63515]: DEBUG nova.compute.manager [None req-66223c7a-f7fb-4712-8ddb-378277f7abf5 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] [instance: 7b0df8d3-1b14-4e00-90c0-3dab3efc153c] Starting instance... {{(pid=63515) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1116.994814] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-622c50e8-e16b-4ce4-ae0a-3e267502d1f6 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.002417] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3c714bf-f76b-456d-be25-d5137b0f0188 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.015483] env[63515]: DEBUG nova.compute.provider_tree [None req-121e8bf6-aa17-4f45-a156-987d42701ccb tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1117.121033] env[63515]: DEBUG nova.compute.manager [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] [instance: d7e8dcdb-41cb-46fb-8b61-d251e7c2d372] Instance has had 0 of 5 cleanup attempts {{(pid=63515) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1117.512522] env[63515]: DEBUG oslo_concurrency.lockutils [None req-66223c7a-f7fb-4712-8ddb-378277f7abf5 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1117.518507] env[63515]: DEBUG nova.scheduler.client.report [None req-121e8bf6-aa17-4f45-a156-987d42701ccb tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1117.623748] env[63515]: DEBUG nova.compute.manager [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] [instance: b4477e66-ae12-4929-90ed-b7b652e0f207] Instance has had 0 of 5 cleanup attempts {{(pid=63515) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1118.126535] env[63515]: DEBUG nova.compute.manager [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] [instance: 37e6c27e-317b-45d2-bd55-2fd78ccf009f] Instance has had 0 of 5 cleanup attempts {{(pid=63515) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1118.529342] env[63515]: DEBUG oslo_concurrency.lockutils [None req-121e8bf6-aa17-4f45-a156-987d42701ccb tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: held 2.677s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1118.535207] env[63515]: DEBUG oslo_concurrency.lockutils [None req-66223c7a-f7fb-4712-8ddb-378277f7abf5 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.023s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1118.536646] env[63515]: INFO nova.compute.claims [None req-66223c7a-f7fb-4712-8ddb-378277f7abf5 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] [instance: 7b0df8d3-1b14-4e00-90c0-3dab3efc153c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1118.629830] env[63515]: DEBUG nova.compute.manager [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] [instance: 452f3ca2-6141-43b2-a77a-c9ab5754192d] Instance has had 0 of 5 cleanup attempts {{(pid=63515) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1119.133122] env[63515]: DEBUG nova.compute.manager [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] [instance: 87c468d9-9594-4804-b461-527f01f6118f] Instance has had 0 of 5 cleanup attempts {{(pid=63515) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1119.633534] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8cf2886-687a-4d3d-8e87-b2121ca31578 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.636194] env[63515]: DEBUG nova.compute.manager [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] [instance: 396e49dd-48c0-496b-a1ec-190c33a22c5e] Instance has had 0 of 5 cleanup attempts {{(pid=63515) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1119.642644] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03b68f4c-4633-40a3-a597-9867451e42ef {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.672794] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b862da93-cb0b-4bea-bdc7-a77dbf4d2cf1 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.680738] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0ceb43f-85e1-4ffc-b4ef-14568daa2e37 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.694058] env[63515]: DEBUG nova.compute.provider_tree [None req-66223c7a-f7fb-4712-8ddb-378277f7abf5 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1120.073398] env[63515]: INFO nova.compute.manager [None req-121e8bf6-aa17-4f45-a156-987d42701ccb tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: aa83d035-9c47-434d-ab1a-140e6b3110b4] Swapping old allocation on dict_keys(['2bd86232-2b6f-44d5-9057-1a3a6b27185a']) held by migration d9a7b867-b152-403e-8f62-d0daeb836c83 for instance [ 1120.094784] env[63515]: DEBUG nova.scheduler.client.report [None req-121e8bf6-aa17-4f45-a156-987d42701ccb tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Overwriting current allocation {'allocations': {'2bd86232-2b6f-44d5-9057-1a3a6b27185a': {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}, 'generation': 158}}, 'project_id': 'b18d4daacbc84e758f9f9ae4f3f09d28', 'user_id': '82fe830c309f41a4ab512a25d8481742', 'consumer_generation': 1} on consumer aa83d035-9c47-434d-ab1a-140e6b3110b4 {{(pid=63515) move_allocations /opt/stack/nova/nova/scheduler/client/report.py:2032}} [ 1120.139377] env[63515]: DEBUG nova.compute.manager [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] [instance: 90f4930b-aaa0-4c4b-9ab8-92aed45e200b] Instance has had 0 of 5 cleanup attempts {{(pid=63515) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1120.174911] env[63515]: DEBUG oslo_concurrency.lockutils [None req-121e8bf6-aa17-4f45-a156-987d42701ccb tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Acquiring lock "refresh_cache-aa83d035-9c47-434d-ab1a-140e6b3110b4" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1120.175129] env[63515]: DEBUG oslo_concurrency.lockutils [None req-121e8bf6-aa17-4f45-a156-987d42701ccb tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Acquired lock "refresh_cache-aa83d035-9c47-434d-ab1a-140e6b3110b4" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1120.175317] env[63515]: DEBUG nova.network.neutron [None req-121e8bf6-aa17-4f45-a156-987d42701ccb tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: aa83d035-9c47-434d-ab1a-140e6b3110b4] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1120.196332] env[63515]: DEBUG nova.scheduler.client.report [None req-66223c7a-f7fb-4712-8ddb-378277f7abf5 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1120.645113] env[63515]: DEBUG nova.compute.manager [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] [instance: f1d01b75-ac9d-458d-8cc2-ae64cffca4e8] Instance has had 0 of 5 cleanup attempts {{(pid=63515) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1120.703436] env[63515]: DEBUG oslo_concurrency.lockutils [None req-66223c7a-f7fb-4712-8ddb-378277f7abf5 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.168s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1120.703692] env[63515]: DEBUG nova.compute.manager [None req-66223c7a-f7fb-4712-8ddb-378277f7abf5 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] [instance: 7b0df8d3-1b14-4e00-90c0-3dab3efc153c] Start building networks asynchronously for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1120.892281] env[63515]: DEBUG nova.network.neutron [None req-121e8bf6-aa17-4f45-a156-987d42701ccb tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: aa83d035-9c47-434d-ab1a-140e6b3110b4] Updating instance_info_cache with network_info: [{"id": "614aed8f-3ab8-4672-ac13-0ae8cb0b233d", "address": "fa:16:3e:98:68:38", "network": {"id": "5758c885-e1c9-45c8-86aa-27a802101f52", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-734458138-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.204", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b18d4daacbc84e758f9f9ae4f3f09d28", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69351262-8d39-441a-85ba-3a78df436d17", "external-id": "nsx-vlan-transportzone-205", "segmentation_id": 205, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap614aed8f-3a", "ovs_interfaceid": "614aed8f-3ab8-4672-ac13-0ae8cb0b233d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1121.148970] env[63515]: DEBUG nova.compute.manager [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] [instance: 5b42f744-fdd6-45b1-8563-896869648c23] Instance has had 0 of 5 cleanup attempts {{(pid=63515) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1121.210726] env[63515]: DEBUG nova.compute.utils [None req-66223c7a-f7fb-4712-8ddb-378277f7abf5 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Using /dev/sd instead of None {{(pid=63515) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1121.212804] env[63515]: DEBUG nova.compute.manager [None req-66223c7a-f7fb-4712-8ddb-378277f7abf5 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] [instance: 7b0df8d3-1b14-4e00-90c0-3dab3efc153c] Allocating IP information in the background. {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1121.212909] env[63515]: DEBUG nova.network.neutron [None req-66223c7a-f7fb-4712-8ddb-378277f7abf5 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] [instance: 7b0df8d3-1b14-4e00-90c0-3dab3efc153c] allocate_for_instance() {{(pid=63515) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1121.266914] env[63515]: DEBUG nova.policy [None req-66223c7a-f7fb-4712-8ddb-378277f7abf5 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e8da40be55174af2a0f4dd3bf2844cb4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'fe873600ed6d4a3eb1ee3d3a19fceaf7', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63515) authorize /opt/stack/nova/nova/policy.py:201}} [ 1121.395275] env[63515]: DEBUG oslo_concurrency.lockutils [None req-121e8bf6-aa17-4f45-a156-987d42701ccb tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Releasing lock "refresh_cache-aa83d035-9c47-434d-ab1a-140e6b3110b4" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1121.396318] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f5a5b9f-273f-4253-abaf-b8e858611144 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.404314] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00af80c2-4e92-4555-abf7-928e6ebd4018 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.511137] env[63515]: DEBUG nova.network.neutron [None req-66223c7a-f7fb-4712-8ddb-378277f7abf5 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] [instance: 7b0df8d3-1b14-4e00-90c0-3dab3efc153c] Successfully created port: fd2511c0-002c-4a92-a30a-49c9e58641b3 {{(pid=63515) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1121.652825] env[63515]: DEBUG nova.compute.manager [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] [instance: b911a5b5-9617-4fb3-9b5e-fb8c492e4931] Instance has had 0 of 5 cleanup attempts {{(pid=63515) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1121.682552] env[63515]: DEBUG oslo_concurrency.lockutils [None req-a2d941f8-9d25-4f98-968e-6daf11b6ab5a tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Acquiring lock "9a864a51-7df3-4afc-8be7-2794ef2dcf1a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1121.683999] env[63515]: DEBUG oslo_concurrency.lockutils [None req-a2d941f8-9d25-4f98-968e-6daf11b6ab5a tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Lock "9a864a51-7df3-4afc-8be7-2794ef2dcf1a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1121.716324] env[63515]: DEBUG nova.compute.manager [None req-66223c7a-f7fb-4712-8ddb-378277f7abf5 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] [instance: 7b0df8d3-1b14-4e00-90c0-3dab3efc153c] Start building block device mappings for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1122.157058] env[63515]: DEBUG oslo_service.periodic_task [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Running periodic task ComputeManager._cleanup_incomplete_migrations {{(pid=63515) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1122.157196] env[63515]: DEBUG nova.compute.manager [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Cleaning up deleted instances with incomplete migration {{(pid=63515) _cleanup_incomplete_migrations /opt/stack/nova/nova/compute/manager.py:11264}} [ 1122.185103] env[63515]: DEBUG nova.compute.manager [None req-a2d941f8-9d25-4f98-968e-6daf11b6ab5a tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: 9a864a51-7df3-4afc-8be7-2794ef2dcf1a] Starting instance... {{(pid=63515) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1122.487129] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-121e8bf6-aa17-4f45-a156-987d42701ccb tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: aa83d035-9c47-434d-ab1a-140e6b3110b4] Powering off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1122.487499] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d84ba428-7ed4-46dc-bd40-513c50279382 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.495641] env[63515]: DEBUG oslo_vmware.api [None req-121e8bf6-aa17-4f45-a156-987d42701ccb tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Waiting for the task: (returnval){ [ 1122.495641] env[63515]: value = "task-1111979" [ 1122.495641] env[63515]: _type = "Task" [ 1122.495641] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1122.506031] env[63515]: DEBUG oslo_vmware.api [None req-121e8bf6-aa17-4f45-a156-987d42701ccb tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Task: {'id': task-1111979, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1122.709863] env[63515]: DEBUG oslo_concurrency.lockutils [None req-a2d941f8-9d25-4f98-968e-6daf11b6ab5a tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1122.709863] env[63515]: DEBUG oslo_concurrency.lockutils [None req-a2d941f8-9d25-4f98-968e-6daf11b6ab5a tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.001s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1122.711298] env[63515]: INFO nova.compute.claims [None req-a2d941f8-9d25-4f98-968e-6daf11b6ab5a tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: 9a864a51-7df3-4afc-8be7-2794ef2dcf1a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1122.724639] env[63515]: DEBUG nova.compute.manager [None req-66223c7a-f7fb-4712-8ddb-378277f7abf5 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] [instance: 7b0df8d3-1b14-4e00-90c0-3dab3efc153c] Start spawning the instance on the hypervisor. {{(pid=63515) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1122.751305] env[63515]: DEBUG nova.virt.hardware [None req-66223c7a-f7fb-4712-8ddb-378277f7abf5 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T02:52:14Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T02:51:56Z,direct_url=,disk_format='vmdk',id=8a120570-cb06-4099-b262-554ca0ad15c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b569255cc43e42e7ae3f2b4ad37c6ef4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T02:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1122.751556] env[63515]: DEBUG nova.virt.hardware [None req-66223c7a-f7fb-4712-8ddb-378277f7abf5 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Flavor limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1122.751723] env[63515]: DEBUG nova.virt.hardware [None req-66223c7a-f7fb-4712-8ddb-378277f7abf5 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Image limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1122.751946] env[63515]: DEBUG nova.virt.hardware [None req-66223c7a-f7fb-4712-8ddb-378277f7abf5 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Flavor pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1122.752079] env[63515]: DEBUG nova.virt.hardware [None req-66223c7a-f7fb-4712-8ddb-378277f7abf5 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Image pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1122.752238] env[63515]: DEBUG nova.virt.hardware [None req-66223c7a-f7fb-4712-8ddb-378277f7abf5 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1122.752453] env[63515]: DEBUG nova.virt.hardware [None req-66223c7a-f7fb-4712-8ddb-378277f7abf5 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1122.752617] env[63515]: DEBUG nova.virt.hardware [None req-66223c7a-f7fb-4712-8ddb-378277f7abf5 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1122.752788] env[63515]: DEBUG nova.virt.hardware [None req-66223c7a-f7fb-4712-8ddb-378277f7abf5 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Got 1 possible topologies {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1122.752955] env[63515]: DEBUG nova.virt.hardware [None req-66223c7a-f7fb-4712-8ddb-378277f7abf5 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1122.753144] env[63515]: DEBUG nova.virt.hardware [None req-66223c7a-f7fb-4712-8ddb-378277f7abf5 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1122.754303] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5903d1c8-f55f-410b-b356-4e4f9ba54b71 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.763051] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02d0f2ab-cbef-45b3-ac1b-b124f82b2968 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.885082] env[63515]: DEBUG nova.compute.manager [req-2111f511-5728-437c-aa43-d69b2ea1be43 req-21e59269-ae92-4021-aee6-eb8f5f305a7c service nova] [instance: 7b0df8d3-1b14-4e00-90c0-3dab3efc153c] Received event network-vif-plugged-fd2511c0-002c-4a92-a30a-49c9e58641b3 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1122.885337] env[63515]: DEBUG oslo_concurrency.lockutils [req-2111f511-5728-437c-aa43-d69b2ea1be43 req-21e59269-ae92-4021-aee6-eb8f5f305a7c service nova] Acquiring lock "7b0df8d3-1b14-4e00-90c0-3dab3efc153c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1122.885556] env[63515]: DEBUG oslo_concurrency.lockutils [req-2111f511-5728-437c-aa43-d69b2ea1be43 req-21e59269-ae92-4021-aee6-eb8f5f305a7c service nova] Lock "7b0df8d3-1b14-4e00-90c0-3dab3efc153c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1122.885729] env[63515]: DEBUG oslo_concurrency.lockutils [req-2111f511-5728-437c-aa43-d69b2ea1be43 req-21e59269-ae92-4021-aee6-eb8f5f305a7c service nova] Lock "7b0df8d3-1b14-4e00-90c0-3dab3efc153c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1122.885947] env[63515]: DEBUG nova.compute.manager [req-2111f511-5728-437c-aa43-d69b2ea1be43 req-21e59269-ae92-4021-aee6-eb8f5f305a7c service nova] [instance: 7b0df8d3-1b14-4e00-90c0-3dab3efc153c] No waiting events found dispatching network-vif-plugged-fd2511c0-002c-4a92-a30a-49c9e58641b3 {{(pid=63515) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1122.886152] env[63515]: WARNING nova.compute.manager [req-2111f511-5728-437c-aa43-d69b2ea1be43 req-21e59269-ae92-4021-aee6-eb8f5f305a7c service nova] [instance: 7b0df8d3-1b14-4e00-90c0-3dab3efc153c] Received unexpected event network-vif-plugged-fd2511c0-002c-4a92-a30a-49c9e58641b3 for instance with vm_state building and task_state spawning. [ 1123.005272] env[63515]: DEBUG oslo_vmware.api [None req-121e8bf6-aa17-4f45-a156-987d42701ccb tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Task: {'id': task-1111979, 'name': PowerOffVM_Task, 'duration_secs': 0.271102} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1123.005553] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-121e8bf6-aa17-4f45-a156-987d42701ccb tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: aa83d035-9c47-434d-ab1a-140e6b3110b4] Powered off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1123.006210] env[63515]: DEBUG nova.virt.hardware [None req-121e8bf6-aa17-4f45-a156-987d42701ccb tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T02:52:14Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=8a120570-cb06-4099-b262-554ca0ad15c5,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1123.006430] env[63515]: DEBUG nova.virt.hardware [None req-121e8bf6-aa17-4f45-a156-987d42701ccb tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Flavor limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1123.006592] env[63515]: DEBUG nova.virt.hardware [None req-121e8bf6-aa17-4f45-a156-987d42701ccb tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Image limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1123.007221] env[63515]: DEBUG nova.virt.hardware [None req-121e8bf6-aa17-4f45-a156-987d42701ccb tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Flavor pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1123.007221] env[63515]: DEBUG nova.virt.hardware [None req-121e8bf6-aa17-4f45-a156-987d42701ccb tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Image pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1123.007221] env[63515]: DEBUG nova.virt.hardware [None req-121e8bf6-aa17-4f45-a156-987d42701ccb tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1123.007394] env[63515]: DEBUG nova.virt.hardware [None req-121e8bf6-aa17-4f45-a156-987d42701ccb tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1123.007429] env[63515]: DEBUG nova.virt.hardware [None req-121e8bf6-aa17-4f45-a156-987d42701ccb tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1123.007593] env[63515]: DEBUG nova.virt.hardware [None req-121e8bf6-aa17-4f45-a156-987d42701ccb tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Got 1 possible topologies {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1123.007760] env[63515]: DEBUG nova.virt.hardware [None req-121e8bf6-aa17-4f45-a156-987d42701ccb tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1123.007933] env[63515]: DEBUG nova.virt.hardware [None req-121e8bf6-aa17-4f45-a156-987d42701ccb tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1123.012873] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3fa89d8e-ad10-408c-812d-849a6ba63cb9 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.028666] env[63515]: DEBUG oslo_vmware.api [None req-121e8bf6-aa17-4f45-a156-987d42701ccb tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Waiting for the task: (returnval){ [ 1123.028666] env[63515]: value = "task-1111980" [ 1123.028666] env[63515]: _type = "Task" [ 1123.028666] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1123.036188] env[63515]: DEBUG oslo_vmware.api [None req-121e8bf6-aa17-4f45-a156-987d42701ccb tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Task: {'id': task-1111980, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1123.151161] env[63515]: DEBUG oslo_service.periodic_task [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=63515) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1123.151850] env[63515]: DEBUG oslo_service.periodic_task [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Running periodic task ComputeManager._sync_power_states {{(pid=63515) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1123.456398] env[63515]: DEBUG nova.network.neutron [None req-66223c7a-f7fb-4712-8ddb-378277f7abf5 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] [instance: 7b0df8d3-1b14-4e00-90c0-3dab3efc153c] Successfully updated port: fd2511c0-002c-4a92-a30a-49c9e58641b3 {{(pid=63515) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1123.496458] env[63515]: DEBUG nova.compute.manager [req-b57c196f-8bbd-45ce-87b3-21fa4ad4d94e req-5fb4dfe7-2376-44a0-993b-6fab243db1f2 service nova] [instance: 7b0df8d3-1b14-4e00-90c0-3dab3efc153c] Received event network-changed-fd2511c0-002c-4a92-a30a-49c9e58641b3 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1123.496686] env[63515]: DEBUG nova.compute.manager [req-b57c196f-8bbd-45ce-87b3-21fa4ad4d94e req-5fb4dfe7-2376-44a0-993b-6fab243db1f2 service nova] [instance: 7b0df8d3-1b14-4e00-90c0-3dab3efc153c] Refreshing instance network info cache due to event network-changed-fd2511c0-002c-4a92-a30a-49c9e58641b3. {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1123.496902] env[63515]: DEBUG oslo_concurrency.lockutils [req-b57c196f-8bbd-45ce-87b3-21fa4ad4d94e req-5fb4dfe7-2376-44a0-993b-6fab243db1f2 service nova] Acquiring lock "refresh_cache-7b0df8d3-1b14-4e00-90c0-3dab3efc153c" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1123.497094] env[63515]: DEBUG oslo_concurrency.lockutils [req-b57c196f-8bbd-45ce-87b3-21fa4ad4d94e req-5fb4dfe7-2376-44a0-993b-6fab243db1f2 service nova] Acquired lock "refresh_cache-7b0df8d3-1b14-4e00-90c0-3dab3efc153c" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1123.497617] env[63515]: DEBUG nova.network.neutron [req-b57c196f-8bbd-45ce-87b3-21fa4ad4d94e req-5fb4dfe7-2376-44a0-993b-6fab243db1f2 service nova] [instance: 7b0df8d3-1b14-4e00-90c0-3dab3efc153c] Refreshing network info cache for port fd2511c0-002c-4a92-a30a-49c9e58641b3 {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1123.538934] env[63515]: DEBUG oslo_vmware.api [None req-121e8bf6-aa17-4f45-a156-987d42701ccb tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Task: {'id': task-1111980, 'name': ReconfigVM_Task, 'duration_secs': 0.182347} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1123.539841] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c09d83c6-494e-4ed9-9436-35be0dcbcf56 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.562177] env[63515]: DEBUG nova.virt.hardware [None req-121e8bf6-aa17-4f45-a156-987d42701ccb tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T02:52:14Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=8a120570-cb06-4099-b262-554ca0ad15c5,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1123.562429] env[63515]: DEBUG nova.virt.hardware [None req-121e8bf6-aa17-4f45-a156-987d42701ccb tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Flavor limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1123.562708] env[63515]: DEBUG nova.virt.hardware [None req-121e8bf6-aa17-4f45-a156-987d42701ccb tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Image limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1123.562835] env[63515]: DEBUG nova.virt.hardware [None req-121e8bf6-aa17-4f45-a156-987d42701ccb tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Flavor pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1123.562990] env[63515]: DEBUG nova.virt.hardware [None req-121e8bf6-aa17-4f45-a156-987d42701ccb tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Image pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1123.563159] env[63515]: DEBUG nova.virt.hardware [None req-121e8bf6-aa17-4f45-a156-987d42701ccb tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1123.563367] env[63515]: DEBUG nova.virt.hardware [None req-121e8bf6-aa17-4f45-a156-987d42701ccb tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1123.563530] env[63515]: DEBUG nova.virt.hardware [None req-121e8bf6-aa17-4f45-a156-987d42701ccb tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1123.563702] env[63515]: DEBUG nova.virt.hardware [None req-121e8bf6-aa17-4f45-a156-987d42701ccb tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Got 1 possible topologies {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1123.563867] env[63515]: DEBUG nova.virt.hardware [None req-121e8bf6-aa17-4f45-a156-987d42701ccb tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1123.564052] env[63515]: DEBUG nova.virt.hardware [None req-121e8bf6-aa17-4f45-a156-987d42701ccb tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1123.564805] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-97bc9c57-68ab-4f35-97a9-f9331b928a56 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.570389] env[63515]: DEBUG oslo_vmware.api [None req-121e8bf6-aa17-4f45-a156-987d42701ccb tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Waiting for the task: (returnval){ [ 1123.570389] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]52b7729f-74d5-6aca-369f-da573332435c" [ 1123.570389] env[63515]: _type = "Task" [ 1123.570389] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1123.578519] env[63515]: DEBUG oslo_vmware.api [None req-121e8bf6-aa17-4f45-a156-987d42701ccb tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52b7729f-74d5-6aca-369f-da573332435c, 'name': SearchDatastore_Task} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1123.583824] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-121e8bf6-aa17-4f45-a156-987d42701ccb tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: aa83d035-9c47-434d-ab1a-140e6b3110b4] Reconfiguring VM instance instance-00000066 to detach disk 2000 {{(pid=63515) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1123.584120] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7cdb1c1b-3db5-4a84-9b9c-48d68d041e2c {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.601916] env[63515]: DEBUG oslo_vmware.api [None req-121e8bf6-aa17-4f45-a156-987d42701ccb tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Waiting for the task: (returnval){ [ 1123.601916] env[63515]: value = "task-1111981" [ 1123.601916] env[63515]: _type = "Task" [ 1123.601916] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1123.609490] env[63515]: DEBUG oslo_vmware.api [None req-121e8bf6-aa17-4f45-a156-987d42701ccb tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Task: {'id': task-1111981, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1123.655924] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Getting list of instances from cluster (obj){ [ 1123.655924] env[63515]: value = "domain-c8" [ 1123.655924] env[63515]: _type = "ClusterComputeResource" [ 1123.655924] env[63515]: } {{(pid=63515) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2116}} [ 1123.656911] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20d90d20-16c8-4ccd-88c7-ffe8e86014a4 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.670736] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Got total of 5 instances {{(pid=63515) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2125}} [ 1123.670898] env[63515]: WARNING nova.compute.manager [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] While synchronizing instance power states, found 7 instances in the database and 5 instances on the hypervisor. [ 1123.671056] env[63515]: DEBUG nova.compute.manager [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Triggering sync for uuid e13da90c-28e6-43d2-99b7-19c5095954ca {{(pid=63515) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10336}} [ 1123.671249] env[63515]: DEBUG nova.compute.manager [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Triggering sync for uuid aa83d035-9c47-434d-ab1a-140e6b3110b4 {{(pid=63515) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10336}} [ 1123.671406] env[63515]: DEBUG nova.compute.manager [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Triggering sync for uuid ea2e0810-c9f1-49a6-8589-7db402847aea {{(pid=63515) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10336}} [ 1123.671558] env[63515]: DEBUG nova.compute.manager [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Triggering sync for uuid c8dbf542-70cb-45be-9308-5e12307c0be6 {{(pid=63515) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10336}} [ 1123.671709] env[63515]: DEBUG nova.compute.manager [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Triggering sync for uuid c9e07a6c-fd14-4071-8c69-6ece62e9109a {{(pid=63515) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10336}} [ 1123.671856] env[63515]: DEBUG nova.compute.manager [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Triggering sync for uuid 7b0df8d3-1b14-4e00-90c0-3dab3efc153c {{(pid=63515) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10336}} [ 1123.672028] env[63515]: DEBUG nova.compute.manager [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Triggering sync for uuid 9a864a51-7df3-4afc-8be7-2794ef2dcf1a {{(pid=63515) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10336}} [ 1123.672375] env[63515]: DEBUG oslo_concurrency.lockutils [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Acquiring lock "e13da90c-28e6-43d2-99b7-19c5095954ca" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1123.672594] env[63515]: DEBUG oslo_concurrency.lockutils [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Lock "e13da90c-28e6-43d2-99b7-19c5095954ca" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1123.672861] env[63515]: DEBUG oslo_concurrency.lockutils [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Acquiring lock "aa83d035-9c47-434d-ab1a-140e6b3110b4" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1123.673061] env[63515]: DEBUG oslo_concurrency.lockutils [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Lock "aa83d035-9c47-434d-ab1a-140e6b3110b4" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1123.673228] env[63515]: INFO nova.compute.manager [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] [instance: aa83d035-9c47-434d-ab1a-140e6b3110b4] During sync_power_state the instance has a pending task (resize_reverting). Skip. [ 1123.673482] env[63515]: DEBUG oslo_concurrency.lockutils [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Lock "aa83d035-9c47-434d-ab1a-140e6b3110b4" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1123.673581] env[63515]: DEBUG oslo_concurrency.lockutils [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Acquiring lock "ea2e0810-c9f1-49a6-8589-7db402847aea" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1123.673759] env[63515]: DEBUG oslo_concurrency.lockutils [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Lock "ea2e0810-c9f1-49a6-8589-7db402847aea" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1123.673993] env[63515]: DEBUG oslo_concurrency.lockutils [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Acquiring lock "c8dbf542-70cb-45be-9308-5e12307c0be6" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1123.674197] env[63515]: DEBUG oslo_concurrency.lockutils [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Lock "c8dbf542-70cb-45be-9308-5e12307c0be6" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1123.674422] env[63515]: DEBUG oslo_concurrency.lockutils [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Acquiring lock "c9e07a6c-fd14-4071-8c69-6ece62e9109a" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1123.674605] env[63515]: DEBUG oslo_concurrency.lockutils [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Lock "c9e07a6c-fd14-4071-8c69-6ece62e9109a" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1123.674832] env[63515]: DEBUG oslo_concurrency.lockutils [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Acquiring lock "7b0df8d3-1b14-4e00-90c0-3dab3efc153c" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1123.675052] env[63515]: DEBUG oslo_concurrency.lockutils [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Acquiring lock "9a864a51-7df3-4afc-8be7-2794ef2dcf1a" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1123.675829] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-709dee35-13b9-487e-a552-3c35532351cd {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.678938] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d972384-0ac1-450c-b542-58e255d0f316 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.681729] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dfe92867-aa9f-43ab-a8e8-a023dc0b1cd9 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.684450] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b0ac7bc-125d-47bd-94ee-0150c2cc17b3 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.812731] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0610be2-2280-41f9-9718-f885e86c66bf {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.820683] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66e1131c-6654-4b21-bc3c-b9b73216a995 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.852425] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ec8ad46-950d-428c-adc8-87bb98175af9 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.860367] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb99c1c8-89f9-4403-b7ef-124f5502b65b {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.873317] env[63515]: DEBUG nova.compute.provider_tree [None req-a2d941f8-9d25-4f98-968e-6daf11b6ab5a tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1123.959528] env[63515]: DEBUG oslo_concurrency.lockutils [None req-66223c7a-f7fb-4712-8ddb-378277f7abf5 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Acquiring lock "refresh_cache-7b0df8d3-1b14-4e00-90c0-3dab3efc153c" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1124.027874] env[63515]: DEBUG nova.network.neutron [req-b57c196f-8bbd-45ce-87b3-21fa4ad4d94e req-5fb4dfe7-2376-44a0-993b-6fab243db1f2 service nova] [instance: 7b0df8d3-1b14-4e00-90c0-3dab3efc153c] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1124.096215] env[63515]: DEBUG nova.network.neutron [req-b57c196f-8bbd-45ce-87b3-21fa4ad4d94e req-5fb4dfe7-2376-44a0-993b-6fab243db1f2 service nova] [instance: 7b0df8d3-1b14-4e00-90c0-3dab3efc153c] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1124.114061] env[63515]: DEBUG oslo_vmware.api [None req-121e8bf6-aa17-4f45-a156-987d42701ccb tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Task: {'id': task-1111981, 'name': ReconfigVM_Task, 'duration_secs': 0.22948} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1124.114061] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-121e8bf6-aa17-4f45-a156-987d42701ccb tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: aa83d035-9c47-434d-ab1a-140e6b3110b4] Reconfigured VM instance instance-00000066 to detach disk 2000 {{(pid=63515) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1124.114499] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4908bd89-8226-47bf-90dc-ae66f5fc67d1 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.139487] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-121e8bf6-aa17-4f45-a156-987d42701ccb tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: aa83d035-9c47-434d-ab1a-140e6b3110b4] Reconfiguring VM instance instance-00000066 to attach disk [datastore2] aa83d035-9c47-434d-ab1a-140e6b3110b4/aa83d035-9c47-434d-ab1a-140e6b3110b4.vmdk or device None with type thin {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1124.139787] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-65ae3219-1db6-4e61-ac83-808f7dfe8f20 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.157888] env[63515]: DEBUG oslo_vmware.api [None req-121e8bf6-aa17-4f45-a156-987d42701ccb tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Waiting for the task: (returnval){ [ 1124.157888] env[63515]: value = "task-1111982" [ 1124.157888] env[63515]: _type = "Task" [ 1124.157888] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1124.166701] env[63515]: DEBUG oslo_vmware.api [None req-121e8bf6-aa17-4f45-a156-987d42701ccb tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Task: {'id': task-1111982, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1124.198107] env[63515]: DEBUG oslo_concurrency.lockutils [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Lock "c9e07a6c-fd14-4071-8c69-6ece62e9109a" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.523s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1124.199822] env[63515]: DEBUG oslo_concurrency.lockutils [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Lock "e13da90c-28e6-43d2-99b7-19c5095954ca" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.527s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1124.200234] env[63515]: DEBUG oslo_concurrency.lockutils [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Lock "ea2e0810-c9f1-49a6-8589-7db402847aea" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.526s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1124.200579] env[63515]: DEBUG oslo_concurrency.lockutils [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Lock "c8dbf542-70cb-45be-9308-5e12307c0be6" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.526s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1124.376525] env[63515]: DEBUG nova.scheduler.client.report [None req-a2d941f8-9d25-4f98-968e-6daf11b6ab5a tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1124.599304] env[63515]: DEBUG oslo_concurrency.lockutils [req-b57c196f-8bbd-45ce-87b3-21fa4ad4d94e req-5fb4dfe7-2376-44a0-993b-6fab243db1f2 service nova] Releasing lock "refresh_cache-7b0df8d3-1b14-4e00-90c0-3dab3efc153c" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1124.599819] env[63515]: DEBUG oslo_concurrency.lockutils [None req-66223c7a-f7fb-4712-8ddb-378277f7abf5 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Acquired lock "refresh_cache-7b0df8d3-1b14-4e00-90c0-3dab3efc153c" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1124.600791] env[63515]: DEBUG nova.network.neutron [None req-66223c7a-f7fb-4712-8ddb-378277f7abf5 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] [instance: 7b0df8d3-1b14-4e00-90c0-3dab3efc153c] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1124.667826] env[63515]: DEBUG oslo_vmware.api [None req-121e8bf6-aa17-4f45-a156-987d42701ccb tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Task: {'id': task-1111982, 'name': ReconfigVM_Task, 'duration_secs': 0.293192} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1124.668129] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-121e8bf6-aa17-4f45-a156-987d42701ccb tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: aa83d035-9c47-434d-ab1a-140e6b3110b4] Reconfigured VM instance instance-00000066 to attach disk [datastore2] aa83d035-9c47-434d-ab1a-140e6b3110b4/aa83d035-9c47-434d-ab1a-140e6b3110b4.vmdk or device None with type thin {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1124.668996] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56e992b1-6538-4e6a-a7a6-455caabf720f {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.688650] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd746933-2b04-4675-b26a-a20bd274b7d7 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.708994] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4cceff86-c63f-4734-8604-e92f9f067d66 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.728839] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-300525cf-ec69-4d88-a4f5-37660f03d1f0 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.734850] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-121e8bf6-aa17-4f45-a156-987d42701ccb tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: aa83d035-9c47-434d-ab1a-140e6b3110b4] Powering on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1124.735076] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8afd4184-7b54-4cab-a06a-7720cd101993 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.740537] env[63515]: DEBUG oslo_vmware.api [None req-121e8bf6-aa17-4f45-a156-987d42701ccb tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Waiting for the task: (returnval){ [ 1124.740537] env[63515]: value = "task-1111983" [ 1124.740537] env[63515]: _type = "Task" [ 1124.740537] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1124.748156] env[63515]: DEBUG oslo_vmware.api [None req-121e8bf6-aa17-4f45-a156-987d42701ccb tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Task: {'id': task-1111983, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1124.881893] env[63515]: DEBUG oslo_concurrency.lockutils [None req-a2d941f8-9d25-4f98-968e-6daf11b6ab5a tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.172s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1124.882285] env[63515]: DEBUG nova.compute.manager [None req-a2d941f8-9d25-4f98-968e-6daf11b6ab5a tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: 9a864a51-7df3-4afc-8be7-2794ef2dcf1a] Start building networks asynchronously for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1125.131681] env[63515]: DEBUG nova.network.neutron [None req-66223c7a-f7fb-4712-8ddb-378277f7abf5 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] [instance: 7b0df8d3-1b14-4e00-90c0-3dab3efc153c] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1125.250500] env[63515]: DEBUG oslo_vmware.api [None req-121e8bf6-aa17-4f45-a156-987d42701ccb tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Task: {'id': task-1111983, 'name': PowerOnVM_Task, 'duration_secs': 0.37683} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1125.251346] env[63515]: DEBUG nova.network.neutron [None req-66223c7a-f7fb-4712-8ddb-378277f7abf5 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] [instance: 7b0df8d3-1b14-4e00-90c0-3dab3efc153c] Updating instance_info_cache with network_info: [{"id": "fd2511c0-002c-4a92-a30a-49c9e58641b3", "address": "fa:16:3e:90:c4:93", "network": {"id": "0e23b3a4-91ea-4bbe-9732-a3539d56b382", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1727906976-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fe873600ed6d4a3eb1ee3d3a19fceaf7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c7d2575f-b92f-44ec-a863-634cb76631a2", "external-id": "nsx-vlan-transportzone-794", "segmentation_id": 794, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfd2511c0-00", "ovs_interfaceid": "fd2511c0-002c-4a92-a30a-49c9e58641b3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1125.252720] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-121e8bf6-aa17-4f45-a156-987d42701ccb tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: aa83d035-9c47-434d-ab1a-140e6b3110b4] Powered on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1125.387477] env[63515]: DEBUG nova.compute.utils [None req-a2d941f8-9d25-4f98-968e-6daf11b6ab5a tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Using /dev/sd instead of None {{(pid=63515) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1125.389330] env[63515]: DEBUG nova.compute.manager [None req-a2d941f8-9d25-4f98-968e-6daf11b6ab5a tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: 9a864a51-7df3-4afc-8be7-2794ef2dcf1a] Allocating IP information in the background. {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1125.389522] env[63515]: DEBUG nova.network.neutron [None req-a2d941f8-9d25-4f98-968e-6daf11b6ab5a tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: 9a864a51-7df3-4afc-8be7-2794ef2dcf1a] allocate_for_instance() {{(pid=63515) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1125.425275] env[63515]: DEBUG nova.policy [None req-a2d941f8-9d25-4f98-968e-6daf11b6ab5a tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ed295ef8f0134bcaa668149ac9f4dd74', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'aef947bfaf7d4180b46b4be662e3d903', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63515) authorize /opt/stack/nova/nova/policy.py:201}} [ 1125.715092] env[63515]: DEBUG nova.network.neutron [None req-a2d941f8-9d25-4f98-968e-6daf11b6ab5a tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: 9a864a51-7df3-4afc-8be7-2794ef2dcf1a] Successfully created port: d96464dd-1638-4d19-8c71-9faeee62242f {{(pid=63515) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1125.762790] env[63515]: DEBUG oslo_concurrency.lockutils [None req-66223c7a-f7fb-4712-8ddb-378277f7abf5 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Releasing lock "refresh_cache-7b0df8d3-1b14-4e00-90c0-3dab3efc153c" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1125.764672] env[63515]: DEBUG nova.compute.manager [None req-66223c7a-f7fb-4712-8ddb-378277f7abf5 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] [instance: 7b0df8d3-1b14-4e00-90c0-3dab3efc153c] Instance network_info: |[{"id": "fd2511c0-002c-4a92-a30a-49c9e58641b3", "address": "fa:16:3e:90:c4:93", "network": {"id": "0e23b3a4-91ea-4bbe-9732-a3539d56b382", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1727906976-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fe873600ed6d4a3eb1ee3d3a19fceaf7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c7d2575f-b92f-44ec-a863-634cb76631a2", "external-id": "nsx-vlan-transportzone-794", "segmentation_id": 794, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfd2511c0-00", "ovs_interfaceid": "fd2511c0-002c-4a92-a30a-49c9e58641b3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1125.764672] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-66223c7a-f7fb-4712-8ddb-378277f7abf5 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] [instance: 7b0df8d3-1b14-4e00-90c0-3dab3efc153c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:90:c4:93', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c7d2575f-b92f-44ec-a863-634cb76631a2', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'fd2511c0-002c-4a92-a30a-49c9e58641b3', 'vif_model': 'vmxnet3'}] {{(pid=63515) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1125.773052] env[63515]: DEBUG oslo.service.loopingcall [None req-66223c7a-f7fb-4712-8ddb-378277f7abf5 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1125.773052] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7b0df8d3-1b14-4e00-90c0-3dab3efc153c] Creating VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1125.773052] env[63515]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-594e793b-ba01-4f27-8f7f-0acede847642 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.795823] env[63515]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1125.795823] env[63515]: value = "task-1111984" [ 1125.795823] env[63515]: _type = "Task" [ 1125.795823] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1125.804636] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1111984, 'name': CreateVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1125.893921] env[63515]: DEBUG nova.compute.manager [None req-a2d941f8-9d25-4f98-968e-6daf11b6ab5a tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: 9a864a51-7df3-4afc-8be7-2794ef2dcf1a] Start building block device mappings for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1126.301396] env[63515]: INFO nova.compute.manager [None req-121e8bf6-aa17-4f45-a156-987d42701ccb tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: aa83d035-9c47-434d-ab1a-140e6b3110b4] Updating instance to original state: 'active' [ 1126.311631] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1111984, 'name': CreateVM_Task, 'duration_secs': 0.345478} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1126.313028] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7b0df8d3-1b14-4e00-90c0-3dab3efc153c] Created VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1126.313028] env[63515]: DEBUG oslo_concurrency.lockutils [None req-66223c7a-f7fb-4712-8ddb-378277f7abf5 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1126.313028] env[63515]: DEBUG oslo_concurrency.lockutils [None req-66223c7a-f7fb-4712-8ddb-378277f7abf5 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1126.313028] env[63515]: DEBUG oslo_concurrency.lockutils [None req-66223c7a-f7fb-4712-8ddb-378277f7abf5 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1126.313290] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3f50683f-dfbe-4983-9212-a913264df0f9 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.317998] env[63515]: DEBUG oslo_vmware.api [None req-66223c7a-f7fb-4712-8ddb-378277f7abf5 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Waiting for the task: (returnval){ [ 1126.317998] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]525aa86a-1724-0bfe-5533-d97f31d2ece9" [ 1126.317998] env[63515]: _type = "Task" [ 1126.317998] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1126.326367] env[63515]: DEBUG oslo_vmware.api [None req-66223c7a-f7fb-4712-8ddb-378277f7abf5 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]525aa86a-1724-0bfe-5533-d97f31d2ece9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1126.399277] env[63515]: INFO nova.virt.block_device [None req-a2d941f8-9d25-4f98-968e-6daf11b6ab5a tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: 9a864a51-7df3-4afc-8be7-2794ef2dcf1a] Booting with volume 4ca406d5-39c8-41c2-b985-4734131e1c3c at /dev/sda [ 1126.430739] env[63515]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-0023e3c3-8da5-42e0-b5bc-b113dd652699 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.440463] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a549db34-53e9-4dbd-8673-e4e5dc3b5e08 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.467127] env[63515]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-2fc91db0-3922-48d6-911b-4607b74b00f5 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.474999] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a366cde3-030b-441b-a4c1-7a6571617b61 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.501401] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60f9d0b5-e8e5-4216-bef7-917244a59506 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.507652] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34d595fb-71ee-4a59-b6ef-f683458e4c3b {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.520008] env[63515]: DEBUG nova.virt.block_device [None req-a2d941f8-9d25-4f98-968e-6daf11b6ab5a tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: 9a864a51-7df3-4afc-8be7-2794ef2dcf1a] Updating existing volume attachment record: 19cf3156-db1a-46e8-9269-11c300d385f7 {{(pid=63515) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1126.828955] env[63515]: DEBUG oslo_vmware.api [None req-66223c7a-f7fb-4712-8ddb-378277f7abf5 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]525aa86a-1724-0bfe-5533-d97f31d2ece9, 'name': SearchDatastore_Task, 'duration_secs': 0.009739} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1126.828955] env[63515]: DEBUG oslo_concurrency.lockutils [None req-66223c7a-f7fb-4712-8ddb-378277f7abf5 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1126.828955] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-66223c7a-f7fb-4712-8ddb-378277f7abf5 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] [instance: 7b0df8d3-1b14-4e00-90c0-3dab3efc153c] Processing image 8a120570-cb06-4099-b262-554ca0ad15c5 {{(pid=63515) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1126.829223] env[63515]: DEBUG oslo_concurrency.lockutils [None req-66223c7a-f7fb-4712-8ddb-378277f7abf5 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1126.829352] env[63515]: DEBUG oslo_concurrency.lockutils [None req-66223c7a-f7fb-4712-8ddb-378277f7abf5 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1126.829589] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-66223c7a-f7fb-4712-8ddb-378277f7abf5 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1126.829856] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-15e1d55d-1bef-4677-9c11-156e5e78eadc {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.837827] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-66223c7a-f7fb-4712-8ddb-378277f7abf5 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1126.838019] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-66223c7a-f7fb-4712-8ddb-378277f7abf5 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63515) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1126.838790] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-56016528-ce25-4e5d-8079-0c2349477635 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.843563] env[63515]: DEBUG oslo_vmware.api [None req-66223c7a-f7fb-4712-8ddb-378277f7abf5 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Waiting for the task: (returnval){ [ 1126.843563] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]52902bf0-5583-b07b-6241-a927af7aa842" [ 1126.843563] env[63515]: _type = "Task" [ 1126.843563] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1126.852071] env[63515]: DEBUG oslo_vmware.api [None req-66223c7a-f7fb-4712-8ddb-378277f7abf5 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52902bf0-5583-b07b-6241-a927af7aa842, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1127.125918] env[63515]: DEBUG nova.compute.manager [req-e90fa89f-a58a-4d0b-aee3-d3e2c79bf027 req-a468f209-8144-434d-84cd-2ba6ed3d542a service nova] [instance: 9a864a51-7df3-4afc-8be7-2794ef2dcf1a] Received event network-vif-plugged-d96464dd-1638-4d19-8c71-9faeee62242f {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1127.126188] env[63515]: DEBUG oslo_concurrency.lockutils [req-e90fa89f-a58a-4d0b-aee3-d3e2c79bf027 req-a468f209-8144-434d-84cd-2ba6ed3d542a service nova] Acquiring lock "9a864a51-7df3-4afc-8be7-2794ef2dcf1a-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1127.126374] env[63515]: DEBUG oslo_concurrency.lockutils [req-e90fa89f-a58a-4d0b-aee3-d3e2c79bf027 req-a468f209-8144-434d-84cd-2ba6ed3d542a service nova] Lock "9a864a51-7df3-4afc-8be7-2794ef2dcf1a-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1127.126555] env[63515]: DEBUG oslo_concurrency.lockutils [req-e90fa89f-a58a-4d0b-aee3-d3e2c79bf027 req-a468f209-8144-434d-84cd-2ba6ed3d542a service nova] Lock "9a864a51-7df3-4afc-8be7-2794ef2dcf1a-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1127.126727] env[63515]: DEBUG nova.compute.manager [req-e90fa89f-a58a-4d0b-aee3-d3e2c79bf027 req-a468f209-8144-434d-84cd-2ba6ed3d542a service nova] [instance: 9a864a51-7df3-4afc-8be7-2794ef2dcf1a] No waiting events found dispatching network-vif-plugged-d96464dd-1638-4d19-8c71-9faeee62242f {{(pid=63515) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1127.126924] env[63515]: WARNING nova.compute.manager [req-e90fa89f-a58a-4d0b-aee3-d3e2c79bf027 req-a468f209-8144-434d-84cd-2ba6ed3d542a service nova] [instance: 9a864a51-7df3-4afc-8be7-2794ef2dcf1a] Received unexpected event network-vif-plugged-d96464dd-1638-4d19-8c71-9faeee62242f for instance with vm_state building and task_state block_device_mapping. [ 1127.229512] env[63515]: DEBUG nova.network.neutron [None req-a2d941f8-9d25-4f98-968e-6daf11b6ab5a tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: 9a864a51-7df3-4afc-8be7-2794ef2dcf1a] Successfully updated port: d96464dd-1638-4d19-8c71-9faeee62242f {{(pid=63515) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1127.353579] env[63515]: DEBUG oslo_vmware.api [None req-66223c7a-f7fb-4712-8ddb-378277f7abf5 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52902bf0-5583-b07b-6241-a927af7aa842, 'name': SearchDatastore_Task, 'duration_secs': 0.008967} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1127.354323] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-93a4ef54-4543-4934-89ea-e264ed249908 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.359485] env[63515]: DEBUG oslo_vmware.api [None req-66223c7a-f7fb-4712-8ddb-378277f7abf5 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Waiting for the task: (returnval){ [ 1127.359485] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]52952d48-f94a-de56-10c9-2ba912f4cc45" [ 1127.359485] env[63515]: _type = "Task" [ 1127.359485] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1127.367101] env[63515]: DEBUG oslo_vmware.api [None req-66223c7a-f7fb-4712-8ddb-378277f7abf5 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52952d48-f94a-de56-10c9-2ba912f4cc45, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1127.664318] env[63515]: DEBUG oslo_concurrency.lockutils [None req-0b105c17-1ad5-45ed-a572-02068cd6b529 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Acquiring lock "aa83d035-9c47-434d-ab1a-140e6b3110b4" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1127.664582] env[63515]: DEBUG oslo_concurrency.lockutils [None req-0b105c17-1ad5-45ed-a572-02068cd6b529 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Lock "aa83d035-9c47-434d-ab1a-140e6b3110b4" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1127.664806] env[63515]: DEBUG oslo_concurrency.lockutils [None req-0b105c17-1ad5-45ed-a572-02068cd6b529 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Acquiring lock "aa83d035-9c47-434d-ab1a-140e6b3110b4-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1127.664985] env[63515]: DEBUG oslo_concurrency.lockutils [None req-0b105c17-1ad5-45ed-a572-02068cd6b529 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Lock "aa83d035-9c47-434d-ab1a-140e6b3110b4-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1127.665181] env[63515]: DEBUG oslo_concurrency.lockutils [None req-0b105c17-1ad5-45ed-a572-02068cd6b529 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Lock "aa83d035-9c47-434d-ab1a-140e6b3110b4-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1127.667531] env[63515]: INFO nova.compute.manager [None req-0b105c17-1ad5-45ed-a572-02068cd6b529 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: aa83d035-9c47-434d-ab1a-140e6b3110b4] Terminating instance [ 1127.669494] env[63515]: DEBUG nova.compute.manager [None req-0b105c17-1ad5-45ed-a572-02068cd6b529 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: aa83d035-9c47-434d-ab1a-140e6b3110b4] Start destroying the instance on the hypervisor. {{(pid=63515) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1127.669718] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-0b105c17-1ad5-45ed-a572-02068cd6b529 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: aa83d035-9c47-434d-ab1a-140e6b3110b4] Powering off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1127.669958] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-11a0ef4a-c4c9-4b2b-9968-a575082e6647 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.677082] env[63515]: DEBUG oslo_vmware.api [None req-0b105c17-1ad5-45ed-a572-02068cd6b529 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Waiting for the task: (returnval){ [ 1127.677082] env[63515]: value = "task-1111985" [ 1127.677082] env[63515]: _type = "Task" [ 1127.677082] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1127.685166] env[63515]: DEBUG oslo_vmware.api [None req-0b105c17-1ad5-45ed-a572-02068cd6b529 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Task: {'id': task-1111985, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1127.732583] env[63515]: DEBUG oslo_concurrency.lockutils [None req-a2d941f8-9d25-4f98-968e-6daf11b6ab5a tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Acquiring lock "refresh_cache-9a864a51-7df3-4afc-8be7-2794ef2dcf1a" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1127.732723] env[63515]: DEBUG oslo_concurrency.lockutils [None req-a2d941f8-9d25-4f98-968e-6daf11b6ab5a tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Acquired lock "refresh_cache-9a864a51-7df3-4afc-8be7-2794ef2dcf1a" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1127.732903] env[63515]: DEBUG nova.network.neutron [None req-a2d941f8-9d25-4f98-968e-6daf11b6ab5a tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: 9a864a51-7df3-4afc-8be7-2794ef2dcf1a] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1127.870394] env[63515]: DEBUG oslo_vmware.api [None req-66223c7a-f7fb-4712-8ddb-378277f7abf5 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52952d48-f94a-de56-10c9-2ba912f4cc45, 'name': SearchDatastore_Task, 'duration_secs': 0.009565} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1127.870675] env[63515]: DEBUG oslo_concurrency.lockutils [None req-66223c7a-f7fb-4712-8ddb-378277f7abf5 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1127.870942] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-66223c7a-f7fb-4712-8ddb-378277f7abf5 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk to [datastore2] 7b0df8d3-1b14-4e00-90c0-3dab3efc153c/7b0df8d3-1b14-4e00-90c0-3dab3efc153c.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1127.871254] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f853aa66-2cda-4c62-bb88-4cca006a926a {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.878577] env[63515]: DEBUG oslo_vmware.api [None req-66223c7a-f7fb-4712-8ddb-378277f7abf5 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Waiting for the task: (returnval){ [ 1127.878577] env[63515]: value = "task-1111986" [ 1127.878577] env[63515]: _type = "Task" [ 1127.878577] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1127.886634] env[63515]: DEBUG oslo_vmware.api [None req-66223c7a-f7fb-4712-8ddb-378277f7abf5 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Task: {'id': task-1111986, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1128.190126] env[63515]: DEBUG oslo_vmware.api [None req-0b105c17-1ad5-45ed-a572-02068cd6b529 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Task: {'id': task-1111985, 'name': PowerOffVM_Task, 'duration_secs': 0.192469} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1128.190571] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-0b105c17-1ad5-45ed-a572-02068cd6b529 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: aa83d035-9c47-434d-ab1a-140e6b3110b4] Powered off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1128.190802] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-0b105c17-1ad5-45ed-a572-02068cd6b529 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: aa83d035-9c47-434d-ab1a-140e6b3110b4] Volume detach. Driver type: vmdk {{(pid=63515) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1128.191154] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-0b105c17-1ad5-45ed-a572-02068cd6b529 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: aa83d035-9c47-434d-ab1a-140e6b3110b4] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-243578', 'volume_id': '81e015c1-d604-49a7-9bad-778a82466e35', 'name': 'volume-81e015c1-d604-49a7-9bad-778a82466e35', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attaching', 'instance': 'aa83d035-9c47-434d-ab1a-140e6b3110b4', 'attached_at': '2024-10-03T03:03:51.000000', 'detached_at': '', 'volume_id': '81e015c1-d604-49a7-9bad-778a82466e35', 'serial': '81e015c1-d604-49a7-9bad-778a82466e35'} {{(pid=63515) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1128.192460] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae859761-f5d2-49c5-b846-5502332bfc31 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.220130] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00eb7791-4282-4a68-a1dc-3e67125e5b67 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.230088] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-194d2adb-d011-4624-bf16-1980745b94d9 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.258438] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ed287b6-449a-4049-8466-1f8c4260a817 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.277808] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-0b105c17-1ad5-45ed-a572-02068cd6b529 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] The volume has not been displaced from its original location: [datastore1] volume-81e015c1-d604-49a7-9bad-778a82466e35/volume-81e015c1-d604-49a7-9bad-778a82466e35.vmdk. No consolidation needed. {{(pid=63515) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1128.284222] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-0b105c17-1ad5-45ed-a572-02068cd6b529 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: aa83d035-9c47-434d-ab1a-140e6b3110b4] Reconfiguring VM instance instance-00000066 to detach disk 2001 {{(pid=63515) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1128.284733] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4b742fed-40f0-432e-8350-f3c954374e04 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.299567] env[63515]: DEBUG nova.network.neutron [None req-a2d941f8-9d25-4f98-968e-6daf11b6ab5a tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: 9a864a51-7df3-4afc-8be7-2794ef2dcf1a] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1128.308755] env[63515]: DEBUG oslo_vmware.api [None req-0b105c17-1ad5-45ed-a572-02068cd6b529 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Waiting for the task: (returnval){ [ 1128.308755] env[63515]: value = "task-1111987" [ 1128.308755] env[63515]: _type = "Task" [ 1128.308755] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1128.317454] env[63515]: DEBUG oslo_vmware.api [None req-0b105c17-1ad5-45ed-a572-02068cd6b529 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Task: {'id': task-1111987, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1128.387831] env[63515]: DEBUG oslo_vmware.api [None req-66223c7a-f7fb-4712-8ddb-378277f7abf5 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Task: {'id': task-1111986, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.464022} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1128.388150] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-66223c7a-f7fb-4712-8ddb-378277f7abf5 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk to [datastore2] 7b0df8d3-1b14-4e00-90c0-3dab3efc153c/7b0df8d3-1b14-4e00-90c0-3dab3efc153c.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1128.388400] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-66223c7a-f7fb-4712-8ddb-378277f7abf5 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] [instance: 7b0df8d3-1b14-4e00-90c0-3dab3efc153c] Extending root virtual disk to 1048576 {{(pid=63515) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1128.388695] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-fda99ede-c137-41e8-850a-ef1df1b31c56 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.394521] env[63515]: DEBUG oslo_vmware.api [None req-66223c7a-f7fb-4712-8ddb-378277f7abf5 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Waiting for the task: (returnval){ [ 1128.394521] env[63515]: value = "task-1111988" [ 1128.394521] env[63515]: _type = "Task" [ 1128.394521] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1128.401879] env[63515]: DEBUG oslo_vmware.api [None req-66223c7a-f7fb-4712-8ddb-378277f7abf5 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Task: {'id': task-1111988, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1128.511267] env[63515]: DEBUG nova.network.neutron [None req-a2d941f8-9d25-4f98-968e-6daf11b6ab5a tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: 9a864a51-7df3-4afc-8be7-2794ef2dcf1a] Updating instance_info_cache with network_info: [{"id": "d96464dd-1638-4d19-8c71-9faeee62242f", "address": "fa:16:3e:36:fc:a7", "network": {"id": "cec614cb-7a5b-4b16-a83d-061ff62d6677", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1501468300-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "aef947bfaf7d4180b46b4be662e3d903", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3a0ddd7d-c321-4187-bdd8-b19044ea2c4a", "external-id": "nsx-vlan-transportzone-747", "segmentation_id": 747, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd96464dd-16", "ovs_interfaceid": "d96464dd-1638-4d19-8c71-9faeee62242f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1128.637733] env[63515]: DEBUG nova.compute.manager [None req-a2d941f8-9d25-4f98-968e-6daf11b6ab5a tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: 9a864a51-7df3-4afc-8be7-2794ef2dcf1a] Start spawning the instance on the hypervisor. {{(pid=63515) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1128.638320] env[63515]: DEBUG nova.virt.hardware [None req-a2d941f8-9d25-4f98-968e-6daf11b6ab5a tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T02:52:14Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=0,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=1073741824,status='active',tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1128.638588] env[63515]: DEBUG nova.virt.hardware [None req-a2d941f8-9d25-4f98-968e-6daf11b6ab5a tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Flavor limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1128.638766] env[63515]: DEBUG nova.virt.hardware [None req-a2d941f8-9d25-4f98-968e-6daf11b6ab5a tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Image limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1128.638959] env[63515]: DEBUG nova.virt.hardware [None req-a2d941f8-9d25-4f98-968e-6daf11b6ab5a tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Flavor pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1128.639133] env[63515]: DEBUG nova.virt.hardware [None req-a2d941f8-9d25-4f98-968e-6daf11b6ab5a tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Image pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1128.639288] env[63515]: DEBUG nova.virt.hardware [None req-a2d941f8-9d25-4f98-968e-6daf11b6ab5a tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1128.639507] env[63515]: DEBUG nova.virt.hardware [None req-a2d941f8-9d25-4f98-968e-6daf11b6ab5a tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1128.639704] env[63515]: DEBUG nova.virt.hardware [None req-a2d941f8-9d25-4f98-968e-6daf11b6ab5a tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1128.639881] env[63515]: DEBUG nova.virt.hardware [None req-a2d941f8-9d25-4f98-968e-6daf11b6ab5a tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Got 1 possible topologies {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1128.640070] env[63515]: DEBUG nova.virt.hardware [None req-a2d941f8-9d25-4f98-968e-6daf11b6ab5a tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1128.640259] env[63515]: DEBUG nova.virt.hardware [None req-a2d941f8-9d25-4f98-968e-6daf11b6ab5a tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1128.641132] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c3026b5-0e85-4f96-9d53-aca02953faaa {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.649489] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d488ebba-5df0-4c8a-a5cc-48bf540bf71f {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.818830] env[63515]: DEBUG oslo_vmware.api [None req-0b105c17-1ad5-45ed-a572-02068cd6b529 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Task: {'id': task-1111987, 'name': ReconfigVM_Task, 'duration_secs': 0.201353} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1128.819166] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-0b105c17-1ad5-45ed-a572-02068cd6b529 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: aa83d035-9c47-434d-ab1a-140e6b3110b4] Reconfigured VM instance instance-00000066 to detach disk 2001 {{(pid=63515) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1128.823733] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1ee7005f-8267-473f-b0f6-f9ba2919bc19 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.838118] env[63515]: DEBUG oslo_vmware.api [None req-0b105c17-1ad5-45ed-a572-02068cd6b529 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Waiting for the task: (returnval){ [ 1128.838118] env[63515]: value = "task-1111989" [ 1128.838118] env[63515]: _type = "Task" [ 1128.838118] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1128.847073] env[63515]: DEBUG oslo_vmware.api [None req-0b105c17-1ad5-45ed-a572-02068cd6b529 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Task: {'id': task-1111989, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1128.903740] env[63515]: DEBUG oslo_vmware.api [None req-66223c7a-f7fb-4712-8ddb-378277f7abf5 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Task: {'id': task-1111988, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.060102} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1128.904040] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-66223c7a-f7fb-4712-8ddb-378277f7abf5 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] [instance: 7b0df8d3-1b14-4e00-90c0-3dab3efc153c] Extended root virtual disk {{(pid=63515) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1128.904811] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e375bb50-9513-4121-8463-da80205d3ae5 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.926102] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-66223c7a-f7fb-4712-8ddb-378277f7abf5 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] [instance: 7b0df8d3-1b14-4e00-90c0-3dab3efc153c] Reconfiguring VM instance instance-0000006d to attach disk [datastore2] 7b0df8d3-1b14-4e00-90c0-3dab3efc153c/7b0df8d3-1b14-4e00-90c0-3dab3efc153c.vmdk or device None with type sparse {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1128.926389] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3e760621-8796-472a-aa06-1609b2282175 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.945606] env[63515]: DEBUG oslo_vmware.api [None req-66223c7a-f7fb-4712-8ddb-378277f7abf5 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Waiting for the task: (returnval){ [ 1128.945606] env[63515]: value = "task-1111990" [ 1128.945606] env[63515]: _type = "Task" [ 1128.945606] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1128.952991] env[63515]: DEBUG oslo_vmware.api [None req-66223c7a-f7fb-4712-8ddb-378277f7abf5 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Task: {'id': task-1111990, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1129.014036] env[63515]: DEBUG oslo_concurrency.lockutils [None req-a2d941f8-9d25-4f98-968e-6daf11b6ab5a tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Releasing lock "refresh_cache-9a864a51-7df3-4afc-8be7-2794ef2dcf1a" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1129.014205] env[63515]: DEBUG nova.compute.manager [None req-a2d941f8-9d25-4f98-968e-6daf11b6ab5a tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: 9a864a51-7df3-4afc-8be7-2794ef2dcf1a] Instance network_info: |[{"id": "d96464dd-1638-4d19-8c71-9faeee62242f", "address": "fa:16:3e:36:fc:a7", "network": {"id": "cec614cb-7a5b-4b16-a83d-061ff62d6677", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1501468300-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "aef947bfaf7d4180b46b4be662e3d903", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3a0ddd7d-c321-4187-bdd8-b19044ea2c4a", "external-id": "nsx-vlan-transportzone-747", "segmentation_id": 747, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd96464dd-16", "ovs_interfaceid": "d96464dd-1638-4d19-8c71-9faeee62242f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1129.014650] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-a2d941f8-9d25-4f98-968e-6daf11b6ab5a tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: 9a864a51-7df3-4afc-8be7-2794ef2dcf1a] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:36:fc:a7', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '3a0ddd7d-c321-4187-bdd8-b19044ea2c4a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'd96464dd-1638-4d19-8c71-9faeee62242f', 'vif_model': 'vmxnet3'}] {{(pid=63515) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1129.022077] env[63515]: DEBUG oslo.service.loopingcall [None req-a2d941f8-9d25-4f98-968e-6daf11b6ab5a tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1129.022306] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9a864a51-7df3-4afc-8be7-2794ef2dcf1a] Creating VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1129.022532] env[63515]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-1164c5ca-b224-44ee-845a-fa578f9028bc {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.042358] env[63515]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1129.042358] env[63515]: value = "task-1111991" [ 1129.042358] env[63515]: _type = "Task" [ 1129.042358] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1129.050370] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1111991, 'name': CreateVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1129.152865] env[63515]: DEBUG nova.compute.manager [req-c0181870-dbbb-474a-b2cb-e61f60c073a4 req-0555bc0a-9eba-476d-ae28-3edeabc49b0c service nova] [instance: 9a864a51-7df3-4afc-8be7-2794ef2dcf1a] Received event network-changed-d96464dd-1638-4d19-8c71-9faeee62242f {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1129.152865] env[63515]: DEBUG nova.compute.manager [req-c0181870-dbbb-474a-b2cb-e61f60c073a4 req-0555bc0a-9eba-476d-ae28-3edeabc49b0c service nova] [instance: 9a864a51-7df3-4afc-8be7-2794ef2dcf1a] Refreshing instance network info cache due to event network-changed-d96464dd-1638-4d19-8c71-9faeee62242f. {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1129.153202] env[63515]: DEBUG oslo_concurrency.lockutils [req-c0181870-dbbb-474a-b2cb-e61f60c073a4 req-0555bc0a-9eba-476d-ae28-3edeabc49b0c service nova] Acquiring lock "refresh_cache-9a864a51-7df3-4afc-8be7-2794ef2dcf1a" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1129.153357] env[63515]: DEBUG oslo_concurrency.lockutils [req-c0181870-dbbb-474a-b2cb-e61f60c073a4 req-0555bc0a-9eba-476d-ae28-3edeabc49b0c service nova] Acquired lock "refresh_cache-9a864a51-7df3-4afc-8be7-2794ef2dcf1a" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1129.153528] env[63515]: DEBUG nova.network.neutron [req-c0181870-dbbb-474a-b2cb-e61f60c073a4 req-0555bc0a-9eba-476d-ae28-3edeabc49b0c service nova] [instance: 9a864a51-7df3-4afc-8be7-2794ef2dcf1a] Refreshing network info cache for port d96464dd-1638-4d19-8c71-9faeee62242f {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1129.347063] env[63515]: DEBUG oslo_vmware.api [None req-0b105c17-1ad5-45ed-a572-02068cd6b529 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Task: {'id': task-1111989, 'name': ReconfigVM_Task, 'duration_secs': 0.148392} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1129.347432] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-0b105c17-1ad5-45ed-a572-02068cd6b529 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: aa83d035-9c47-434d-ab1a-140e6b3110b4] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-243578', 'volume_id': '81e015c1-d604-49a7-9bad-778a82466e35', 'name': 'volume-81e015c1-d604-49a7-9bad-778a82466e35', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attaching', 'instance': 'aa83d035-9c47-434d-ab1a-140e6b3110b4', 'attached_at': '2024-10-03T03:03:51.000000', 'detached_at': '', 'volume_id': '81e015c1-d604-49a7-9bad-778a82466e35', 'serial': '81e015c1-d604-49a7-9bad-778a82466e35'} {{(pid=63515) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1129.347603] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-0b105c17-1ad5-45ed-a572-02068cd6b529 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: aa83d035-9c47-434d-ab1a-140e6b3110b4] Destroying instance {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1129.348371] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bea1ec8c-2886-47c1-b704-bf8a7cde9f40 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.354919] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-0b105c17-1ad5-45ed-a572-02068cd6b529 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: aa83d035-9c47-434d-ab1a-140e6b3110b4] Unregistering the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1129.355172] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-1adccd67-3100-4896-92b8-f5c08ce40664 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.420498] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-0b105c17-1ad5-45ed-a572-02068cd6b529 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: aa83d035-9c47-434d-ab1a-140e6b3110b4] Unregistered the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1129.420744] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-0b105c17-1ad5-45ed-a572-02068cd6b529 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: aa83d035-9c47-434d-ab1a-140e6b3110b4] Deleting contents of the VM from datastore datastore2 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1129.421116] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-0b105c17-1ad5-45ed-a572-02068cd6b529 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Deleting the datastore file [datastore2] aa83d035-9c47-434d-ab1a-140e6b3110b4 {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1129.421428] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c3e7a065-853a-4322-9c7d-a5250283d845 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.428573] env[63515]: DEBUG oslo_vmware.api [None req-0b105c17-1ad5-45ed-a572-02068cd6b529 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Waiting for the task: (returnval){ [ 1129.428573] env[63515]: value = "task-1111993" [ 1129.428573] env[63515]: _type = "Task" [ 1129.428573] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1129.436067] env[63515]: DEBUG oslo_vmware.api [None req-0b105c17-1ad5-45ed-a572-02068cd6b529 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Task: {'id': task-1111993, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1129.454610] env[63515]: DEBUG oslo_vmware.api [None req-66223c7a-f7fb-4712-8ddb-378277f7abf5 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Task: {'id': task-1111990, 'name': ReconfigVM_Task, 'duration_secs': 0.272165} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1129.454918] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-66223c7a-f7fb-4712-8ddb-378277f7abf5 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] [instance: 7b0df8d3-1b14-4e00-90c0-3dab3efc153c] Reconfigured VM instance instance-0000006d to attach disk [datastore2] 7b0df8d3-1b14-4e00-90c0-3dab3efc153c/7b0df8d3-1b14-4e00-90c0-3dab3efc153c.vmdk or device None with type sparse {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1129.455621] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-135df2d3-d0fe-4e35-814f-6d7243188a63 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.461645] env[63515]: DEBUG oslo_vmware.api [None req-66223c7a-f7fb-4712-8ddb-378277f7abf5 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Waiting for the task: (returnval){ [ 1129.461645] env[63515]: value = "task-1111994" [ 1129.461645] env[63515]: _type = "Task" [ 1129.461645] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1129.469079] env[63515]: DEBUG oslo_vmware.api [None req-66223c7a-f7fb-4712-8ddb-378277f7abf5 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Task: {'id': task-1111994, 'name': Rename_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1129.552164] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1111991, 'name': CreateVM_Task, 'duration_secs': 0.343846} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1129.552341] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9a864a51-7df3-4afc-8be7-2794ef2dcf1a] Created VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1129.553157] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-a2d941f8-9d25-4f98-968e-6daf11b6ab5a tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: 9a864a51-7df3-4afc-8be7-2794ef2dcf1a] Block device information present: {'root_device_name': '/dev/sda', 'image': [], 'ephemerals': [], 'block_device_mapping': [{'connection_info': {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-243587', 'volume_id': '4ca406d5-39c8-41c2-b985-4734131e1c3c', 'name': 'volume-4ca406d5-39c8-41c2-b985-4734131e1c3c', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '9a864a51-7df3-4afc-8be7-2794ef2dcf1a', 'attached_at': '', 'detached_at': '', 'volume_id': '4ca406d5-39c8-41c2-b985-4734131e1c3c', 'serial': '4ca406d5-39c8-41c2-b985-4734131e1c3c'}, 'disk_bus': None, 'attachment_id': '19cf3156-db1a-46e8-9269-11c300d385f7', 'guest_format': None, 'mount_device': '/dev/sda', 'delete_on_termination': True, 'device_type': None, 'boot_index': 0, 'volume_type': None}], 'swap': None} {{(pid=63515) spawn /opt/stack/nova/nova/virt/vmwareapi/vmops.py:799}} [ 1129.553316] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-a2d941f8-9d25-4f98-968e-6daf11b6ab5a tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: 9a864a51-7df3-4afc-8be7-2794ef2dcf1a] Root volume attach. Driver type: vmdk {{(pid=63515) attach_root_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:661}} [ 1129.554193] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b460290-6fb9-4dbe-a3f5-9d4b3a7c8238 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.561647] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0aa979ab-21b2-48b2-851c-e0a3dab35c79 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.567956] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85d9f1ec-32ca-4bce-9dce-5e26e8036b2a {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.575306] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.RelocateVM_Task with opID=oslo.vmware-c0df0621-cb05-4563-a523-af97b3cdb97d {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.582364] env[63515]: DEBUG oslo_vmware.api [None req-a2d941f8-9d25-4f98-968e-6daf11b6ab5a tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Waiting for the task: (returnval){ [ 1129.582364] env[63515]: value = "task-1111995" [ 1129.582364] env[63515]: _type = "Task" [ 1129.582364] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1129.590588] env[63515]: DEBUG oslo_vmware.api [None req-a2d941f8-9d25-4f98-968e-6daf11b6ab5a tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Task: {'id': task-1111995, 'name': RelocateVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1129.889327] env[63515]: DEBUG nova.network.neutron [req-c0181870-dbbb-474a-b2cb-e61f60c073a4 req-0555bc0a-9eba-476d-ae28-3edeabc49b0c service nova] [instance: 9a864a51-7df3-4afc-8be7-2794ef2dcf1a] Updated VIF entry in instance network info cache for port d96464dd-1638-4d19-8c71-9faeee62242f. {{(pid=63515) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1129.889844] env[63515]: DEBUG nova.network.neutron [req-c0181870-dbbb-474a-b2cb-e61f60c073a4 req-0555bc0a-9eba-476d-ae28-3edeabc49b0c service nova] [instance: 9a864a51-7df3-4afc-8be7-2794ef2dcf1a] Updating instance_info_cache with network_info: [{"id": "d96464dd-1638-4d19-8c71-9faeee62242f", "address": "fa:16:3e:36:fc:a7", "network": {"id": "cec614cb-7a5b-4b16-a83d-061ff62d6677", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1501468300-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "aef947bfaf7d4180b46b4be662e3d903", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3a0ddd7d-c321-4187-bdd8-b19044ea2c4a", "external-id": "nsx-vlan-transportzone-747", "segmentation_id": 747, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd96464dd-16", "ovs_interfaceid": "d96464dd-1638-4d19-8c71-9faeee62242f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1129.941651] env[63515]: DEBUG oslo_vmware.api [None req-0b105c17-1ad5-45ed-a572-02068cd6b529 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Task: {'id': task-1111993, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.326202} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1129.941956] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-0b105c17-1ad5-45ed-a572-02068cd6b529 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Deleted the datastore file {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1129.942213] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-0b105c17-1ad5-45ed-a572-02068cd6b529 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: aa83d035-9c47-434d-ab1a-140e6b3110b4] Deleted contents of the VM from datastore datastore2 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1129.942426] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-0b105c17-1ad5-45ed-a572-02068cd6b529 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: aa83d035-9c47-434d-ab1a-140e6b3110b4] Instance destroyed {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1129.942643] env[63515]: INFO nova.compute.manager [None req-0b105c17-1ad5-45ed-a572-02068cd6b529 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: aa83d035-9c47-434d-ab1a-140e6b3110b4] Took 2.27 seconds to destroy the instance on the hypervisor. [ 1129.942957] env[63515]: DEBUG oslo.service.loopingcall [None req-0b105c17-1ad5-45ed-a572-02068cd6b529 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1129.943219] env[63515]: DEBUG nova.compute.manager [-] [instance: aa83d035-9c47-434d-ab1a-140e6b3110b4] Deallocating network for instance {{(pid=63515) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1129.943319] env[63515]: DEBUG nova.network.neutron [-] [instance: aa83d035-9c47-434d-ab1a-140e6b3110b4] deallocate_for_instance() {{(pid=63515) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1129.972687] env[63515]: DEBUG oslo_vmware.api [None req-66223c7a-f7fb-4712-8ddb-378277f7abf5 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Task: {'id': task-1111994, 'name': Rename_Task, 'duration_secs': 0.166737} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1129.973026] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-66223c7a-f7fb-4712-8ddb-378277f7abf5 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] [instance: 7b0df8d3-1b14-4e00-90c0-3dab3efc153c] Powering on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1129.973290] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e218cad7-9978-458c-9e04-a6090ca620f1 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.982794] env[63515]: DEBUG oslo_vmware.api [None req-66223c7a-f7fb-4712-8ddb-378277f7abf5 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Waiting for the task: (returnval){ [ 1129.982794] env[63515]: value = "task-1111996" [ 1129.982794] env[63515]: _type = "Task" [ 1129.982794] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1129.992054] env[63515]: DEBUG oslo_vmware.api [None req-66223c7a-f7fb-4712-8ddb-378277f7abf5 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Task: {'id': task-1111996, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1130.093127] env[63515]: DEBUG oslo_vmware.api [None req-a2d941f8-9d25-4f98-968e-6daf11b6ab5a tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Task: {'id': task-1111995, 'name': RelocateVM_Task} progress is 40%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1130.393663] env[63515]: DEBUG oslo_concurrency.lockutils [req-c0181870-dbbb-474a-b2cb-e61f60c073a4 req-0555bc0a-9eba-476d-ae28-3edeabc49b0c service nova] Releasing lock "refresh_cache-9a864a51-7df3-4afc-8be7-2794ef2dcf1a" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1130.495026] env[63515]: DEBUG oslo_vmware.api [None req-66223c7a-f7fb-4712-8ddb-378277f7abf5 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Task: {'id': task-1111996, 'name': PowerOnVM_Task} progress is 89%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1130.594565] env[63515]: DEBUG oslo_vmware.api [None req-a2d941f8-9d25-4f98-968e-6daf11b6ab5a tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Task: {'id': task-1111995, 'name': RelocateVM_Task} progress is 54%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1130.976877] env[63515]: DEBUG nova.network.neutron [-] [instance: aa83d035-9c47-434d-ab1a-140e6b3110b4] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1130.994665] env[63515]: DEBUG oslo_vmware.api [None req-66223c7a-f7fb-4712-8ddb-378277f7abf5 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Task: {'id': task-1111996, 'name': PowerOnVM_Task, 'duration_secs': 0.624454} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1130.994998] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-66223c7a-f7fb-4712-8ddb-378277f7abf5 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] [instance: 7b0df8d3-1b14-4e00-90c0-3dab3efc153c] Powered on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1130.995216] env[63515]: INFO nova.compute.manager [None req-66223c7a-f7fb-4712-8ddb-378277f7abf5 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] [instance: 7b0df8d3-1b14-4e00-90c0-3dab3efc153c] Took 8.27 seconds to spawn the instance on the hypervisor. [ 1130.995403] env[63515]: DEBUG nova.compute.manager [None req-66223c7a-f7fb-4712-8ddb-378277f7abf5 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] [instance: 7b0df8d3-1b14-4e00-90c0-3dab3efc153c] Checking state {{(pid=63515) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1130.996230] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f34bc75f-4bdc-4270-8f39-aaab84f171f8 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.093906] env[63515]: DEBUG oslo_vmware.api [None req-a2d941f8-9d25-4f98-968e-6daf11b6ab5a tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Task: {'id': task-1111995, 'name': RelocateVM_Task} progress is 69%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1131.181879] env[63515]: DEBUG nova.compute.manager [req-1cbdc8b5-c144-4c19-9807-f4121d76ba11 req-b995e611-bb9c-4a83-9dc3-0c48f0b77169 service nova] [instance: aa83d035-9c47-434d-ab1a-140e6b3110b4] Received event network-vif-deleted-614aed8f-3ab8-4672-ac13-0ae8cb0b233d {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1131.479917] env[63515]: INFO nova.compute.manager [-] [instance: aa83d035-9c47-434d-ab1a-140e6b3110b4] Took 1.54 seconds to deallocate network for instance. [ 1131.514050] env[63515]: INFO nova.compute.manager [None req-66223c7a-f7fb-4712-8ddb-378277f7abf5 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] [instance: 7b0df8d3-1b14-4e00-90c0-3dab3efc153c] Took 14.01 seconds to build instance. [ 1131.593883] env[63515]: DEBUG oslo_vmware.api [None req-a2d941f8-9d25-4f98-968e-6daf11b6ab5a tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Task: {'id': task-1111995, 'name': RelocateVM_Task} progress is 84%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1132.003338] env[63515]: DEBUG oslo_concurrency.lockutils [None req-aa64b967-4018-4675-9821-814923c53bec tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Acquiring lock "7b0df8d3-1b14-4e00-90c0-3dab3efc153c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1132.016375] env[63515]: DEBUG oslo_concurrency.lockutils [None req-66223c7a-f7fb-4712-8ddb-378277f7abf5 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Lock "7b0df8d3-1b14-4e00-90c0-3dab3efc153c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.544s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1132.016596] env[63515]: DEBUG oslo_concurrency.lockutils [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Lock "7b0df8d3-1b14-4e00-90c0-3dab3efc153c" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 8.342s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1132.017476] env[63515]: INFO nova.compute.manager [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] [instance: 7b0df8d3-1b14-4e00-90c0-3dab3efc153c] During sync_power_state the instance has a pending task (spawning). Skip. [ 1132.017476] env[63515]: DEBUG oslo_concurrency.lockutils [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Lock "7b0df8d3-1b14-4e00-90c0-3dab3efc153c" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.001s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1132.017476] env[63515]: DEBUG oslo_concurrency.lockutils [None req-aa64b967-4018-4675-9821-814923c53bec tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Lock "7b0df8d3-1b14-4e00-90c0-3dab3efc153c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.014s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1132.017476] env[63515]: DEBUG oslo_concurrency.lockutils [None req-aa64b967-4018-4675-9821-814923c53bec tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Acquiring lock "7b0df8d3-1b14-4e00-90c0-3dab3efc153c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1132.017763] env[63515]: DEBUG oslo_concurrency.lockutils [None req-aa64b967-4018-4675-9821-814923c53bec tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Lock "7b0df8d3-1b14-4e00-90c0-3dab3efc153c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1132.017950] env[63515]: DEBUG oslo_concurrency.lockutils [None req-aa64b967-4018-4675-9821-814923c53bec tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Lock "7b0df8d3-1b14-4e00-90c0-3dab3efc153c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1132.021126] env[63515]: INFO nova.compute.manager [None req-aa64b967-4018-4675-9821-814923c53bec tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] [instance: 7b0df8d3-1b14-4e00-90c0-3dab3efc153c] Terminating instance [ 1132.024046] env[63515]: DEBUG nova.compute.manager [None req-aa64b967-4018-4675-9821-814923c53bec tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] [instance: 7b0df8d3-1b14-4e00-90c0-3dab3efc153c] Start destroying the instance on the hypervisor. {{(pid=63515) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1132.024046] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-aa64b967-4018-4675-9821-814923c53bec tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] [instance: 7b0df8d3-1b14-4e00-90c0-3dab3efc153c] Destroying instance {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1132.024705] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc883fbf-72ef-4de0-a32c-7a915d55e48f {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.029800] env[63515]: INFO nova.compute.manager [None req-0b105c17-1ad5-45ed-a572-02068cd6b529 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: aa83d035-9c47-434d-ab1a-140e6b3110b4] Took 0.55 seconds to detach 1 volumes for instance. [ 1132.038677] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-aa64b967-4018-4675-9821-814923c53bec tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] [instance: 7b0df8d3-1b14-4e00-90c0-3dab3efc153c] Powering off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1132.039191] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d11da57b-ef40-4ec5-b0da-770357e5ab22 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.045858] env[63515]: DEBUG oslo_vmware.api [None req-aa64b967-4018-4675-9821-814923c53bec tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Waiting for the task: (returnval){ [ 1132.045858] env[63515]: value = "task-1111997" [ 1132.045858] env[63515]: _type = "Task" [ 1132.045858] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1132.055368] env[63515]: DEBUG oslo_vmware.api [None req-aa64b967-4018-4675-9821-814923c53bec tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Task: {'id': task-1111997, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1132.095939] env[63515]: DEBUG oslo_vmware.api [None req-a2d941f8-9d25-4f98-968e-6daf11b6ab5a tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Task: {'id': task-1111995, 'name': RelocateVM_Task} progress is 92%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1132.540697] env[63515]: DEBUG oslo_concurrency.lockutils [None req-0b105c17-1ad5-45ed-a572-02068cd6b529 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1132.541219] env[63515]: DEBUG oslo_concurrency.lockutils [None req-0b105c17-1ad5-45ed-a572-02068cd6b529 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1132.541445] env[63515]: DEBUG oslo_concurrency.lockutils [None req-0b105c17-1ad5-45ed-a572-02068cd6b529 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1132.559641] env[63515]: DEBUG oslo_vmware.api [None req-aa64b967-4018-4675-9821-814923c53bec tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Task: {'id': task-1111997, 'name': PowerOffVM_Task, 'duration_secs': 0.402596} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1132.560056] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-aa64b967-4018-4675-9821-814923c53bec tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] [instance: 7b0df8d3-1b14-4e00-90c0-3dab3efc153c] Powered off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1132.560366] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-aa64b967-4018-4675-9821-814923c53bec tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] [instance: 7b0df8d3-1b14-4e00-90c0-3dab3efc153c] Unregistering the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1132.560748] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e45b1dae-8aee-44de-a88a-3cd6a2a9ac93 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.567705] env[63515]: INFO nova.scheduler.client.report [None req-0b105c17-1ad5-45ed-a572-02068cd6b529 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Deleted allocations for instance aa83d035-9c47-434d-ab1a-140e6b3110b4 [ 1132.595240] env[63515]: DEBUG oslo_vmware.api [None req-a2d941f8-9d25-4f98-968e-6daf11b6ab5a tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Task: {'id': task-1111995, 'name': RelocateVM_Task} progress is 97%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1132.633194] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-aa64b967-4018-4675-9821-814923c53bec tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] [instance: 7b0df8d3-1b14-4e00-90c0-3dab3efc153c] Unregistered the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1132.633194] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-aa64b967-4018-4675-9821-814923c53bec tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] [instance: 7b0df8d3-1b14-4e00-90c0-3dab3efc153c] Deleting contents of the VM from datastore datastore2 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1132.633194] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-aa64b967-4018-4675-9821-814923c53bec tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Deleting the datastore file [datastore2] 7b0df8d3-1b14-4e00-90c0-3dab3efc153c {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1132.633571] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-72b22b70-9bc3-4661-acd4-52204eb17137 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.642111] env[63515]: DEBUG oslo_vmware.api [None req-aa64b967-4018-4675-9821-814923c53bec tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Waiting for the task: (returnval){ [ 1132.642111] env[63515]: value = "task-1111999" [ 1132.642111] env[63515]: _type = "Task" [ 1132.642111] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1132.650139] env[63515]: DEBUG oslo_vmware.api [None req-aa64b967-4018-4675-9821-814923c53bec tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Task: {'id': task-1111999, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1133.077180] env[63515]: DEBUG oslo_concurrency.lockutils [None req-0b105c17-1ad5-45ed-a572-02068cd6b529 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Lock "aa83d035-9c47-434d-ab1a-140e6b3110b4" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.412s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1133.097052] env[63515]: DEBUG oslo_vmware.api [None req-a2d941f8-9d25-4f98-968e-6daf11b6ab5a tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Task: {'id': task-1111995, 'name': RelocateVM_Task} progress is 98%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1133.151939] env[63515]: DEBUG oslo_vmware.api [None req-aa64b967-4018-4675-9821-814923c53bec tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Task: {'id': task-1111999, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.19129} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1133.152264] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-aa64b967-4018-4675-9821-814923c53bec tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Deleted the datastore file {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1133.152474] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-aa64b967-4018-4675-9821-814923c53bec tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] [instance: 7b0df8d3-1b14-4e00-90c0-3dab3efc153c] Deleted contents of the VM from datastore datastore2 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1133.152702] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-aa64b967-4018-4675-9821-814923c53bec tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] [instance: 7b0df8d3-1b14-4e00-90c0-3dab3efc153c] Instance destroyed {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1133.152926] env[63515]: INFO nova.compute.manager [None req-aa64b967-4018-4675-9821-814923c53bec tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] [instance: 7b0df8d3-1b14-4e00-90c0-3dab3efc153c] Took 1.13 seconds to destroy the instance on the hypervisor. [ 1133.153242] env[63515]: DEBUG oslo.service.loopingcall [None req-aa64b967-4018-4675-9821-814923c53bec tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1133.153420] env[63515]: DEBUG nova.compute.manager [-] [instance: 7b0df8d3-1b14-4e00-90c0-3dab3efc153c] Deallocating network for instance {{(pid=63515) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1133.153537] env[63515]: DEBUG nova.network.neutron [-] [instance: 7b0df8d3-1b14-4e00-90c0-3dab3efc153c] deallocate_for_instance() {{(pid=63515) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1133.457218] env[63515]: DEBUG nova.compute.manager [req-c5eb11ad-9de7-4293-bba7-ee6ab7f8607c req-078a1bb5-603d-4444-9ef6-cd4c6d093b96 service nova] [instance: 7b0df8d3-1b14-4e00-90c0-3dab3efc153c] Received event network-vif-deleted-fd2511c0-002c-4a92-a30a-49c9e58641b3 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1133.457319] env[63515]: INFO nova.compute.manager [req-c5eb11ad-9de7-4293-bba7-ee6ab7f8607c req-078a1bb5-603d-4444-9ef6-cd4c6d093b96 service nova] [instance: 7b0df8d3-1b14-4e00-90c0-3dab3efc153c] Neutron deleted interface fd2511c0-002c-4a92-a30a-49c9e58641b3; detaching it from the instance and deleting it from the info cache [ 1133.457529] env[63515]: DEBUG nova.network.neutron [req-c5eb11ad-9de7-4293-bba7-ee6ab7f8607c req-078a1bb5-603d-4444-9ef6-cd4c6d093b96 service nova] [instance: 7b0df8d3-1b14-4e00-90c0-3dab3efc153c] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1133.597335] env[63515]: DEBUG oslo_vmware.api [None req-a2d941f8-9d25-4f98-968e-6daf11b6ab5a tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Task: {'id': task-1111995, 'name': RelocateVM_Task, 'duration_secs': 3.733888} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1133.597789] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-a2d941f8-9d25-4f98-968e-6daf11b6ab5a tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: 9a864a51-7df3-4afc-8be7-2794ef2dcf1a] Volume attach. Driver type: vmdk {{(pid=63515) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1133.597996] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-a2d941f8-9d25-4f98-968e-6daf11b6ab5a tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: 9a864a51-7df3-4afc-8be7-2794ef2dcf1a] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-243587', 'volume_id': '4ca406d5-39c8-41c2-b985-4734131e1c3c', 'name': 'volume-4ca406d5-39c8-41c2-b985-4734131e1c3c', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '9a864a51-7df3-4afc-8be7-2794ef2dcf1a', 'attached_at': '', 'detached_at': '', 'volume_id': '4ca406d5-39c8-41c2-b985-4734131e1c3c', 'serial': '4ca406d5-39c8-41c2-b985-4734131e1c3c'} {{(pid=63515) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1133.598775] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56ac98d9-3052-4f8a-9328-9ad1a60c6bc9 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.613603] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14ebc7f9-fcf8-4dc1-900b-d63683cc5f36 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.634165] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-a2d941f8-9d25-4f98-968e-6daf11b6ab5a tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: 9a864a51-7df3-4afc-8be7-2794ef2dcf1a] Reconfiguring VM instance instance-0000006e to attach disk [datastore2] volume-4ca406d5-39c8-41c2-b985-4734131e1c3c/volume-4ca406d5-39c8-41c2-b985-4734131e1c3c.vmdk or device None with type thin {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1133.634384] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c2a0c849-2bd9-4c0e-8061-30a6f415b9aa {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.652111] env[63515]: DEBUG oslo_vmware.api [None req-a2d941f8-9d25-4f98-968e-6daf11b6ab5a tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Waiting for the task: (returnval){ [ 1133.652111] env[63515]: value = "task-1112000" [ 1133.652111] env[63515]: _type = "Task" [ 1133.652111] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1133.659997] env[63515]: DEBUG oslo_vmware.api [None req-a2d941f8-9d25-4f98-968e-6daf11b6ab5a tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Task: {'id': task-1112000, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1133.861355] env[63515]: DEBUG oslo_concurrency.lockutils [None req-fdc6b8c1-6273-4192-b97d-29e8d79474af tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Acquiring lock "24a79a6c-5210-4778-a6ed-482bfe14933c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1133.861758] env[63515]: DEBUG oslo_concurrency.lockutils [None req-fdc6b8c1-6273-4192-b97d-29e8d79474af tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Lock "24a79a6c-5210-4778-a6ed-482bfe14933c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1133.929612] env[63515]: DEBUG nova.network.neutron [-] [instance: 7b0df8d3-1b14-4e00-90c0-3dab3efc153c] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1133.959776] env[63515]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-1ea693cf-4a37-480c-b505-7b0d433c5ca1 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.969323] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-366620d1-cd1d-482f-a223-a0ac6cff3be2 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.995127] env[63515]: DEBUG nova.compute.manager [req-c5eb11ad-9de7-4293-bba7-ee6ab7f8607c req-078a1bb5-603d-4444-9ef6-cd4c6d093b96 service nova] [instance: 7b0df8d3-1b14-4e00-90c0-3dab3efc153c] Detach interface failed, port_id=fd2511c0-002c-4a92-a30a-49c9e58641b3, reason: Instance 7b0df8d3-1b14-4e00-90c0-3dab3efc153c could not be found. {{(pid=63515) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1134.162278] env[63515]: DEBUG oslo_vmware.api [None req-a2d941f8-9d25-4f98-968e-6daf11b6ab5a tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Task: {'id': task-1112000, 'name': ReconfigVM_Task, 'duration_secs': 0.264872} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1134.162571] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-a2d941f8-9d25-4f98-968e-6daf11b6ab5a tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: 9a864a51-7df3-4afc-8be7-2794ef2dcf1a] Reconfigured VM instance instance-0000006e to attach disk [datastore2] volume-4ca406d5-39c8-41c2-b985-4734131e1c3c/volume-4ca406d5-39c8-41c2-b985-4734131e1c3c.vmdk or device None with type thin {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1134.167149] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4d2a8afb-5656-4579-879b-b014057ceb47 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.183641] env[63515]: DEBUG oslo_vmware.api [None req-a2d941f8-9d25-4f98-968e-6daf11b6ab5a tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Waiting for the task: (returnval){ [ 1134.183641] env[63515]: value = "task-1112001" [ 1134.183641] env[63515]: _type = "Task" [ 1134.183641] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1134.193443] env[63515]: DEBUG oslo_vmware.api [None req-a2d941f8-9d25-4f98-968e-6daf11b6ab5a tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Task: {'id': task-1112001, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1134.364690] env[63515]: DEBUG nova.compute.manager [None req-fdc6b8c1-6273-4192-b97d-29e8d79474af tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 24a79a6c-5210-4778-a6ed-482bfe14933c] Starting instance... {{(pid=63515) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1134.432491] env[63515]: INFO nova.compute.manager [-] [instance: 7b0df8d3-1b14-4e00-90c0-3dab3efc153c] Took 1.28 seconds to deallocate network for instance. [ 1134.693487] env[63515]: DEBUG oslo_vmware.api [None req-a2d941f8-9d25-4f98-968e-6daf11b6ab5a tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Task: {'id': task-1112001, 'name': ReconfigVM_Task, 'duration_secs': 0.174339} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1134.693858] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-a2d941f8-9d25-4f98-968e-6daf11b6ab5a tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: 9a864a51-7df3-4afc-8be7-2794ef2dcf1a] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-243587', 'volume_id': '4ca406d5-39c8-41c2-b985-4734131e1c3c', 'name': 'volume-4ca406d5-39c8-41c2-b985-4734131e1c3c', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '9a864a51-7df3-4afc-8be7-2794ef2dcf1a', 'attached_at': '', 'detached_at': '', 'volume_id': '4ca406d5-39c8-41c2-b985-4734131e1c3c', 'serial': '4ca406d5-39c8-41c2-b985-4734131e1c3c'} {{(pid=63515) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1134.694276] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-104e1227-3e7b-4cf3-8d21-80fe80de4132 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.700450] env[63515]: DEBUG oslo_vmware.api [None req-a2d941f8-9d25-4f98-968e-6daf11b6ab5a tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Waiting for the task: (returnval){ [ 1134.700450] env[63515]: value = "task-1112002" [ 1134.700450] env[63515]: _type = "Task" [ 1134.700450] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1134.708057] env[63515]: DEBUG oslo_vmware.api [None req-a2d941f8-9d25-4f98-968e-6daf11b6ab5a tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Task: {'id': task-1112002, 'name': Rename_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1134.889735] env[63515]: DEBUG oslo_concurrency.lockutils [None req-fdc6b8c1-6273-4192-b97d-29e8d79474af tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1134.890017] env[63515]: DEBUG oslo_concurrency.lockutils [None req-fdc6b8c1-6273-4192-b97d-29e8d79474af tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1134.891527] env[63515]: INFO nova.compute.claims [None req-fdc6b8c1-6273-4192-b97d-29e8d79474af tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 24a79a6c-5210-4778-a6ed-482bfe14933c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1134.938561] env[63515]: DEBUG oslo_concurrency.lockutils [None req-aa64b967-4018-4675-9821-814923c53bec tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1135.210099] env[63515]: DEBUG oslo_vmware.api [None req-a2d941f8-9d25-4f98-968e-6daf11b6ab5a tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Task: {'id': task-1112002, 'name': Rename_Task} progress is 14%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1135.710921] env[63515]: DEBUG oslo_vmware.api [None req-a2d941f8-9d25-4f98-968e-6daf11b6ab5a tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Task: {'id': task-1112002, 'name': Rename_Task, 'duration_secs': 0.766027} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1135.711295] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-a2d941f8-9d25-4f98-968e-6daf11b6ab5a tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: 9a864a51-7df3-4afc-8be7-2794ef2dcf1a] Powering on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1135.711565] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f17d23e5-e77a-4c8a-82cb-28ef06677c44 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.717586] env[63515]: DEBUG oslo_vmware.api [None req-a2d941f8-9d25-4f98-968e-6daf11b6ab5a tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Waiting for the task: (returnval){ [ 1135.717586] env[63515]: value = "task-1112003" [ 1135.717586] env[63515]: _type = "Task" [ 1135.717586] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1135.724762] env[63515]: DEBUG oslo_vmware.api [None req-a2d941f8-9d25-4f98-968e-6daf11b6ab5a tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Task: {'id': task-1112003, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1135.989723] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01733ab8-c6fd-4b31-a13e-4dc3712a5835 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.997450] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b0f166c-507d-4ec1-8cb3-f270536d2ae0 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1136.028618] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54018250-cb5d-4621-8dd9-beb10578f12b {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1136.036392] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74a02f5e-8803-4470-8b61-e7d80616626b {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1136.049535] env[63515]: DEBUG nova.compute.provider_tree [None req-fdc6b8c1-6273-4192-b97d-29e8d79474af tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1136.228030] env[63515]: DEBUG oslo_vmware.api [None req-a2d941f8-9d25-4f98-968e-6daf11b6ab5a tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Task: {'id': task-1112003, 'name': PowerOnVM_Task, 'duration_secs': 0.444541} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1136.228261] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-a2d941f8-9d25-4f98-968e-6daf11b6ab5a tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: 9a864a51-7df3-4afc-8be7-2794ef2dcf1a] Powered on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1136.228467] env[63515]: INFO nova.compute.manager [None req-a2d941f8-9d25-4f98-968e-6daf11b6ab5a tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: 9a864a51-7df3-4afc-8be7-2794ef2dcf1a] Took 7.59 seconds to spawn the instance on the hypervisor. [ 1136.228679] env[63515]: DEBUG nova.compute.manager [None req-a2d941f8-9d25-4f98-968e-6daf11b6ab5a tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: 9a864a51-7df3-4afc-8be7-2794ef2dcf1a] Checking state {{(pid=63515) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1136.229442] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-861f0a07-d200-4aa6-bbba-31a00f276a17 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1136.552374] env[63515]: DEBUG nova.scheduler.client.report [None req-fdc6b8c1-6273-4192-b97d-29e8d79474af tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1136.747912] env[63515]: INFO nova.compute.manager [None req-a2d941f8-9d25-4f98-968e-6daf11b6ab5a tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: 9a864a51-7df3-4afc-8be7-2794ef2dcf1a] Took 14.06 seconds to build instance. [ 1137.057364] env[63515]: DEBUG oslo_concurrency.lockutils [None req-fdc6b8c1-6273-4192-b97d-29e8d79474af tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.167s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1137.058016] env[63515]: DEBUG nova.compute.manager [None req-fdc6b8c1-6273-4192-b97d-29e8d79474af tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 24a79a6c-5210-4778-a6ed-482bfe14933c] Start building networks asynchronously for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1137.060732] env[63515]: DEBUG oslo_concurrency.lockutils [None req-aa64b967-4018-4675-9821-814923c53bec tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.122s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1137.064099] env[63515]: DEBUG nova.objects.instance [None req-aa64b967-4018-4675-9821-814923c53bec tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Lazy-loading 'resources' on Instance uuid 7b0df8d3-1b14-4e00-90c0-3dab3efc153c {{(pid=63515) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1137.249966] env[63515]: DEBUG oslo_concurrency.lockutils [None req-a2d941f8-9d25-4f98-968e-6daf11b6ab5a tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Lock "9a864a51-7df3-4afc-8be7-2794ef2dcf1a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.567s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1137.250278] env[63515]: DEBUG oslo_concurrency.lockutils [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Lock "9a864a51-7df3-4afc-8be7-2794ef2dcf1a" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 13.575s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1137.251197] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26e099b5-1317-4135-9b4f-5f06509bc7ff {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.511595] env[63515]: DEBUG nova.compute.manager [req-0a392f2f-a391-4e4c-a4eb-9d5ec28ceb76 req-23f44871-ec7d-4e72-b1d3-b57f74ee53d7 service nova] [instance: e13da90c-28e6-43d2-99b7-19c5095954ca] Received event network-changed-d9363b80-630e-4262-901f-c13e5c8b8c2a {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1137.511804] env[63515]: DEBUG nova.compute.manager [req-0a392f2f-a391-4e4c-a4eb-9d5ec28ceb76 req-23f44871-ec7d-4e72-b1d3-b57f74ee53d7 service nova] [instance: e13da90c-28e6-43d2-99b7-19c5095954ca] Refreshing instance network info cache due to event network-changed-d9363b80-630e-4262-901f-c13e5c8b8c2a. {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1137.512034] env[63515]: DEBUG oslo_concurrency.lockutils [req-0a392f2f-a391-4e4c-a4eb-9d5ec28ceb76 req-23f44871-ec7d-4e72-b1d3-b57f74ee53d7 service nova] Acquiring lock "refresh_cache-e13da90c-28e6-43d2-99b7-19c5095954ca" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1137.512182] env[63515]: DEBUG oslo_concurrency.lockutils [req-0a392f2f-a391-4e4c-a4eb-9d5ec28ceb76 req-23f44871-ec7d-4e72-b1d3-b57f74ee53d7 service nova] Acquired lock "refresh_cache-e13da90c-28e6-43d2-99b7-19c5095954ca" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1137.512342] env[63515]: DEBUG nova.network.neutron [req-0a392f2f-a391-4e4c-a4eb-9d5ec28ceb76 req-23f44871-ec7d-4e72-b1d3-b57f74ee53d7 service nova] [instance: e13da90c-28e6-43d2-99b7-19c5095954ca] Refreshing network info cache for port d9363b80-630e-4262-901f-c13e5c8b8c2a {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1137.563358] env[63515]: DEBUG nova.compute.utils [None req-fdc6b8c1-6273-4192-b97d-29e8d79474af tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Using /dev/sd instead of None {{(pid=63515) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1137.569020] env[63515]: DEBUG nova.compute.manager [None req-fdc6b8c1-6273-4192-b97d-29e8d79474af tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 24a79a6c-5210-4778-a6ed-482bfe14933c] Allocating IP information in the background. {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1137.569020] env[63515]: DEBUG nova.network.neutron [None req-fdc6b8c1-6273-4192-b97d-29e8d79474af tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 24a79a6c-5210-4778-a6ed-482bfe14933c] allocate_for_instance() {{(pid=63515) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1137.630315] env[63515]: DEBUG nova.policy [None req-fdc6b8c1-6273-4192-b97d-29e8d79474af tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '82fe830c309f41a4ab512a25d8481742', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b18d4daacbc84e758f9f9ae4f3f09d28', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63515) authorize /opt/stack/nova/nova/policy.py:201}} [ 1137.673597] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7fa1e49-a0fa-49bc-829a-db32b56c1799 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.683141] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ef3f047-cff3-4b6e-8883-05b108102104 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.717727] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d4f280d-d19f-48ac-aa14-2a396b0c9bd7 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.725264] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0256df9f-faf1-42d5-a399-bd9cca70004c {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.738188] env[63515]: DEBUG nova.compute.provider_tree [None req-aa64b967-4018-4675-9821-814923c53bec tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1137.760491] env[63515]: DEBUG oslo_concurrency.lockutils [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Lock "9a864a51-7df3-4afc-8be7-2794ef2dcf1a" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.510s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1137.969805] env[63515]: DEBUG nova.network.neutron [None req-fdc6b8c1-6273-4192-b97d-29e8d79474af tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 24a79a6c-5210-4778-a6ed-482bfe14933c] Successfully created port: f338b564-7636-4839-b034-73c3233d7da6 {{(pid=63515) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1138.059975] env[63515]: DEBUG oslo_service.periodic_task [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=63515) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1138.072618] env[63515]: DEBUG nova.compute.manager [None req-fdc6b8c1-6273-4192-b97d-29e8d79474af tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 24a79a6c-5210-4778-a6ed-482bfe14933c] Start building block device mappings for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1138.217271] env[63515]: DEBUG nova.network.neutron [req-0a392f2f-a391-4e4c-a4eb-9d5ec28ceb76 req-23f44871-ec7d-4e72-b1d3-b57f74ee53d7 service nova] [instance: e13da90c-28e6-43d2-99b7-19c5095954ca] Updated VIF entry in instance network info cache for port d9363b80-630e-4262-901f-c13e5c8b8c2a. {{(pid=63515) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1138.217692] env[63515]: DEBUG nova.network.neutron [req-0a392f2f-a391-4e4c-a4eb-9d5ec28ceb76 req-23f44871-ec7d-4e72-b1d3-b57f74ee53d7 service nova] [instance: e13da90c-28e6-43d2-99b7-19c5095954ca] Updating instance_info_cache with network_info: [{"id": "d9363b80-630e-4262-901f-c13e5c8b8c2a", "address": "fa:16:3e:dd:fe:76", "network": {"id": "cec614cb-7a5b-4b16-a83d-061ff62d6677", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1501468300-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "aef947bfaf7d4180b46b4be662e3d903", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3a0ddd7d-c321-4187-bdd8-b19044ea2c4a", "external-id": "nsx-vlan-transportzone-747", "segmentation_id": 747, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd9363b80-63", "ovs_interfaceid": "d9363b80-630e-4262-901f-c13e5c8b8c2a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1138.242913] env[63515]: DEBUG nova.scheduler.client.report [None req-aa64b967-4018-4675-9821-814923c53bec tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1138.534563] env[63515]: DEBUG oslo_service.periodic_task [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=63515) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1138.534994] env[63515]: DEBUG oslo_service.periodic_task [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=63515) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1138.535102] env[63515]: DEBUG nova.compute.manager [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=63515) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10545}} [ 1138.721036] env[63515]: DEBUG oslo_concurrency.lockutils [req-0a392f2f-a391-4e4c-a4eb-9d5ec28ceb76 req-23f44871-ec7d-4e72-b1d3-b57f74ee53d7 service nova] Releasing lock "refresh_cache-e13da90c-28e6-43d2-99b7-19c5095954ca" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1138.747459] env[63515]: DEBUG oslo_concurrency.lockutils [None req-aa64b967-4018-4675-9821-814923c53bec tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.686s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1138.771657] env[63515]: INFO nova.scheduler.client.report [None req-aa64b967-4018-4675-9821-814923c53bec tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Deleted allocations for instance 7b0df8d3-1b14-4e00-90c0-3dab3efc153c [ 1138.946095] env[63515]: DEBUG nova.compute.manager [None req-6435b2a2-c59b-4028-9161-1e2243f4b9d6 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: 9a864a51-7df3-4afc-8be7-2794ef2dcf1a] Stashing vm_state: active {{(pid=63515) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 1139.087512] env[63515]: DEBUG nova.compute.manager [None req-fdc6b8c1-6273-4192-b97d-29e8d79474af tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 24a79a6c-5210-4778-a6ed-482bfe14933c] Start spawning the instance on the hypervisor. {{(pid=63515) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1139.113018] env[63515]: DEBUG nova.virt.hardware [None req-fdc6b8c1-6273-4192-b97d-29e8d79474af tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T02:52:14Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T02:51:56Z,direct_url=,disk_format='vmdk',id=8a120570-cb06-4099-b262-554ca0ad15c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b569255cc43e42e7ae3f2b4ad37c6ef4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T02:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1139.113018] env[63515]: DEBUG nova.virt.hardware [None req-fdc6b8c1-6273-4192-b97d-29e8d79474af tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Flavor limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1139.113018] env[63515]: DEBUG nova.virt.hardware [None req-fdc6b8c1-6273-4192-b97d-29e8d79474af tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Image limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1139.113278] env[63515]: DEBUG nova.virt.hardware [None req-fdc6b8c1-6273-4192-b97d-29e8d79474af tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Flavor pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1139.113315] env[63515]: DEBUG nova.virt.hardware [None req-fdc6b8c1-6273-4192-b97d-29e8d79474af tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Image pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1139.113466] env[63515]: DEBUG nova.virt.hardware [None req-fdc6b8c1-6273-4192-b97d-29e8d79474af tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1139.113684] env[63515]: DEBUG nova.virt.hardware [None req-fdc6b8c1-6273-4192-b97d-29e8d79474af tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1139.113848] env[63515]: DEBUG nova.virt.hardware [None req-fdc6b8c1-6273-4192-b97d-29e8d79474af tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1139.114028] env[63515]: DEBUG nova.virt.hardware [None req-fdc6b8c1-6273-4192-b97d-29e8d79474af tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Got 1 possible topologies {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1139.114202] env[63515]: DEBUG nova.virt.hardware [None req-fdc6b8c1-6273-4192-b97d-29e8d79474af tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1139.114382] env[63515]: DEBUG nova.virt.hardware [None req-fdc6b8c1-6273-4192-b97d-29e8d79474af tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1139.115278] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33d3e306-2009-45b8-af83-5cae45c42c45 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.124179] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8eafeaca-3ef1-4ed1-89f3-6cbc9a430968 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.279444] env[63515]: DEBUG oslo_concurrency.lockutils [None req-aa64b967-4018-4675-9821-814923c53bec tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Lock "7b0df8d3-1b14-4e00-90c0-3dab3efc153c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.262s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1139.469958] env[63515]: DEBUG oslo_concurrency.lockutils [None req-6435b2a2-c59b-4028-9161-1e2243f4b9d6 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1139.470273] env[63515]: DEBUG oslo_concurrency.lockutils [None req-6435b2a2-c59b-4028-9161-1e2243f4b9d6 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 0.001s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1139.553147] env[63515]: DEBUG nova.network.neutron [None req-fdc6b8c1-6273-4192-b97d-29e8d79474af tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 24a79a6c-5210-4778-a6ed-482bfe14933c] Successfully updated port: f338b564-7636-4839-b034-73c3233d7da6 {{(pid=63515) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1139.566026] env[63515]: DEBUG nova.compute.manager [req-dcb4bab0-feda-4916-a9f0-3b4d10abba0f req-18d46ca5-3c9f-4144-8fb8-6f6233645170 service nova] [instance: 9a864a51-7df3-4afc-8be7-2794ef2dcf1a] Received event network-changed-d96464dd-1638-4d19-8c71-9faeee62242f {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1139.566338] env[63515]: DEBUG nova.compute.manager [req-dcb4bab0-feda-4916-a9f0-3b4d10abba0f req-18d46ca5-3c9f-4144-8fb8-6f6233645170 service nova] [instance: 9a864a51-7df3-4afc-8be7-2794ef2dcf1a] Refreshing instance network info cache due to event network-changed-d96464dd-1638-4d19-8c71-9faeee62242f. {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1139.566738] env[63515]: DEBUG oslo_concurrency.lockutils [req-dcb4bab0-feda-4916-a9f0-3b4d10abba0f req-18d46ca5-3c9f-4144-8fb8-6f6233645170 service nova] Acquiring lock "refresh_cache-9a864a51-7df3-4afc-8be7-2794ef2dcf1a" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1139.567023] env[63515]: DEBUG oslo_concurrency.lockutils [req-dcb4bab0-feda-4916-a9f0-3b4d10abba0f req-18d46ca5-3c9f-4144-8fb8-6f6233645170 service nova] Acquired lock "refresh_cache-9a864a51-7df3-4afc-8be7-2794ef2dcf1a" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1139.567023] env[63515]: DEBUG nova.network.neutron [req-dcb4bab0-feda-4916-a9f0-3b4d10abba0f req-18d46ca5-3c9f-4144-8fb8-6f6233645170 service nova] [instance: 9a864a51-7df3-4afc-8be7-2794ef2dcf1a] Refreshing network info cache for port d96464dd-1638-4d19-8c71-9faeee62242f {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1139.975994] env[63515]: INFO nova.compute.claims [None req-6435b2a2-c59b-4028-9161-1e2243f4b9d6 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: 9a864a51-7df3-4afc-8be7-2794ef2dcf1a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1140.055463] env[63515]: DEBUG oslo_concurrency.lockutils [None req-fdc6b8c1-6273-4192-b97d-29e8d79474af tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Acquiring lock "refresh_cache-24a79a6c-5210-4778-a6ed-482bfe14933c" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1140.055511] env[63515]: DEBUG oslo_concurrency.lockutils [None req-fdc6b8c1-6273-4192-b97d-29e8d79474af tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Acquired lock "refresh_cache-24a79a6c-5210-4778-a6ed-482bfe14933c" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1140.055679] env[63515]: DEBUG nova.network.neutron [None req-fdc6b8c1-6273-4192-b97d-29e8d79474af tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 24a79a6c-5210-4778-a6ed-482bfe14933c] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1140.294019] env[63515]: DEBUG nova.network.neutron [req-dcb4bab0-feda-4916-a9f0-3b4d10abba0f req-18d46ca5-3c9f-4144-8fb8-6f6233645170 service nova] [instance: 9a864a51-7df3-4afc-8be7-2794ef2dcf1a] Updated VIF entry in instance network info cache for port d96464dd-1638-4d19-8c71-9faeee62242f. {{(pid=63515) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1140.294396] env[63515]: DEBUG nova.network.neutron [req-dcb4bab0-feda-4916-a9f0-3b4d10abba0f req-18d46ca5-3c9f-4144-8fb8-6f6233645170 service nova] [instance: 9a864a51-7df3-4afc-8be7-2794ef2dcf1a] Updating instance_info_cache with network_info: [{"id": "d96464dd-1638-4d19-8c71-9faeee62242f", "address": "fa:16:3e:36:fc:a7", "network": {"id": "cec614cb-7a5b-4b16-a83d-061ff62d6677", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1501468300-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.155", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "aef947bfaf7d4180b46b4be662e3d903", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3a0ddd7d-c321-4187-bdd8-b19044ea2c4a", "external-id": "nsx-vlan-transportzone-747", "segmentation_id": 747, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd96464dd-16", "ovs_interfaceid": "d96464dd-1638-4d19-8c71-9faeee62242f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1140.483370] env[63515]: INFO nova.compute.resource_tracker [None req-6435b2a2-c59b-4028-9161-1e2243f4b9d6 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: 9a864a51-7df3-4afc-8be7-2794ef2dcf1a] Updating resource usage from migration 2ac80a52-5354-43ed-b08a-87a7e019e350 [ 1140.572794] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-457fac4f-e22a-43a4-9a9c-6b5577e24ecc {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1140.580949] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fff631ba-648b-41d2-aba7-2b0229275921 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1140.585820] env[63515]: DEBUG nova.network.neutron [None req-fdc6b8c1-6273-4192-b97d-29e8d79474af tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 24a79a6c-5210-4778-a6ed-482bfe14933c] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1140.613715] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3996c86-4892-49a2-ba3e-2a542059224e {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1140.621019] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e9f2899-2282-4f4d-94ef-d65661197aa4 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1140.635352] env[63515]: DEBUG nova.compute.provider_tree [None req-6435b2a2-c59b-4028-9161-1e2243f4b9d6 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1140.738706] env[63515]: DEBUG nova.network.neutron [None req-fdc6b8c1-6273-4192-b97d-29e8d79474af tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 24a79a6c-5210-4778-a6ed-482bfe14933c] Updating instance_info_cache with network_info: [{"id": "f338b564-7636-4839-b034-73c3233d7da6", "address": "fa:16:3e:bc:5a:12", "network": {"id": "5758c885-e1c9-45c8-86aa-27a802101f52", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-734458138-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b18d4daacbc84e758f9f9ae4f3f09d28", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69351262-8d39-441a-85ba-3a78df436d17", "external-id": "nsx-vlan-transportzone-205", "segmentation_id": 205, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf338b564-76", "ovs_interfaceid": "f338b564-7636-4839-b034-73c3233d7da6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1140.798052] env[63515]: DEBUG oslo_concurrency.lockutils [req-dcb4bab0-feda-4916-a9f0-3b4d10abba0f req-18d46ca5-3c9f-4144-8fb8-6f6233645170 service nova] Releasing lock "refresh_cache-9a864a51-7df3-4afc-8be7-2794ef2dcf1a" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1140.798052] env[63515]: DEBUG nova.compute.manager [req-dcb4bab0-feda-4916-a9f0-3b4d10abba0f req-18d46ca5-3c9f-4144-8fb8-6f6233645170 service nova] [instance: 24a79a6c-5210-4778-a6ed-482bfe14933c] Received event network-vif-plugged-f338b564-7636-4839-b034-73c3233d7da6 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1140.798052] env[63515]: DEBUG oslo_concurrency.lockutils [req-dcb4bab0-feda-4916-a9f0-3b4d10abba0f req-18d46ca5-3c9f-4144-8fb8-6f6233645170 service nova] Acquiring lock "24a79a6c-5210-4778-a6ed-482bfe14933c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1140.798052] env[63515]: DEBUG oslo_concurrency.lockutils [req-dcb4bab0-feda-4916-a9f0-3b4d10abba0f req-18d46ca5-3c9f-4144-8fb8-6f6233645170 service nova] Lock "24a79a6c-5210-4778-a6ed-482bfe14933c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1140.798052] env[63515]: DEBUG oslo_concurrency.lockutils [req-dcb4bab0-feda-4916-a9f0-3b4d10abba0f req-18d46ca5-3c9f-4144-8fb8-6f6233645170 service nova] Lock "24a79a6c-5210-4778-a6ed-482bfe14933c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1140.798309] env[63515]: DEBUG nova.compute.manager [req-dcb4bab0-feda-4916-a9f0-3b4d10abba0f req-18d46ca5-3c9f-4144-8fb8-6f6233645170 service nova] [instance: 24a79a6c-5210-4778-a6ed-482bfe14933c] No waiting events found dispatching network-vif-plugged-f338b564-7636-4839-b034-73c3233d7da6 {{(pid=63515) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1140.798309] env[63515]: WARNING nova.compute.manager [req-dcb4bab0-feda-4916-a9f0-3b4d10abba0f req-18d46ca5-3c9f-4144-8fb8-6f6233645170 service nova] [instance: 24a79a6c-5210-4778-a6ed-482bfe14933c] Received unexpected event network-vif-plugged-f338b564-7636-4839-b034-73c3233d7da6 for instance with vm_state building and task_state spawning. [ 1141.141022] env[63515]: DEBUG nova.scheduler.client.report [None req-6435b2a2-c59b-4028-9161-1e2243f4b9d6 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1141.241944] env[63515]: DEBUG oslo_concurrency.lockutils [None req-fdc6b8c1-6273-4192-b97d-29e8d79474af tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Releasing lock "refresh_cache-24a79a6c-5210-4778-a6ed-482bfe14933c" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1141.242124] env[63515]: DEBUG nova.compute.manager [None req-fdc6b8c1-6273-4192-b97d-29e8d79474af tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 24a79a6c-5210-4778-a6ed-482bfe14933c] Instance network_info: |[{"id": "f338b564-7636-4839-b034-73c3233d7da6", "address": "fa:16:3e:bc:5a:12", "network": {"id": "5758c885-e1c9-45c8-86aa-27a802101f52", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-734458138-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b18d4daacbc84e758f9f9ae4f3f09d28", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69351262-8d39-441a-85ba-3a78df436d17", "external-id": "nsx-vlan-transportzone-205", "segmentation_id": 205, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf338b564-76", "ovs_interfaceid": "f338b564-7636-4839-b034-73c3233d7da6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1141.243046] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-fdc6b8c1-6273-4192-b97d-29e8d79474af tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 24a79a6c-5210-4778-a6ed-482bfe14933c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:bc:5a:12', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '69351262-8d39-441a-85ba-3a78df436d17', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f338b564-7636-4839-b034-73c3233d7da6', 'vif_model': 'vmxnet3'}] {{(pid=63515) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1141.250157] env[63515]: DEBUG oslo.service.loopingcall [None req-fdc6b8c1-6273-4192-b97d-29e8d79474af tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1141.250336] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 24a79a6c-5210-4778-a6ed-482bfe14933c] Creating VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1141.251054] env[63515]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-7cff0eb6-2c7f-4ee1-980a-7442f6594276 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.270248] env[63515]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1141.270248] env[63515]: value = "task-1112004" [ 1141.270248] env[63515]: _type = "Task" [ 1141.270248] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1141.277812] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1112004, 'name': CreateVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1141.530796] env[63515]: DEBUG oslo_service.periodic_task [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=63515) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1141.534353] env[63515]: DEBUG oslo_service.periodic_task [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=63515) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1141.592730] env[63515]: DEBUG nova.compute.manager [req-7392f68f-1c8d-4efc-bfa4-e33fcc2fe93e req-96829cb0-8e78-4f08-8fbe-13ae25d686e4 service nova] [instance: 24a79a6c-5210-4778-a6ed-482bfe14933c] Received event network-changed-f338b564-7636-4839-b034-73c3233d7da6 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1141.592945] env[63515]: DEBUG nova.compute.manager [req-7392f68f-1c8d-4efc-bfa4-e33fcc2fe93e req-96829cb0-8e78-4f08-8fbe-13ae25d686e4 service nova] [instance: 24a79a6c-5210-4778-a6ed-482bfe14933c] Refreshing instance network info cache due to event network-changed-f338b564-7636-4839-b034-73c3233d7da6. {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1141.593204] env[63515]: DEBUG oslo_concurrency.lockutils [req-7392f68f-1c8d-4efc-bfa4-e33fcc2fe93e req-96829cb0-8e78-4f08-8fbe-13ae25d686e4 service nova] Acquiring lock "refresh_cache-24a79a6c-5210-4778-a6ed-482bfe14933c" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1141.593359] env[63515]: DEBUG oslo_concurrency.lockutils [req-7392f68f-1c8d-4efc-bfa4-e33fcc2fe93e req-96829cb0-8e78-4f08-8fbe-13ae25d686e4 service nova] Acquired lock "refresh_cache-24a79a6c-5210-4778-a6ed-482bfe14933c" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1141.593528] env[63515]: DEBUG nova.network.neutron [req-7392f68f-1c8d-4efc-bfa4-e33fcc2fe93e req-96829cb0-8e78-4f08-8fbe-13ae25d686e4 service nova] [instance: 24a79a6c-5210-4778-a6ed-482bfe14933c] Refreshing network info cache for port f338b564-7636-4839-b034-73c3233d7da6 {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1141.643647] env[63515]: DEBUG oslo_concurrency.lockutils [None req-6435b2a2-c59b-4028-9161-1e2243f4b9d6 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.173s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1141.643831] env[63515]: INFO nova.compute.manager [None req-6435b2a2-c59b-4028-9161-1e2243f4b9d6 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: 9a864a51-7df3-4afc-8be7-2794ef2dcf1a] Migrating [ 1141.781381] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1112004, 'name': CreateVM_Task, 'duration_secs': 0.309845} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1141.781550] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 24a79a6c-5210-4778-a6ed-482bfe14933c] Created VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1141.782269] env[63515]: DEBUG oslo_concurrency.lockutils [None req-fdc6b8c1-6273-4192-b97d-29e8d79474af tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1141.782490] env[63515]: DEBUG oslo_concurrency.lockutils [None req-fdc6b8c1-6273-4192-b97d-29e8d79474af tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1141.782889] env[63515]: DEBUG oslo_concurrency.lockutils [None req-fdc6b8c1-6273-4192-b97d-29e8d79474af tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1141.783176] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-76985e09-1569-404d-b824-456d05f6a7da {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.788732] env[63515]: DEBUG oslo_vmware.api [None req-fdc6b8c1-6273-4192-b97d-29e8d79474af tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Waiting for the task: (returnval){ [ 1141.788732] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]525e580b-f880-f0fe-32ff-fb05ff82c742" [ 1141.788732] env[63515]: _type = "Task" [ 1141.788732] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1141.796035] env[63515]: DEBUG oslo_vmware.api [None req-fdc6b8c1-6273-4192-b97d-29e8d79474af tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]525e580b-f880-f0fe-32ff-fb05ff82c742, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1142.157201] env[63515]: DEBUG oslo_concurrency.lockutils [None req-6435b2a2-c59b-4028-9161-1e2243f4b9d6 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Acquiring lock "refresh_cache-9a864a51-7df3-4afc-8be7-2794ef2dcf1a" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1142.157488] env[63515]: DEBUG oslo_concurrency.lockutils [None req-6435b2a2-c59b-4028-9161-1e2243f4b9d6 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Acquired lock "refresh_cache-9a864a51-7df3-4afc-8be7-2794ef2dcf1a" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1142.157488] env[63515]: DEBUG nova.network.neutron [None req-6435b2a2-c59b-4028-9161-1e2243f4b9d6 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: 9a864a51-7df3-4afc-8be7-2794ef2dcf1a] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1142.299104] env[63515]: DEBUG oslo_vmware.api [None req-fdc6b8c1-6273-4192-b97d-29e8d79474af tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]525e580b-f880-f0fe-32ff-fb05ff82c742, 'name': SearchDatastore_Task, 'duration_secs': 0.009693} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1142.299410] env[63515]: DEBUG oslo_concurrency.lockutils [None req-fdc6b8c1-6273-4192-b97d-29e8d79474af tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1142.299645] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-fdc6b8c1-6273-4192-b97d-29e8d79474af tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 24a79a6c-5210-4778-a6ed-482bfe14933c] Processing image 8a120570-cb06-4099-b262-554ca0ad15c5 {{(pid=63515) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1142.299942] env[63515]: DEBUG oslo_concurrency.lockutils [None req-fdc6b8c1-6273-4192-b97d-29e8d79474af tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1142.300335] env[63515]: DEBUG oslo_concurrency.lockutils [None req-fdc6b8c1-6273-4192-b97d-29e8d79474af tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1142.300335] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-fdc6b8c1-6273-4192-b97d-29e8d79474af tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1142.300551] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-9fd64216-ca1d-4abd-970b-af6158722537 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1142.309146] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-fdc6b8c1-6273-4192-b97d-29e8d79474af tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1142.309326] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-fdc6b8c1-6273-4192-b97d-29e8d79474af tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63515) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1142.310048] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-410f7f70-8d4b-4cbc-a35c-a4e63e422d8c {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1142.314864] env[63515]: DEBUG oslo_vmware.api [None req-fdc6b8c1-6273-4192-b97d-29e8d79474af tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Waiting for the task: (returnval){ [ 1142.314864] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]52922893-0733-a736-e493-34703b0b71ea" [ 1142.314864] env[63515]: _type = "Task" [ 1142.314864] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1142.323112] env[63515]: DEBUG oslo_vmware.api [None req-fdc6b8c1-6273-4192-b97d-29e8d79474af tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52922893-0733-a736-e493-34703b0b71ea, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1142.337137] env[63515]: DEBUG nova.network.neutron [req-7392f68f-1c8d-4efc-bfa4-e33fcc2fe93e req-96829cb0-8e78-4f08-8fbe-13ae25d686e4 service nova] [instance: 24a79a6c-5210-4778-a6ed-482bfe14933c] Updated VIF entry in instance network info cache for port f338b564-7636-4839-b034-73c3233d7da6. {{(pid=63515) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1142.337394] env[63515]: DEBUG nova.network.neutron [req-7392f68f-1c8d-4efc-bfa4-e33fcc2fe93e req-96829cb0-8e78-4f08-8fbe-13ae25d686e4 service nova] [instance: 24a79a6c-5210-4778-a6ed-482bfe14933c] Updating instance_info_cache with network_info: [{"id": "f338b564-7636-4839-b034-73c3233d7da6", "address": "fa:16:3e:bc:5a:12", "network": {"id": "5758c885-e1c9-45c8-86aa-27a802101f52", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-734458138-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b18d4daacbc84e758f9f9ae4f3f09d28", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69351262-8d39-441a-85ba-3a78df436d17", "external-id": "nsx-vlan-transportzone-205", "segmentation_id": 205, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf338b564-76", "ovs_interfaceid": "f338b564-7636-4839-b034-73c3233d7da6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1142.535460] env[63515]: DEBUG oslo_service.periodic_task [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=63515) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1142.825113] env[63515]: DEBUG oslo_vmware.api [None req-fdc6b8c1-6273-4192-b97d-29e8d79474af tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52922893-0733-a736-e493-34703b0b71ea, 'name': SearchDatastore_Task, 'duration_secs': 0.009525} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1142.826008] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4fea4d9f-cba4-4d65-8155-ce32d754d6d4 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1142.831216] env[63515]: DEBUG oslo_vmware.api [None req-fdc6b8c1-6273-4192-b97d-29e8d79474af tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Waiting for the task: (returnval){ [ 1142.831216] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]52bf9b08-757d-8b61-5cde-aae0011cf4c4" [ 1142.831216] env[63515]: _type = "Task" [ 1142.831216] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1142.840354] env[63515]: DEBUG oslo_concurrency.lockutils [req-7392f68f-1c8d-4efc-bfa4-e33fcc2fe93e req-96829cb0-8e78-4f08-8fbe-13ae25d686e4 service nova] Releasing lock "refresh_cache-24a79a6c-5210-4778-a6ed-482bfe14933c" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1142.840722] env[63515]: DEBUG oslo_vmware.api [None req-fdc6b8c1-6273-4192-b97d-29e8d79474af tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52bf9b08-757d-8b61-5cde-aae0011cf4c4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1142.905468] env[63515]: DEBUG nova.network.neutron [None req-6435b2a2-c59b-4028-9161-1e2243f4b9d6 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: 9a864a51-7df3-4afc-8be7-2794ef2dcf1a] Updating instance_info_cache with network_info: [{"id": "d96464dd-1638-4d19-8c71-9faeee62242f", "address": "fa:16:3e:36:fc:a7", "network": {"id": "cec614cb-7a5b-4b16-a83d-061ff62d6677", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1501468300-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.155", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "aef947bfaf7d4180b46b4be662e3d903", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3a0ddd7d-c321-4187-bdd8-b19044ea2c4a", "external-id": "nsx-vlan-transportzone-747", "segmentation_id": 747, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd96464dd-16", "ovs_interfaceid": "d96464dd-1638-4d19-8c71-9faeee62242f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1143.343448] env[63515]: DEBUG oslo_vmware.api [None req-fdc6b8c1-6273-4192-b97d-29e8d79474af tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52bf9b08-757d-8b61-5cde-aae0011cf4c4, 'name': SearchDatastore_Task, 'duration_secs': 0.009425} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1143.343721] env[63515]: DEBUG oslo_concurrency.lockutils [None req-fdc6b8c1-6273-4192-b97d-29e8d79474af tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1143.344308] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-fdc6b8c1-6273-4192-b97d-29e8d79474af tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk to [datastore2] 24a79a6c-5210-4778-a6ed-482bfe14933c/24a79a6c-5210-4778-a6ed-482bfe14933c.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1143.344308] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-cdf40efc-42e7-4f41-82fa-674ff7922c9a {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.351789] env[63515]: DEBUG oslo_vmware.api [None req-fdc6b8c1-6273-4192-b97d-29e8d79474af tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Waiting for the task: (returnval){ [ 1143.351789] env[63515]: value = "task-1112005" [ 1143.351789] env[63515]: _type = "Task" [ 1143.351789] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1143.359878] env[63515]: DEBUG oslo_vmware.api [None req-fdc6b8c1-6273-4192-b97d-29e8d79474af tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Task: {'id': task-1112005, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1143.407815] env[63515]: DEBUG oslo_concurrency.lockutils [None req-6435b2a2-c59b-4028-9161-1e2243f4b9d6 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Releasing lock "refresh_cache-9a864a51-7df3-4afc-8be7-2794ef2dcf1a" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1143.416640] env[63515]: DEBUG oslo_concurrency.lockutils [None req-bd647ce8-c9a2-4ac0-8bba-7604eb48d8bd tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Acquiring lock "c8dbf542-70cb-45be-9308-5e12307c0be6" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1143.416922] env[63515]: DEBUG oslo_concurrency.lockutils [None req-bd647ce8-c9a2-4ac0-8bba-7604eb48d8bd tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Lock "c8dbf542-70cb-45be-9308-5e12307c0be6" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1143.537319] env[63515]: DEBUG oslo_service.periodic_task [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=63515) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1143.537511] env[63515]: DEBUG nova.compute.manager [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Starting heal instance info cache {{(pid=63515) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9926}} [ 1143.537634] env[63515]: DEBUG nova.compute.manager [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Rebuilding the list of instances to heal {{(pid=63515) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9930}} [ 1143.861941] env[63515]: DEBUG oslo_vmware.api [None req-fdc6b8c1-6273-4192-b97d-29e8d79474af tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Task: {'id': task-1112005, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.461489} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1143.862237] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-fdc6b8c1-6273-4192-b97d-29e8d79474af tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk to [datastore2] 24a79a6c-5210-4778-a6ed-482bfe14933c/24a79a6c-5210-4778-a6ed-482bfe14933c.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1143.862456] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-fdc6b8c1-6273-4192-b97d-29e8d79474af tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 24a79a6c-5210-4778-a6ed-482bfe14933c] Extending root virtual disk to 1048576 {{(pid=63515) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1143.862719] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-2f31180a-7ada-461b-bcf2-c44570524cff {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.868812] env[63515]: DEBUG oslo_vmware.api [None req-fdc6b8c1-6273-4192-b97d-29e8d79474af tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Waiting for the task: (returnval){ [ 1143.868812] env[63515]: value = "task-1112006" [ 1143.868812] env[63515]: _type = "Task" [ 1143.868812] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1143.875945] env[63515]: DEBUG oslo_vmware.api [None req-fdc6b8c1-6273-4192-b97d-29e8d79474af tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Task: {'id': task-1112006, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1143.919981] env[63515]: DEBUG nova.compute.utils [None req-bd647ce8-c9a2-4ac0-8bba-7604eb48d8bd tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Using /dev/sd instead of None {{(pid=63515) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1144.041613] env[63515]: DEBUG nova.compute.manager [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] [instance: 24a79a6c-5210-4778-a6ed-482bfe14933c] Skipping network cache update for instance because it is Building. {{(pid=63515) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 1144.072374] env[63515]: DEBUG oslo_concurrency.lockutils [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Acquiring lock "refresh_cache-e13da90c-28e6-43d2-99b7-19c5095954ca" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1144.072670] env[63515]: DEBUG oslo_concurrency.lockutils [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Acquired lock "refresh_cache-e13da90c-28e6-43d2-99b7-19c5095954ca" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1144.072917] env[63515]: DEBUG nova.network.neutron [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] [instance: e13da90c-28e6-43d2-99b7-19c5095954ca] Forcefully refreshing network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2004}} [ 1144.073193] env[63515]: DEBUG nova.objects.instance [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Lazy-loading 'info_cache' on Instance uuid e13da90c-28e6-43d2-99b7-19c5095954ca {{(pid=63515) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1144.379206] env[63515]: DEBUG oslo_vmware.api [None req-fdc6b8c1-6273-4192-b97d-29e8d79474af tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Task: {'id': task-1112006, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.214891} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1144.379499] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-fdc6b8c1-6273-4192-b97d-29e8d79474af tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 24a79a6c-5210-4778-a6ed-482bfe14933c] Extended root virtual disk {{(pid=63515) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1144.380349] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5f03894-a42c-4b4f-b8b6-3f4efaede9e6 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1144.401759] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-fdc6b8c1-6273-4192-b97d-29e8d79474af tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 24a79a6c-5210-4778-a6ed-482bfe14933c] Reconfiguring VM instance instance-0000006f to attach disk [datastore2] 24a79a6c-5210-4778-a6ed-482bfe14933c/24a79a6c-5210-4778-a6ed-482bfe14933c.vmdk or device None with type sparse {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1144.402036] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d82c1afb-62a2-49a2-88a2-843612cf7d2b {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1144.423200] env[63515]: DEBUG oslo_concurrency.lockutils [None req-bd647ce8-c9a2-4ac0-8bba-7604eb48d8bd tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Lock "c8dbf542-70cb-45be-9308-5e12307c0be6" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.006s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1144.423761] env[63515]: DEBUG oslo_vmware.api [None req-fdc6b8c1-6273-4192-b97d-29e8d79474af tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Waiting for the task: (returnval){ [ 1144.423761] env[63515]: value = "task-1112007" [ 1144.423761] env[63515]: _type = "Task" [ 1144.423761] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1144.433243] env[63515]: DEBUG oslo_vmware.api [None req-fdc6b8c1-6273-4192-b97d-29e8d79474af tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Task: {'id': task-1112007, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1144.924324] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a2e9bd4-9a4a-4b8f-88e8-2b1a25119535 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1144.935924] env[63515]: DEBUG oslo_vmware.api [None req-fdc6b8c1-6273-4192-b97d-29e8d79474af tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Task: {'id': task-1112007, 'name': ReconfigVM_Task, 'duration_secs': 0.304683} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1144.950366] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-fdc6b8c1-6273-4192-b97d-29e8d79474af tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 24a79a6c-5210-4778-a6ed-482bfe14933c] Reconfigured VM instance instance-0000006f to attach disk [datastore2] 24a79a6c-5210-4778-a6ed-482bfe14933c/24a79a6c-5210-4778-a6ed-482bfe14933c.vmdk or device None with type sparse {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1144.951230] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-6435b2a2-c59b-4028-9161-1e2243f4b9d6 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: 9a864a51-7df3-4afc-8be7-2794ef2dcf1a] Updating instance '9a864a51-7df3-4afc-8be7-2794ef2dcf1a' progress to 0 {{(pid=63515) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1144.954417] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-12cc494e-3d88-4651-af90-d676d108eaf4 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1144.960674] env[63515]: DEBUG oslo_vmware.api [None req-fdc6b8c1-6273-4192-b97d-29e8d79474af tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Waiting for the task: (returnval){ [ 1144.960674] env[63515]: value = "task-1112008" [ 1144.960674] env[63515]: _type = "Task" [ 1144.960674] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1144.970957] env[63515]: DEBUG oslo_vmware.api [None req-fdc6b8c1-6273-4192-b97d-29e8d79474af tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Task: {'id': task-1112008, 'name': Rename_Task} progress is 5%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1145.458659] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-6435b2a2-c59b-4028-9161-1e2243f4b9d6 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: 9a864a51-7df3-4afc-8be7-2794ef2dcf1a] Powering off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1145.458659] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7bc9a2da-a89c-4e40-89f5-a26bfb43ee3c {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.467053] env[63515]: DEBUG oslo_vmware.api [None req-6435b2a2-c59b-4028-9161-1e2243f4b9d6 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Waiting for the task: (returnval){ [ 1145.467053] env[63515]: value = "task-1112009" [ 1145.467053] env[63515]: _type = "Task" [ 1145.467053] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1145.472624] env[63515]: DEBUG oslo_vmware.api [None req-fdc6b8c1-6273-4192-b97d-29e8d79474af tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Task: {'id': task-1112008, 'name': Rename_Task, 'duration_secs': 0.1324} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1145.473249] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-fdc6b8c1-6273-4192-b97d-29e8d79474af tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 24a79a6c-5210-4778-a6ed-482bfe14933c] Powering on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1145.473495] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-325d81d4-4262-4690-93ec-dd5f76c77b81 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.477691] env[63515]: DEBUG oslo_vmware.api [None req-6435b2a2-c59b-4028-9161-1e2243f4b9d6 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Task: {'id': task-1112009, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1145.482278] env[63515]: DEBUG oslo_vmware.api [None req-fdc6b8c1-6273-4192-b97d-29e8d79474af tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Waiting for the task: (returnval){ [ 1145.482278] env[63515]: value = "task-1112010" [ 1145.482278] env[63515]: _type = "Task" [ 1145.482278] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1145.490024] env[63515]: DEBUG oslo_concurrency.lockutils [None req-bd647ce8-c9a2-4ac0-8bba-7604eb48d8bd tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Acquiring lock "c8dbf542-70cb-45be-9308-5e12307c0be6" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1145.490331] env[63515]: DEBUG oslo_concurrency.lockutils [None req-bd647ce8-c9a2-4ac0-8bba-7604eb48d8bd tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Lock "c8dbf542-70cb-45be-9308-5e12307c0be6" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1145.490567] env[63515]: INFO nova.compute.manager [None req-bd647ce8-c9a2-4ac0-8bba-7604eb48d8bd tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] [instance: c8dbf542-70cb-45be-9308-5e12307c0be6] Attaching volume b9b77a26-5a59-41f2-b991-e09802684a66 to /dev/sdb [ 1145.492285] env[63515]: DEBUG oslo_vmware.api [None req-fdc6b8c1-6273-4192-b97d-29e8d79474af tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Task: {'id': task-1112010, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1145.524634] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f89c4bce-e753-4a11-a4c2-4fea219ca7db {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.531709] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83d34f38-f1a7-474a-88ed-4575d7be3716 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.545623] env[63515]: DEBUG nova.virt.block_device [None req-bd647ce8-c9a2-4ac0-8bba-7604eb48d8bd tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] [instance: c8dbf542-70cb-45be-9308-5e12307c0be6] Updating existing volume attachment record: a2f01946-54c6-483a-bef6-d6136a693187 {{(pid=63515) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1145.868472] env[63515]: DEBUG nova.network.neutron [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] [instance: e13da90c-28e6-43d2-99b7-19c5095954ca] Updating instance_info_cache with network_info: [{"id": "d9363b80-630e-4262-901f-c13e5c8b8c2a", "address": "fa:16:3e:dd:fe:76", "network": {"id": "cec614cb-7a5b-4b16-a83d-061ff62d6677", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1501468300-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "aef947bfaf7d4180b46b4be662e3d903", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3a0ddd7d-c321-4187-bdd8-b19044ea2c4a", "external-id": "nsx-vlan-transportzone-747", "segmentation_id": 747, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd9363b80-63", "ovs_interfaceid": "d9363b80-630e-4262-901f-c13e5c8b8c2a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1145.975557] env[63515]: DEBUG oslo_vmware.api [None req-6435b2a2-c59b-4028-9161-1e2243f4b9d6 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Task: {'id': task-1112009, 'name': PowerOffVM_Task, 'duration_secs': 0.183377} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1145.975826] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-6435b2a2-c59b-4028-9161-1e2243f4b9d6 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: 9a864a51-7df3-4afc-8be7-2794ef2dcf1a] Powered off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1145.976040] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-6435b2a2-c59b-4028-9161-1e2243f4b9d6 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: 9a864a51-7df3-4afc-8be7-2794ef2dcf1a] Updating instance '9a864a51-7df3-4afc-8be7-2794ef2dcf1a' progress to 17 {{(pid=63515) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1145.991766] env[63515]: DEBUG oslo_vmware.api [None req-fdc6b8c1-6273-4192-b97d-29e8d79474af tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Task: {'id': task-1112010, 'name': PowerOnVM_Task, 'duration_secs': 0.439753} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1145.992053] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-fdc6b8c1-6273-4192-b97d-29e8d79474af tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 24a79a6c-5210-4778-a6ed-482bfe14933c] Powered on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1145.992270] env[63515]: INFO nova.compute.manager [None req-fdc6b8c1-6273-4192-b97d-29e8d79474af tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 24a79a6c-5210-4778-a6ed-482bfe14933c] Took 6.90 seconds to spawn the instance on the hypervisor. [ 1145.992460] env[63515]: DEBUG nova.compute.manager [None req-fdc6b8c1-6273-4192-b97d-29e8d79474af tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 24a79a6c-5210-4778-a6ed-482bfe14933c] Checking state {{(pid=63515) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1145.993293] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25cd322f-29bd-4e57-901e-fe793c01fa94 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1146.371258] env[63515]: DEBUG oslo_concurrency.lockutils [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Releasing lock "refresh_cache-e13da90c-28e6-43d2-99b7-19c5095954ca" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1146.371490] env[63515]: DEBUG nova.compute.manager [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] [instance: e13da90c-28e6-43d2-99b7-19c5095954ca] Updated the network info_cache for instance {{(pid=63515) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9997}} [ 1146.371711] env[63515]: DEBUG oslo_service.periodic_task [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=63515) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1146.371884] env[63515]: DEBUG oslo_service.periodic_task [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Running periodic task ComputeManager.update_available_resource {{(pid=63515) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1146.482638] env[63515]: DEBUG nova.virt.hardware [None req-6435b2a2-c59b-4028-9161-1e2243f4b9d6 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T02:52:16Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1146.483057] env[63515]: DEBUG nova.virt.hardware [None req-6435b2a2-c59b-4028-9161-1e2243f4b9d6 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Flavor limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1146.483171] env[63515]: DEBUG nova.virt.hardware [None req-6435b2a2-c59b-4028-9161-1e2243f4b9d6 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Image limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1146.483423] env[63515]: DEBUG nova.virt.hardware [None req-6435b2a2-c59b-4028-9161-1e2243f4b9d6 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Flavor pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1146.483617] env[63515]: DEBUG nova.virt.hardware [None req-6435b2a2-c59b-4028-9161-1e2243f4b9d6 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Image pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1146.483807] env[63515]: DEBUG nova.virt.hardware [None req-6435b2a2-c59b-4028-9161-1e2243f4b9d6 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1146.484067] env[63515]: DEBUG nova.virt.hardware [None req-6435b2a2-c59b-4028-9161-1e2243f4b9d6 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1146.484296] env[63515]: DEBUG nova.virt.hardware [None req-6435b2a2-c59b-4028-9161-1e2243f4b9d6 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1146.484485] env[63515]: DEBUG nova.virt.hardware [None req-6435b2a2-c59b-4028-9161-1e2243f4b9d6 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Got 1 possible topologies {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1146.484656] env[63515]: DEBUG nova.virt.hardware [None req-6435b2a2-c59b-4028-9161-1e2243f4b9d6 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1146.484831] env[63515]: DEBUG nova.virt.hardware [None req-6435b2a2-c59b-4028-9161-1e2243f4b9d6 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1146.490646] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2139c17d-121c-4959-99f9-e55d80598491 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1146.508211] env[63515]: DEBUG oslo_vmware.api [None req-6435b2a2-c59b-4028-9161-1e2243f4b9d6 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Waiting for the task: (returnval){ [ 1146.508211] env[63515]: value = "task-1112014" [ 1146.508211] env[63515]: _type = "Task" [ 1146.508211] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1146.516174] env[63515]: INFO nova.compute.manager [None req-fdc6b8c1-6273-4192-b97d-29e8d79474af tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 24a79a6c-5210-4778-a6ed-482bfe14933c] Took 11.65 seconds to build instance. [ 1146.522662] env[63515]: DEBUG oslo_vmware.api [None req-6435b2a2-c59b-4028-9161-1e2243f4b9d6 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Task: {'id': task-1112014, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1146.876639] env[63515]: DEBUG oslo_concurrency.lockutils [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1146.876639] env[63515]: DEBUG oslo_concurrency.lockutils [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1146.876639] env[63515]: DEBUG oslo_concurrency.lockutils [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1146.876639] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=63515) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1146.876639] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-864d7c0e-5e58-4638-a16e-444efb27f87a {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1146.885192] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09208226-a437-42e6-b0a3-5e02fc62fb41 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1146.899114] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b15f4e45-0480-4697-b7e9-f0bb6931f5a3 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1146.905910] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5cc0e415-5876-402b-940f-312b7efafaf8 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1146.936288] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181121MB free_disk=171GB free_vcpus=48 pci_devices=None {{(pid=63515) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1146.936459] env[63515]: DEBUG oslo_concurrency.lockutils [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1146.936663] env[63515]: DEBUG oslo_concurrency.lockutils [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1147.020441] env[63515]: DEBUG oslo_concurrency.lockutils [None req-fdc6b8c1-6273-4192-b97d-29e8d79474af tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Lock "24a79a6c-5210-4778-a6ed-482bfe14933c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.159s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1147.020735] env[63515]: DEBUG oslo_vmware.api [None req-6435b2a2-c59b-4028-9161-1e2243f4b9d6 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Task: {'id': task-1112014, 'name': ReconfigVM_Task, 'duration_secs': 0.164524} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1147.021029] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-6435b2a2-c59b-4028-9161-1e2243f4b9d6 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: 9a864a51-7df3-4afc-8be7-2794ef2dcf1a] Updating instance '9a864a51-7df3-4afc-8be7-2794ef2dcf1a' progress to 33 {{(pid=63515) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1147.242828] env[63515]: DEBUG nova.compute.manager [req-8d61832f-3286-4835-bd4a-a62aacf4edbd req-29c53d48-c042-45da-9db4-62a419511a71 service nova] [instance: 24a79a6c-5210-4778-a6ed-482bfe14933c] Received event network-changed-f338b564-7636-4839-b034-73c3233d7da6 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1147.243052] env[63515]: DEBUG nova.compute.manager [req-8d61832f-3286-4835-bd4a-a62aacf4edbd req-29c53d48-c042-45da-9db4-62a419511a71 service nova] [instance: 24a79a6c-5210-4778-a6ed-482bfe14933c] Refreshing instance network info cache due to event network-changed-f338b564-7636-4839-b034-73c3233d7da6. {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1147.243282] env[63515]: DEBUG oslo_concurrency.lockutils [req-8d61832f-3286-4835-bd4a-a62aacf4edbd req-29c53d48-c042-45da-9db4-62a419511a71 service nova] Acquiring lock "refresh_cache-24a79a6c-5210-4778-a6ed-482bfe14933c" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1147.243435] env[63515]: DEBUG oslo_concurrency.lockutils [req-8d61832f-3286-4835-bd4a-a62aacf4edbd req-29c53d48-c042-45da-9db4-62a419511a71 service nova] Acquired lock "refresh_cache-24a79a6c-5210-4778-a6ed-482bfe14933c" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1147.243598] env[63515]: DEBUG nova.network.neutron [req-8d61832f-3286-4835-bd4a-a62aacf4edbd req-29c53d48-c042-45da-9db4-62a419511a71 service nova] [instance: 24a79a6c-5210-4778-a6ed-482bfe14933c] Refreshing network info cache for port f338b564-7636-4839-b034-73c3233d7da6 {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1147.528107] env[63515]: DEBUG nova.virt.hardware [None req-6435b2a2-c59b-4028-9161-1e2243f4b9d6 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T02:52:14Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1147.528776] env[63515]: DEBUG nova.virt.hardware [None req-6435b2a2-c59b-4028-9161-1e2243f4b9d6 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Flavor limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1147.528776] env[63515]: DEBUG nova.virt.hardware [None req-6435b2a2-c59b-4028-9161-1e2243f4b9d6 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Image limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1147.529054] env[63515]: DEBUG nova.virt.hardware [None req-6435b2a2-c59b-4028-9161-1e2243f4b9d6 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Flavor pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1147.529691] env[63515]: DEBUG nova.virt.hardware [None req-6435b2a2-c59b-4028-9161-1e2243f4b9d6 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Image pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1147.529691] env[63515]: DEBUG nova.virt.hardware [None req-6435b2a2-c59b-4028-9161-1e2243f4b9d6 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1147.529691] env[63515]: DEBUG nova.virt.hardware [None req-6435b2a2-c59b-4028-9161-1e2243f4b9d6 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1147.529904] env[63515]: DEBUG nova.virt.hardware [None req-6435b2a2-c59b-4028-9161-1e2243f4b9d6 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1147.530160] env[63515]: DEBUG nova.virt.hardware [None req-6435b2a2-c59b-4028-9161-1e2243f4b9d6 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Got 1 possible topologies {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1147.530477] env[63515]: DEBUG nova.virt.hardware [None req-6435b2a2-c59b-4028-9161-1e2243f4b9d6 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1147.530785] env[63515]: DEBUG nova.virt.hardware [None req-6435b2a2-c59b-4028-9161-1e2243f4b9d6 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1147.536357] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-6435b2a2-c59b-4028-9161-1e2243f4b9d6 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: 9a864a51-7df3-4afc-8be7-2794ef2dcf1a] Reconfiguring VM instance instance-0000006e to detach disk 2000 {{(pid=63515) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1147.536711] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-67522b70-1fa9-4221-9b11-1a88d799bbb1 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.558011] env[63515]: DEBUG oslo_vmware.api [None req-6435b2a2-c59b-4028-9161-1e2243f4b9d6 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Waiting for the task: (returnval){ [ 1147.558011] env[63515]: value = "task-1112015" [ 1147.558011] env[63515]: _type = "Task" [ 1147.558011] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1147.566058] env[63515]: DEBUG oslo_vmware.api [None req-6435b2a2-c59b-4028-9161-1e2243f4b9d6 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Task: {'id': task-1112015, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1147.946516] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Applying migration context for instance 9a864a51-7df3-4afc-8be7-2794ef2dcf1a as it has an incoming, in-progress migration 2ac80a52-5354-43ed-b08a-87a7e019e350. Migration status is migrating {{(pid=63515) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1016}} [ 1147.947558] env[63515]: INFO nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] [instance: 9a864a51-7df3-4afc-8be7-2794ef2dcf1a] Updating resource usage from migration 2ac80a52-5354-43ed-b08a-87a7e019e350 [ 1147.969572] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Instance e13da90c-28e6-43d2-99b7-19c5095954ca actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63515) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1147.971109] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Instance ea2e0810-c9f1-49a6-8589-7db402847aea actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63515) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1147.971318] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Instance c8dbf542-70cb-45be-9308-5e12307c0be6 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63515) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1147.971453] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Instance c9e07a6c-fd14-4071-8c69-6ece62e9109a actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63515) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1147.971571] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Instance 24a79a6c-5210-4778-a6ed-482bfe14933c actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63515) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1147.971685] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Migration 2ac80a52-5354-43ed-b08a-87a7e019e350 is active on this compute host and has allocations in placement: {'resources': {'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63515) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1712}} [ 1147.971797] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Instance 9a864a51-7df3-4afc-8be7-2794ef2dcf1a actively managed on this compute host and has allocations in placement: {'resources': {'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=63515) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1147.971985] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Total usable vcpus: 48, total allocated vcpus: 7 {{(pid=63515) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1147.972135] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1920MB phys_disk=200GB used_disk=5GB total_vcpus=48 used_vcpus=7 pci_stats=[] {{(pid=63515) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1148.058985] env[63515]: DEBUG nova.network.neutron [req-8d61832f-3286-4835-bd4a-a62aacf4edbd req-29c53d48-c042-45da-9db4-62a419511a71 service nova] [instance: 24a79a6c-5210-4778-a6ed-482bfe14933c] Updated VIF entry in instance network info cache for port f338b564-7636-4839-b034-73c3233d7da6. {{(pid=63515) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1148.059431] env[63515]: DEBUG nova.network.neutron [req-8d61832f-3286-4835-bd4a-a62aacf4edbd req-29c53d48-c042-45da-9db4-62a419511a71 service nova] [instance: 24a79a6c-5210-4778-a6ed-482bfe14933c] Updating instance_info_cache with network_info: [{"id": "f338b564-7636-4839-b034-73c3233d7da6", "address": "fa:16:3e:bc:5a:12", "network": {"id": "5758c885-e1c9-45c8-86aa-27a802101f52", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-734458138-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.204", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b18d4daacbc84e758f9f9ae4f3f09d28", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69351262-8d39-441a-85ba-3a78df436d17", "external-id": "nsx-vlan-transportzone-205", "segmentation_id": 205, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf338b564-76", "ovs_interfaceid": "f338b564-7636-4839-b034-73c3233d7da6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1148.071391] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c96c4345-4528-4f95-91db-6b53d57409b8 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1148.074055] env[63515]: DEBUG oslo_vmware.api [None req-6435b2a2-c59b-4028-9161-1e2243f4b9d6 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Task: {'id': task-1112015, 'name': ReconfigVM_Task, 'duration_secs': 0.192197} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1148.074450] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-6435b2a2-c59b-4028-9161-1e2243f4b9d6 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: 9a864a51-7df3-4afc-8be7-2794ef2dcf1a] Reconfigured VM instance instance-0000006e to detach disk 2000 {{(pid=63515) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1148.075565] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8578a22-d5fa-41c5-ba84-676418c33d8c {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1148.080926] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-882efcce-d841-4bde-9576-680216250c8e {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1148.101549] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-6435b2a2-c59b-4028-9161-1e2243f4b9d6 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: 9a864a51-7df3-4afc-8be7-2794ef2dcf1a] Reconfiguring VM instance instance-0000006e to attach disk [datastore2] volume-4ca406d5-39c8-41c2-b985-4734131e1c3c/volume-4ca406d5-39c8-41c2-b985-4734131e1c3c.vmdk or device None with type thin {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1148.102060] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-632ddf10-b564-461e-884f-62d5e0bc5d76 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1148.138674] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b770f1a-c625-4743-a7f6-d7efa04dd4d8 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1148.142625] env[63515]: DEBUG oslo_vmware.api [None req-6435b2a2-c59b-4028-9161-1e2243f4b9d6 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Waiting for the task: (returnval){ [ 1148.142625] env[63515]: value = "task-1112017" [ 1148.142625] env[63515]: _type = "Task" [ 1148.142625] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1148.147879] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dece741c-bae4-417a-ba7c-e30ac81ee8b2 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1148.157486] env[63515]: DEBUG oslo_vmware.api [None req-6435b2a2-c59b-4028-9161-1e2243f4b9d6 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Task: {'id': task-1112017, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1148.165024] env[63515]: DEBUG nova.compute.provider_tree [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1148.565384] env[63515]: DEBUG oslo_concurrency.lockutils [req-8d61832f-3286-4835-bd4a-a62aacf4edbd req-29c53d48-c042-45da-9db4-62a419511a71 service nova] Releasing lock "refresh_cache-24a79a6c-5210-4778-a6ed-482bfe14933c" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1148.653804] env[63515]: DEBUG oslo_vmware.api [None req-6435b2a2-c59b-4028-9161-1e2243f4b9d6 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Task: {'id': task-1112017, 'name': ReconfigVM_Task, 'duration_secs': 0.504282} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1148.654239] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-6435b2a2-c59b-4028-9161-1e2243f4b9d6 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: 9a864a51-7df3-4afc-8be7-2794ef2dcf1a] Reconfigured VM instance instance-0000006e to attach disk [datastore2] volume-4ca406d5-39c8-41c2-b985-4734131e1c3c/volume-4ca406d5-39c8-41c2-b985-4734131e1c3c.vmdk or device None with type thin {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1148.654579] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-6435b2a2-c59b-4028-9161-1e2243f4b9d6 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: 9a864a51-7df3-4afc-8be7-2794ef2dcf1a] Updating instance '9a864a51-7df3-4afc-8be7-2794ef2dcf1a' progress to 50 {{(pid=63515) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1148.668519] env[63515]: DEBUG nova.scheduler.client.report [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1149.162494] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03a7461b-2350-4dd7-bc97-009b6aef334e {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.181169] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=63515) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1149.181521] env[63515]: DEBUG oslo_concurrency.lockutils [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.245s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1149.182440] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-989b6e14-f662-4603-b760-4debd6457951 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.200068] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-6435b2a2-c59b-4028-9161-1e2243f4b9d6 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: 9a864a51-7df3-4afc-8be7-2794ef2dcf1a] Updating instance '9a864a51-7df3-4afc-8be7-2794ef2dcf1a' progress to 67 {{(pid=63515) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1150.095052] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-bd647ce8-c9a2-4ac0-8bba-7604eb48d8bd tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] [instance: c8dbf542-70cb-45be-9308-5e12307c0be6] Volume attach. Driver type: vmdk {{(pid=63515) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1150.095052] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-bd647ce8-c9a2-4ac0-8bba-7604eb48d8bd tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] [instance: c8dbf542-70cb-45be-9308-5e12307c0be6] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-243592', 'volume_id': 'b9b77a26-5a59-41f2-b991-e09802684a66', 'name': 'volume-b9b77a26-5a59-41f2-b991-e09802684a66', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'c8dbf542-70cb-45be-9308-5e12307c0be6', 'attached_at': '', 'detached_at': '', 'volume_id': 'b9b77a26-5a59-41f2-b991-e09802684a66', 'serial': 'b9b77a26-5a59-41f2-b991-e09802684a66'} {{(pid=63515) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1150.096488] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b279557a-3b62-4a56-bbf1-a949001070be {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.121537] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c3aff6e-015f-4ca7-be28-b5cf398c49a2 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.164591] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-bd647ce8-c9a2-4ac0-8bba-7604eb48d8bd tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] [instance: c8dbf542-70cb-45be-9308-5e12307c0be6] Reconfiguring VM instance instance-0000006b to attach disk [datastore2] volume-b9b77a26-5a59-41f2-b991-e09802684a66/volume-b9b77a26-5a59-41f2-b991-e09802684a66.vmdk or device None with type thin {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1150.165087] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c0bea1ba-517e-4a64-94c2-4791f497dc8b {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.193787] env[63515]: DEBUG oslo_vmware.api [None req-bd647ce8-c9a2-4ac0-8bba-7604eb48d8bd tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Waiting for the task: (returnval){ [ 1150.193787] env[63515]: value = "task-1112018" [ 1150.193787] env[63515]: _type = "Task" [ 1150.193787] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1150.206484] env[63515]: DEBUG oslo_vmware.api [None req-bd647ce8-c9a2-4ac0-8bba-7604eb48d8bd tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Task: {'id': task-1112018, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1150.703758] env[63515]: DEBUG oslo_vmware.api [None req-bd647ce8-c9a2-4ac0-8bba-7604eb48d8bd tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Task: {'id': task-1112018, 'name': ReconfigVM_Task, 'duration_secs': 0.358826} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1150.704117] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-bd647ce8-c9a2-4ac0-8bba-7604eb48d8bd tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] [instance: c8dbf542-70cb-45be-9308-5e12307c0be6] Reconfigured VM instance instance-0000006b to attach disk [datastore2] volume-b9b77a26-5a59-41f2-b991-e09802684a66/volume-b9b77a26-5a59-41f2-b991-e09802684a66.vmdk or device None with type thin {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1150.708923] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-58edb515-07c7-4c40-a57f-5067f7428584 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.723399] env[63515]: DEBUG oslo_vmware.api [None req-bd647ce8-c9a2-4ac0-8bba-7604eb48d8bd tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Waiting for the task: (returnval){ [ 1150.723399] env[63515]: value = "task-1112019" [ 1150.723399] env[63515]: _type = "Task" [ 1150.723399] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1150.731073] env[63515]: DEBUG oslo_vmware.api [None req-bd647ce8-c9a2-4ac0-8bba-7604eb48d8bd tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Task: {'id': task-1112019, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1150.831309] env[63515]: DEBUG oslo_concurrency.lockutils [None req-2f3128dc-4680-45d2-84d9-34c83ca4e4b1 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Acquiring lock "c9e07a6c-fd14-4071-8c69-6ece62e9109a" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1150.831600] env[63515]: DEBUG oslo_concurrency.lockutils [None req-2f3128dc-4680-45d2-84d9-34c83ca4e4b1 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Lock "c9e07a6c-fd14-4071-8c69-6ece62e9109a" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.001s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1150.831787] env[63515]: INFO nova.compute.manager [None req-2f3128dc-4680-45d2-84d9-34c83ca4e4b1 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] [instance: c9e07a6c-fd14-4071-8c69-6ece62e9109a] Shelving [ 1150.852140] env[63515]: DEBUG nova.network.neutron [None req-6435b2a2-c59b-4028-9161-1e2243f4b9d6 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: 9a864a51-7df3-4afc-8be7-2794ef2dcf1a] Port d96464dd-1638-4d19-8c71-9faeee62242f binding to destination host cpu-1 is already ACTIVE {{(pid=63515) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 1151.235853] env[63515]: DEBUG oslo_vmware.api [None req-bd647ce8-c9a2-4ac0-8bba-7604eb48d8bd tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Task: {'id': task-1112019, 'name': ReconfigVM_Task, 'duration_secs': 0.139023} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1151.236184] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-bd647ce8-c9a2-4ac0-8bba-7604eb48d8bd tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] [instance: c8dbf542-70cb-45be-9308-5e12307c0be6] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-243592', 'volume_id': 'b9b77a26-5a59-41f2-b991-e09802684a66', 'name': 'volume-b9b77a26-5a59-41f2-b991-e09802684a66', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'c8dbf542-70cb-45be-9308-5e12307c0be6', 'attached_at': '', 'detached_at': '', 'volume_id': 'b9b77a26-5a59-41f2-b991-e09802684a66', 'serial': 'b9b77a26-5a59-41f2-b991-e09802684a66'} {{(pid=63515) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1151.339314] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-2f3128dc-4680-45d2-84d9-34c83ca4e4b1 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] [instance: c9e07a6c-fd14-4071-8c69-6ece62e9109a] Powering off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1151.339576] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e0737593-dc62-475c-92f2-d3e0098b3e5b {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1151.347129] env[63515]: DEBUG oslo_vmware.api [None req-2f3128dc-4680-45d2-84d9-34c83ca4e4b1 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Waiting for the task: (returnval){ [ 1151.347129] env[63515]: value = "task-1112020" [ 1151.347129] env[63515]: _type = "Task" [ 1151.347129] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1151.358354] env[63515]: DEBUG oslo_vmware.api [None req-2f3128dc-4680-45d2-84d9-34c83ca4e4b1 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Task: {'id': task-1112020, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1151.856836] env[63515]: DEBUG oslo_vmware.api [None req-2f3128dc-4680-45d2-84d9-34c83ca4e4b1 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Task: {'id': task-1112020, 'name': PowerOffVM_Task, 'duration_secs': 0.19854} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1151.857070] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-2f3128dc-4680-45d2-84d9-34c83ca4e4b1 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] [instance: c9e07a6c-fd14-4071-8c69-6ece62e9109a] Powered off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1151.857856] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d542294-8a16-45ef-89c7-daa19e1a8004 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1151.876024] env[63515]: DEBUG oslo_concurrency.lockutils [None req-6435b2a2-c59b-4028-9161-1e2243f4b9d6 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Acquiring lock "9a864a51-7df3-4afc-8be7-2794ef2dcf1a-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1151.876024] env[63515]: DEBUG oslo_concurrency.lockutils [None req-6435b2a2-c59b-4028-9161-1e2243f4b9d6 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Lock "9a864a51-7df3-4afc-8be7-2794ef2dcf1a-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1151.876024] env[63515]: DEBUG oslo_concurrency.lockutils [None req-6435b2a2-c59b-4028-9161-1e2243f4b9d6 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Lock "9a864a51-7df3-4afc-8be7-2794ef2dcf1a-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1151.890394] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f88acff-5c9a-4693-a20c-1cd85b922beb {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1152.271143] env[63515]: DEBUG nova.objects.instance [None req-bd647ce8-c9a2-4ac0-8bba-7604eb48d8bd tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Lazy-loading 'flavor' on Instance uuid c8dbf542-70cb-45be-9308-5e12307c0be6 {{(pid=63515) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1152.400601] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-2f3128dc-4680-45d2-84d9-34c83ca4e4b1 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] [instance: c9e07a6c-fd14-4071-8c69-6ece62e9109a] Creating Snapshot of the VM instance {{(pid=63515) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1152.400929] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-bdb8f4de-378a-43ae-8d22-9b07a8fd0195 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1152.409080] env[63515]: DEBUG oslo_vmware.api [None req-2f3128dc-4680-45d2-84d9-34c83ca4e4b1 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Waiting for the task: (returnval){ [ 1152.409080] env[63515]: value = "task-1112021" [ 1152.409080] env[63515]: _type = "Task" [ 1152.409080] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1152.417615] env[63515]: DEBUG oslo_vmware.api [None req-2f3128dc-4680-45d2-84d9-34c83ca4e4b1 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Task: {'id': task-1112021, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1152.776560] env[63515]: DEBUG oslo_concurrency.lockutils [None req-bd647ce8-c9a2-4ac0-8bba-7604eb48d8bd tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Lock "c8dbf542-70cb-45be-9308-5e12307c0be6" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.286s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1152.918977] env[63515]: DEBUG oslo_vmware.api [None req-2f3128dc-4680-45d2-84d9-34c83ca4e4b1 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Task: {'id': task-1112021, 'name': CreateSnapshot_Task, 'duration_secs': 0.418591} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1152.919363] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-2f3128dc-4680-45d2-84d9-34c83ca4e4b1 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] [instance: c9e07a6c-fd14-4071-8c69-6ece62e9109a] Created Snapshot of the VM instance {{(pid=63515) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1152.920144] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc2b2cb2-2fc7-47b0-a56b-3e2573132fc1 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1152.933355] env[63515]: DEBUG oslo_concurrency.lockutils [None req-6435b2a2-c59b-4028-9161-1e2243f4b9d6 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Acquiring lock "refresh_cache-9a864a51-7df3-4afc-8be7-2794ef2dcf1a" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1152.933355] env[63515]: DEBUG oslo_concurrency.lockutils [None req-6435b2a2-c59b-4028-9161-1e2243f4b9d6 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Acquired lock "refresh_cache-9a864a51-7df3-4afc-8be7-2794ef2dcf1a" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1152.933355] env[63515]: DEBUG nova.network.neutron [None req-6435b2a2-c59b-4028-9161-1e2243f4b9d6 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: 9a864a51-7df3-4afc-8be7-2794ef2dcf1a] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1153.438734] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-2f3128dc-4680-45d2-84d9-34c83ca4e4b1 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] [instance: c9e07a6c-fd14-4071-8c69-6ece62e9109a] Creating linked-clone VM from snapshot {{(pid=63515) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1153.441988] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-922eba4d-0524-48cf-9cd3-aadc1ab1f84d {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1153.449622] env[63515]: DEBUG oslo_vmware.api [None req-2f3128dc-4680-45d2-84d9-34c83ca4e4b1 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Waiting for the task: (returnval){ [ 1153.449622] env[63515]: value = "task-1112022" [ 1153.449622] env[63515]: _type = "Task" [ 1153.449622] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1153.460790] env[63515]: DEBUG oslo_vmware.api [None req-2f3128dc-4680-45d2-84d9-34c83ca4e4b1 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Task: {'id': task-1112022, 'name': CloneVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1153.684226] env[63515]: DEBUG nova.network.neutron [None req-6435b2a2-c59b-4028-9161-1e2243f4b9d6 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: 9a864a51-7df3-4afc-8be7-2794ef2dcf1a] Updating instance_info_cache with network_info: [{"id": "d96464dd-1638-4d19-8c71-9faeee62242f", "address": "fa:16:3e:36:fc:a7", "network": {"id": "cec614cb-7a5b-4b16-a83d-061ff62d6677", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1501468300-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.155", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "aef947bfaf7d4180b46b4be662e3d903", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3a0ddd7d-c321-4187-bdd8-b19044ea2c4a", "external-id": "nsx-vlan-transportzone-747", "segmentation_id": 747, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd96464dd-16", "ovs_interfaceid": "d96464dd-1638-4d19-8c71-9faeee62242f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1153.959711] env[63515]: DEBUG oslo_vmware.api [None req-2f3128dc-4680-45d2-84d9-34c83ca4e4b1 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Task: {'id': task-1112022, 'name': CloneVM_Task} progress is 94%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1154.000170] env[63515]: DEBUG oslo_concurrency.lockutils [None req-8edc157d-485a-48e3-9c94-dc21771fbb70 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Acquiring lock "bd3a77ee-a78b-4719-9353-307ae8b166b6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1154.000435] env[63515]: DEBUG oslo_concurrency.lockutils [None req-8edc157d-485a-48e3-9c94-dc21771fbb70 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Lock "bd3a77ee-a78b-4719-9353-307ae8b166b6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1154.187139] env[63515]: DEBUG oslo_concurrency.lockutils [None req-6435b2a2-c59b-4028-9161-1e2243f4b9d6 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Releasing lock "refresh_cache-9a864a51-7df3-4afc-8be7-2794ef2dcf1a" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1154.460863] env[63515]: DEBUG oslo_vmware.api [None req-2f3128dc-4680-45d2-84d9-34c83ca4e4b1 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Task: {'id': task-1112022, 'name': CloneVM_Task} progress is 94%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1154.503133] env[63515]: DEBUG nova.compute.manager [None req-8edc157d-485a-48e3-9c94-dc21771fbb70 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] [instance: bd3a77ee-a78b-4719-9353-307ae8b166b6] Starting instance... {{(pid=63515) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1154.697568] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c72146eb-dc4a-45b1-b2b9-37174e889241 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.704753] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d7fc545-1582-4ad0-aefa-d5dadee6492d {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.960628] env[63515]: DEBUG oslo_vmware.api [None req-2f3128dc-4680-45d2-84d9-34c83ca4e4b1 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Task: {'id': task-1112022, 'name': CloneVM_Task, 'duration_secs': 1.348433} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1154.960973] env[63515]: INFO nova.virt.vmwareapi.vmops [None req-2f3128dc-4680-45d2-84d9-34c83ca4e4b1 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] [instance: c9e07a6c-fd14-4071-8c69-6ece62e9109a] Created linked-clone VM from snapshot [ 1154.961749] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b35958fc-4c9b-4701-990d-d91b81514997 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.969032] env[63515]: DEBUG nova.virt.vmwareapi.images [None req-2f3128dc-4680-45d2-84d9-34c83ca4e4b1 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] [instance: c9e07a6c-fd14-4071-8c69-6ece62e9109a] Uploading image bea6df9b-0878-417c-8cb9-f77d080de2d6 {{(pid=63515) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 1154.994165] env[63515]: DEBUG oslo_vmware.rw_handles [None req-2f3128dc-4680-45d2-84d9-34c83ca4e4b1 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 1154.994165] env[63515]: value = "vm-243594" [ 1154.994165] env[63515]: _type = "VirtualMachine" [ 1154.994165] env[63515]: }. {{(pid=63515) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 1154.994477] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-e682803c-3299-446f-a5cc-b84c4f18399d {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1155.002012] env[63515]: DEBUG oslo_vmware.rw_handles [None req-2f3128dc-4680-45d2-84d9-34c83ca4e4b1 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Lease: (returnval){ [ 1155.002012] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]52576ecf-9852-239c-1934-19ecbc96f31b" [ 1155.002012] env[63515]: _type = "HttpNfcLease" [ 1155.002012] env[63515]: } obtained for exporting VM: (result){ [ 1155.002012] env[63515]: value = "vm-243594" [ 1155.002012] env[63515]: _type = "VirtualMachine" [ 1155.002012] env[63515]: }. {{(pid=63515) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 1155.002328] env[63515]: DEBUG oslo_vmware.api [None req-2f3128dc-4680-45d2-84d9-34c83ca4e4b1 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Waiting for the lease: (returnval){ [ 1155.002328] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]52576ecf-9852-239c-1934-19ecbc96f31b" [ 1155.002328] env[63515]: _type = "HttpNfcLease" [ 1155.002328] env[63515]: } to be ready. {{(pid=63515) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1155.010820] env[63515]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1155.010820] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]52576ecf-9852-239c-1934-19ecbc96f31b" [ 1155.010820] env[63515]: _type = "HttpNfcLease" [ 1155.010820] env[63515]: } is initializing. {{(pid=63515) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1155.026616] env[63515]: DEBUG oslo_concurrency.lockutils [None req-8edc157d-485a-48e3-9c94-dc21771fbb70 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1155.026890] env[63515]: DEBUG oslo_concurrency.lockutils [None req-8edc157d-485a-48e3-9c94-dc21771fbb70 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1155.028498] env[63515]: INFO nova.compute.claims [None req-8edc157d-485a-48e3-9c94-dc21771fbb70 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] [instance: bd3a77ee-a78b-4719-9353-307ae8b166b6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1155.510293] env[63515]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1155.510293] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]52576ecf-9852-239c-1934-19ecbc96f31b" [ 1155.510293] env[63515]: _type = "HttpNfcLease" [ 1155.510293] env[63515]: } is ready. {{(pid=63515) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1155.510904] env[63515]: DEBUG oslo_vmware.rw_handles [None req-2f3128dc-4680-45d2-84d9-34c83ca4e4b1 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1155.510904] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]52576ecf-9852-239c-1934-19ecbc96f31b" [ 1155.510904] env[63515]: _type = "HttpNfcLease" [ 1155.510904] env[63515]: }. {{(pid=63515) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 1155.511165] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46a5cbec-ae80-42f0-8d68-28013724b5b0 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1155.518689] env[63515]: DEBUG oslo_vmware.rw_handles [None req-2f3128dc-4680-45d2-84d9-34c83ca4e4b1 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52a6b98c-dd1e-547a-6e19-d80ce28fc3e4/disk-0.vmdk from lease info. {{(pid=63515) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1155.518877] env[63515]: DEBUG oslo_vmware.rw_handles [None req-2f3128dc-4680-45d2-84d9-34c83ca4e4b1 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52a6b98c-dd1e-547a-6e19-d80ce28fc3e4/disk-0.vmdk for reading. {{(pid=63515) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 1155.612565] env[63515]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-43116e1c-5060-4091-81fa-1e94ef8f6c7f {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1155.813685] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0589bce5-2476-45b6-bf8b-749168b9d7ab {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1155.847592] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba925ba8-065a-4c74-bfb4-66e87920b7f8 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1155.858611] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-6435b2a2-c59b-4028-9161-1e2243f4b9d6 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: 9a864a51-7df3-4afc-8be7-2794ef2dcf1a] Updating instance '9a864a51-7df3-4afc-8be7-2794ef2dcf1a' progress to 83 {{(pid=63515) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1156.194885] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf55065f-a7df-4d11-b8e3-3c2e25884aee {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1156.203206] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42e74cae-84f8-4a11-aff7-72e3976c7395 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1156.233862] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bcda400f-f741-4c61-9159-4a1eb79f8e06 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1156.242399] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7989b635-cf70-482b-9b29-8142a8072b8c {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1156.258699] env[63515]: DEBUG nova.compute.provider_tree [None req-8edc157d-485a-48e3-9c94-dc21771fbb70 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1156.366680] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-6435b2a2-c59b-4028-9161-1e2243f4b9d6 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: 9a864a51-7df3-4afc-8be7-2794ef2dcf1a] Powering on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1156.367623] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-cea54d8a-fa2e-49b7-b872-ead61b16edac {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1156.376697] env[63515]: DEBUG oslo_vmware.api [None req-6435b2a2-c59b-4028-9161-1e2243f4b9d6 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Waiting for the task: (returnval){ [ 1156.376697] env[63515]: value = "task-1112024" [ 1156.376697] env[63515]: _type = "Task" [ 1156.376697] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1156.386072] env[63515]: DEBUG oslo_vmware.api [None req-6435b2a2-c59b-4028-9161-1e2243f4b9d6 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Task: {'id': task-1112024, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1156.762163] env[63515]: DEBUG nova.scheduler.client.report [None req-8edc157d-485a-48e3-9c94-dc21771fbb70 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1156.887246] env[63515]: DEBUG oslo_vmware.api [None req-6435b2a2-c59b-4028-9161-1e2243f4b9d6 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Task: {'id': task-1112024, 'name': PowerOnVM_Task, 'duration_secs': 0.416099} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1156.887246] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-6435b2a2-c59b-4028-9161-1e2243f4b9d6 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: 9a864a51-7df3-4afc-8be7-2794ef2dcf1a] Powered on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1156.887402] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-6435b2a2-c59b-4028-9161-1e2243f4b9d6 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: 9a864a51-7df3-4afc-8be7-2794ef2dcf1a] Updating instance '9a864a51-7df3-4afc-8be7-2794ef2dcf1a' progress to 100 {{(pid=63515) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1157.269406] env[63515]: DEBUG oslo_concurrency.lockutils [None req-8edc157d-485a-48e3-9c94-dc21771fbb70 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.241s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1157.269406] env[63515]: DEBUG nova.compute.manager [None req-8edc157d-485a-48e3-9c94-dc21771fbb70 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] [instance: bd3a77ee-a78b-4719-9353-307ae8b166b6] Start building networks asynchronously for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1157.773940] env[63515]: DEBUG nova.compute.utils [None req-8edc157d-485a-48e3-9c94-dc21771fbb70 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Using /dev/sd instead of None {{(pid=63515) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1157.775603] env[63515]: DEBUG nova.compute.manager [None req-8edc157d-485a-48e3-9c94-dc21771fbb70 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] [instance: bd3a77ee-a78b-4719-9353-307ae8b166b6] Allocating IP information in the background. {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1157.775777] env[63515]: DEBUG nova.network.neutron [None req-8edc157d-485a-48e3-9c94-dc21771fbb70 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] [instance: bd3a77ee-a78b-4719-9353-307ae8b166b6] allocate_for_instance() {{(pid=63515) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1157.818265] env[63515]: DEBUG nova.policy [None req-8edc157d-485a-48e3-9c94-dc21771fbb70 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'cb4f41c6bfb94aa8ab60135e1d729a6a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '97f2e119708746038f38359572c99438', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63515) authorize /opt/stack/nova/nova/policy.py:201}} [ 1158.123830] env[63515]: DEBUG nova.network.neutron [None req-8edc157d-485a-48e3-9c94-dc21771fbb70 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] [instance: bd3a77ee-a78b-4719-9353-307ae8b166b6] Successfully created port: 764f6a9d-e89a-44b6-804c-70000c555f95 {{(pid=63515) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1158.279751] env[63515]: DEBUG nova.compute.manager [None req-8edc157d-485a-48e3-9c94-dc21771fbb70 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] [instance: bd3a77ee-a78b-4719-9353-307ae8b166b6] Start building block device mappings for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1158.824196] env[63515]: DEBUG oslo_concurrency.lockutils [None req-b86bd653-3ed8-48a7-a3cc-0c8e7c3ff20e tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Acquiring lock "9a864a51-7df3-4afc-8be7-2794ef2dcf1a" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1158.824492] env[63515]: DEBUG oslo_concurrency.lockutils [None req-b86bd653-3ed8-48a7-a3cc-0c8e7c3ff20e tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Lock "9a864a51-7df3-4afc-8be7-2794ef2dcf1a" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1158.824608] env[63515]: DEBUG nova.compute.manager [None req-b86bd653-3ed8-48a7-a3cc-0c8e7c3ff20e tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: 9a864a51-7df3-4afc-8be7-2794ef2dcf1a] Going to confirm migration 7 {{(pid=63515) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 1159.289805] env[63515]: DEBUG nova.compute.manager [None req-8edc157d-485a-48e3-9c94-dc21771fbb70 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] [instance: bd3a77ee-a78b-4719-9353-307ae8b166b6] Start spawning the instance on the hypervisor. {{(pid=63515) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1159.316506] env[63515]: DEBUG nova.virt.hardware [None req-8edc157d-485a-48e3-9c94-dc21771fbb70 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T02:52:14Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T02:51:56Z,direct_url=,disk_format='vmdk',id=8a120570-cb06-4099-b262-554ca0ad15c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b569255cc43e42e7ae3f2b4ad37c6ef4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T02:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1159.316776] env[63515]: DEBUG nova.virt.hardware [None req-8edc157d-485a-48e3-9c94-dc21771fbb70 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Flavor limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1159.316937] env[63515]: DEBUG nova.virt.hardware [None req-8edc157d-485a-48e3-9c94-dc21771fbb70 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Image limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1159.317178] env[63515]: DEBUG nova.virt.hardware [None req-8edc157d-485a-48e3-9c94-dc21771fbb70 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Flavor pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1159.317335] env[63515]: DEBUG nova.virt.hardware [None req-8edc157d-485a-48e3-9c94-dc21771fbb70 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Image pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1159.317488] env[63515]: DEBUG nova.virt.hardware [None req-8edc157d-485a-48e3-9c94-dc21771fbb70 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1159.317694] env[63515]: DEBUG nova.virt.hardware [None req-8edc157d-485a-48e3-9c94-dc21771fbb70 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1159.317858] env[63515]: DEBUG nova.virt.hardware [None req-8edc157d-485a-48e3-9c94-dc21771fbb70 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1159.318488] env[63515]: DEBUG nova.virt.hardware [None req-8edc157d-485a-48e3-9c94-dc21771fbb70 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Got 1 possible topologies {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1159.318488] env[63515]: DEBUG nova.virt.hardware [None req-8edc157d-485a-48e3-9c94-dc21771fbb70 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1159.318488] env[63515]: DEBUG nova.virt.hardware [None req-8edc157d-485a-48e3-9c94-dc21771fbb70 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1159.319362] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14498cea-df0b-4aaf-8380-187e24352525 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1159.327957] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b790b0e8-c8ee-42c8-9930-52ef846cb277 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1159.401951] env[63515]: DEBUG oslo_concurrency.lockutils [None req-b86bd653-3ed8-48a7-a3cc-0c8e7c3ff20e tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Acquiring lock "refresh_cache-9a864a51-7df3-4afc-8be7-2794ef2dcf1a" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1159.402169] env[63515]: DEBUG oslo_concurrency.lockutils [None req-b86bd653-3ed8-48a7-a3cc-0c8e7c3ff20e tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Acquired lock "refresh_cache-9a864a51-7df3-4afc-8be7-2794ef2dcf1a" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1159.402359] env[63515]: DEBUG nova.network.neutron [None req-b86bd653-3ed8-48a7-a3cc-0c8e7c3ff20e tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: 9a864a51-7df3-4afc-8be7-2794ef2dcf1a] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1159.402561] env[63515]: DEBUG nova.objects.instance [None req-b86bd653-3ed8-48a7-a3cc-0c8e7c3ff20e tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Lazy-loading 'info_cache' on Instance uuid 9a864a51-7df3-4afc-8be7-2794ef2dcf1a {{(pid=63515) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1159.669084] env[63515]: DEBUG nova.compute.manager [req-f48728ef-84aa-4b8e-acf3-36c32b16570d req-6c1cdb98-2b1c-437a-aa86-172b2ad6415e service nova] [instance: bd3a77ee-a78b-4719-9353-307ae8b166b6] Received event network-vif-plugged-764f6a9d-e89a-44b6-804c-70000c555f95 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1159.669386] env[63515]: DEBUG oslo_concurrency.lockutils [req-f48728ef-84aa-4b8e-acf3-36c32b16570d req-6c1cdb98-2b1c-437a-aa86-172b2ad6415e service nova] Acquiring lock "bd3a77ee-a78b-4719-9353-307ae8b166b6-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1159.669626] env[63515]: DEBUG oslo_concurrency.lockutils [req-f48728ef-84aa-4b8e-acf3-36c32b16570d req-6c1cdb98-2b1c-437a-aa86-172b2ad6415e service nova] Lock "bd3a77ee-a78b-4719-9353-307ae8b166b6-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1159.669890] env[63515]: DEBUG oslo_concurrency.lockutils [req-f48728ef-84aa-4b8e-acf3-36c32b16570d req-6c1cdb98-2b1c-437a-aa86-172b2ad6415e service nova] Lock "bd3a77ee-a78b-4719-9353-307ae8b166b6-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1159.670166] env[63515]: DEBUG nova.compute.manager [req-f48728ef-84aa-4b8e-acf3-36c32b16570d req-6c1cdb98-2b1c-437a-aa86-172b2ad6415e service nova] [instance: bd3a77ee-a78b-4719-9353-307ae8b166b6] No waiting events found dispatching network-vif-plugged-764f6a9d-e89a-44b6-804c-70000c555f95 {{(pid=63515) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1159.670366] env[63515]: WARNING nova.compute.manager [req-f48728ef-84aa-4b8e-acf3-36c32b16570d req-6c1cdb98-2b1c-437a-aa86-172b2ad6415e service nova] [instance: bd3a77ee-a78b-4719-9353-307ae8b166b6] Received unexpected event network-vif-plugged-764f6a9d-e89a-44b6-804c-70000c555f95 for instance with vm_state building and task_state spawning. [ 1159.760478] env[63515]: DEBUG nova.network.neutron [None req-8edc157d-485a-48e3-9c94-dc21771fbb70 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] [instance: bd3a77ee-a78b-4719-9353-307ae8b166b6] Successfully updated port: 764f6a9d-e89a-44b6-804c-70000c555f95 {{(pid=63515) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1160.266803] env[63515]: DEBUG oslo_concurrency.lockutils [None req-8edc157d-485a-48e3-9c94-dc21771fbb70 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Acquiring lock "refresh_cache-bd3a77ee-a78b-4719-9353-307ae8b166b6" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1160.267232] env[63515]: DEBUG oslo_concurrency.lockutils [None req-8edc157d-485a-48e3-9c94-dc21771fbb70 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Acquired lock "refresh_cache-bd3a77ee-a78b-4719-9353-307ae8b166b6" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1160.267232] env[63515]: DEBUG nova.network.neutron [None req-8edc157d-485a-48e3-9c94-dc21771fbb70 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] [instance: bd3a77ee-a78b-4719-9353-307ae8b166b6] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1160.657795] env[63515]: DEBUG nova.network.neutron [None req-b86bd653-3ed8-48a7-a3cc-0c8e7c3ff20e tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: 9a864a51-7df3-4afc-8be7-2794ef2dcf1a] Updating instance_info_cache with network_info: [{"id": "d96464dd-1638-4d19-8c71-9faeee62242f", "address": "fa:16:3e:36:fc:a7", "network": {"id": "cec614cb-7a5b-4b16-a83d-061ff62d6677", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-1501468300-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.155", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "aef947bfaf7d4180b46b4be662e3d903", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3a0ddd7d-c321-4187-bdd8-b19044ea2c4a", "external-id": "nsx-vlan-transportzone-747", "segmentation_id": 747, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd96464dd-16", "ovs_interfaceid": "d96464dd-1638-4d19-8c71-9faeee62242f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1160.802018] env[63515]: DEBUG nova.network.neutron [None req-8edc157d-485a-48e3-9c94-dc21771fbb70 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] [instance: bd3a77ee-a78b-4719-9353-307ae8b166b6] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1160.939418] env[63515]: DEBUG nova.network.neutron [None req-8edc157d-485a-48e3-9c94-dc21771fbb70 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] [instance: bd3a77ee-a78b-4719-9353-307ae8b166b6] Updating instance_info_cache with network_info: [{"id": "764f6a9d-e89a-44b6-804c-70000c555f95", "address": "fa:16:3e:56:19:d2", "network": {"id": "d6da2650-9ba9-4ef6-8035-e58b635a64b2", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-391564968-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "97f2e119708746038f38359572c99438", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "043ca97b-0fca-4b54-8be6-027123fa76d0", "external-id": "nsx-vlan-transportzone-702", "segmentation_id": 702, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap764f6a9d-e8", "ovs_interfaceid": "764f6a9d-e89a-44b6-804c-70000c555f95", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1161.161023] env[63515]: DEBUG oslo_concurrency.lockutils [None req-b86bd653-3ed8-48a7-a3cc-0c8e7c3ff20e tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Releasing lock "refresh_cache-9a864a51-7df3-4afc-8be7-2794ef2dcf1a" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1161.161342] env[63515]: DEBUG nova.objects.instance [None req-b86bd653-3ed8-48a7-a3cc-0c8e7c3ff20e tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Lazy-loading 'migration_context' on Instance uuid 9a864a51-7df3-4afc-8be7-2794ef2dcf1a {{(pid=63515) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1161.441932] env[63515]: DEBUG oslo_concurrency.lockutils [None req-8edc157d-485a-48e3-9c94-dc21771fbb70 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Releasing lock "refresh_cache-bd3a77ee-a78b-4719-9353-307ae8b166b6" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1161.442354] env[63515]: DEBUG nova.compute.manager [None req-8edc157d-485a-48e3-9c94-dc21771fbb70 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] [instance: bd3a77ee-a78b-4719-9353-307ae8b166b6] Instance network_info: |[{"id": "764f6a9d-e89a-44b6-804c-70000c555f95", "address": "fa:16:3e:56:19:d2", "network": {"id": "d6da2650-9ba9-4ef6-8035-e58b635a64b2", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-391564968-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "97f2e119708746038f38359572c99438", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "043ca97b-0fca-4b54-8be6-027123fa76d0", "external-id": "nsx-vlan-transportzone-702", "segmentation_id": 702, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap764f6a9d-e8", "ovs_interfaceid": "764f6a9d-e89a-44b6-804c-70000c555f95", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1161.442857] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-8edc157d-485a-48e3-9c94-dc21771fbb70 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] [instance: bd3a77ee-a78b-4719-9353-307ae8b166b6] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:56:19:d2', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '043ca97b-0fca-4b54-8be6-027123fa76d0', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '764f6a9d-e89a-44b6-804c-70000c555f95', 'vif_model': 'vmxnet3'}] {{(pid=63515) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1161.451148] env[63515]: DEBUG oslo.service.loopingcall [None req-8edc157d-485a-48e3-9c94-dc21771fbb70 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1161.451374] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bd3a77ee-a78b-4719-9353-307ae8b166b6] Creating VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1161.452054] env[63515]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-7a9418d2-0ddc-4414-be9a-717e9999d236 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.473352] env[63515]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1161.473352] env[63515]: value = "task-1112025" [ 1161.473352] env[63515]: _type = "Task" [ 1161.473352] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1161.481115] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1112025, 'name': CreateVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1161.664473] env[63515]: DEBUG nova.objects.base [None req-b86bd653-3ed8-48a7-a3cc-0c8e7c3ff20e tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Object Instance<9a864a51-7df3-4afc-8be7-2794ef2dcf1a> lazy-loaded attributes: info_cache,migration_context {{(pid=63515) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1161.665491] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f44aaa42-1ca1-4a02-ae79-cda0b67e175e {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.688423] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-29181ed7-a342-4e9e-86a1-35603f805db8 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.694204] env[63515]: DEBUG oslo_vmware.api [None req-b86bd653-3ed8-48a7-a3cc-0c8e7c3ff20e tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Waiting for the task: (returnval){ [ 1161.694204] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]52a9f0c2-353c-df1a-a1a2-ec40913bc672" [ 1161.694204] env[63515]: _type = "Task" [ 1161.694204] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1161.699901] env[63515]: DEBUG nova.compute.manager [req-2145ad35-85f9-41a1-b6b4-e4102c7eaa5b req-15619a96-e991-48a7-8267-8e17ee42ee26 service nova] [instance: bd3a77ee-a78b-4719-9353-307ae8b166b6] Received event network-changed-764f6a9d-e89a-44b6-804c-70000c555f95 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1161.699901] env[63515]: DEBUG nova.compute.manager [req-2145ad35-85f9-41a1-b6b4-e4102c7eaa5b req-15619a96-e991-48a7-8267-8e17ee42ee26 service nova] [instance: bd3a77ee-a78b-4719-9353-307ae8b166b6] Refreshing instance network info cache due to event network-changed-764f6a9d-e89a-44b6-804c-70000c555f95. {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1161.700176] env[63515]: DEBUG oslo_concurrency.lockutils [req-2145ad35-85f9-41a1-b6b4-e4102c7eaa5b req-15619a96-e991-48a7-8267-8e17ee42ee26 service nova] Acquiring lock "refresh_cache-bd3a77ee-a78b-4719-9353-307ae8b166b6" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1161.700332] env[63515]: DEBUG oslo_concurrency.lockutils [req-2145ad35-85f9-41a1-b6b4-e4102c7eaa5b req-15619a96-e991-48a7-8267-8e17ee42ee26 service nova] Acquired lock "refresh_cache-bd3a77ee-a78b-4719-9353-307ae8b166b6" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1161.700499] env[63515]: DEBUG nova.network.neutron [req-2145ad35-85f9-41a1-b6b4-e4102c7eaa5b req-15619a96-e991-48a7-8267-8e17ee42ee26 service nova] [instance: bd3a77ee-a78b-4719-9353-307ae8b166b6] Refreshing network info cache for port 764f6a9d-e89a-44b6-804c-70000c555f95 {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1161.707079] env[63515]: DEBUG oslo_vmware.api [None req-b86bd653-3ed8-48a7-a3cc-0c8e7c3ff20e tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52a9f0c2-353c-df1a-a1a2-ec40913bc672, 'name': SearchDatastore_Task, 'duration_secs': 0.008368} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1161.707370] env[63515]: DEBUG oslo_concurrency.lockutils [None req-b86bd653-3ed8-48a7-a3cc-0c8e7c3ff20e tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1161.707656] env[63515]: DEBUG oslo_concurrency.lockutils [None req-b86bd653-3ed8-48a7-a3cc-0c8e7c3ff20e tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1161.984538] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1112025, 'name': CreateVM_Task, 'duration_secs': 0.431772} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1161.984713] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bd3a77ee-a78b-4719-9353-307ae8b166b6] Created VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1161.985517] env[63515]: DEBUG oslo_concurrency.lockutils [None req-8edc157d-485a-48e3-9c94-dc21771fbb70 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1161.985738] env[63515]: DEBUG oslo_concurrency.lockutils [None req-8edc157d-485a-48e3-9c94-dc21771fbb70 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1161.986108] env[63515]: DEBUG oslo_concurrency.lockutils [None req-8edc157d-485a-48e3-9c94-dc21771fbb70 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1161.986392] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-14c55b41-beb4-48f1-8c27-951dc193e1a6 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.991161] env[63515]: DEBUG oslo_vmware.api [None req-8edc157d-485a-48e3-9c94-dc21771fbb70 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Waiting for the task: (returnval){ [ 1161.991161] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]52e1b942-7ca1-0e50-dbe4-a14bf24a9bc8" [ 1161.991161] env[63515]: _type = "Task" [ 1161.991161] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1162.000102] env[63515]: DEBUG oslo_vmware.api [None req-8edc157d-485a-48e3-9c94-dc21771fbb70 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52e1b942-7ca1-0e50-dbe4-a14bf24a9bc8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1162.322510] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4160625d-5958-4021-886b-50f02a95ec85 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.330874] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9df7b182-a250-432b-af23-5325e1648c6f {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.362573] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff7237f6-d974-455d-beec-83dd417d264f {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.370492] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60b16aa9-2163-4042-94a5-6b67d8ae12a4 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.385192] env[63515]: DEBUG nova.compute.provider_tree [None req-b86bd653-3ed8-48a7-a3cc-0c8e7c3ff20e tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1162.436605] env[63515]: DEBUG nova.network.neutron [req-2145ad35-85f9-41a1-b6b4-e4102c7eaa5b req-15619a96-e991-48a7-8267-8e17ee42ee26 service nova] [instance: bd3a77ee-a78b-4719-9353-307ae8b166b6] Updated VIF entry in instance network info cache for port 764f6a9d-e89a-44b6-804c-70000c555f95. {{(pid=63515) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1162.436962] env[63515]: DEBUG nova.network.neutron [req-2145ad35-85f9-41a1-b6b4-e4102c7eaa5b req-15619a96-e991-48a7-8267-8e17ee42ee26 service nova] [instance: bd3a77ee-a78b-4719-9353-307ae8b166b6] Updating instance_info_cache with network_info: [{"id": "764f6a9d-e89a-44b6-804c-70000c555f95", "address": "fa:16:3e:56:19:d2", "network": {"id": "d6da2650-9ba9-4ef6-8035-e58b635a64b2", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-391564968-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "97f2e119708746038f38359572c99438", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "043ca97b-0fca-4b54-8be6-027123fa76d0", "external-id": "nsx-vlan-transportzone-702", "segmentation_id": 702, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap764f6a9d-e8", "ovs_interfaceid": "764f6a9d-e89a-44b6-804c-70000c555f95", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1162.501704] env[63515]: DEBUG oslo_vmware.api [None req-8edc157d-485a-48e3-9c94-dc21771fbb70 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52e1b942-7ca1-0e50-dbe4-a14bf24a9bc8, 'name': SearchDatastore_Task, 'duration_secs': 0.011286} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1162.502181] env[63515]: DEBUG oslo_concurrency.lockutils [None req-8edc157d-485a-48e3-9c94-dc21771fbb70 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1162.502237] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-8edc157d-485a-48e3-9c94-dc21771fbb70 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] [instance: bd3a77ee-a78b-4719-9353-307ae8b166b6] Processing image 8a120570-cb06-4099-b262-554ca0ad15c5 {{(pid=63515) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1162.502448] env[63515]: DEBUG oslo_concurrency.lockutils [None req-8edc157d-485a-48e3-9c94-dc21771fbb70 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1162.502598] env[63515]: DEBUG oslo_concurrency.lockutils [None req-8edc157d-485a-48e3-9c94-dc21771fbb70 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Acquired lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1162.502782] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-8edc157d-485a-48e3-9c94-dc21771fbb70 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1162.503077] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-321b851b-2e08-4843-92e0-f0e7a3e99549 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.511318] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-8edc157d-485a-48e3-9c94-dc21771fbb70 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1162.511503] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-8edc157d-485a-48e3-9c94-dc21771fbb70 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63515) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1162.512532] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ee32b152-e772-404e-b112-c2bb1b0df03a {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.517995] env[63515]: DEBUG oslo_vmware.api [None req-8edc157d-485a-48e3-9c94-dc21771fbb70 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Waiting for the task: (returnval){ [ 1162.517995] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]52970c90-1940-aec7-568e-d3f8e7a67bcc" [ 1162.517995] env[63515]: _type = "Task" [ 1162.517995] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1162.525336] env[63515]: DEBUG oslo_vmware.api [None req-8edc157d-485a-48e3-9c94-dc21771fbb70 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52970c90-1940-aec7-568e-d3f8e7a67bcc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1162.888646] env[63515]: DEBUG nova.scheduler.client.report [None req-b86bd653-3ed8-48a7-a3cc-0c8e7c3ff20e tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1162.940178] env[63515]: DEBUG oslo_concurrency.lockutils [req-2145ad35-85f9-41a1-b6b4-e4102c7eaa5b req-15619a96-e991-48a7-8267-8e17ee42ee26 service nova] Releasing lock "refresh_cache-bd3a77ee-a78b-4719-9353-307ae8b166b6" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1163.032057] env[63515]: DEBUG oslo_vmware.api [None req-8edc157d-485a-48e3-9c94-dc21771fbb70 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52970c90-1940-aec7-568e-d3f8e7a67bcc, 'name': SearchDatastore_Task, 'duration_secs': 0.009777} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1163.033158] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-07ce71f8-dad8-4d5e-9c23-65bf72ea0aab {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.040110] env[63515]: DEBUG oslo_vmware.api [None req-8edc157d-485a-48e3-9c94-dc21771fbb70 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Waiting for the task: (returnval){ [ 1163.040110] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]52e1bda1-590d-9410-4514-97c02f92a813" [ 1163.040110] env[63515]: _type = "Task" [ 1163.040110] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1163.050780] env[63515]: DEBUG oslo_vmware.api [None req-8edc157d-485a-48e3-9c94-dc21771fbb70 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52e1bda1-590d-9410-4514-97c02f92a813, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1163.549932] env[63515]: DEBUG oslo_vmware.api [None req-8edc157d-485a-48e3-9c94-dc21771fbb70 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52e1bda1-590d-9410-4514-97c02f92a813, 'name': SearchDatastore_Task, 'duration_secs': 0.013111} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1163.550330] env[63515]: DEBUG oslo_concurrency.lockutils [None req-8edc157d-485a-48e3-9c94-dc21771fbb70 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Releasing lock "[datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1163.550482] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-8edc157d-485a-48e3-9c94-dc21771fbb70 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk to [datastore2] bd3a77ee-a78b-4719-9353-307ae8b166b6/bd3a77ee-a78b-4719-9353-307ae8b166b6.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1163.550709] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-9073476b-fa7b-4859-b80e-b88e5cf2954a {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.557560] env[63515]: DEBUG oslo_vmware.api [None req-8edc157d-485a-48e3-9c94-dc21771fbb70 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Waiting for the task: (returnval){ [ 1163.557560] env[63515]: value = "task-1112026" [ 1163.557560] env[63515]: _type = "Task" [ 1163.557560] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1163.564845] env[63515]: DEBUG oslo_vmware.api [None req-8edc157d-485a-48e3-9c94-dc21771fbb70 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Task: {'id': task-1112026, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1163.900157] env[63515]: DEBUG oslo_concurrency.lockutils [None req-b86bd653-3ed8-48a7-a3cc-0c8e7c3ff20e tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.192s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1163.993957] env[63515]: DEBUG oslo_vmware.rw_handles [None req-2f3128dc-4680-45d2-84d9-34c83ca4e4b1 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52a6b98c-dd1e-547a-6e19-d80ce28fc3e4/disk-0.vmdk. {{(pid=63515) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1163.995020] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43fc0c80-6558-47e8-8d71-694d0303eaa4 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1164.004650] env[63515]: DEBUG oslo_vmware.rw_handles [None req-2f3128dc-4680-45d2-84d9-34c83ca4e4b1 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52a6b98c-dd1e-547a-6e19-d80ce28fc3e4/disk-0.vmdk is in state: ready. {{(pid=63515) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1164.004954] env[63515]: ERROR oslo_vmware.rw_handles [None req-2f3128dc-4680-45d2-84d9-34c83ca4e4b1 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52a6b98c-dd1e-547a-6e19-d80ce28fc3e4/disk-0.vmdk due to incomplete transfer. [ 1164.005329] env[63515]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-2181b325-90a3-432c-9170-8b58e1fbabaf {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1164.013803] env[63515]: DEBUG oslo_vmware.rw_handles [None req-2f3128dc-4680-45d2-84d9-34c83ca4e4b1 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52a6b98c-dd1e-547a-6e19-d80ce28fc3e4/disk-0.vmdk. {{(pid=63515) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 1164.014077] env[63515]: DEBUG nova.virt.vmwareapi.images [None req-2f3128dc-4680-45d2-84d9-34c83ca4e4b1 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] [instance: c9e07a6c-fd14-4071-8c69-6ece62e9109a] Uploaded image bea6df9b-0878-417c-8cb9-f77d080de2d6 to the Glance image server {{(pid=63515) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 1164.016768] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-2f3128dc-4680-45d2-84d9-34c83ca4e4b1 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] [instance: c9e07a6c-fd14-4071-8c69-6ece62e9109a] Destroying the VM {{(pid=63515) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 1164.017195] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-747e5839-2dcb-42ac-8c1c-ea35c4df668f {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1164.027262] env[63515]: DEBUG oslo_vmware.api [None req-2f3128dc-4680-45d2-84d9-34c83ca4e4b1 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Waiting for the task: (returnval){ [ 1164.027262] env[63515]: value = "task-1112027" [ 1164.027262] env[63515]: _type = "Task" [ 1164.027262] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1164.035819] env[63515]: DEBUG oslo_vmware.api [None req-2f3128dc-4680-45d2-84d9-34c83ca4e4b1 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Task: {'id': task-1112027, 'name': Destroy_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1164.066439] env[63515]: DEBUG oslo_vmware.api [None req-8edc157d-485a-48e3-9c94-dc21771fbb70 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Task: {'id': task-1112026, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.498847} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1164.066688] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-8edc157d-485a-48e3-9c94-dc21771fbb70 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk to [datastore2] bd3a77ee-a78b-4719-9353-307ae8b166b6/bd3a77ee-a78b-4719-9353-307ae8b166b6.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1164.067803] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-8edc157d-485a-48e3-9c94-dc21771fbb70 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] [instance: bd3a77ee-a78b-4719-9353-307ae8b166b6] Extending root virtual disk to 1048576 {{(pid=63515) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1164.067803] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-e6c6fbbe-7a88-4d78-a326-914e00cd0f5d {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1164.074215] env[63515]: DEBUG oslo_vmware.api [None req-8edc157d-485a-48e3-9c94-dc21771fbb70 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Waiting for the task: (returnval){ [ 1164.074215] env[63515]: value = "task-1112028" [ 1164.074215] env[63515]: _type = "Task" [ 1164.074215] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1164.083402] env[63515]: DEBUG oslo_vmware.api [None req-8edc157d-485a-48e3-9c94-dc21771fbb70 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Task: {'id': task-1112028, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1164.229284] env[63515]: INFO nova.compute.manager [None req-40a36ff3-c179-44d6-bea8-756041846789 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: 9a864a51-7df3-4afc-8be7-2794ef2dcf1a] Get console output [ 1164.229691] env[63515]: WARNING nova.virt.vmwareapi.driver [None req-40a36ff3-c179-44d6-bea8-756041846789 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: 9a864a51-7df3-4afc-8be7-2794ef2dcf1a] The console log is missing. Check your VSPC configuration [ 1164.457192] env[63515]: INFO nova.scheduler.client.report [None req-b86bd653-3ed8-48a7-a3cc-0c8e7c3ff20e tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Deleted allocation for migration 2ac80a52-5354-43ed-b08a-87a7e019e350 [ 1164.537797] env[63515]: DEBUG oslo_vmware.api [None req-2f3128dc-4680-45d2-84d9-34c83ca4e4b1 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Task: {'id': task-1112027, 'name': Destroy_Task, 'duration_secs': 0.364081} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1164.537797] env[63515]: INFO nova.virt.vmwareapi.vm_util [None req-2f3128dc-4680-45d2-84d9-34c83ca4e4b1 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] [instance: c9e07a6c-fd14-4071-8c69-6ece62e9109a] Destroyed the VM [ 1164.537797] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-2f3128dc-4680-45d2-84d9-34c83ca4e4b1 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] [instance: c9e07a6c-fd14-4071-8c69-6ece62e9109a] Deleting Snapshot of the VM instance {{(pid=63515) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1164.538098] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-93d66439-e35d-4f56-9c7a-54d94323c0c1 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1164.545528] env[63515]: DEBUG oslo_vmware.api [None req-2f3128dc-4680-45d2-84d9-34c83ca4e4b1 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Waiting for the task: (returnval){ [ 1164.545528] env[63515]: value = "task-1112029" [ 1164.545528] env[63515]: _type = "Task" [ 1164.545528] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1164.553069] env[63515]: DEBUG oslo_vmware.api [None req-2f3128dc-4680-45d2-84d9-34c83ca4e4b1 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Task: {'id': task-1112029, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1164.583472] env[63515]: DEBUG oslo_vmware.api [None req-8edc157d-485a-48e3-9c94-dc21771fbb70 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Task: {'id': task-1112028, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.06178} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1164.583754] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-8edc157d-485a-48e3-9c94-dc21771fbb70 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] [instance: bd3a77ee-a78b-4719-9353-307ae8b166b6] Extended root virtual disk {{(pid=63515) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1164.584600] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-176a4354-0f7f-4777-9c71-084ba3a218bc {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1164.606729] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-8edc157d-485a-48e3-9c94-dc21771fbb70 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] [instance: bd3a77ee-a78b-4719-9353-307ae8b166b6] Reconfiguring VM instance instance-00000070 to attach disk [datastore2] bd3a77ee-a78b-4719-9353-307ae8b166b6/bd3a77ee-a78b-4719-9353-307ae8b166b6.vmdk or device None with type sparse {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1164.607039] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2d6f8b0f-547e-4658-887e-36c9dd9d11c2 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1164.626424] env[63515]: DEBUG oslo_vmware.api [None req-8edc157d-485a-48e3-9c94-dc21771fbb70 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Waiting for the task: (returnval){ [ 1164.626424] env[63515]: value = "task-1112030" [ 1164.626424] env[63515]: _type = "Task" [ 1164.626424] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1164.634021] env[63515]: DEBUG oslo_vmware.api [None req-8edc157d-485a-48e3-9c94-dc21771fbb70 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Task: {'id': task-1112030, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1164.962907] env[63515]: DEBUG oslo_concurrency.lockutils [None req-b86bd653-3ed8-48a7-a3cc-0c8e7c3ff20e tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Lock "9a864a51-7df3-4afc-8be7-2794ef2dcf1a" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 6.138s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1165.057083] env[63515]: DEBUG oslo_vmware.api [None req-2f3128dc-4680-45d2-84d9-34c83ca4e4b1 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Task: {'id': task-1112029, 'name': RemoveSnapshot_Task, 'duration_secs': 0.368941} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1165.057383] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-2f3128dc-4680-45d2-84d9-34c83ca4e4b1 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] [instance: c9e07a6c-fd14-4071-8c69-6ece62e9109a] Deleted Snapshot of the VM instance {{(pid=63515) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1165.057666] env[63515]: DEBUG nova.compute.manager [None req-2f3128dc-4680-45d2-84d9-34c83ca4e4b1 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] [instance: c9e07a6c-fd14-4071-8c69-6ece62e9109a] Checking state {{(pid=63515) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1165.058442] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d16ae7f3-9255-4779-ac2d-e30337e51dec {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.137110] env[63515]: DEBUG oslo_vmware.api [None req-8edc157d-485a-48e3-9c94-dc21771fbb70 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Task: {'id': task-1112030, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1165.570349] env[63515]: INFO nova.compute.manager [None req-2f3128dc-4680-45d2-84d9-34c83ca4e4b1 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] [instance: c9e07a6c-fd14-4071-8c69-6ece62e9109a] Shelve offloading [ 1165.572015] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-2f3128dc-4680-45d2-84d9-34c83ca4e4b1 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] [instance: c9e07a6c-fd14-4071-8c69-6ece62e9109a] Powering off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1165.572271] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8f70c30f-9b4b-4f0f-9769-c3c514b44c8f {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.578854] env[63515]: DEBUG oslo_vmware.api [None req-2f3128dc-4680-45d2-84d9-34c83ca4e4b1 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Waiting for the task: (returnval){ [ 1165.578854] env[63515]: value = "task-1112031" [ 1165.578854] env[63515]: _type = "Task" [ 1165.578854] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1165.586515] env[63515]: DEBUG oslo_vmware.api [None req-2f3128dc-4680-45d2-84d9-34c83ca4e4b1 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Task: {'id': task-1112031, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1165.636527] env[63515]: DEBUG oslo_vmware.api [None req-8edc157d-485a-48e3-9c94-dc21771fbb70 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Task: {'id': task-1112030, 'name': ReconfigVM_Task, 'duration_secs': 0.89062} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1165.636870] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-8edc157d-485a-48e3-9c94-dc21771fbb70 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] [instance: bd3a77ee-a78b-4719-9353-307ae8b166b6] Reconfigured VM instance instance-00000070 to attach disk [datastore2] bd3a77ee-a78b-4719-9353-307ae8b166b6/bd3a77ee-a78b-4719-9353-307ae8b166b6.vmdk or device None with type sparse {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1165.637500] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-36ac4513-a12c-4a93-9b4d-e7fce1d34900 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.643030] env[63515]: DEBUG oslo_vmware.api [None req-8edc157d-485a-48e3-9c94-dc21771fbb70 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Waiting for the task: (returnval){ [ 1165.643030] env[63515]: value = "task-1112032" [ 1165.643030] env[63515]: _type = "Task" [ 1165.643030] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1165.650015] env[63515]: DEBUG oslo_vmware.api [None req-8edc157d-485a-48e3-9c94-dc21771fbb70 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Task: {'id': task-1112032, 'name': Rename_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1166.089207] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-2f3128dc-4680-45d2-84d9-34c83ca4e4b1 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] [instance: c9e07a6c-fd14-4071-8c69-6ece62e9109a] VM already powered off {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1166.089384] env[63515]: DEBUG nova.compute.manager [None req-2f3128dc-4680-45d2-84d9-34c83ca4e4b1 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] [instance: c9e07a6c-fd14-4071-8c69-6ece62e9109a] Checking state {{(pid=63515) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1166.090168] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79ebe1c8-018c-427b-8856-67f46366292e {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.096900] env[63515]: DEBUG oslo_concurrency.lockutils [None req-2f3128dc-4680-45d2-84d9-34c83ca4e4b1 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Acquiring lock "refresh_cache-c9e07a6c-fd14-4071-8c69-6ece62e9109a" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1166.097084] env[63515]: DEBUG oslo_concurrency.lockutils [None req-2f3128dc-4680-45d2-84d9-34c83ca4e4b1 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Acquired lock "refresh_cache-c9e07a6c-fd14-4071-8c69-6ece62e9109a" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1166.097260] env[63515]: DEBUG nova.network.neutron [None req-2f3128dc-4680-45d2-84d9-34c83ca4e4b1 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] [instance: c9e07a6c-fd14-4071-8c69-6ece62e9109a] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1166.151646] env[63515]: DEBUG oslo_vmware.api [None req-8edc157d-485a-48e3-9c94-dc21771fbb70 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Task: {'id': task-1112032, 'name': Rename_Task, 'duration_secs': 0.144283} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1166.151889] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-8edc157d-485a-48e3-9c94-dc21771fbb70 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] [instance: bd3a77ee-a78b-4719-9353-307ae8b166b6] Powering on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1166.152131] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-fe29c626-3086-4693-87a5-48742895a4c4 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.158217] env[63515]: DEBUG oslo_vmware.api [None req-8edc157d-485a-48e3-9c94-dc21771fbb70 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Waiting for the task: (returnval){ [ 1166.158217] env[63515]: value = "task-1112033" [ 1166.158217] env[63515]: _type = "Task" [ 1166.158217] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1166.165020] env[63515]: DEBUG oslo_vmware.api [None req-8edc157d-485a-48e3-9c94-dc21771fbb70 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Task: {'id': task-1112033, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1166.669707] env[63515]: DEBUG oslo_vmware.api [None req-8edc157d-485a-48e3-9c94-dc21771fbb70 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Task: {'id': task-1112033, 'name': PowerOnVM_Task, 'duration_secs': 0.440201} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1166.670028] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-8edc157d-485a-48e3-9c94-dc21771fbb70 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] [instance: bd3a77ee-a78b-4719-9353-307ae8b166b6] Powered on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1166.670191] env[63515]: INFO nova.compute.manager [None req-8edc157d-485a-48e3-9c94-dc21771fbb70 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] [instance: bd3a77ee-a78b-4719-9353-307ae8b166b6] Took 7.38 seconds to spawn the instance on the hypervisor. [ 1166.670373] env[63515]: DEBUG nova.compute.manager [None req-8edc157d-485a-48e3-9c94-dc21771fbb70 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] [instance: bd3a77ee-a78b-4719-9353-307ae8b166b6] Checking state {{(pid=63515) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1166.671136] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dacbba23-052f-44e9-a216-bb6299ba83fd {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.793674] env[63515]: DEBUG nova.network.neutron [None req-2f3128dc-4680-45d2-84d9-34c83ca4e4b1 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] [instance: c9e07a6c-fd14-4071-8c69-6ece62e9109a] Updating instance_info_cache with network_info: [{"id": "bd316695-ec41-4f7d-b042-5cd5d319aa55", "address": "fa:16:3e:da:4b:8e", "network": {"id": "0e23b3a4-91ea-4bbe-9732-a3539d56b382", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1727906976-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fe873600ed6d4a3eb1ee3d3a19fceaf7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c7d2575f-b92f-44ec-a863-634cb76631a2", "external-id": "nsx-vlan-transportzone-794", "segmentation_id": 794, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbd316695-ec", "ovs_interfaceid": "bd316695-ec41-4f7d-b042-5cd5d319aa55", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1167.189562] env[63515]: INFO nova.compute.manager [None req-8edc157d-485a-48e3-9c94-dc21771fbb70 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] [instance: bd3a77ee-a78b-4719-9353-307ae8b166b6] Took 12.18 seconds to build instance. [ 1167.296060] env[63515]: DEBUG oslo_concurrency.lockutils [None req-2f3128dc-4680-45d2-84d9-34c83ca4e4b1 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Releasing lock "refresh_cache-c9e07a6c-fd14-4071-8c69-6ece62e9109a" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1167.691738] env[63515]: DEBUG oslo_concurrency.lockutils [None req-8edc157d-485a-48e3-9c94-dc21771fbb70 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Lock "bd3a77ee-a78b-4719-9353-307ae8b166b6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.691s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1167.745034] env[63515]: DEBUG nova.compute.manager [req-f7877240-582a-486a-8857-0d874e47123e req-8f96160a-83fe-45fe-91c9-174b6f6e7402 service nova] [instance: c9e07a6c-fd14-4071-8c69-6ece62e9109a] Received event network-vif-unplugged-bd316695-ec41-4f7d-b042-5cd5d319aa55 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1167.745034] env[63515]: DEBUG oslo_concurrency.lockutils [req-f7877240-582a-486a-8857-0d874e47123e req-8f96160a-83fe-45fe-91c9-174b6f6e7402 service nova] Acquiring lock "c9e07a6c-fd14-4071-8c69-6ece62e9109a-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1167.745034] env[63515]: DEBUG oslo_concurrency.lockutils [req-f7877240-582a-486a-8857-0d874e47123e req-8f96160a-83fe-45fe-91c9-174b6f6e7402 service nova] Lock "c9e07a6c-fd14-4071-8c69-6ece62e9109a-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1167.745034] env[63515]: DEBUG oslo_concurrency.lockutils [req-f7877240-582a-486a-8857-0d874e47123e req-8f96160a-83fe-45fe-91c9-174b6f6e7402 service nova] Lock "c9e07a6c-fd14-4071-8c69-6ece62e9109a-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1167.745034] env[63515]: DEBUG nova.compute.manager [req-f7877240-582a-486a-8857-0d874e47123e req-8f96160a-83fe-45fe-91c9-174b6f6e7402 service nova] [instance: c9e07a6c-fd14-4071-8c69-6ece62e9109a] No waiting events found dispatching network-vif-unplugged-bd316695-ec41-4f7d-b042-5cd5d319aa55 {{(pid=63515) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1167.745034] env[63515]: WARNING nova.compute.manager [req-f7877240-582a-486a-8857-0d874e47123e req-8f96160a-83fe-45fe-91c9-174b6f6e7402 service nova] [instance: c9e07a6c-fd14-4071-8c69-6ece62e9109a] Received unexpected event network-vif-unplugged-bd316695-ec41-4f7d-b042-5cd5d319aa55 for instance with vm_state shelved and task_state shelving_offloading. [ 1167.762343] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-2f3128dc-4680-45d2-84d9-34c83ca4e4b1 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] [instance: c9e07a6c-fd14-4071-8c69-6ece62e9109a] Destroying instance {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1167.763337] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9c7f0c4-a60a-409d-817e-cbed7c035e86 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1167.771743] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-2f3128dc-4680-45d2-84d9-34c83ca4e4b1 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] [instance: c9e07a6c-fd14-4071-8c69-6ece62e9109a] Unregistering the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1167.772036] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-509d8543-2ef8-44d5-9c6e-4a8e9bc5f4c6 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1168.123838] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-2f3128dc-4680-45d2-84d9-34c83ca4e4b1 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] [instance: c9e07a6c-fd14-4071-8c69-6ece62e9109a] Unregistered the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1168.124103] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-2f3128dc-4680-45d2-84d9-34c83ca4e4b1 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] [instance: c9e07a6c-fd14-4071-8c69-6ece62e9109a] Deleting contents of the VM from datastore datastore1 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1168.124301] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-2f3128dc-4680-45d2-84d9-34c83ca4e4b1 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Deleting the datastore file [datastore1] c9e07a6c-fd14-4071-8c69-6ece62e9109a {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1168.124579] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7b4babde-d352-4a18-bb63-a1a3ac9d75f9 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1168.130723] env[63515]: DEBUG oslo_vmware.api [None req-2f3128dc-4680-45d2-84d9-34c83ca4e4b1 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Waiting for the task: (returnval){ [ 1168.130723] env[63515]: value = "task-1112035" [ 1168.130723] env[63515]: _type = "Task" [ 1168.130723] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1168.138412] env[63515]: DEBUG oslo_vmware.api [None req-2f3128dc-4680-45d2-84d9-34c83ca4e4b1 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Task: {'id': task-1112035, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1168.640188] env[63515]: DEBUG oslo_vmware.api [None req-2f3128dc-4680-45d2-84d9-34c83ca4e4b1 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Task: {'id': task-1112035, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.138487} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1168.640449] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-2f3128dc-4680-45d2-84d9-34c83ca4e4b1 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Deleted the datastore file {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1168.640638] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-2f3128dc-4680-45d2-84d9-34c83ca4e4b1 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] [instance: c9e07a6c-fd14-4071-8c69-6ece62e9109a] Deleted contents of the VM from datastore datastore1 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1168.640819] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-2f3128dc-4680-45d2-84d9-34c83ca4e4b1 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] [instance: c9e07a6c-fd14-4071-8c69-6ece62e9109a] Instance destroyed {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1168.663204] env[63515]: INFO nova.scheduler.client.report [None req-2f3128dc-4680-45d2-84d9-34c83ca4e4b1 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Deleted allocations for instance c9e07a6c-fd14-4071-8c69-6ece62e9109a [ 1169.167663] env[63515]: DEBUG oslo_concurrency.lockutils [None req-2f3128dc-4680-45d2-84d9-34c83ca4e4b1 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1169.168157] env[63515]: DEBUG oslo_concurrency.lockutils [None req-2f3128dc-4680-45d2-84d9-34c83ca4e4b1 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1169.168219] env[63515]: DEBUG nova.objects.instance [None req-2f3128dc-4680-45d2-84d9-34c83ca4e4b1 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Lazy-loading 'resources' on Instance uuid c9e07a6c-fd14-4071-8c69-6ece62e9109a {{(pid=63515) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1169.671535] env[63515]: DEBUG nova.objects.instance [None req-2f3128dc-4680-45d2-84d9-34c83ca4e4b1 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Lazy-loading 'numa_topology' on Instance uuid c9e07a6c-fd14-4071-8c69-6ece62e9109a {{(pid=63515) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1169.774518] env[63515]: DEBUG nova.compute.manager [req-cad3bd0a-0ae5-432b-a0a3-31187e38b1c7 req-668bec36-de1d-4522-a5d6-4df2cd2fd46f service nova] [instance: c9e07a6c-fd14-4071-8c69-6ece62e9109a] Received event network-changed-bd316695-ec41-4f7d-b042-5cd5d319aa55 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1169.774730] env[63515]: DEBUG nova.compute.manager [req-cad3bd0a-0ae5-432b-a0a3-31187e38b1c7 req-668bec36-de1d-4522-a5d6-4df2cd2fd46f service nova] [instance: c9e07a6c-fd14-4071-8c69-6ece62e9109a] Refreshing instance network info cache due to event network-changed-bd316695-ec41-4f7d-b042-5cd5d319aa55. {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1169.774965] env[63515]: DEBUG oslo_concurrency.lockutils [req-cad3bd0a-0ae5-432b-a0a3-31187e38b1c7 req-668bec36-de1d-4522-a5d6-4df2cd2fd46f service nova] Acquiring lock "refresh_cache-c9e07a6c-fd14-4071-8c69-6ece62e9109a" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1169.775134] env[63515]: DEBUG oslo_concurrency.lockutils [req-cad3bd0a-0ae5-432b-a0a3-31187e38b1c7 req-668bec36-de1d-4522-a5d6-4df2cd2fd46f service nova] Acquired lock "refresh_cache-c9e07a6c-fd14-4071-8c69-6ece62e9109a" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1169.775305] env[63515]: DEBUG nova.network.neutron [req-cad3bd0a-0ae5-432b-a0a3-31187e38b1c7 req-668bec36-de1d-4522-a5d6-4df2cd2fd46f service nova] [instance: c9e07a6c-fd14-4071-8c69-6ece62e9109a] Refreshing network info cache for port bd316695-ec41-4f7d-b042-5cd5d319aa55 {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1170.174147] env[63515]: DEBUG nova.objects.base [None req-2f3128dc-4680-45d2-84d9-34c83ca4e4b1 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Object Instance lazy-loaded attributes: resources,numa_topology {{(pid=63515) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1170.255301] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8194a76-5243-41ab-b660-4191adc974f5 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.264586] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4bfd8621-7e8f-4995-9af4-30a55a22d9e7 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.295935] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5808c9b-ef19-45b2-8fa4-1f31914e31fc {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.303103] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fbca4999-472f-4f49-baac-d0c729c4dd8d {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.315984] env[63515]: DEBUG nova.compute.provider_tree [None req-2f3128dc-4680-45d2-84d9-34c83ca4e4b1 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1170.493086] env[63515]: DEBUG nova.network.neutron [req-cad3bd0a-0ae5-432b-a0a3-31187e38b1c7 req-668bec36-de1d-4522-a5d6-4df2cd2fd46f service nova] [instance: c9e07a6c-fd14-4071-8c69-6ece62e9109a] Updated VIF entry in instance network info cache for port bd316695-ec41-4f7d-b042-5cd5d319aa55. {{(pid=63515) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1170.493454] env[63515]: DEBUG nova.network.neutron [req-cad3bd0a-0ae5-432b-a0a3-31187e38b1c7 req-668bec36-de1d-4522-a5d6-4df2cd2fd46f service nova] [instance: c9e07a6c-fd14-4071-8c69-6ece62e9109a] Updating instance_info_cache with network_info: [{"id": "bd316695-ec41-4f7d-b042-5cd5d319aa55", "address": "fa:16:3e:da:4b:8e", "network": {"id": "0e23b3a4-91ea-4bbe-9732-a3539d56b382", "bridge": null, "label": "tempest-ServersNegativeTestJSON-1727906976-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fe873600ed6d4a3eb1ee3d3a19fceaf7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tapbd316695-ec", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1170.818824] env[63515]: DEBUG nova.scheduler.client.report [None req-2f3128dc-4680-45d2-84d9-34c83ca4e4b1 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1170.996124] env[63515]: DEBUG oslo_concurrency.lockutils [req-cad3bd0a-0ae5-432b-a0a3-31187e38b1c7 req-668bec36-de1d-4522-a5d6-4df2cd2fd46f service nova] Releasing lock "refresh_cache-c9e07a6c-fd14-4071-8c69-6ece62e9109a" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1170.996451] env[63515]: DEBUG nova.compute.manager [req-cad3bd0a-0ae5-432b-a0a3-31187e38b1c7 req-668bec36-de1d-4522-a5d6-4df2cd2fd46f service nova] [instance: bd3a77ee-a78b-4719-9353-307ae8b166b6] Received event network-changed-764f6a9d-e89a-44b6-804c-70000c555f95 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1170.996627] env[63515]: DEBUG nova.compute.manager [req-cad3bd0a-0ae5-432b-a0a3-31187e38b1c7 req-668bec36-de1d-4522-a5d6-4df2cd2fd46f service nova] [instance: bd3a77ee-a78b-4719-9353-307ae8b166b6] Refreshing instance network info cache due to event network-changed-764f6a9d-e89a-44b6-804c-70000c555f95. {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1170.996846] env[63515]: DEBUG oslo_concurrency.lockutils [req-cad3bd0a-0ae5-432b-a0a3-31187e38b1c7 req-668bec36-de1d-4522-a5d6-4df2cd2fd46f service nova] Acquiring lock "refresh_cache-bd3a77ee-a78b-4719-9353-307ae8b166b6" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1170.996990] env[63515]: DEBUG oslo_concurrency.lockutils [req-cad3bd0a-0ae5-432b-a0a3-31187e38b1c7 req-668bec36-de1d-4522-a5d6-4df2cd2fd46f service nova] Acquired lock "refresh_cache-bd3a77ee-a78b-4719-9353-307ae8b166b6" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1170.997245] env[63515]: DEBUG nova.network.neutron [req-cad3bd0a-0ae5-432b-a0a3-31187e38b1c7 req-668bec36-de1d-4522-a5d6-4df2cd2fd46f service nova] [instance: bd3a77ee-a78b-4719-9353-307ae8b166b6] Refreshing network info cache for port 764f6a9d-e89a-44b6-804c-70000c555f95 {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1171.289562] env[63515]: DEBUG oslo_concurrency.lockutils [None req-6684c2a7-1d4f-492e-a50c-c770e1d1c230 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Acquiring lock "c9e07a6c-fd14-4071-8c69-6ece62e9109a" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1171.325083] env[63515]: DEBUG oslo_concurrency.lockutils [None req-2f3128dc-4680-45d2-84d9-34c83ca4e4b1 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.157s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1171.704654] env[63515]: DEBUG nova.network.neutron [req-cad3bd0a-0ae5-432b-a0a3-31187e38b1c7 req-668bec36-de1d-4522-a5d6-4df2cd2fd46f service nova] [instance: bd3a77ee-a78b-4719-9353-307ae8b166b6] Updated VIF entry in instance network info cache for port 764f6a9d-e89a-44b6-804c-70000c555f95. {{(pid=63515) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1171.705044] env[63515]: DEBUG nova.network.neutron [req-cad3bd0a-0ae5-432b-a0a3-31187e38b1c7 req-668bec36-de1d-4522-a5d6-4df2cd2fd46f service nova] [instance: bd3a77ee-a78b-4719-9353-307ae8b166b6] Updating instance_info_cache with network_info: [{"id": "764f6a9d-e89a-44b6-804c-70000c555f95", "address": "fa:16:3e:56:19:d2", "network": {"id": "d6da2650-9ba9-4ef6-8035-e58b635a64b2", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-391564968-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.251", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "97f2e119708746038f38359572c99438", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "043ca97b-0fca-4b54-8be6-027123fa76d0", "external-id": "nsx-vlan-transportzone-702", "segmentation_id": 702, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap764f6a9d-e8", "ovs_interfaceid": "764f6a9d-e89a-44b6-804c-70000c555f95", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1171.833305] env[63515]: DEBUG oslo_concurrency.lockutils [None req-2f3128dc-4680-45d2-84d9-34c83ca4e4b1 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Lock "c9e07a6c-fd14-4071-8c69-6ece62e9109a" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 21.002s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1171.834138] env[63515]: DEBUG oslo_concurrency.lockutils [None req-6684c2a7-1d4f-492e-a50c-c770e1d1c230 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Lock "c9e07a6c-fd14-4071-8c69-6ece62e9109a" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 0.545s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1171.834323] env[63515]: INFO nova.compute.manager [None req-6684c2a7-1d4f-492e-a50c-c770e1d1c230 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] [instance: c9e07a6c-fd14-4071-8c69-6ece62e9109a] Unshelving [ 1172.207350] env[63515]: DEBUG oslo_concurrency.lockutils [req-cad3bd0a-0ae5-432b-a0a3-31187e38b1c7 req-668bec36-de1d-4522-a5d6-4df2cd2fd46f service nova] Releasing lock "refresh_cache-bd3a77ee-a78b-4719-9353-307ae8b166b6" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1172.858628] env[63515]: DEBUG oslo_concurrency.lockutils [None req-6684c2a7-1d4f-492e-a50c-c770e1d1c230 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1172.858898] env[63515]: DEBUG oslo_concurrency.lockutils [None req-6684c2a7-1d4f-492e-a50c-c770e1d1c230 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1172.859140] env[63515]: DEBUG nova.objects.instance [None req-6684c2a7-1d4f-492e-a50c-c770e1d1c230 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Lazy-loading 'pci_requests' on Instance uuid c9e07a6c-fd14-4071-8c69-6ece62e9109a {{(pid=63515) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1173.362989] env[63515]: DEBUG nova.objects.instance [None req-6684c2a7-1d4f-492e-a50c-c770e1d1c230 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Lazy-loading 'numa_topology' on Instance uuid c9e07a6c-fd14-4071-8c69-6ece62e9109a {{(pid=63515) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1173.865533] env[63515]: INFO nova.compute.claims [None req-6684c2a7-1d4f-492e-a50c-c770e1d1c230 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] [instance: c9e07a6c-fd14-4071-8c69-6ece62e9109a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1174.958106] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4aa5d555-ba4a-4669-bf01-ec9da04149db {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1174.965543] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19bc8b29-59e4-4ca8-82fe-aad807df28b6 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1174.996923] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8024c05-14a7-46bd-a194-a96b962b8aa9 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1175.003897] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99b7645e-3211-43b2-8bb9-a4489e64bc43 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1175.016724] env[63515]: DEBUG nova.compute.provider_tree [None req-6684c2a7-1d4f-492e-a50c-c770e1d1c230 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1175.519874] env[63515]: DEBUG nova.scheduler.client.report [None req-6684c2a7-1d4f-492e-a50c-c770e1d1c230 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1176.025470] env[63515]: DEBUG oslo_concurrency.lockutils [None req-6684c2a7-1d4f-492e-a50c-c770e1d1c230 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.166s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1176.054429] env[63515]: INFO nova.network.neutron [None req-6684c2a7-1d4f-492e-a50c-c770e1d1c230 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] [instance: c9e07a6c-fd14-4071-8c69-6ece62e9109a] Updating port bd316695-ec41-4f7d-b042-5cd5d319aa55 with attributes {'binding:host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 1177.401891] env[63515]: DEBUG nova.compute.manager [req-693f386c-a7d5-481e-82ad-db608b37d005 req-4038dbce-0e0f-44b4-97c5-cc170e31da3b service nova] [instance: c9e07a6c-fd14-4071-8c69-6ece62e9109a] Received event network-vif-plugged-bd316695-ec41-4f7d-b042-5cd5d319aa55 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1177.402937] env[63515]: DEBUG oslo_concurrency.lockutils [req-693f386c-a7d5-481e-82ad-db608b37d005 req-4038dbce-0e0f-44b4-97c5-cc170e31da3b service nova] Acquiring lock "c9e07a6c-fd14-4071-8c69-6ece62e9109a-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1177.402937] env[63515]: DEBUG oslo_concurrency.lockutils [req-693f386c-a7d5-481e-82ad-db608b37d005 req-4038dbce-0e0f-44b4-97c5-cc170e31da3b service nova] Lock "c9e07a6c-fd14-4071-8c69-6ece62e9109a-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1177.402937] env[63515]: DEBUG oslo_concurrency.lockutils [req-693f386c-a7d5-481e-82ad-db608b37d005 req-4038dbce-0e0f-44b4-97c5-cc170e31da3b service nova] Lock "c9e07a6c-fd14-4071-8c69-6ece62e9109a-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1177.402937] env[63515]: DEBUG nova.compute.manager [req-693f386c-a7d5-481e-82ad-db608b37d005 req-4038dbce-0e0f-44b4-97c5-cc170e31da3b service nova] [instance: c9e07a6c-fd14-4071-8c69-6ece62e9109a] No waiting events found dispatching network-vif-plugged-bd316695-ec41-4f7d-b042-5cd5d319aa55 {{(pid=63515) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1177.402937] env[63515]: WARNING nova.compute.manager [req-693f386c-a7d5-481e-82ad-db608b37d005 req-4038dbce-0e0f-44b4-97c5-cc170e31da3b service nova] [instance: c9e07a6c-fd14-4071-8c69-6ece62e9109a] Received unexpected event network-vif-plugged-bd316695-ec41-4f7d-b042-5cd5d319aa55 for instance with vm_state shelved_offloaded and task_state spawning. [ 1177.478812] env[63515]: DEBUG oslo_concurrency.lockutils [None req-6684c2a7-1d4f-492e-a50c-c770e1d1c230 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Acquiring lock "refresh_cache-c9e07a6c-fd14-4071-8c69-6ece62e9109a" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1177.479030] env[63515]: DEBUG oslo_concurrency.lockutils [None req-6684c2a7-1d4f-492e-a50c-c770e1d1c230 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Acquired lock "refresh_cache-c9e07a6c-fd14-4071-8c69-6ece62e9109a" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1177.479262] env[63515]: DEBUG nova.network.neutron [None req-6684c2a7-1d4f-492e-a50c-c770e1d1c230 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] [instance: c9e07a6c-fd14-4071-8c69-6ece62e9109a] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1178.162284] env[63515]: DEBUG nova.network.neutron [None req-6684c2a7-1d4f-492e-a50c-c770e1d1c230 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] [instance: c9e07a6c-fd14-4071-8c69-6ece62e9109a] Updating instance_info_cache with network_info: [{"id": "bd316695-ec41-4f7d-b042-5cd5d319aa55", "address": "fa:16:3e:da:4b:8e", "network": {"id": "0e23b3a4-91ea-4bbe-9732-a3539d56b382", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1727906976-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fe873600ed6d4a3eb1ee3d3a19fceaf7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c7d2575f-b92f-44ec-a863-634cb76631a2", "external-id": "nsx-vlan-transportzone-794", "segmentation_id": 794, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbd316695-ec", "ovs_interfaceid": "bd316695-ec41-4f7d-b042-5cd5d319aa55", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1178.665359] env[63515]: DEBUG oslo_concurrency.lockutils [None req-6684c2a7-1d4f-492e-a50c-c770e1d1c230 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Releasing lock "refresh_cache-c9e07a6c-fd14-4071-8c69-6ece62e9109a" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1178.691599] env[63515]: DEBUG nova.virt.hardware [None req-6684c2a7-1d4f-492e-a50c-c770e1d1c230 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T02:52:14Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='9b82a9af615a189be90b058c30569070',container_format='bare',created_at=2024-10-03T03:04:20Z,direct_url=,disk_format='vmdk',id=bea6df9b-0878-417c-8cb9-f77d080de2d6,min_disk=1,min_ram=0,name='tempest-ServersNegativeTestJSON-server-889517681-shelved',owner='fe873600ed6d4a3eb1ee3d3a19fceaf7',properties=ImageMetaProps,protected=,size=31661056,status='active',tags=,updated_at=2024-10-03T03:04:34Z,virtual_size=,visibility=), allow threads: False {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1178.691860] env[63515]: DEBUG nova.virt.hardware [None req-6684c2a7-1d4f-492e-a50c-c770e1d1c230 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Flavor limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1178.692035] env[63515]: DEBUG nova.virt.hardware [None req-6684c2a7-1d4f-492e-a50c-c770e1d1c230 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Image limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1178.692230] env[63515]: DEBUG nova.virt.hardware [None req-6684c2a7-1d4f-492e-a50c-c770e1d1c230 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Flavor pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1178.692380] env[63515]: DEBUG nova.virt.hardware [None req-6684c2a7-1d4f-492e-a50c-c770e1d1c230 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Image pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1178.692529] env[63515]: DEBUG nova.virt.hardware [None req-6684c2a7-1d4f-492e-a50c-c770e1d1c230 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1178.692736] env[63515]: DEBUG nova.virt.hardware [None req-6684c2a7-1d4f-492e-a50c-c770e1d1c230 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1178.692898] env[63515]: DEBUG nova.virt.hardware [None req-6684c2a7-1d4f-492e-a50c-c770e1d1c230 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1178.693078] env[63515]: DEBUG nova.virt.hardware [None req-6684c2a7-1d4f-492e-a50c-c770e1d1c230 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Got 1 possible topologies {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1178.693250] env[63515]: DEBUG nova.virt.hardware [None req-6684c2a7-1d4f-492e-a50c-c770e1d1c230 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1178.693429] env[63515]: DEBUG nova.virt.hardware [None req-6684c2a7-1d4f-492e-a50c-c770e1d1c230 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1178.694360] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3498df8-d60d-44ac-b7dd-8d5a037583e5 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1178.702045] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ff80384-3a85-459c-aae5-d4666d39f8f4 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1178.715928] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-6684c2a7-1d4f-492e-a50c-c770e1d1c230 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] [instance: c9e07a6c-fd14-4071-8c69-6ece62e9109a] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:da:4b:8e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c7d2575f-b92f-44ec-a863-634cb76631a2', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'bd316695-ec41-4f7d-b042-5cd5d319aa55', 'vif_model': 'vmxnet3'}] {{(pid=63515) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1178.723210] env[63515]: DEBUG oslo.service.loopingcall [None req-6684c2a7-1d4f-492e-a50c-c770e1d1c230 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1178.723441] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c9e07a6c-fd14-4071-8c69-6ece62e9109a] Creating VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1178.723642] env[63515]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-1a8ea689-9c7c-4cde-b381-d6c47d49070d {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1178.743045] env[63515]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1178.743045] env[63515]: value = "task-1112036" [ 1178.743045] env[63515]: _type = "Task" [ 1178.743045] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1178.750652] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1112036, 'name': CreateVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1179.253198] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1112036, 'name': CreateVM_Task, 'duration_secs': 0.315693} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1179.253419] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c9e07a6c-fd14-4071-8c69-6ece62e9109a] Created VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1179.254053] env[63515]: DEBUG oslo_concurrency.lockutils [None req-6684c2a7-1d4f-492e-a50c-c770e1d1c230 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/bea6df9b-0878-417c-8cb9-f77d080de2d6" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1179.254233] env[63515]: DEBUG oslo_concurrency.lockutils [None req-6684c2a7-1d4f-492e-a50c-c770e1d1c230 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Acquired lock "[datastore2] devstack-image-cache_base/bea6df9b-0878-417c-8cb9-f77d080de2d6" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1179.254611] env[63515]: DEBUG oslo_concurrency.lockutils [None req-6684c2a7-1d4f-492e-a50c-c770e1d1c230 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/bea6df9b-0878-417c-8cb9-f77d080de2d6" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1179.254870] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fcec0138-40e1-4246-880a-1d30206ac156 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1179.259323] env[63515]: DEBUG oslo_vmware.api [None req-6684c2a7-1d4f-492e-a50c-c770e1d1c230 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Waiting for the task: (returnval){ [ 1179.259323] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]52388d89-51d2-de3e-be85-60ac807f1a5f" [ 1179.259323] env[63515]: _type = "Task" [ 1179.259323] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1179.266644] env[63515]: DEBUG oslo_vmware.api [None req-6684c2a7-1d4f-492e-a50c-c770e1d1c230 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52388d89-51d2-de3e-be85-60ac807f1a5f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1179.431088] env[63515]: DEBUG nova.compute.manager [req-50d1b8e6-4ab4-440e-aa19-2ab49edb63b1 req-9ee74131-5849-45c9-9568-40309de6f321 service nova] [instance: c9e07a6c-fd14-4071-8c69-6ece62e9109a] Received event network-changed-bd316695-ec41-4f7d-b042-5cd5d319aa55 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1179.431310] env[63515]: DEBUG nova.compute.manager [req-50d1b8e6-4ab4-440e-aa19-2ab49edb63b1 req-9ee74131-5849-45c9-9568-40309de6f321 service nova] [instance: c9e07a6c-fd14-4071-8c69-6ece62e9109a] Refreshing instance network info cache due to event network-changed-bd316695-ec41-4f7d-b042-5cd5d319aa55. {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1179.431528] env[63515]: DEBUG oslo_concurrency.lockutils [req-50d1b8e6-4ab4-440e-aa19-2ab49edb63b1 req-9ee74131-5849-45c9-9568-40309de6f321 service nova] Acquiring lock "refresh_cache-c9e07a6c-fd14-4071-8c69-6ece62e9109a" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1179.431697] env[63515]: DEBUG oslo_concurrency.lockutils [req-50d1b8e6-4ab4-440e-aa19-2ab49edb63b1 req-9ee74131-5849-45c9-9568-40309de6f321 service nova] Acquired lock "refresh_cache-c9e07a6c-fd14-4071-8c69-6ece62e9109a" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1179.431866] env[63515]: DEBUG nova.network.neutron [req-50d1b8e6-4ab4-440e-aa19-2ab49edb63b1 req-9ee74131-5849-45c9-9568-40309de6f321 service nova] [instance: c9e07a6c-fd14-4071-8c69-6ece62e9109a] Refreshing network info cache for port bd316695-ec41-4f7d-b042-5cd5d319aa55 {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1179.769399] env[63515]: DEBUG oslo_concurrency.lockutils [None req-6684c2a7-1d4f-492e-a50c-c770e1d1c230 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Releasing lock "[datastore2] devstack-image-cache_base/bea6df9b-0878-417c-8cb9-f77d080de2d6" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1179.769803] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-6684c2a7-1d4f-492e-a50c-c770e1d1c230 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] [instance: c9e07a6c-fd14-4071-8c69-6ece62e9109a] Processing image bea6df9b-0878-417c-8cb9-f77d080de2d6 {{(pid=63515) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1179.769866] env[63515]: DEBUG oslo_concurrency.lockutils [None req-6684c2a7-1d4f-492e-a50c-c770e1d1c230 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/bea6df9b-0878-417c-8cb9-f77d080de2d6/bea6df9b-0878-417c-8cb9-f77d080de2d6.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1179.770017] env[63515]: DEBUG oslo_concurrency.lockutils [None req-6684c2a7-1d4f-492e-a50c-c770e1d1c230 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Acquired lock "[datastore2] devstack-image-cache_base/bea6df9b-0878-417c-8cb9-f77d080de2d6/bea6df9b-0878-417c-8cb9-f77d080de2d6.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1179.770211] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-6684c2a7-1d4f-492e-a50c-c770e1d1c230 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1179.770454] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-4ddf75be-bdd4-45e2-a662-65eb0367ae0b {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1179.787829] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-6684c2a7-1d4f-492e-a50c-c770e1d1c230 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1179.788027] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-6684c2a7-1d4f-492e-a50c-c770e1d1c230 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=63515) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1179.788719] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0d7faa80-18db-441d-83ad-1f23d190318d {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1179.793608] env[63515]: DEBUG oslo_vmware.api [None req-6684c2a7-1d4f-492e-a50c-c770e1d1c230 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Waiting for the task: (returnval){ [ 1179.793608] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]523a3ccb-83a9-32b2-e81e-b65f9057ddf3" [ 1179.793608] env[63515]: _type = "Task" [ 1179.793608] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1179.801537] env[63515]: DEBUG oslo_vmware.api [None req-6684c2a7-1d4f-492e-a50c-c770e1d1c230 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]523a3ccb-83a9-32b2-e81e-b65f9057ddf3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1180.114423] env[63515]: DEBUG nova.network.neutron [req-50d1b8e6-4ab4-440e-aa19-2ab49edb63b1 req-9ee74131-5849-45c9-9568-40309de6f321 service nova] [instance: c9e07a6c-fd14-4071-8c69-6ece62e9109a] Updated VIF entry in instance network info cache for port bd316695-ec41-4f7d-b042-5cd5d319aa55. {{(pid=63515) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1180.114784] env[63515]: DEBUG nova.network.neutron [req-50d1b8e6-4ab4-440e-aa19-2ab49edb63b1 req-9ee74131-5849-45c9-9568-40309de6f321 service nova] [instance: c9e07a6c-fd14-4071-8c69-6ece62e9109a] Updating instance_info_cache with network_info: [{"id": "bd316695-ec41-4f7d-b042-5cd5d319aa55", "address": "fa:16:3e:da:4b:8e", "network": {"id": "0e23b3a4-91ea-4bbe-9732-a3539d56b382", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1727906976-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fe873600ed6d4a3eb1ee3d3a19fceaf7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c7d2575f-b92f-44ec-a863-634cb76631a2", "external-id": "nsx-vlan-transportzone-794", "segmentation_id": 794, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbd316695-ec", "ovs_interfaceid": "bd316695-ec41-4f7d-b042-5cd5d319aa55", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1180.303991] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-6684c2a7-1d4f-492e-a50c-c770e1d1c230 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] [instance: c9e07a6c-fd14-4071-8c69-6ece62e9109a] Preparing fetch location {{(pid=63515) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 1180.304261] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-6684c2a7-1d4f-492e-a50c-c770e1d1c230 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] [instance: c9e07a6c-fd14-4071-8c69-6ece62e9109a] Fetch image to [datastore2] OSTACK_IMG_6c984334-e5c3-4dee-978d-9a69acf3c3a6/OSTACK_IMG_6c984334-e5c3-4dee-978d-9a69acf3c3a6.vmdk {{(pid=63515) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 1180.304455] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-6684c2a7-1d4f-492e-a50c-c770e1d1c230 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] [instance: c9e07a6c-fd14-4071-8c69-6ece62e9109a] Downloading stream optimized image bea6df9b-0878-417c-8cb9-f77d080de2d6 to [datastore2] OSTACK_IMG_6c984334-e5c3-4dee-978d-9a69acf3c3a6/OSTACK_IMG_6c984334-e5c3-4dee-978d-9a69acf3c3a6.vmdk on the data store datastore2 as vApp {{(pid=63515) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 1180.304631] env[63515]: DEBUG nova.virt.vmwareapi.images [None req-6684c2a7-1d4f-492e-a50c-c770e1d1c230 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] [instance: c9e07a6c-fd14-4071-8c69-6ece62e9109a] Downloading image file data bea6df9b-0878-417c-8cb9-f77d080de2d6 to the ESX as VM named 'OSTACK_IMG_6c984334-e5c3-4dee-978d-9a69acf3c3a6' {{(pid=63515) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 1180.369117] env[63515]: DEBUG oslo_vmware.rw_handles [None req-6684c2a7-1d4f-492e-a50c-c770e1d1c230 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 1180.369117] env[63515]: value = "resgroup-9" [ 1180.369117] env[63515]: _type = "ResourcePool" [ 1180.369117] env[63515]: }. {{(pid=63515) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 1180.369500] env[63515]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-a6312d25-783a-408a-bf2a-df0fcd82c754 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1180.390365] env[63515]: DEBUG oslo_vmware.rw_handles [None req-6684c2a7-1d4f-492e-a50c-c770e1d1c230 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Lease: (returnval){ [ 1180.390365] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]52a7d464-05f7-3350-769b-de68163f49a1" [ 1180.390365] env[63515]: _type = "HttpNfcLease" [ 1180.390365] env[63515]: } obtained for vApp import into resource pool (val){ [ 1180.390365] env[63515]: value = "resgroup-9" [ 1180.390365] env[63515]: _type = "ResourcePool" [ 1180.390365] env[63515]: }. {{(pid=63515) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 1180.390632] env[63515]: DEBUG oslo_vmware.api [None req-6684c2a7-1d4f-492e-a50c-c770e1d1c230 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Waiting for the lease: (returnval){ [ 1180.390632] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]52a7d464-05f7-3350-769b-de68163f49a1" [ 1180.390632] env[63515]: _type = "HttpNfcLease" [ 1180.390632] env[63515]: } to be ready. {{(pid=63515) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1180.399010] env[63515]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1180.399010] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]52a7d464-05f7-3350-769b-de68163f49a1" [ 1180.399010] env[63515]: _type = "HttpNfcLease" [ 1180.399010] env[63515]: } is initializing. {{(pid=63515) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1180.617744] env[63515]: DEBUG oslo_concurrency.lockutils [req-50d1b8e6-4ab4-440e-aa19-2ab49edb63b1 req-9ee74131-5849-45c9-9568-40309de6f321 service nova] Releasing lock "refresh_cache-c9e07a6c-fd14-4071-8c69-6ece62e9109a" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1180.898878] env[63515]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1180.898878] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]52a7d464-05f7-3350-769b-de68163f49a1" [ 1180.898878] env[63515]: _type = "HttpNfcLease" [ 1180.898878] env[63515]: } is ready. {{(pid=63515) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1180.898878] env[63515]: DEBUG oslo_vmware.rw_handles [None req-6684c2a7-1d4f-492e-a50c-c770e1d1c230 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1180.898878] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]52a7d464-05f7-3350-769b-de68163f49a1" [ 1180.898878] env[63515]: _type = "HttpNfcLease" [ 1180.898878] env[63515]: }. {{(pid=63515) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 1180.899625] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8d0546d-59ab-4214-beba-47fec18f51db {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1180.906507] env[63515]: DEBUG oslo_vmware.rw_handles [None req-6684c2a7-1d4f-492e-a50c-c770e1d1c230 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52380882-0bdc-fed7-21d2-a3cb1af6dabf/disk-0.vmdk from lease info. {{(pid=63515) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1180.906687] env[63515]: DEBUG oslo_vmware.rw_handles [None req-6684c2a7-1d4f-492e-a50c-c770e1d1c230 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Creating HTTP connection to write to file with size = 31661056 and URL = https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52380882-0bdc-fed7-21d2-a3cb1af6dabf/disk-0.vmdk. {{(pid=63515) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 1180.970352] env[63515]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-56bdb963-b742-4e53-9d79-e8de2ad3c12f {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1182.094886] env[63515]: DEBUG oslo_vmware.rw_handles [None req-6684c2a7-1d4f-492e-a50c-c770e1d1c230 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Completed reading data from the image iterator. {{(pid=63515) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 1182.095271] env[63515]: DEBUG oslo_vmware.rw_handles [None req-6684c2a7-1d4f-492e-a50c-c770e1d1c230 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52380882-0bdc-fed7-21d2-a3cb1af6dabf/disk-0.vmdk. {{(pid=63515) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1182.096156] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf32f322-94b0-45d8-b82c-04a0d8a80b3d {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1182.102983] env[63515]: DEBUG oslo_vmware.rw_handles [None req-6684c2a7-1d4f-492e-a50c-c770e1d1c230 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52380882-0bdc-fed7-21d2-a3cb1af6dabf/disk-0.vmdk is in state: ready. {{(pid=63515) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1182.103220] env[63515]: DEBUG oslo_vmware.rw_handles [None req-6684c2a7-1d4f-492e-a50c-c770e1d1c230 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Releasing lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52380882-0bdc-fed7-21d2-a3cb1af6dabf/disk-0.vmdk. {{(pid=63515) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 1182.103495] env[63515]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-39f03227-0ba0-464f-aa8f-303cbfd26381 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1182.365653] env[63515]: DEBUG oslo_vmware.rw_handles [None req-6684c2a7-1d4f-492e-a50c-c770e1d1c230 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Closed VMDK write handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52380882-0bdc-fed7-21d2-a3cb1af6dabf/disk-0.vmdk. {{(pid=63515) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 1182.365940] env[63515]: INFO nova.virt.vmwareapi.images [None req-6684c2a7-1d4f-492e-a50c-c770e1d1c230 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] [instance: c9e07a6c-fd14-4071-8c69-6ece62e9109a] Downloaded image file data bea6df9b-0878-417c-8cb9-f77d080de2d6 [ 1182.366810] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa579554-6093-4ad9-927d-abb89fabaa7d {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1182.383250] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-0e875872-fad5-4486-b5c0-b404f16f7227 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1182.432645] env[63515]: INFO nova.virt.vmwareapi.images [None req-6684c2a7-1d4f-492e-a50c-c770e1d1c230 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] [instance: c9e07a6c-fd14-4071-8c69-6ece62e9109a] The imported VM was unregistered [ 1182.434085] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-6684c2a7-1d4f-492e-a50c-c770e1d1c230 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] [instance: c9e07a6c-fd14-4071-8c69-6ece62e9109a] Caching image {{(pid=63515) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 1182.434312] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-6684c2a7-1d4f-492e-a50c-c770e1d1c230 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Creating directory with path [datastore2] devstack-image-cache_base/bea6df9b-0878-417c-8cb9-f77d080de2d6 {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1182.434578] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-131a89e2-68c7-4af7-8707-518e71647520 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1182.456777] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-6684c2a7-1d4f-492e-a50c-c770e1d1c230 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Created directory with path [datastore2] devstack-image-cache_base/bea6df9b-0878-417c-8cb9-f77d080de2d6 {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1182.457018] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-6684c2a7-1d4f-492e-a50c-c770e1d1c230 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Moving virtual disk from [datastore2] OSTACK_IMG_6c984334-e5c3-4dee-978d-9a69acf3c3a6/OSTACK_IMG_6c984334-e5c3-4dee-978d-9a69acf3c3a6.vmdk to [datastore2] devstack-image-cache_base/bea6df9b-0878-417c-8cb9-f77d080de2d6/bea6df9b-0878-417c-8cb9-f77d080de2d6.vmdk. {{(pid=63515) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 1182.458215] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-daebe767-3a08-4a9e-975d-2359f4cd3589 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1182.464186] env[63515]: DEBUG oslo_vmware.api [None req-6684c2a7-1d4f-492e-a50c-c770e1d1c230 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Waiting for the task: (returnval){ [ 1182.464186] env[63515]: value = "task-1112039" [ 1182.464186] env[63515]: _type = "Task" [ 1182.464186] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1182.471886] env[63515]: DEBUG oslo_vmware.api [None req-6684c2a7-1d4f-492e-a50c-c770e1d1c230 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Task: {'id': task-1112039, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1182.975372] env[63515]: DEBUG oslo_vmware.api [None req-6684c2a7-1d4f-492e-a50c-c770e1d1c230 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Task: {'id': task-1112039, 'name': MoveVirtualDisk_Task} progress is 12%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1183.476521] env[63515]: DEBUG oslo_vmware.api [None req-6684c2a7-1d4f-492e-a50c-c770e1d1c230 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Task: {'id': task-1112039, 'name': MoveVirtualDisk_Task} progress is 35%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1183.979141] env[63515]: DEBUG oslo_vmware.api [None req-6684c2a7-1d4f-492e-a50c-c770e1d1c230 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Task: {'id': task-1112039, 'name': MoveVirtualDisk_Task} progress is 57%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1184.478279] env[63515]: DEBUG oslo_vmware.api [None req-6684c2a7-1d4f-492e-a50c-c770e1d1c230 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Task: {'id': task-1112039, 'name': MoveVirtualDisk_Task} progress is 80%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1184.976589] env[63515]: DEBUG oslo_vmware.api [None req-6684c2a7-1d4f-492e-a50c-c770e1d1c230 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Task: {'id': task-1112039, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.494777} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1184.977679] env[63515]: INFO nova.virt.vmwareapi.ds_util [None req-6684c2a7-1d4f-492e-a50c-c770e1d1c230 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Moved virtual disk from [datastore2] OSTACK_IMG_6c984334-e5c3-4dee-978d-9a69acf3c3a6/OSTACK_IMG_6c984334-e5c3-4dee-978d-9a69acf3c3a6.vmdk to [datastore2] devstack-image-cache_base/bea6df9b-0878-417c-8cb9-f77d080de2d6/bea6df9b-0878-417c-8cb9-f77d080de2d6.vmdk. [ 1184.977679] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-6684c2a7-1d4f-492e-a50c-c770e1d1c230 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] [instance: c9e07a6c-fd14-4071-8c69-6ece62e9109a] Cleaning up location [datastore2] OSTACK_IMG_6c984334-e5c3-4dee-978d-9a69acf3c3a6 {{(pid=63515) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 1184.977679] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-6684c2a7-1d4f-492e-a50c-c770e1d1c230 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Deleting the datastore file [datastore2] OSTACK_IMG_6c984334-e5c3-4dee-978d-9a69acf3c3a6 {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1184.977679] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-9a0751e0-2d84-41b5-aca7-b51967f34aca {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1184.983685] env[63515]: DEBUG oslo_vmware.api [None req-6684c2a7-1d4f-492e-a50c-c770e1d1c230 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Waiting for the task: (returnval){ [ 1184.983685] env[63515]: value = "task-1112040" [ 1184.983685] env[63515]: _type = "Task" [ 1184.983685] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1184.991264] env[63515]: DEBUG oslo_vmware.api [None req-6684c2a7-1d4f-492e-a50c-c770e1d1c230 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Task: {'id': task-1112040, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1185.494114] env[63515]: DEBUG oslo_vmware.api [None req-6684c2a7-1d4f-492e-a50c-c770e1d1c230 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Task: {'id': task-1112040, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.050235} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1185.494471] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-6684c2a7-1d4f-492e-a50c-c770e1d1c230 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Deleted the datastore file {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1185.494471] env[63515]: DEBUG oslo_concurrency.lockutils [None req-6684c2a7-1d4f-492e-a50c-c770e1d1c230 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Releasing lock "[datastore2] devstack-image-cache_base/bea6df9b-0878-417c-8cb9-f77d080de2d6/bea6df9b-0878-417c-8cb9-f77d080de2d6.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1185.494662] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-6684c2a7-1d4f-492e-a50c-c770e1d1c230 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/bea6df9b-0878-417c-8cb9-f77d080de2d6/bea6df9b-0878-417c-8cb9-f77d080de2d6.vmdk to [datastore2] c9e07a6c-fd14-4071-8c69-6ece62e9109a/c9e07a6c-fd14-4071-8c69-6ece62e9109a.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1185.494927] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-aac9406a-b24e-45df-833d-321d26fef2a6 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1185.501877] env[63515]: DEBUG oslo_vmware.api [None req-6684c2a7-1d4f-492e-a50c-c770e1d1c230 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Waiting for the task: (returnval){ [ 1185.501877] env[63515]: value = "task-1112041" [ 1185.501877] env[63515]: _type = "Task" [ 1185.501877] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1185.509153] env[63515]: DEBUG oslo_vmware.api [None req-6684c2a7-1d4f-492e-a50c-c770e1d1c230 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Task: {'id': task-1112041, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1185.639015] env[63515]: DEBUG oslo_concurrency.lockutils [None req-c6329246-02d8-41a2-98f4-7ff34c5957c8 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Acquiring lock "24a79a6c-5210-4778-a6ed-482bfe14933c" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1185.639375] env[63515]: DEBUG oslo_concurrency.lockutils [None req-c6329246-02d8-41a2-98f4-7ff34c5957c8 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Lock "24a79a6c-5210-4778-a6ed-482bfe14933c" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.001s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1185.639610] env[63515]: INFO nova.compute.manager [None req-c6329246-02d8-41a2-98f4-7ff34c5957c8 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 24a79a6c-5210-4778-a6ed-482bfe14933c] Shelving [ 1186.015010] env[63515]: DEBUG oslo_vmware.api [None req-6684c2a7-1d4f-492e-a50c-c770e1d1c230 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Task: {'id': task-1112041, 'name': CopyVirtualDisk_Task} progress is 24%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1186.147465] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-c6329246-02d8-41a2-98f4-7ff34c5957c8 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 24a79a6c-5210-4778-a6ed-482bfe14933c] Powering off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1186.147736] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-fd5c4119-cbe0-4b4c-974e-24b67d719e2d {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1186.156156] env[63515]: DEBUG oslo_vmware.api [None req-c6329246-02d8-41a2-98f4-7ff34c5957c8 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Waiting for the task: (returnval){ [ 1186.156156] env[63515]: value = "task-1112042" [ 1186.156156] env[63515]: _type = "Task" [ 1186.156156] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1186.167927] env[63515]: DEBUG oslo_vmware.api [None req-c6329246-02d8-41a2-98f4-7ff34c5957c8 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Task: {'id': task-1112042, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1186.514332] env[63515]: DEBUG oslo_vmware.api [None req-6684c2a7-1d4f-492e-a50c-c770e1d1c230 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Task: {'id': task-1112041, 'name': CopyVirtualDisk_Task} progress is 46%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1186.668080] env[63515]: DEBUG oslo_vmware.api [None req-c6329246-02d8-41a2-98f4-7ff34c5957c8 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Task: {'id': task-1112042, 'name': PowerOffVM_Task, 'duration_secs': 0.2032} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1186.668382] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-c6329246-02d8-41a2-98f4-7ff34c5957c8 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 24a79a6c-5210-4778-a6ed-482bfe14933c] Powered off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1186.669243] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51db347e-e56d-4319-9453-744c0e9cd24b {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1186.691441] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d45ce9de-1f40-4b6a-9aa6-99647794e01d {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1187.014152] env[63515]: DEBUG oslo_vmware.api [None req-6684c2a7-1d4f-492e-a50c-c770e1d1c230 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Task: {'id': task-1112041, 'name': CopyVirtualDisk_Task} progress is 69%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1187.204576] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-c6329246-02d8-41a2-98f4-7ff34c5957c8 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 24a79a6c-5210-4778-a6ed-482bfe14933c] Creating Snapshot of the VM instance {{(pid=63515) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1187.204865] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-56105cb2-c8d8-4f20-aefc-842998be7b6f {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1187.214901] env[63515]: DEBUG oslo_vmware.api [None req-c6329246-02d8-41a2-98f4-7ff34c5957c8 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Waiting for the task: (returnval){ [ 1187.214901] env[63515]: value = "task-1112043" [ 1187.214901] env[63515]: _type = "Task" [ 1187.214901] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1187.225635] env[63515]: DEBUG oslo_vmware.api [None req-c6329246-02d8-41a2-98f4-7ff34c5957c8 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Task: {'id': task-1112043, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1187.515613] env[63515]: DEBUG oslo_vmware.api [None req-6684c2a7-1d4f-492e-a50c-c770e1d1c230 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Task: {'id': task-1112041, 'name': CopyVirtualDisk_Task} progress is 91%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1187.724257] env[63515]: DEBUG oslo_vmware.api [None req-c6329246-02d8-41a2-98f4-7ff34c5957c8 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Task: {'id': task-1112043, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1188.013899] env[63515]: DEBUG oslo_vmware.api [None req-6684c2a7-1d4f-492e-a50c-c770e1d1c230 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Task: {'id': task-1112041, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.182959} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1188.014193] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-6684c2a7-1d4f-492e-a50c-c770e1d1c230 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/bea6df9b-0878-417c-8cb9-f77d080de2d6/bea6df9b-0878-417c-8cb9-f77d080de2d6.vmdk to [datastore2] c9e07a6c-fd14-4071-8c69-6ece62e9109a/c9e07a6c-fd14-4071-8c69-6ece62e9109a.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1188.014975] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-765b858c-ddd9-463f-a981-e8d543aacf7f {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.036206] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-6684c2a7-1d4f-492e-a50c-c770e1d1c230 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] [instance: c9e07a6c-fd14-4071-8c69-6ece62e9109a] Reconfiguring VM instance instance-0000006c to attach disk [datastore2] c9e07a6c-fd14-4071-8c69-6ece62e9109a/c9e07a6c-fd14-4071-8c69-6ece62e9109a.vmdk or device None with type streamOptimized {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1188.036484] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b8d8b540-85cd-4664-8996-d126e947361a {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.055353] env[63515]: DEBUG oslo_vmware.api [None req-6684c2a7-1d4f-492e-a50c-c770e1d1c230 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Waiting for the task: (returnval){ [ 1188.055353] env[63515]: value = "task-1112044" [ 1188.055353] env[63515]: _type = "Task" [ 1188.055353] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1188.063077] env[63515]: DEBUG oslo_vmware.api [None req-6684c2a7-1d4f-492e-a50c-c770e1d1c230 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Task: {'id': task-1112044, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1188.227441] env[63515]: DEBUG oslo_vmware.api [None req-c6329246-02d8-41a2-98f4-7ff34c5957c8 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Task: {'id': task-1112043, 'name': CreateSnapshot_Task, 'duration_secs': 0.841343} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1188.227729] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-c6329246-02d8-41a2-98f4-7ff34c5957c8 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 24a79a6c-5210-4778-a6ed-482bfe14933c] Created Snapshot of the VM instance {{(pid=63515) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1188.228463] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3f9c212-c812-49fd-b0db-6379147fc469 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.565243] env[63515]: DEBUG oslo_vmware.api [None req-6684c2a7-1d4f-492e-a50c-c770e1d1c230 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Task: {'id': task-1112044, 'name': ReconfigVM_Task, 'duration_secs': 0.308996} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1188.565554] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-6684c2a7-1d4f-492e-a50c-c770e1d1c230 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] [instance: c9e07a6c-fd14-4071-8c69-6ece62e9109a] Reconfigured VM instance instance-0000006c to attach disk [datastore2] c9e07a6c-fd14-4071-8c69-6ece62e9109a/c9e07a6c-fd14-4071-8c69-6ece62e9109a.vmdk or device None with type streamOptimized {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1188.566223] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e0eab1ca-2d91-4e49-b39f-a676c3b77cce {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.572584] env[63515]: DEBUG oslo_vmware.api [None req-6684c2a7-1d4f-492e-a50c-c770e1d1c230 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Waiting for the task: (returnval){ [ 1188.572584] env[63515]: value = "task-1112045" [ 1188.572584] env[63515]: _type = "Task" [ 1188.572584] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1188.580124] env[63515]: DEBUG oslo_vmware.api [None req-6684c2a7-1d4f-492e-a50c-c770e1d1c230 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Task: {'id': task-1112045, 'name': Rename_Task} progress is 5%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1188.745529] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-c6329246-02d8-41a2-98f4-7ff34c5957c8 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 24a79a6c-5210-4778-a6ed-482bfe14933c] Creating linked-clone VM from snapshot {{(pid=63515) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1188.745856] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-0592807b-79fd-49cb-a335-db4ecef9bdc1 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.753977] env[63515]: DEBUG oslo_vmware.api [None req-c6329246-02d8-41a2-98f4-7ff34c5957c8 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Waiting for the task: (returnval){ [ 1188.753977] env[63515]: value = "task-1112046" [ 1188.753977] env[63515]: _type = "Task" [ 1188.753977] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1188.761698] env[63515]: DEBUG oslo_vmware.api [None req-c6329246-02d8-41a2-98f4-7ff34c5957c8 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Task: {'id': task-1112046, 'name': CloneVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1189.082474] env[63515]: DEBUG oslo_vmware.api [None req-6684c2a7-1d4f-492e-a50c-c770e1d1c230 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Task: {'id': task-1112045, 'name': Rename_Task, 'duration_secs': 0.12896} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1189.082750] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-6684c2a7-1d4f-492e-a50c-c770e1d1c230 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] [instance: c9e07a6c-fd14-4071-8c69-6ece62e9109a] Powering on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1189.082997] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-cbd32f4b-c7ca-45d0-a39b-1cb6d95613b6 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1189.089826] env[63515]: DEBUG oslo_vmware.api [None req-6684c2a7-1d4f-492e-a50c-c770e1d1c230 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Waiting for the task: (returnval){ [ 1189.089826] env[63515]: value = "task-1112047" [ 1189.089826] env[63515]: _type = "Task" [ 1189.089826] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1189.097683] env[63515]: DEBUG oslo_vmware.api [None req-6684c2a7-1d4f-492e-a50c-c770e1d1c230 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Task: {'id': task-1112047, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1189.263793] env[63515]: DEBUG oslo_vmware.api [None req-c6329246-02d8-41a2-98f4-7ff34c5957c8 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Task: {'id': task-1112046, 'name': CloneVM_Task} progress is 94%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1189.599548] env[63515]: DEBUG oslo_vmware.api [None req-6684c2a7-1d4f-492e-a50c-c770e1d1c230 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Task: {'id': task-1112047, 'name': PowerOnVM_Task, 'duration_secs': 0.403444} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1189.599883] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-6684c2a7-1d4f-492e-a50c-c770e1d1c230 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] [instance: c9e07a6c-fd14-4071-8c69-6ece62e9109a] Powered on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1189.711207] env[63515]: DEBUG nova.compute.manager [None req-6684c2a7-1d4f-492e-a50c-c770e1d1c230 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] [instance: c9e07a6c-fd14-4071-8c69-6ece62e9109a] Checking state {{(pid=63515) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1189.712164] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0bb72b45-5693-47af-9553-92709a25e249 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1189.764885] env[63515]: DEBUG oslo_vmware.api [None req-c6329246-02d8-41a2-98f4-7ff34c5957c8 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Task: {'id': task-1112046, 'name': CloneVM_Task} progress is 100%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1190.228543] env[63515]: DEBUG oslo_concurrency.lockutils [None req-6684c2a7-1d4f-492e-a50c-c770e1d1c230 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Lock "c9e07a6c-fd14-4071-8c69-6ece62e9109a" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 18.394s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1190.266771] env[63515]: DEBUG oslo_vmware.api [None req-c6329246-02d8-41a2-98f4-7ff34c5957c8 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Task: {'id': task-1112046, 'name': CloneVM_Task, 'duration_secs': 1.074958} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1190.267081] env[63515]: INFO nova.virt.vmwareapi.vmops [None req-c6329246-02d8-41a2-98f4-7ff34c5957c8 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 24a79a6c-5210-4778-a6ed-482bfe14933c] Created linked-clone VM from snapshot [ 1190.267885] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac4adc28-08a8-417d-a1b8-2903880df1ac {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1190.277492] env[63515]: DEBUG nova.virt.vmwareapi.images [None req-c6329246-02d8-41a2-98f4-7ff34c5957c8 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 24a79a6c-5210-4778-a6ed-482bfe14933c] Uploading image af53d566-21dc-48bf-aacf-f6a96a111935 {{(pid=63515) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 1190.300401] env[63515]: DEBUG oslo_vmware.rw_handles [None req-c6329246-02d8-41a2-98f4-7ff34c5957c8 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 1190.300401] env[63515]: value = "vm-243599" [ 1190.300401] env[63515]: _type = "VirtualMachine" [ 1190.300401] env[63515]: }. {{(pid=63515) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 1190.300666] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-03439ae1-05e2-4d6c-87a7-aa8680a9e877 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1190.307812] env[63515]: DEBUG oslo_vmware.rw_handles [None req-c6329246-02d8-41a2-98f4-7ff34c5957c8 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Lease: (returnval){ [ 1190.307812] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]523c1a95-ae7c-ef9c-7216-b8b882930ed3" [ 1190.307812] env[63515]: _type = "HttpNfcLease" [ 1190.307812] env[63515]: } obtained for exporting VM: (result){ [ 1190.307812] env[63515]: value = "vm-243599" [ 1190.307812] env[63515]: _type = "VirtualMachine" [ 1190.307812] env[63515]: }. {{(pid=63515) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 1190.308105] env[63515]: DEBUG oslo_vmware.api [None req-c6329246-02d8-41a2-98f4-7ff34c5957c8 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Waiting for the lease: (returnval){ [ 1190.308105] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]523c1a95-ae7c-ef9c-7216-b8b882930ed3" [ 1190.308105] env[63515]: _type = "HttpNfcLease" [ 1190.308105] env[63515]: } to be ready. {{(pid=63515) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1190.314591] env[63515]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1190.314591] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]523c1a95-ae7c-ef9c-7216-b8b882930ed3" [ 1190.314591] env[63515]: _type = "HttpNfcLease" [ 1190.314591] env[63515]: } is initializing. {{(pid=63515) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1190.817031] env[63515]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1190.817031] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]523c1a95-ae7c-ef9c-7216-b8b882930ed3" [ 1190.817031] env[63515]: _type = "HttpNfcLease" [ 1190.817031] env[63515]: } is ready. {{(pid=63515) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1190.817510] env[63515]: DEBUG oslo_vmware.rw_handles [None req-c6329246-02d8-41a2-98f4-7ff34c5957c8 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1190.817510] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]523c1a95-ae7c-ef9c-7216-b8b882930ed3" [ 1190.817510] env[63515]: _type = "HttpNfcLease" [ 1190.817510] env[63515]: }. {{(pid=63515) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 1190.817993] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d30e9927-4fc4-4ce3-b627-c881eb1f3cd6 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1190.825482] env[63515]: DEBUG oslo_vmware.rw_handles [None req-c6329246-02d8-41a2-98f4-7ff34c5957c8 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52da35ac-01a8-79bf-c553-be629f05aa03/disk-0.vmdk from lease info. {{(pid=63515) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1190.825670] env[63515]: DEBUG oslo_vmware.rw_handles [None req-c6329246-02d8-41a2-98f4-7ff34c5957c8 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52da35ac-01a8-79bf-c553-be629f05aa03/disk-0.vmdk for reading. {{(pid=63515) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 1190.925355] env[63515]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-62360ef4-323c-42ea-829e-fa4fede8edef {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1191.172791] env[63515]: DEBUG oslo_concurrency.lockutils [None req-97bda4ee-4c81-40c5-af18-5e3d6e4c9b03 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Acquiring lock "9a864a51-7df3-4afc-8be7-2794ef2dcf1a" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1191.173327] env[63515]: DEBUG oslo_concurrency.lockutils [None req-97bda4ee-4c81-40c5-af18-5e3d6e4c9b03 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Lock "9a864a51-7df3-4afc-8be7-2794ef2dcf1a" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1191.173676] env[63515]: DEBUG oslo_concurrency.lockutils [None req-97bda4ee-4c81-40c5-af18-5e3d6e4c9b03 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Acquiring lock "9a864a51-7df3-4afc-8be7-2794ef2dcf1a-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1191.174023] env[63515]: DEBUG oslo_concurrency.lockutils [None req-97bda4ee-4c81-40c5-af18-5e3d6e4c9b03 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Lock "9a864a51-7df3-4afc-8be7-2794ef2dcf1a-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1191.174282] env[63515]: DEBUG oslo_concurrency.lockutils [None req-97bda4ee-4c81-40c5-af18-5e3d6e4c9b03 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Lock "9a864a51-7df3-4afc-8be7-2794ef2dcf1a-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1191.176797] env[63515]: INFO nova.compute.manager [None req-97bda4ee-4c81-40c5-af18-5e3d6e4c9b03 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: 9a864a51-7df3-4afc-8be7-2794ef2dcf1a] Terminating instance [ 1191.178874] env[63515]: DEBUG nova.compute.manager [None req-97bda4ee-4c81-40c5-af18-5e3d6e4c9b03 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: 9a864a51-7df3-4afc-8be7-2794ef2dcf1a] Start destroying the instance on the hypervisor. {{(pid=63515) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1191.179110] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-97bda4ee-4c81-40c5-af18-5e3d6e4c9b03 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: 9a864a51-7df3-4afc-8be7-2794ef2dcf1a] Powering off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1191.179554] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-be927894-8d53-418f-a828-0ee04f6b046f {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1191.188887] env[63515]: DEBUG oslo_vmware.api [None req-97bda4ee-4c81-40c5-af18-5e3d6e4c9b03 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Waiting for the task: (returnval){ [ 1191.188887] env[63515]: value = "task-1112049" [ 1191.188887] env[63515]: _type = "Task" [ 1191.188887] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1191.197485] env[63515]: DEBUG oslo_vmware.api [None req-97bda4ee-4c81-40c5-af18-5e3d6e4c9b03 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Task: {'id': task-1112049, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1191.660864] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a21f489d-01f5-4c98-a74e-c6acfa855153 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1191.669578] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-3de9f5db-38ba-4b6f-a224-bf047926302e tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] [instance: c9e07a6c-fd14-4071-8c69-6ece62e9109a] Suspending the VM {{(pid=63515) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1162}} [ 1191.669996] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-852cf30e-7f7b-4d8e-9c54-96dd788aca27 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1191.676966] env[63515]: DEBUG oslo_vmware.api [None req-3de9f5db-38ba-4b6f-a224-bf047926302e tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Waiting for the task: (returnval){ [ 1191.676966] env[63515]: value = "task-1112050" [ 1191.676966] env[63515]: _type = "Task" [ 1191.676966] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1191.688530] env[63515]: DEBUG oslo_vmware.api [None req-3de9f5db-38ba-4b6f-a224-bf047926302e tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Task: {'id': task-1112050, 'name': SuspendVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1191.698695] env[63515]: DEBUG oslo_vmware.api [None req-97bda4ee-4c81-40c5-af18-5e3d6e4c9b03 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Task: {'id': task-1112049, 'name': PowerOffVM_Task, 'duration_secs': 0.195433} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1191.698981] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-97bda4ee-4c81-40c5-af18-5e3d6e4c9b03 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: 9a864a51-7df3-4afc-8be7-2794ef2dcf1a] Powered off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1191.699358] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-97bda4ee-4c81-40c5-af18-5e3d6e4c9b03 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: 9a864a51-7df3-4afc-8be7-2794ef2dcf1a] Volume detach. Driver type: vmdk {{(pid=63515) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1191.699755] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-97bda4ee-4c81-40c5-af18-5e3d6e4c9b03 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: 9a864a51-7df3-4afc-8be7-2794ef2dcf1a] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-243587', 'volume_id': '4ca406d5-39c8-41c2-b985-4734131e1c3c', 'name': 'volume-4ca406d5-39c8-41c2-b985-4734131e1c3c', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attaching', 'instance': '9a864a51-7df3-4afc-8be7-2794ef2dcf1a', 'attached_at': '2024-10-03T03:04:25.000000', 'detached_at': '', 'volume_id': '4ca406d5-39c8-41c2-b985-4734131e1c3c', 'serial': '4ca406d5-39c8-41c2-b985-4734131e1c3c'} {{(pid=63515) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1191.700707] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5625f39-056b-488f-985a-2717e0300ba1 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1191.719214] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc9aa08e-cbd0-4eef-a25f-9eff616eb2b9 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1191.726048] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21b4e1ef-b6d4-4f25-b7a1-062054b02d92 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1191.745600] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17054441-cbee-4ea9-a6c9-8c440d864289 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1191.761391] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-97bda4ee-4c81-40c5-af18-5e3d6e4c9b03 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] The volume has not been displaced from its original location: [datastore2] volume-4ca406d5-39c8-41c2-b985-4734131e1c3c/volume-4ca406d5-39c8-41c2-b985-4734131e1c3c.vmdk. No consolidation needed. {{(pid=63515) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1191.766727] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-97bda4ee-4c81-40c5-af18-5e3d6e4c9b03 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: 9a864a51-7df3-4afc-8be7-2794ef2dcf1a] Reconfiguring VM instance instance-0000006e to detach disk 2000 {{(pid=63515) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1191.767210] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-95a9c877-8b72-430d-bf55-de4fe3d88a2f {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1191.785657] env[63515]: DEBUG oslo_vmware.api [None req-97bda4ee-4c81-40c5-af18-5e3d6e4c9b03 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Waiting for the task: (returnval){ [ 1191.785657] env[63515]: value = "task-1112051" [ 1191.785657] env[63515]: _type = "Task" [ 1191.785657] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1191.794190] env[63515]: DEBUG oslo_vmware.api [None req-97bda4ee-4c81-40c5-af18-5e3d6e4c9b03 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Task: {'id': task-1112051, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1192.186698] env[63515]: DEBUG oslo_vmware.api [None req-3de9f5db-38ba-4b6f-a224-bf047926302e tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Task: {'id': task-1112050, 'name': SuspendVM_Task} progress is 66%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1192.296160] env[63515]: DEBUG oslo_vmware.api [None req-97bda4ee-4c81-40c5-af18-5e3d6e4c9b03 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Task: {'id': task-1112051, 'name': ReconfigVM_Task, 'duration_secs': 0.263481} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1192.296585] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-97bda4ee-4c81-40c5-af18-5e3d6e4c9b03 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: 9a864a51-7df3-4afc-8be7-2794ef2dcf1a] Reconfigured VM instance instance-0000006e to detach disk 2000 {{(pid=63515) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1192.301527] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-748ecf15-294c-4476-a09d-9e4cd273aaab {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1192.317088] env[63515]: DEBUG oslo_vmware.api [None req-97bda4ee-4c81-40c5-af18-5e3d6e4c9b03 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Waiting for the task: (returnval){ [ 1192.317088] env[63515]: value = "task-1112052" [ 1192.317088] env[63515]: _type = "Task" [ 1192.317088] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1192.325461] env[63515]: DEBUG oslo_vmware.api [None req-97bda4ee-4c81-40c5-af18-5e3d6e4c9b03 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Task: {'id': task-1112052, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1192.687677] env[63515]: DEBUG oslo_vmware.api [None req-3de9f5db-38ba-4b6f-a224-bf047926302e tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Task: {'id': task-1112050, 'name': SuspendVM_Task, 'duration_secs': 0.650443} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1192.688062] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-3de9f5db-38ba-4b6f-a224-bf047926302e tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] [instance: c9e07a6c-fd14-4071-8c69-6ece62e9109a] Suspended the VM {{(pid=63515) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1166}} [ 1192.688332] env[63515]: DEBUG nova.compute.manager [None req-3de9f5db-38ba-4b6f-a224-bf047926302e tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] [instance: c9e07a6c-fd14-4071-8c69-6ece62e9109a] Checking state {{(pid=63515) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1192.689220] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-111c11af-2564-49f1-b0c9-92c0ad9106f4 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1192.827805] env[63515]: DEBUG oslo_vmware.api [None req-97bda4ee-4c81-40c5-af18-5e3d6e4c9b03 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Task: {'id': task-1112052, 'name': ReconfigVM_Task} progress is 99%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1193.328348] env[63515]: DEBUG oslo_vmware.api [None req-97bda4ee-4c81-40c5-af18-5e3d6e4c9b03 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Task: {'id': task-1112052, 'name': ReconfigVM_Task} progress is 99%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1193.830223] env[63515]: DEBUG oslo_vmware.api [None req-97bda4ee-4c81-40c5-af18-5e3d6e4c9b03 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Task: {'id': task-1112052, 'name': ReconfigVM_Task, 'duration_secs': 1.222259} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1193.830223] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-97bda4ee-4c81-40c5-af18-5e3d6e4c9b03 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: 9a864a51-7df3-4afc-8be7-2794ef2dcf1a] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-243587', 'volume_id': '4ca406d5-39c8-41c2-b985-4734131e1c3c', 'name': 'volume-4ca406d5-39c8-41c2-b985-4734131e1c3c', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attaching', 'instance': '9a864a51-7df3-4afc-8be7-2794ef2dcf1a', 'attached_at': '2024-10-03T03:04:25.000000', 'detached_at': '', 'volume_id': '4ca406d5-39c8-41c2-b985-4734131e1c3c', 'serial': '4ca406d5-39c8-41c2-b985-4734131e1c3c'} {{(pid=63515) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1193.830223] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-97bda4ee-4c81-40c5-af18-5e3d6e4c9b03 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: 9a864a51-7df3-4afc-8be7-2794ef2dcf1a] Destroying instance {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1193.830869] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d380c5e3-ee14-495b-bdb3-045157fb6e74 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1193.837455] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-97bda4ee-4c81-40c5-af18-5e3d6e4c9b03 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: 9a864a51-7df3-4afc-8be7-2794ef2dcf1a] Unregistering the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1193.837693] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d09d7296-cd36-4980-96e8-35afff512d55 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1193.901929] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-97bda4ee-4c81-40c5-af18-5e3d6e4c9b03 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: 9a864a51-7df3-4afc-8be7-2794ef2dcf1a] Unregistered the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1193.902154] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-97bda4ee-4c81-40c5-af18-5e3d6e4c9b03 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: 9a864a51-7df3-4afc-8be7-2794ef2dcf1a] Deleting contents of the VM from datastore datastore2 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1193.902319] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-97bda4ee-4c81-40c5-af18-5e3d6e4c9b03 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Deleting the datastore file [datastore2] 9a864a51-7df3-4afc-8be7-2794ef2dcf1a {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1193.902626] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-9b66bb4b-27f3-4bc6-8183-0937dbd5de59 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1193.910005] env[63515]: DEBUG oslo_vmware.api [None req-97bda4ee-4c81-40c5-af18-5e3d6e4c9b03 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Waiting for the task: (returnval){ [ 1193.910005] env[63515]: value = "task-1112054" [ 1193.910005] env[63515]: _type = "Task" [ 1193.910005] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1193.917708] env[63515]: DEBUG oslo_vmware.api [None req-97bda4ee-4c81-40c5-af18-5e3d6e4c9b03 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Task: {'id': task-1112054, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1194.068709] env[63515]: INFO nova.compute.manager [None req-88a6ce31-47ce-4881-b21f-d6c99c0e7dd7 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] [instance: c9e07a6c-fd14-4071-8c69-6ece62e9109a] Resuming [ 1194.069487] env[63515]: DEBUG nova.objects.instance [None req-88a6ce31-47ce-4881-b21f-d6c99c0e7dd7 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Lazy-loading 'flavor' on Instance uuid c9e07a6c-fd14-4071-8c69-6ece62e9109a {{(pid=63515) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1194.420763] env[63515]: DEBUG oslo_vmware.api [None req-97bda4ee-4c81-40c5-af18-5e3d6e4c9b03 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Task: {'id': task-1112054, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.085174} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1194.421127] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-97bda4ee-4c81-40c5-af18-5e3d6e4c9b03 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Deleted the datastore file {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1194.421296] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-97bda4ee-4c81-40c5-af18-5e3d6e4c9b03 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: 9a864a51-7df3-4afc-8be7-2794ef2dcf1a] Deleted contents of the VM from datastore datastore2 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1194.421489] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-97bda4ee-4c81-40c5-af18-5e3d6e4c9b03 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: 9a864a51-7df3-4afc-8be7-2794ef2dcf1a] Instance destroyed {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1194.421700] env[63515]: INFO nova.compute.manager [None req-97bda4ee-4c81-40c5-af18-5e3d6e4c9b03 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: 9a864a51-7df3-4afc-8be7-2794ef2dcf1a] Took 3.24 seconds to destroy the instance on the hypervisor. [ 1194.421959] env[63515]: DEBUG oslo.service.loopingcall [None req-97bda4ee-4c81-40c5-af18-5e3d6e4c9b03 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1194.422173] env[63515]: DEBUG nova.compute.manager [-] [instance: 9a864a51-7df3-4afc-8be7-2794ef2dcf1a] Deallocating network for instance {{(pid=63515) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1194.422270] env[63515]: DEBUG nova.network.neutron [-] [instance: 9a864a51-7df3-4afc-8be7-2794ef2dcf1a] deallocate_for_instance() {{(pid=63515) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1194.912818] env[63515]: DEBUG nova.compute.manager [req-70b42b8e-bf12-4dfb-9afb-08a4537ebc67 req-8480de33-08cc-48ce-b7d6-7770337b48d2 service nova] [instance: 9a864a51-7df3-4afc-8be7-2794ef2dcf1a] Received event network-vif-deleted-d96464dd-1638-4d19-8c71-9faeee62242f {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1194.913101] env[63515]: INFO nova.compute.manager [req-70b42b8e-bf12-4dfb-9afb-08a4537ebc67 req-8480de33-08cc-48ce-b7d6-7770337b48d2 service nova] [instance: 9a864a51-7df3-4afc-8be7-2794ef2dcf1a] Neutron deleted interface d96464dd-1638-4d19-8c71-9faeee62242f; detaching it from the instance and deleting it from the info cache [ 1194.913266] env[63515]: DEBUG nova.network.neutron [req-70b42b8e-bf12-4dfb-9afb-08a4537ebc67 req-8480de33-08cc-48ce-b7d6-7770337b48d2 service nova] [instance: 9a864a51-7df3-4afc-8be7-2794ef2dcf1a] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1195.077890] env[63515]: DEBUG oslo_concurrency.lockutils [None req-88a6ce31-47ce-4881-b21f-d6c99c0e7dd7 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Acquiring lock "refresh_cache-c9e07a6c-fd14-4071-8c69-6ece62e9109a" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1195.078119] env[63515]: DEBUG oslo_concurrency.lockutils [None req-88a6ce31-47ce-4881-b21f-d6c99c0e7dd7 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Acquired lock "refresh_cache-c9e07a6c-fd14-4071-8c69-6ece62e9109a" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1195.078310] env[63515]: DEBUG nova.network.neutron [None req-88a6ce31-47ce-4881-b21f-d6c99c0e7dd7 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] [instance: c9e07a6c-fd14-4071-8c69-6ece62e9109a] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1195.389160] env[63515]: DEBUG nova.network.neutron [-] [instance: 9a864a51-7df3-4afc-8be7-2794ef2dcf1a] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1195.415768] env[63515]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-cc326a40-622c-42f6-95d4-5f9548d8ba05 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1195.425492] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9aa04150-55b6-4b12-89a8-a7e6ac3d7c6f {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1195.452532] env[63515]: DEBUG nova.compute.manager [req-70b42b8e-bf12-4dfb-9afb-08a4537ebc67 req-8480de33-08cc-48ce-b7d6-7770337b48d2 service nova] [instance: 9a864a51-7df3-4afc-8be7-2794ef2dcf1a] Detach interface failed, port_id=d96464dd-1638-4d19-8c71-9faeee62242f, reason: Instance 9a864a51-7df3-4afc-8be7-2794ef2dcf1a could not be found. {{(pid=63515) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1195.781411] env[63515]: DEBUG nova.network.neutron [None req-88a6ce31-47ce-4881-b21f-d6c99c0e7dd7 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] [instance: c9e07a6c-fd14-4071-8c69-6ece62e9109a] Updating instance_info_cache with network_info: [{"id": "bd316695-ec41-4f7d-b042-5cd5d319aa55", "address": "fa:16:3e:da:4b:8e", "network": {"id": "0e23b3a4-91ea-4bbe-9732-a3539d56b382", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1727906976-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fe873600ed6d4a3eb1ee3d3a19fceaf7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c7d2575f-b92f-44ec-a863-634cb76631a2", "external-id": "nsx-vlan-transportzone-794", "segmentation_id": 794, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbd316695-ec", "ovs_interfaceid": "bd316695-ec41-4f7d-b042-5cd5d319aa55", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1195.892678] env[63515]: INFO nova.compute.manager [-] [instance: 9a864a51-7df3-4afc-8be7-2794ef2dcf1a] Took 1.47 seconds to deallocate network for instance. [ 1196.284815] env[63515]: DEBUG oslo_concurrency.lockutils [None req-88a6ce31-47ce-4881-b21f-d6c99c0e7dd7 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Releasing lock "refresh_cache-c9e07a6c-fd14-4071-8c69-6ece62e9109a" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1196.285883] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60bbaaa2-5470-4cb0-bdb3-9f6ead1c5c68 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.294611] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-88a6ce31-47ce-4881-b21f-d6c99c0e7dd7 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] [instance: c9e07a6c-fd14-4071-8c69-6ece62e9109a] Resuming the VM {{(pid=63515) resume /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1183}} [ 1196.294874] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-87be8fe0-08e2-4f62-8c79-9e075625515e {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.301152] env[63515]: DEBUG oslo_vmware.api [None req-88a6ce31-47ce-4881-b21f-d6c99c0e7dd7 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Waiting for the task: (returnval){ [ 1196.301152] env[63515]: value = "task-1112055" [ 1196.301152] env[63515]: _type = "Task" [ 1196.301152] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1196.310311] env[63515]: DEBUG oslo_vmware.api [None req-88a6ce31-47ce-4881-b21f-d6c99c0e7dd7 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Task: {'id': task-1112055, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1196.441624] env[63515]: INFO nova.compute.manager [None req-97bda4ee-4c81-40c5-af18-5e3d6e4c9b03 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: 9a864a51-7df3-4afc-8be7-2794ef2dcf1a] Took 0.55 seconds to detach 1 volumes for instance. [ 1196.443932] env[63515]: DEBUG nova.compute.manager [None req-97bda4ee-4c81-40c5-af18-5e3d6e4c9b03 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: 9a864a51-7df3-4afc-8be7-2794ef2dcf1a] Deleting volume: 4ca406d5-39c8-41c2-b985-4734131e1c3c {{(pid=63515) _cleanup_volumes /opt/stack/nova/nova/compute/manager.py:3247}} [ 1196.814177] env[63515]: DEBUG oslo_vmware.api [None req-88a6ce31-47ce-4881-b21f-d6c99c0e7dd7 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Task: {'id': task-1112055, 'name': PowerOnVM_Task} progress is 100%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1196.987378] env[63515]: DEBUG oslo_concurrency.lockutils [None req-97bda4ee-4c81-40c5-af18-5e3d6e4c9b03 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1196.987772] env[63515]: DEBUG oslo_concurrency.lockutils [None req-97bda4ee-4c81-40c5-af18-5e3d6e4c9b03 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1196.987992] env[63515]: DEBUG oslo_concurrency.lockutils [None req-97bda4ee-4c81-40c5-af18-5e3d6e4c9b03 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1197.016578] env[63515]: INFO nova.scheduler.client.report [None req-97bda4ee-4c81-40c5-af18-5e3d6e4c9b03 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Deleted allocations for instance 9a864a51-7df3-4afc-8be7-2794ef2dcf1a [ 1197.313882] env[63515]: DEBUG oslo_vmware.api [None req-88a6ce31-47ce-4881-b21f-d6c99c0e7dd7 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Task: {'id': task-1112055, 'name': PowerOnVM_Task, 'duration_secs': 0.541379} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1197.314278] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-88a6ce31-47ce-4881-b21f-d6c99c0e7dd7 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] [instance: c9e07a6c-fd14-4071-8c69-6ece62e9109a] Resumed the VM {{(pid=63515) resume /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1188}} [ 1197.314501] env[63515]: DEBUG nova.compute.manager [None req-88a6ce31-47ce-4881-b21f-d6c99c0e7dd7 tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] [instance: c9e07a6c-fd14-4071-8c69-6ece62e9109a] Checking state {{(pid=63515) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1197.315561] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-466389ca-afc8-4907-adf1-a01a28c05fc1 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1197.524558] env[63515]: DEBUG oslo_concurrency.lockutils [None req-97bda4ee-4c81-40c5-af18-5e3d6e4c9b03 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Lock "9a864a51-7df3-4afc-8be7-2794ef2dcf1a" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.351s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1197.923830] env[63515]: DEBUG oslo_concurrency.lockutils [None req-860547e3-635a-49cd-a65d-8ec8ccee5ac4 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Acquiring lock "ea2e0810-c9f1-49a6-8589-7db402847aea" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1197.924247] env[63515]: DEBUG oslo_concurrency.lockutils [None req-860547e3-635a-49cd-a65d-8ec8ccee5ac4 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Lock "ea2e0810-c9f1-49a6-8589-7db402847aea" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1197.924473] env[63515]: DEBUG oslo_concurrency.lockutils [None req-860547e3-635a-49cd-a65d-8ec8ccee5ac4 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Acquiring lock "ea2e0810-c9f1-49a6-8589-7db402847aea-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1197.924666] env[63515]: DEBUG oslo_concurrency.lockutils [None req-860547e3-635a-49cd-a65d-8ec8ccee5ac4 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Lock "ea2e0810-c9f1-49a6-8589-7db402847aea-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1197.924857] env[63515]: DEBUG oslo_concurrency.lockutils [None req-860547e3-635a-49cd-a65d-8ec8ccee5ac4 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Lock "ea2e0810-c9f1-49a6-8589-7db402847aea-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1197.927089] env[63515]: INFO nova.compute.manager [None req-860547e3-635a-49cd-a65d-8ec8ccee5ac4 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: ea2e0810-c9f1-49a6-8589-7db402847aea] Terminating instance [ 1197.928982] env[63515]: DEBUG nova.compute.manager [None req-860547e3-635a-49cd-a65d-8ec8ccee5ac4 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: ea2e0810-c9f1-49a6-8589-7db402847aea] Start destroying the instance on the hypervisor. {{(pid=63515) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1197.929198] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-860547e3-635a-49cd-a65d-8ec8ccee5ac4 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: ea2e0810-c9f1-49a6-8589-7db402847aea] Destroying instance {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1197.930100] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8c0f92a-7629-46a8-83e0-7b8a994982eb {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1197.937832] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-860547e3-635a-49cd-a65d-8ec8ccee5ac4 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: ea2e0810-c9f1-49a6-8589-7db402847aea] Powering off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1197.938075] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8b49212c-7326-4030-8dd8-cbf79f30dc67 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1197.944286] env[63515]: DEBUG oslo_vmware.api [None req-860547e3-635a-49cd-a65d-8ec8ccee5ac4 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Waiting for the task: (returnval){ [ 1197.944286] env[63515]: value = "task-1112057" [ 1197.944286] env[63515]: _type = "Task" [ 1197.944286] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1197.953644] env[63515]: DEBUG oslo_vmware.api [None req-860547e3-635a-49cd-a65d-8ec8ccee5ac4 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Task: {'id': task-1112057, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1198.454577] env[63515]: DEBUG oslo_vmware.api [None req-860547e3-635a-49cd-a65d-8ec8ccee5ac4 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Task: {'id': task-1112057, 'name': PowerOffVM_Task, 'duration_secs': 0.280755} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1198.454837] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-860547e3-635a-49cd-a65d-8ec8ccee5ac4 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: ea2e0810-c9f1-49a6-8589-7db402847aea] Powered off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1198.455022] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-860547e3-635a-49cd-a65d-8ec8ccee5ac4 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: ea2e0810-c9f1-49a6-8589-7db402847aea] Unregistering the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1198.455321] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-61be7473-1633-4bb8-bfa7-97126888a625 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1198.592575] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-860547e3-635a-49cd-a65d-8ec8ccee5ac4 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: ea2e0810-c9f1-49a6-8589-7db402847aea] Unregistered the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1198.592575] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-860547e3-635a-49cd-a65d-8ec8ccee5ac4 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: ea2e0810-c9f1-49a6-8589-7db402847aea] Deleting contents of the VM from datastore datastore1 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1198.592575] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-860547e3-635a-49cd-a65d-8ec8ccee5ac4 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Deleting the datastore file [datastore1] ea2e0810-c9f1-49a6-8589-7db402847aea {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1198.593083] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-49623247-0c90-494b-a93f-b8958504e11a {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1198.599274] env[63515]: DEBUG oslo_vmware.api [None req-860547e3-635a-49cd-a65d-8ec8ccee5ac4 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Waiting for the task: (returnval){ [ 1198.599274] env[63515]: value = "task-1112059" [ 1198.599274] env[63515]: _type = "Task" [ 1198.599274] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1198.607333] env[63515]: DEBUG oslo_vmware.api [None req-860547e3-635a-49cd-a65d-8ec8ccee5ac4 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Task: {'id': task-1112059, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1199.112469] env[63515]: DEBUG oslo_vmware.api [None req-860547e3-635a-49cd-a65d-8ec8ccee5ac4 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Task: {'id': task-1112059, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.1586} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1199.112743] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-860547e3-635a-49cd-a65d-8ec8ccee5ac4 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Deleted the datastore file {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1199.112949] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-860547e3-635a-49cd-a65d-8ec8ccee5ac4 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: ea2e0810-c9f1-49a6-8589-7db402847aea] Deleted contents of the VM from datastore datastore1 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1199.113169] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-860547e3-635a-49cd-a65d-8ec8ccee5ac4 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: ea2e0810-c9f1-49a6-8589-7db402847aea] Instance destroyed {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1199.113370] env[63515]: INFO nova.compute.manager [None req-860547e3-635a-49cd-a65d-8ec8ccee5ac4 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: ea2e0810-c9f1-49a6-8589-7db402847aea] Took 1.18 seconds to destroy the instance on the hypervisor. [ 1199.113655] env[63515]: DEBUG oslo.service.loopingcall [None req-860547e3-635a-49cd-a65d-8ec8ccee5ac4 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1199.113861] env[63515]: DEBUG nova.compute.manager [-] [instance: ea2e0810-c9f1-49a6-8589-7db402847aea] Deallocating network for instance {{(pid=63515) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1199.113959] env[63515]: DEBUG nova.network.neutron [-] [instance: ea2e0810-c9f1-49a6-8589-7db402847aea] deallocate_for_instance() {{(pid=63515) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1199.481388] env[63515]: DEBUG nova.compute.manager [req-760a030c-542f-4c5b-b895-d7f9e9f3acf4 req-95f5d18b-cfc6-4e1d-935f-e5c77f371721 service nova] [instance: ea2e0810-c9f1-49a6-8589-7db402847aea] Received event network-vif-deleted-4e1d011e-2609-4c71-8317-2685198b6a8b {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1199.481388] env[63515]: INFO nova.compute.manager [req-760a030c-542f-4c5b-b895-d7f9e9f3acf4 req-95f5d18b-cfc6-4e1d-935f-e5c77f371721 service nova] [instance: ea2e0810-c9f1-49a6-8589-7db402847aea] Neutron deleted interface 4e1d011e-2609-4c71-8317-2685198b6a8b; detaching it from the instance and deleting it from the info cache [ 1199.481607] env[63515]: DEBUG nova.network.neutron [req-760a030c-542f-4c5b-b895-d7f9e9f3acf4 req-95f5d18b-cfc6-4e1d-935f-e5c77f371721 service nova] [instance: ea2e0810-c9f1-49a6-8589-7db402847aea] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1199.961183] env[63515]: DEBUG nova.network.neutron [-] [instance: ea2e0810-c9f1-49a6-8589-7db402847aea] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1199.984555] env[63515]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-fcaa2e3b-9778-46d7-a1c6-db00d385a388 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1199.993776] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c3f5feb-6f7b-4612-a577-33c2481bba72 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1200.021751] env[63515]: DEBUG nova.compute.manager [req-760a030c-542f-4c5b-b895-d7f9e9f3acf4 req-95f5d18b-cfc6-4e1d-935f-e5c77f371721 service nova] [instance: ea2e0810-c9f1-49a6-8589-7db402847aea] Detach interface failed, port_id=4e1d011e-2609-4c71-8317-2685198b6a8b, reason: Instance ea2e0810-c9f1-49a6-8589-7db402847aea could not be found. {{(pid=63515) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1200.344657] env[63515]: DEBUG oslo_service.periodic_task [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=63515) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1200.346699] env[63515]: DEBUG oslo_service.periodic_task [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=63515) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1200.352532] env[63515]: DEBUG oslo_vmware.rw_handles [None req-c6329246-02d8-41a2-98f4-7ff34c5957c8 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52da35ac-01a8-79bf-c553-be629f05aa03/disk-0.vmdk. {{(pid=63515) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1200.353378] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-397dcac6-4542-42ad-84b9-e4c2ed1702e4 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1200.359586] env[63515]: DEBUG oslo_vmware.rw_handles [None req-c6329246-02d8-41a2-98f4-7ff34c5957c8 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52da35ac-01a8-79bf-c553-be629f05aa03/disk-0.vmdk is in state: ready. {{(pid=63515) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1200.359767] env[63515]: ERROR oslo_vmware.rw_handles [None req-c6329246-02d8-41a2-98f4-7ff34c5957c8 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52da35ac-01a8-79bf-c553-be629f05aa03/disk-0.vmdk due to incomplete transfer. [ 1200.359989] env[63515]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-146508d7-29f9-4973-a534-86cc335245eb {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1200.366484] env[63515]: DEBUG oslo_vmware.rw_handles [None req-c6329246-02d8-41a2-98f4-7ff34c5957c8 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52da35ac-01a8-79bf-c553-be629f05aa03/disk-0.vmdk. {{(pid=63515) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 1200.366687] env[63515]: DEBUG nova.virt.vmwareapi.images [None req-c6329246-02d8-41a2-98f4-7ff34c5957c8 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 24a79a6c-5210-4778-a6ed-482bfe14933c] Uploaded image af53d566-21dc-48bf-aacf-f6a96a111935 to the Glance image server {{(pid=63515) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 1200.369015] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-c6329246-02d8-41a2-98f4-7ff34c5957c8 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 24a79a6c-5210-4778-a6ed-482bfe14933c] Destroying the VM {{(pid=63515) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 1200.369258] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-79e0abe7-1569-4cdb-8a96-f43bcfbe63df {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1200.375093] env[63515]: DEBUG oslo_vmware.api [None req-c6329246-02d8-41a2-98f4-7ff34c5957c8 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Waiting for the task: (returnval){ [ 1200.375093] env[63515]: value = "task-1112060" [ 1200.375093] env[63515]: _type = "Task" [ 1200.375093] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1200.383041] env[63515]: DEBUG oslo_vmware.api [None req-c6329246-02d8-41a2-98f4-7ff34c5957c8 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Task: {'id': task-1112060, 'name': Destroy_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1200.464121] env[63515]: INFO nova.compute.manager [-] [instance: ea2e0810-c9f1-49a6-8589-7db402847aea] Took 1.35 seconds to deallocate network for instance. [ 1200.535061] env[63515]: DEBUG oslo_service.periodic_task [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=63515) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1200.535061] env[63515]: DEBUG nova.compute.manager [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=63515) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10545}} [ 1200.885829] env[63515]: DEBUG oslo_vmware.api [None req-c6329246-02d8-41a2-98f4-7ff34c5957c8 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Task: {'id': task-1112060, 'name': Destroy_Task, 'duration_secs': 0.353012} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1200.886160] env[63515]: INFO nova.virt.vmwareapi.vm_util [None req-c6329246-02d8-41a2-98f4-7ff34c5957c8 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 24a79a6c-5210-4778-a6ed-482bfe14933c] Destroyed the VM [ 1200.886422] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-c6329246-02d8-41a2-98f4-7ff34c5957c8 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 24a79a6c-5210-4778-a6ed-482bfe14933c] Deleting Snapshot of the VM instance {{(pid=63515) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1200.886712] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-4013b6ee-8218-4015-97d9-571851f119d6 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1200.892964] env[63515]: DEBUG oslo_vmware.api [None req-c6329246-02d8-41a2-98f4-7ff34c5957c8 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Waiting for the task: (returnval){ [ 1200.892964] env[63515]: value = "task-1112061" [ 1200.892964] env[63515]: _type = "Task" [ 1200.892964] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1200.900304] env[63515]: DEBUG oslo_vmware.api [None req-c6329246-02d8-41a2-98f4-7ff34c5957c8 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Task: {'id': task-1112061, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1200.972670] env[63515]: DEBUG oslo_concurrency.lockutils [None req-860547e3-635a-49cd-a65d-8ec8ccee5ac4 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1200.972942] env[63515]: DEBUG oslo_concurrency.lockutils [None req-860547e3-635a-49cd-a65d-8ec8ccee5ac4 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1200.973158] env[63515]: DEBUG nova.objects.instance [None req-860547e3-635a-49cd-a65d-8ec8ccee5ac4 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Lazy-loading 'resources' on Instance uuid ea2e0810-c9f1-49a6-8589-7db402847aea {{(pid=63515) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1201.402786] env[63515]: DEBUG oslo_vmware.api [None req-c6329246-02d8-41a2-98f4-7ff34c5957c8 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Task: {'id': task-1112061, 'name': RemoveSnapshot_Task, 'duration_secs': 0.470601} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1201.403080] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-c6329246-02d8-41a2-98f4-7ff34c5957c8 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 24a79a6c-5210-4778-a6ed-482bfe14933c] Deleted Snapshot of the VM instance {{(pid=63515) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1201.403358] env[63515]: DEBUG nova.compute.manager [None req-c6329246-02d8-41a2-98f4-7ff34c5957c8 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 24a79a6c-5210-4778-a6ed-482bfe14933c] Checking state {{(pid=63515) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1201.404135] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53556d05-ee17-4560-b395-97a556557c9a {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1201.552556] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1990bbee-7ed9-4ce2-8b59-bb54a39f6a88 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1201.559934] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90e8f3d6-902e-43b8-ba82-ca23c0d7dacb {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1201.589254] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0016721a-353c-4254-8db7-60e76790f888 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1201.595935] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fec0d1a2-d9cd-444c-90d7-f39f26e8215f {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1201.609105] env[63515]: DEBUG nova.compute.provider_tree [None req-860547e3-635a-49cd-a65d-8ec8ccee5ac4 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1201.916272] env[63515]: INFO nova.compute.manager [None req-c6329246-02d8-41a2-98f4-7ff34c5957c8 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 24a79a6c-5210-4778-a6ed-482bfe14933c] Shelve offloading [ 1201.918313] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-c6329246-02d8-41a2-98f4-7ff34c5957c8 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 24a79a6c-5210-4778-a6ed-482bfe14933c] Powering off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1201.918559] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a2172371-ad61-44a1-ab99-571a1e38224a {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1201.927396] env[63515]: DEBUG oslo_vmware.api [None req-c6329246-02d8-41a2-98f4-7ff34c5957c8 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Waiting for the task: (returnval){ [ 1201.927396] env[63515]: value = "task-1112062" [ 1201.927396] env[63515]: _type = "Task" [ 1201.927396] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1201.934620] env[63515]: DEBUG oslo_vmware.api [None req-c6329246-02d8-41a2-98f4-7ff34c5957c8 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Task: {'id': task-1112062, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1202.112946] env[63515]: DEBUG nova.scheduler.client.report [None req-860547e3-635a-49cd-a65d-8ec8ccee5ac4 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1202.439869] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-c6329246-02d8-41a2-98f4-7ff34c5957c8 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 24a79a6c-5210-4778-a6ed-482bfe14933c] VM already powered off {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1202.440103] env[63515]: DEBUG nova.compute.manager [None req-c6329246-02d8-41a2-98f4-7ff34c5957c8 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 24a79a6c-5210-4778-a6ed-482bfe14933c] Checking state {{(pid=63515) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1202.440856] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5a9114d-ba92-4b03-aaee-561b0b84009d {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1202.446521] env[63515]: DEBUG oslo_concurrency.lockutils [None req-c6329246-02d8-41a2-98f4-7ff34c5957c8 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Acquiring lock "refresh_cache-24a79a6c-5210-4778-a6ed-482bfe14933c" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1202.446688] env[63515]: DEBUG oslo_concurrency.lockutils [None req-c6329246-02d8-41a2-98f4-7ff34c5957c8 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Acquired lock "refresh_cache-24a79a6c-5210-4778-a6ed-482bfe14933c" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1202.446857] env[63515]: DEBUG nova.network.neutron [None req-c6329246-02d8-41a2-98f4-7ff34c5957c8 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 24a79a6c-5210-4778-a6ed-482bfe14933c] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1202.535537] env[63515]: DEBUG oslo_service.periodic_task [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=63515) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1202.617997] env[63515]: DEBUG oslo_concurrency.lockutils [None req-860547e3-635a-49cd-a65d-8ec8ccee5ac4 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.645s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1202.635536] env[63515]: INFO nova.scheduler.client.report [None req-860547e3-635a-49cd-a65d-8ec8ccee5ac4 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Deleted allocations for instance ea2e0810-c9f1-49a6-8589-7db402847aea [ 1202.663846] env[63515]: DEBUG oslo_concurrency.lockutils [None req-9869ad3b-ff07-4839-acd3-a83e20c15f1b tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Acquiring lock "c9e07a6c-fd14-4071-8c69-6ece62e9109a" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1202.664120] env[63515]: DEBUG oslo_concurrency.lockutils [None req-9869ad3b-ff07-4839-acd3-a83e20c15f1b tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Lock "c9e07a6c-fd14-4071-8c69-6ece62e9109a" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1202.664332] env[63515]: DEBUG oslo_concurrency.lockutils [None req-9869ad3b-ff07-4839-acd3-a83e20c15f1b tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Acquiring lock "c9e07a6c-fd14-4071-8c69-6ece62e9109a-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1202.665042] env[63515]: DEBUG oslo_concurrency.lockutils [None req-9869ad3b-ff07-4839-acd3-a83e20c15f1b tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Lock "c9e07a6c-fd14-4071-8c69-6ece62e9109a-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1202.665042] env[63515]: DEBUG oslo_concurrency.lockutils [None req-9869ad3b-ff07-4839-acd3-a83e20c15f1b tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Lock "c9e07a6c-fd14-4071-8c69-6ece62e9109a-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1202.666613] env[63515]: INFO nova.compute.manager [None req-9869ad3b-ff07-4839-acd3-a83e20c15f1b tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] [instance: c9e07a6c-fd14-4071-8c69-6ece62e9109a] Terminating instance [ 1202.668321] env[63515]: DEBUG nova.compute.manager [None req-9869ad3b-ff07-4839-acd3-a83e20c15f1b tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] [instance: c9e07a6c-fd14-4071-8c69-6ece62e9109a] Start destroying the instance on the hypervisor. {{(pid=63515) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1202.668521] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-9869ad3b-ff07-4839-acd3-a83e20c15f1b tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] [instance: c9e07a6c-fd14-4071-8c69-6ece62e9109a] Destroying instance {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1202.669346] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-320d1df8-8fca-476c-b95a-6fcdb0fe0960 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1202.677014] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-9869ad3b-ff07-4839-acd3-a83e20c15f1b tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] [instance: c9e07a6c-fd14-4071-8c69-6ece62e9109a] Powering off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1202.677738] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-95dcbc2c-cf5b-4904-80cc-85dfe63b7278 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1202.684280] env[63515]: DEBUG oslo_vmware.api [None req-9869ad3b-ff07-4839-acd3-a83e20c15f1b tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Waiting for the task: (returnval){ [ 1202.684280] env[63515]: value = "task-1112063" [ 1202.684280] env[63515]: _type = "Task" [ 1202.684280] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1202.692258] env[63515]: DEBUG oslo_vmware.api [None req-9869ad3b-ff07-4839-acd3-a83e20c15f1b tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Task: {'id': task-1112063, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1203.142594] env[63515]: DEBUG oslo_concurrency.lockutils [None req-860547e3-635a-49cd-a65d-8ec8ccee5ac4 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Lock "ea2e0810-c9f1-49a6-8589-7db402847aea" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.218s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1203.178257] env[63515]: DEBUG nova.network.neutron [None req-c6329246-02d8-41a2-98f4-7ff34c5957c8 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 24a79a6c-5210-4778-a6ed-482bfe14933c] Updating instance_info_cache with network_info: [{"id": "f338b564-7636-4839-b034-73c3233d7da6", "address": "fa:16:3e:bc:5a:12", "network": {"id": "5758c885-e1c9-45c8-86aa-27a802101f52", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-734458138-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.204", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b18d4daacbc84e758f9f9ae4f3f09d28", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69351262-8d39-441a-85ba-3a78df436d17", "external-id": "nsx-vlan-transportzone-205", "segmentation_id": 205, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf338b564-76", "ovs_interfaceid": "f338b564-7636-4839-b034-73c3233d7da6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1203.194401] env[63515]: DEBUG oslo_vmware.api [None req-9869ad3b-ff07-4839-acd3-a83e20c15f1b tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Task: {'id': task-1112063, 'name': PowerOffVM_Task, 'duration_secs': 0.170708} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1203.194726] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-9869ad3b-ff07-4839-acd3-a83e20c15f1b tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] [instance: c9e07a6c-fd14-4071-8c69-6ece62e9109a] Powered off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1203.194830] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-9869ad3b-ff07-4839-acd3-a83e20c15f1b tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] [instance: c9e07a6c-fd14-4071-8c69-6ece62e9109a] Unregistering the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1203.195086] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-fb47e2d3-1a3b-46af-8cf2-6e2a6df3f98a {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1203.254255] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-9869ad3b-ff07-4839-acd3-a83e20c15f1b tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] [instance: c9e07a6c-fd14-4071-8c69-6ece62e9109a] Unregistered the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1203.254485] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-9869ad3b-ff07-4839-acd3-a83e20c15f1b tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] [instance: c9e07a6c-fd14-4071-8c69-6ece62e9109a] Deleting contents of the VM from datastore datastore2 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1203.254672] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-9869ad3b-ff07-4839-acd3-a83e20c15f1b tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Deleting the datastore file [datastore2] c9e07a6c-fd14-4071-8c69-6ece62e9109a {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1203.254933] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-9cf53f9e-6029-432f-81b0-45e65636a840 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1203.261752] env[63515]: DEBUG oslo_vmware.api [None req-9869ad3b-ff07-4839-acd3-a83e20c15f1b tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Waiting for the task: (returnval){ [ 1203.261752] env[63515]: value = "task-1112066" [ 1203.261752] env[63515]: _type = "Task" [ 1203.261752] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1203.268970] env[63515]: DEBUG oslo_vmware.api [None req-9869ad3b-ff07-4839-acd3-a83e20c15f1b tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Task: {'id': task-1112066, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1203.531096] env[63515]: DEBUG oslo_service.periodic_task [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=63515) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1203.534743] env[63515]: DEBUG oslo_service.periodic_task [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=63515) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1203.681407] env[63515]: DEBUG oslo_concurrency.lockutils [None req-c6329246-02d8-41a2-98f4-7ff34c5957c8 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Releasing lock "refresh_cache-24a79a6c-5210-4778-a6ed-482bfe14933c" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1203.772315] env[63515]: DEBUG oslo_vmware.api [None req-9869ad3b-ff07-4839-acd3-a83e20c15f1b tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Task: {'id': task-1112066, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.122405} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1203.772594] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-9869ad3b-ff07-4839-acd3-a83e20c15f1b tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Deleted the datastore file {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1203.772797] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-9869ad3b-ff07-4839-acd3-a83e20c15f1b tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] [instance: c9e07a6c-fd14-4071-8c69-6ece62e9109a] Deleted contents of the VM from datastore datastore2 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1203.772978] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-9869ad3b-ff07-4839-acd3-a83e20c15f1b tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] [instance: c9e07a6c-fd14-4071-8c69-6ece62e9109a] Instance destroyed {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1203.773172] env[63515]: INFO nova.compute.manager [None req-9869ad3b-ff07-4839-acd3-a83e20c15f1b tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] [instance: c9e07a6c-fd14-4071-8c69-6ece62e9109a] Took 1.10 seconds to destroy the instance on the hypervisor. [ 1203.773413] env[63515]: DEBUG oslo.service.loopingcall [None req-9869ad3b-ff07-4839-acd3-a83e20c15f1b tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1203.773606] env[63515]: DEBUG nova.compute.manager [-] [instance: c9e07a6c-fd14-4071-8c69-6ece62e9109a] Deallocating network for instance {{(pid=63515) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1203.773701] env[63515]: DEBUG nova.network.neutron [-] [instance: c9e07a6c-fd14-4071-8c69-6ece62e9109a] deallocate_for_instance() {{(pid=63515) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1203.932169] env[63515]: DEBUG nova.compute.manager [req-38103527-bd07-4e2a-829f-0e6ba321decd req-14e3c7df-2e26-4ee1-998c-452f11bf72fe service nova] [instance: 24a79a6c-5210-4778-a6ed-482bfe14933c] Received event network-vif-unplugged-f338b564-7636-4839-b034-73c3233d7da6 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1203.932409] env[63515]: DEBUG oslo_concurrency.lockutils [req-38103527-bd07-4e2a-829f-0e6ba321decd req-14e3c7df-2e26-4ee1-998c-452f11bf72fe service nova] Acquiring lock "24a79a6c-5210-4778-a6ed-482bfe14933c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1203.932634] env[63515]: DEBUG oslo_concurrency.lockutils [req-38103527-bd07-4e2a-829f-0e6ba321decd req-14e3c7df-2e26-4ee1-998c-452f11bf72fe service nova] Lock "24a79a6c-5210-4778-a6ed-482bfe14933c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1203.936013] env[63515]: DEBUG oslo_concurrency.lockutils [req-38103527-bd07-4e2a-829f-0e6ba321decd req-14e3c7df-2e26-4ee1-998c-452f11bf72fe service nova] Lock "24a79a6c-5210-4778-a6ed-482bfe14933c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1203.936769] env[63515]: DEBUG nova.compute.manager [req-38103527-bd07-4e2a-829f-0e6ba321decd req-14e3c7df-2e26-4ee1-998c-452f11bf72fe service nova] [instance: 24a79a6c-5210-4778-a6ed-482bfe14933c] No waiting events found dispatching network-vif-unplugged-f338b564-7636-4839-b034-73c3233d7da6 {{(pid=63515) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1203.936769] env[63515]: WARNING nova.compute.manager [req-38103527-bd07-4e2a-829f-0e6ba321decd req-14e3c7df-2e26-4ee1-998c-452f11bf72fe service nova] [instance: 24a79a6c-5210-4778-a6ed-482bfe14933c] Received unexpected event network-vif-unplugged-f338b564-7636-4839-b034-73c3233d7da6 for instance with vm_state shelved and task_state shelving_offloading. [ 1204.008581] env[63515]: DEBUG nova.compute.manager [req-2d886c02-1ecc-418f-96a3-9774916db591 req-1cb83427-0810-4cfe-b6a1-2c1eecc617c0 service nova] [instance: c9e07a6c-fd14-4071-8c69-6ece62e9109a] Received event network-vif-deleted-bd316695-ec41-4f7d-b042-5cd5d319aa55 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1204.008891] env[63515]: INFO nova.compute.manager [req-2d886c02-1ecc-418f-96a3-9774916db591 req-1cb83427-0810-4cfe-b6a1-2c1eecc617c0 service nova] [instance: c9e07a6c-fd14-4071-8c69-6ece62e9109a] Neutron deleted interface bd316695-ec41-4f7d-b042-5cd5d319aa55; detaching it from the instance and deleting it from the info cache [ 1204.008988] env[63515]: DEBUG nova.network.neutron [req-2d886c02-1ecc-418f-96a3-9774916db591 req-1cb83427-0810-4cfe-b6a1-2c1eecc617c0 service nova] [instance: c9e07a6c-fd14-4071-8c69-6ece62e9109a] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1204.019987] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-c6329246-02d8-41a2-98f4-7ff34c5957c8 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 24a79a6c-5210-4778-a6ed-482bfe14933c] Destroying instance {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1204.020934] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1a9b48b-e8e4-4fb5-8308-914301bdda31 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1204.028931] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-c6329246-02d8-41a2-98f4-7ff34c5957c8 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 24a79a6c-5210-4778-a6ed-482bfe14933c] Unregistering the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1204.029212] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-232bb8ec-9f42-40a7-83e6-f8994af8eeb1 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1204.094608] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-c6329246-02d8-41a2-98f4-7ff34c5957c8 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 24a79a6c-5210-4778-a6ed-482bfe14933c] Unregistered the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1204.094907] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-c6329246-02d8-41a2-98f4-7ff34c5957c8 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 24a79a6c-5210-4778-a6ed-482bfe14933c] Deleting contents of the VM from datastore datastore2 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1204.095021] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-c6329246-02d8-41a2-98f4-7ff34c5957c8 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Deleting the datastore file [datastore2] 24a79a6c-5210-4778-a6ed-482bfe14933c {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1204.095320] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-afc820dc-fcc6-4a80-a4eb-271dfe1328f6 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1204.103019] env[63515]: DEBUG oslo_vmware.api [None req-c6329246-02d8-41a2-98f4-7ff34c5957c8 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Waiting for the task: (returnval){ [ 1204.103019] env[63515]: value = "task-1112068" [ 1204.103019] env[63515]: _type = "Task" [ 1204.103019] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1204.110168] env[63515]: DEBUG oslo_vmware.api [None req-c6329246-02d8-41a2-98f4-7ff34c5957c8 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Task: {'id': task-1112068, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1204.486982] env[63515]: DEBUG nova.network.neutron [-] [instance: c9e07a6c-fd14-4071-8c69-6ece62e9109a] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1204.512064] env[63515]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-99a1f27e-f74e-45c3-ba40-1d4e52e23a82 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1204.520668] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7331458-04c8-4224-96d1-4fe9fd7b0c68 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1204.534907] env[63515]: DEBUG oslo_service.periodic_task [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=63515) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1204.535125] env[63515]: DEBUG nova.compute.manager [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Starting heal instance info cache {{(pid=63515) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9926}} [ 1204.544572] env[63515]: DEBUG nova.compute.manager [req-2d886c02-1ecc-418f-96a3-9774916db591 req-1cb83427-0810-4cfe-b6a1-2c1eecc617c0 service nova] [instance: c9e07a6c-fd14-4071-8c69-6ece62e9109a] Detach interface failed, port_id=bd316695-ec41-4f7d-b042-5cd5d319aa55, reason: Instance c9e07a6c-fd14-4071-8c69-6ece62e9109a could not be found. {{(pid=63515) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1204.612272] env[63515]: DEBUG oslo_vmware.api [None req-c6329246-02d8-41a2-98f4-7ff34c5957c8 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Task: {'id': task-1112068, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.126557} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1204.612531] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-c6329246-02d8-41a2-98f4-7ff34c5957c8 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Deleted the datastore file {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1204.612741] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-c6329246-02d8-41a2-98f4-7ff34c5957c8 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 24a79a6c-5210-4778-a6ed-482bfe14933c] Deleted contents of the VM from datastore datastore2 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1204.612932] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-c6329246-02d8-41a2-98f4-7ff34c5957c8 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 24a79a6c-5210-4778-a6ed-482bfe14933c] Instance destroyed {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1204.635530] env[63515]: INFO nova.scheduler.client.report [None req-c6329246-02d8-41a2-98f4-7ff34c5957c8 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Deleted allocations for instance 24a79a6c-5210-4778-a6ed-482bfe14933c [ 1204.990225] env[63515]: INFO nova.compute.manager [-] [instance: c9e07a6c-fd14-4071-8c69-6ece62e9109a] Took 1.22 seconds to deallocate network for instance. [ 1205.140099] env[63515]: DEBUG oslo_concurrency.lockutils [None req-c6329246-02d8-41a2-98f4-7ff34c5957c8 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1205.140387] env[63515]: DEBUG oslo_concurrency.lockutils [None req-c6329246-02d8-41a2-98f4-7ff34c5957c8 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1205.140628] env[63515]: DEBUG nova.objects.instance [None req-c6329246-02d8-41a2-98f4-7ff34c5957c8 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Lazy-loading 'resources' on Instance uuid 24a79a6c-5210-4778-a6ed-482bfe14933c {{(pid=63515) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1205.318891] env[63515]: DEBUG oslo_concurrency.lockutils [None req-d016eeb6-236f-4833-8110-8ea8364ec248 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Acquiring lock "bd3a77ee-a78b-4719-9353-307ae8b166b6" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1205.319137] env[63515]: DEBUG oslo_concurrency.lockutils [None req-d016eeb6-236f-4833-8110-8ea8364ec248 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Lock "bd3a77ee-a78b-4719-9353-307ae8b166b6" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1205.319361] env[63515]: DEBUG oslo_concurrency.lockutils [None req-d016eeb6-236f-4833-8110-8ea8364ec248 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Acquiring lock "bd3a77ee-a78b-4719-9353-307ae8b166b6-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1205.319553] env[63515]: DEBUG oslo_concurrency.lockutils [None req-d016eeb6-236f-4833-8110-8ea8364ec248 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Lock "bd3a77ee-a78b-4719-9353-307ae8b166b6-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1205.319785] env[63515]: DEBUG oslo_concurrency.lockutils [None req-d016eeb6-236f-4833-8110-8ea8364ec248 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Lock "bd3a77ee-a78b-4719-9353-307ae8b166b6-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1205.322380] env[63515]: INFO nova.compute.manager [None req-d016eeb6-236f-4833-8110-8ea8364ec248 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] [instance: bd3a77ee-a78b-4719-9353-307ae8b166b6] Terminating instance [ 1205.324274] env[63515]: DEBUG nova.compute.manager [None req-d016eeb6-236f-4833-8110-8ea8364ec248 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] [instance: bd3a77ee-a78b-4719-9353-307ae8b166b6] Start destroying the instance on the hypervisor. {{(pid=63515) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1205.324478] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-d016eeb6-236f-4833-8110-8ea8364ec248 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] [instance: bd3a77ee-a78b-4719-9353-307ae8b166b6] Destroying instance {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1205.325339] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d0404f9-20b8-47a4-893a-3a4899a69d76 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1205.333358] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-d016eeb6-236f-4833-8110-8ea8364ec248 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] [instance: bd3a77ee-a78b-4719-9353-307ae8b166b6] Powering off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1205.333582] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-fe3ce748-1e88-4f43-8374-2329f31e3da5 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1205.340285] env[63515]: DEBUG oslo_vmware.api [None req-d016eeb6-236f-4833-8110-8ea8364ec248 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Waiting for the task: (returnval){ [ 1205.340285] env[63515]: value = "task-1112069" [ 1205.340285] env[63515]: _type = "Task" [ 1205.340285] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1205.348316] env[63515]: DEBUG oslo_vmware.api [None req-d016eeb6-236f-4833-8110-8ea8364ec248 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Task: {'id': task-1112069, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1205.497289] env[63515]: DEBUG oslo_concurrency.lockutils [None req-9869ad3b-ff07-4839-acd3-a83e20c15f1b tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1205.568221] env[63515]: DEBUG oslo_concurrency.lockutils [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Acquiring lock "refresh_cache-c8dbf542-70cb-45be-9308-5e12307c0be6" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1205.568382] env[63515]: DEBUG oslo_concurrency.lockutils [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Acquired lock "refresh_cache-c8dbf542-70cb-45be-9308-5e12307c0be6" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1205.568548] env[63515]: DEBUG nova.network.neutron [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] [instance: c8dbf542-70cb-45be-9308-5e12307c0be6] Forcefully refreshing network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2004}} [ 1205.643038] env[63515]: DEBUG nova.objects.instance [None req-c6329246-02d8-41a2-98f4-7ff34c5957c8 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Lazy-loading 'numa_topology' on Instance uuid 24a79a6c-5210-4778-a6ed-482bfe14933c {{(pid=63515) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1205.850727] env[63515]: DEBUG oslo_vmware.api [None req-d016eeb6-236f-4833-8110-8ea8364ec248 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Task: {'id': task-1112069, 'name': PowerOffVM_Task, 'duration_secs': 0.178136} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1205.851048] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-d016eeb6-236f-4833-8110-8ea8364ec248 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] [instance: bd3a77ee-a78b-4719-9353-307ae8b166b6] Powered off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1205.851236] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-d016eeb6-236f-4833-8110-8ea8364ec248 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] [instance: bd3a77ee-a78b-4719-9353-307ae8b166b6] Unregistering the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1205.851487] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-db2750e8-b70d-4442-96ec-2b05bd663649 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1205.905490] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-d016eeb6-236f-4833-8110-8ea8364ec248 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] [instance: bd3a77ee-a78b-4719-9353-307ae8b166b6] Unregistered the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1205.905753] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-d016eeb6-236f-4833-8110-8ea8364ec248 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] [instance: bd3a77ee-a78b-4719-9353-307ae8b166b6] Deleting contents of the VM from datastore datastore2 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1205.905963] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-d016eeb6-236f-4833-8110-8ea8364ec248 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Deleting the datastore file [datastore2] bd3a77ee-a78b-4719-9353-307ae8b166b6 {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1205.906245] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-679b1ed9-980f-453d-a931-48889848ed7a {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1205.912674] env[63515]: DEBUG oslo_vmware.api [None req-d016eeb6-236f-4833-8110-8ea8364ec248 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Waiting for the task: (returnval){ [ 1205.912674] env[63515]: value = "task-1112071" [ 1205.912674] env[63515]: _type = "Task" [ 1205.912674] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1205.920137] env[63515]: DEBUG oslo_vmware.api [None req-d016eeb6-236f-4833-8110-8ea8364ec248 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Task: {'id': task-1112071, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1205.957684] env[63515]: DEBUG nova.compute.manager [req-790e1af7-8923-401d-aef7-020fe28b1c3a req-ae3e6fab-4400-499c-b743-7e20323a90bf service nova] [instance: 24a79a6c-5210-4778-a6ed-482bfe14933c] Received event network-changed-f338b564-7636-4839-b034-73c3233d7da6 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1205.957868] env[63515]: DEBUG nova.compute.manager [req-790e1af7-8923-401d-aef7-020fe28b1c3a req-ae3e6fab-4400-499c-b743-7e20323a90bf service nova] [instance: 24a79a6c-5210-4778-a6ed-482bfe14933c] Refreshing instance network info cache due to event network-changed-f338b564-7636-4839-b034-73c3233d7da6. {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1205.958097] env[63515]: DEBUG oslo_concurrency.lockutils [req-790e1af7-8923-401d-aef7-020fe28b1c3a req-ae3e6fab-4400-499c-b743-7e20323a90bf service nova] Acquiring lock "refresh_cache-24a79a6c-5210-4778-a6ed-482bfe14933c" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1205.958245] env[63515]: DEBUG oslo_concurrency.lockutils [req-790e1af7-8923-401d-aef7-020fe28b1c3a req-ae3e6fab-4400-499c-b743-7e20323a90bf service nova] Acquired lock "refresh_cache-24a79a6c-5210-4778-a6ed-482bfe14933c" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1205.958411] env[63515]: DEBUG nova.network.neutron [req-790e1af7-8923-401d-aef7-020fe28b1c3a req-ae3e6fab-4400-499c-b743-7e20323a90bf service nova] [instance: 24a79a6c-5210-4778-a6ed-482bfe14933c] Refreshing network info cache for port f338b564-7636-4839-b034-73c3233d7da6 {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1206.147644] env[63515]: DEBUG nova.objects.base [None req-c6329246-02d8-41a2-98f4-7ff34c5957c8 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Object Instance<24a79a6c-5210-4778-a6ed-482bfe14933c> lazy-loaded attributes: resources,numa_topology {{(pid=63515) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1206.216229] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4850525-ad9a-4733-b0c9-84140fbad8d5 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1206.223847] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4393651-abd5-4fe8-818d-8b14d0ea64ef {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1206.254323] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b2175bd-d08e-433a-b667-3ba9896ee879 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1206.261359] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8fe35fc-75ce-4dd0-ab74-f0f31b504ab3 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1206.278019] env[63515]: DEBUG nova.compute.provider_tree [None req-c6329246-02d8-41a2-98f4-7ff34c5957c8 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1206.422145] env[63515]: DEBUG oslo_vmware.api [None req-d016eeb6-236f-4833-8110-8ea8364ec248 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Task: {'id': task-1112071, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.136368} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1206.422429] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-d016eeb6-236f-4833-8110-8ea8364ec248 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Deleted the datastore file {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1206.422624] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-d016eeb6-236f-4833-8110-8ea8364ec248 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] [instance: bd3a77ee-a78b-4719-9353-307ae8b166b6] Deleted contents of the VM from datastore datastore2 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1206.422809] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-d016eeb6-236f-4833-8110-8ea8364ec248 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] [instance: bd3a77ee-a78b-4719-9353-307ae8b166b6] Instance destroyed {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1206.422982] env[63515]: INFO nova.compute.manager [None req-d016eeb6-236f-4833-8110-8ea8364ec248 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] [instance: bd3a77ee-a78b-4719-9353-307ae8b166b6] Took 1.10 seconds to destroy the instance on the hypervisor. [ 1206.423245] env[63515]: DEBUG oslo.service.loopingcall [None req-d016eeb6-236f-4833-8110-8ea8364ec248 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1206.423437] env[63515]: DEBUG nova.compute.manager [-] [instance: bd3a77ee-a78b-4719-9353-307ae8b166b6] Deallocating network for instance {{(pid=63515) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1206.423533] env[63515]: DEBUG nova.network.neutron [-] [instance: bd3a77ee-a78b-4719-9353-307ae8b166b6] deallocate_for_instance() {{(pid=63515) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1206.537211] env[63515]: DEBUG oslo_concurrency.lockutils [None req-e870b2fc-bda2-43f3-9911-405eca0785b8 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Acquiring lock "e13da90c-28e6-43d2-99b7-19c5095954ca" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1206.537448] env[63515]: DEBUG oslo_concurrency.lockutils [None req-e870b2fc-bda2-43f3-9911-405eca0785b8 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Lock "e13da90c-28e6-43d2-99b7-19c5095954ca" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1206.537662] env[63515]: DEBUG oslo_concurrency.lockutils [None req-e870b2fc-bda2-43f3-9911-405eca0785b8 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Acquiring lock "e13da90c-28e6-43d2-99b7-19c5095954ca-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1206.537872] env[63515]: DEBUG oslo_concurrency.lockutils [None req-e870b2fc-bda2-43f3-9911-405eca0785b8 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Lock "e13da90c-28e6-43d2-99b7-19c5095954ca-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1206.538198] env[63515]: DEBUG oslo_concurrency.lockutils [None req-e870b2fc-bda2-43f3-9911-405eca0785b8 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Lock "e13da90c-28e6-43d2-99b7-19c5095954ca-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1206.540255] env[63515]: INFO nova.compute.manager [None req-e870b2fc-bda2-43f3-9911-405eca0785b8 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: e13da90c-28e6-43d2-99b7-19c5095954ca] Terminating instance [ 1206.542091] env[63515]: DEBUG nova.compute.manager [None req-e870b2fc-bda2-43f3-9911-405eca0785b8 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: e13da90c-28e6-43d2-99b7-19c5095954ca] Start destroying the instance on the hypervisor. {{(pid=63515) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1206.542292] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-e870b2fc-bda2-43f3-9911-405eca0785b8 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: e13da90c-28e6-43d2-99b7-19c5095954ca] Destroying instance {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1206.543110] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8aa20c43-8d76-4b63-9143-4bdcb25d976e {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1206.550712] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-e870b2fc-bda2-43f3-9911-405eca0785b8 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: e13da90c-28e6-43d2-99b7-19c5095954ca] Powering off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1206.552897] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-42330615-ab71-4e36-b000-e53ff2f9cd6f {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1206.559725] env[63515]: DEBUG oslo_vmware.api [None req-e870b2fc-bda2-43f3-9911-405eca0785b8 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Waiting for the task: (returnval){ [ 1206.559725] env[63515]: value = "task-1112072" [ 1206.559725] env[63515]: _type = "Task" [ 1206.559725] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1206.566888] env[63515]: DEBUG oslo_vmware.api [None req-e870b2fc-bda2-43f3-9911-405eca0785b8 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Task: {'id': task-1112072, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1206.688132] env[63515]: DEBUG nova.network.neutron [req-790e1af7-8923-401d-aef7-020fe28b1c3a req-ae3e6fab-4400-499c-b743-7e20323a90bf service nova] [instance: 24a79a6c-5210-4778-a6ed-482bfe14933c] Updated VIF entry in instance network info cache for port f338b564-7636-4839-b034-73c3233d7da6. {{(pid=63515) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1206.688524] env[63515]: DEBUG nova.network.neutron [req-790e1af7-8923-401d-aef7-020fe28b1c3a req-ae3e6fab-4400-499c-b743-7e20323a90bf service nova] [instance: 24a79a6c-5210-4778-a6ed-482bfe14933c] Updating instance_info_cache with network_info: [{"id": "f338b564-7636-4839-b034-73c3233d7da6", "address": "fa:16:3e:bc:5a:12", "network": {"id": "5758c885-e1c9-45c8-86aa-27a802101f52", "bridge": null, "label": "tempest-ServerActionsTestOtherB-734458138-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.204", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b18d4daacbc84e758f9f9ae4f3f09d28", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tapf338b564-76", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1206.770559] env[63515]: DEBUG oslo_concurrency.lockutils [None req-8fa768a4-1e0e-4664-bc77-1d533a4f42d6 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Acquiring lock "24a79a6c-5210-4778-a6ed-482bfe14933c" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1206.781553] env[63515]: DEBUG nova.scheduler.client.report [None req-c6329246-02d8-41a2-98f4-7ff34c5957c8 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1207.069354] env[63515]: DEBUG oslo_vmware.api [None req-e870b2fc-bda2-43f3-9911-405eca0785b8 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Task: {'id': task-1112072, 'name': PowerOffVM_Task, 'duration_secs': 0.217666} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1207.069496] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-e870b2fc-bda2-43f3-9911-405eca0785b8 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: e13da90c-28e6-43d2-99b7-19c5095954ca] Powered off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1207.069687] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-e870b2fc-bda2-43f3-9911-405eca0785b8 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: e13da90c-28e6-43d2-99b7-19c5095954ca] Unregistering the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1207.069998] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-18a3ac12-21db-437b-9221-f25fa899168a {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1207.134458] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-e870b2fc-bda2-43f3-9911-405eca0785b8 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: e13da90c-28e6-43d2-99b7-19c5095954ca] Unregistered the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1207.134810] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-e870b2fc-bda2-43f3-9911-405eca0785b8 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: e13da90c-28e6-43d2-99b7-19c5095954ca] Deleting contents of the VM from datastore datastore1 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1207.134907] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-e870b2fc-bda2-43f3-9911-405eca0785b8 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Deleting the datastore file [datastore1] e13da90c-28e6-43d2-99b7-19c5095954ca {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1207.135148] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-993c209f-84e6-49a1-ac17-20bdec31bb3d {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1207.141884] env[63515]: DEBUG oslo_vmware.api [None req-e870b2fc-bda2-43f3-9911-405eca0785b8 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Waiting for the task: (returnval){ [ 1207.141884] env[63515]: value = "task-1112074" [ 1207.141884] env[63515]: _type = "Task" [ 1207.141884] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1207.149793] env[63515]: DEBUG oslo_vmware.api [None req-e870b2fc-bda2-43f3-9911-405eca0785b8 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Task: {'id': task-1112074, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1207.182782] env[63515]: DEBUG nova.network.neutron [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] [instance: c8dbf542-70cb-45be-9308-5e12307c0be6] Updating instance_info_cache with network_info: [{"id": "5569cb8c-d9c8-4cb5-abdc-5890ad7d7ede", "address": "fa:16:3e:8d:b8:b1", "network": {"id": "d6da2650-9ba9-4ef6-8035-e58b635a64b2", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-391564968-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.243", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "97f2e119708746038f38359572c99438", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "043ca97b-0fca-4b54-8be6-027123fa76d0", "external-id": "nsx-vlan-transportzone-702", "segmentation_id": 702, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5569cb8c-d9", "ovs_interfaceid": "5569cb8c-d9c8-4cb5-abdc-5890ad7d7ede", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1207.193061] env[63515]: DEBUG oslo_concurrency.lockutils [req-790e1af7-8923-401d-aef7-020fe28b1c3a req-ae3e6fab-4400-499c-b743-7e20323a90bf service nova] Releasing lock "refresh_cache-24a79a6c-5210-4778-a6ed-482bfe14933c" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1207.288526] env[63515]: DEBUG oslo_concurrency.lockutils [None req-c6329246-02d8-41a2-98f4-7ff34c5957c8 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.148s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1207.291013] env[63515]: DEBUG oslo_concurrency.lockutils [None req-9869ad3b-ff07-4839-acd3-a83e20c15f1b tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.794s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1207.291348] env[63515]: DEBUG nova.objects.instance [None req-9869ad3b-ff07-4839-acd3-a83e20c15f1b tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Lazy-loading 'resources' on Instance uuid c9e07a6c-fd14-4071-8c69-6ece62e9109a {{(pid=63515) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1207.651954] env[63515]: DEBUG oslo_vmware.api [None req-e870b2fc-bda2-43f3-9911-405eca0785b8 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Task: {'id': task-1112074, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.129027} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1207.652362] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-e870b2fc-bda2-43f3-9911-405eca0785b8 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Deleted the datastore file {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1207.652565] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-e870b2fc-bda2-43f3-9911-405eca0785b8 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: e13da90c-28e6-43d2-99b7-19c5095954ca] Deleted contents of the VM from datastore datastore1 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1207.652749] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-e870b2fc-bda2-43f3-9911-405eca0785b8 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: e13da90c-28e6-43d2-99b7-19c5095954ca] Instance destroyed {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1207.652967] env[63515]: INFO nova.compute.manager [None req-e870b2fc-bda2-43f3-9911-405eca0785b8 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] [instance: e13da90c-28e6-43d2-99b7-19c5095954ca] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1207.653192] env[63515]: DEBUG oslo.service.loopingcall [None req-e870b2fc-bda2-43f3-9911-405eca0785b8 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1207.653398] env[63515]: DEBUG nova.compute.manager [-] [instance: e13da90c-28e6-43d2-99b7-19c5095954ca] Deallocating network for instance {{(pid=63515) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1207.653494] env[63515]: DEBUG nova.network.neutron [-] [instance: e13da90c-28e6-43d2-99b7-19c5095954ca] deallocate_for_instance() {{(pid=63515) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1207.685513] env[63515]: DEBUG oslo_concurrency.lockutils [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Releasing lock "refresh_cache-c8dbf542-70cb-45be-9308-5e12307c0be6" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1207.688028] env[63515]: DEBUG nova.compute.manager [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] [instance: c8dbf542-70cb-45be-9308-5e12307c0be6] Updated the network info_cache for instance {{(pid=63515) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9997}} [ 1207.688028] env[63515]: DEBUG oslo_service.periodic_task [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=63515) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1207.688028] env[63515]: DEBUG oslo_service.periodic_task [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Running periodic task ComputeManager.update_available_resource {{(pid=63515) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1207.704519] env[63515]: DEBUG nova.network.neutron [-] [instance: bd3a77ee-a78b-4719-9353-307ae8b166b6] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1207.799768] env[63515]: DEBUG oslo_concurrency.lockutils [None req-c6329246-02d8-41a2-98f4-7ff34c5957c8 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Lock "24a79a6c-5210-4778-a6ed-482bfe14933c" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 22.160s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1207.800797] env[63515]: DEBUG oslo_concurrency.lockutils [None req-8fa768a4-1e0e-4664-bc77-1d533a4f42d6 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Lock "24a79a6c-5210-4778-a6ed-482bfe14933c" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 1.030s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1207.801087] env[63515]: INFO nova.compute.manager [None req-8fa768a4-1e0e-4664-bc77-1d533a4f42d6 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 24a79a6c-5210-4778-a6ed-482bfe14933c] Unshelving [ 1207.875299] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc8bac48-eb0c-435a-a001-006cc8207667 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1207.884754] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-733cda56-769b-4578-ac9b-7e6e0e1afdf3 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1207.914010] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48f8841b-fe8d-442c-8cd6-b4e473b6adf2 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1207.921222] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2560c1c-35fa-44f8-a243-33c221713903 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1207.934453] env[63515]: DEBUG nova.compute.provider_tree [None req-9869ad3b-ff07-4839-acd3-a83e20c15f1b tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1207.986274] env[63515]: DEBUG nova.compute.manager [req-03b31ccb-97ad-4b3b-abae-b7c1f3d7da2c req-326ba9f6-d577-44a0-abcb-388e63e0c7a1 service nova] [instance: bd3a77ee-a78b-4719-9353-307ae8b166b6] Received event network-vif-deleted-764f6a9d-e89a-44b6-804c-70000c555f95 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1207.986482] env[63515]: DEBUG nova.compute.manager [req-03b31ccb-97ad-4b3b-abae-b7c1f3d7da2c req-326ba9f6-d577-44a0-abcb-388e63e0c7a1 service nova] [instance: e13da90c-28e6-43d2-99b7-19c5095954ca] Received event network-vif-deleted-d9363b80-630e-4262-901f-c13e5c8b8c2a {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1207.986678] env[63515]: INFO nova.compute.manager [req-03b31ccb-97ad-4b3b-abae-b7c1f3d7da2c req-326ba9f6-d577-44a0-abcb-388e63e0c7a1 service nova] [instance: e13da90c-28e6-43d2-99b7-19c5095954ca] Neutron deleted interface d9363b80-630e-4262-901f-c13e5c8b8c2a; detaching it from the instance and deleting it from the info cache [ 1207.986853] env[63515]: DEBUG nova.network.neutron [req-03b31ccb-97ad-4b3b-abae-b7c1f3d7da2c req-326ba9f6-d577-44a0-abcb-388e63e0c7a1 service nova] [instance: e13da90c-28e6-43d2-99b7-19c5095954ca] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1208.190206] env[63515]: DEBUG oslo_concurrency.lockutils [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1208.206974] env[63515]: INFO nova.compute.manager [-] [instance: bd3a77ee-a78b-4719-9353-307ae8b166b6] Took 1.78 seconds to deallocate network for instance. [ 1208.385175] env[63515]: DEBUG nova.network.neutron [-] [instance: e13da90c-28e6-43d2-99b7-19c5095954ca] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1208.437607] env[63515]: DEBUG nova.scheduler.client.report [None req-9869ad3b-ff07-4839-acd3-a83e20c15f1b tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1208.490805] env[63515]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-bd0cb685-ebd7-4ccc-9281-12e4fb819361 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1208.500520] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c68a166-025f-49e1-93ba-f01fa9bd9768 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1208.523502] env[63515]: DEBUG nova.compute.manager [req-03b31ccb-97ad-4b3b-abae-b7c1f3d7da2c req-326ba9f6-d577-44a0-abcb-388e63e0c7a1 service nova] [instance: e13da90c-28e6-43d2-99b7-19c5095954ca] Detach interface failed, port_id=d9363b80-630e-4262-901f-c13e5c8b8c2a, reason: Instance e13da90c-28e6-43d2-99b7-19c5095954ca could not be found. {{(pid=63515) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1208.714099] env[63515]: DEBUG oslo_concurrency.lockutils [None req-d016eeb6-236f-4833-8110-8ea8364ec248 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1208.828854] env[63515]: DEBUG oslo_concurrency.lockutils [None req-8fa768a4-1e0e-4664-bc77-1d533a4f42d6 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1208.887542] env[63515]: INFO nova.compute.manager [-] [instance: e13da90c-28e6-43d2-99b7-19c5095954ca] Took 1.23 seconds to deallocate network for instance. [ 1208.942198] env[63515]: DEBUG oslo_concurrency.lockutils [None req-9869ad3b-ff07-4839-acd3-a83e20c15f1b tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.651s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1208.944304] env[63515]: DEBUG oslo_concurrency.lockutils [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.754s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1208.944468] env[63515]: DEBUG oslo_concurrency.lockutils [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1208.944621] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=63515) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1208.944994] env[63515]: DEBUG oslo_concurrency.lockutils [None req-d016eeb6-236f-4833-8110-8ea8364ec248 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.231s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1208.945176] env[63515]: DEBUG nova.objects.instance [None req-d016eeb6-236f-4833-8110-8ea8364ec248 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Lazy-loading 'resources' on Instance uuid bd3a77ee-a78b-4719-9353-307ae8b166b6 {{(pid=63515) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1208.947030] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f825d43-ebfa-416b-80e7-df52587f6865 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1208.955209] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b26937e-a7f6-46c9-b3cd-b8b36c58de03 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1208.959722] env[63515]: INFO nova.scheduler.client.report [None req-9869ad3b-ff07-4839-acd3-a83e20c15f1b tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Deleted allocations for instance c9e07a6c-fd14-4071-8c69-6ece62e9109a [ 1208.972515] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c922bf67-f622-4c09-9e26-25337a119b82 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1208.978961] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c9b21ca-6220-45f9-8d47-205dbf1f2181 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1209.009516] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180573MB free_disk=171GB free_vcpus=48 pci_devices=None {{(pid=63515) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1209.009729] env[63515]: DEBUG oslo_concurrency.lockutils [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1209.394501] env[63515]: DEBUG oslo_concurrency.lockutils [None req-e870b2fc-bda2-43f3-9911-405eca0785b8 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1209.469364] env[63515]: DEBUG oslo_concurrency.lockutils [None req-9869ad3b-ff07-4839-acd3-a83e20c15f1b tempest-ServersNegativeTestJSON-1276404621 tempest-ServersNegativeTestJSON-1276404621-project-member] Lock "c9e07a6c-fd14-4071-8c69-6ece62e9109a" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.805s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1209.512970] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79feda04-349f-4824-a298-910bd892a779 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1209.520915] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69e39b75-0aa6-4ec5-bd3d-33a2948f941d {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1209.549504] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12695b36-cb1b-47e4-a89c-46b690554fed {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1209.556099] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46a296ae-a3d7-4295-bee5-501fec7f7cf6 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1209.568662] env[63515]: DEBUG nova.compute.provider_tree [None req-d016eeb6-236f-4833-8110-8ea8364ec248 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1210.072083] env[63515]: DEBUG nova.scheduler.client.report [None req-d016eeb6-236f-4833-8110-8ea8364ec248 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1210.578087] env[63515]: DEBUG oslo_concurrency.lockutils [None req-d016eeb6-236f-4833-8110-8ea8364ec248 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.633s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1210.580325] env[63515]: DEBUG oslo_concurrency.lockutils [None req-8fa768a4-1e0e-4664-bc77-1d533a4f42d6 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.752s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1210.580554] env[63515]: DEBUG nova.objects.instance [None req-8fa768a4-1e0e-4664-bc77-1d533a4f42d6 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Lazy-loading 'pci_requests' on Instance uuid 24a79a6c-5210-4778-a6ed-482bfe14933c {{(pid=63515) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1210.601605] env[63515]: INFO nova.scheduler.client.report [None req-d016eeb6-236f-4833-8110-8ea8364ec248 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Deleted allocations for instance bd3a77ee-a78b-4719-9353-307ae8b166b6 [ 1211.087626] env[63515]: DEBUG nova.objects.instance [None req-8fa768a4-1e0e-4664-bc77-1d533a4f42d6 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Lazy-loading 'numa_topology' on Instance uuid 24a79a6c-5210-4778-a6ed-482bfe14933c {{(pid=63515) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1211.108970] env[63515]: DEBUG oslo_concurrency.lockutils [None req-d016eeb6-236f-4833-8110-8ea8364ec248 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Lock "bd3a77ee-a78b-4719-9353-307ae8b166b6" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.790s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1211.591735] env[63515]: INFO nova.compute.claims [None req-8fa768a4-1e0e-4664-bc77-1d533a4f42d6 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 24a79a6c-5210-4778-a6ed-482bfe14933c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1212.428559] env[63515]: DEBUG oslo_concurrency.lockutils [None req-862e3b90-0ef2-4d13-a294-feaef7d27492 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Acquiring lock "c8dbf542-70cb-45be-9308-5e12307c0be6" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1212.428921] env[63515]: DEBUG oslo_concurrency.lockutils [None req-862e3b90-0ef2-4d13-a294-feaef7d27492 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Lock "c8dbf542-70cb-45be-9308-5e12307c0be6" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.001s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1212.665201] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1d859b8-7e97-4460-9bd6-2594e00a84d7 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1212.673109] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-159a5303-3da3-496f-a86c-df63fcd0ae8f {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1212.703368] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb9fa710-ca6f-413c-b692-77dee19059fb {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1212.710613] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95a708c3-361b-4f1b-a023-a894fee5c6cd {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1212.723496] env[63515]: DEBUG nova.compute.provider_tree [None req-8fa768a4-1e0e-4664-bc77-1d533a4f42d6 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1212.932369] env[63515]: INFO nova.compute.manager [None req-862e3b90-0ef2-4d13-a294-feaef7d27492 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] [instance: c8dbf542-70cb-45be-9308-5e12307c0be6] Detaching volume b9b77a26-5a59-41f2-b991-e09802684a66 [ 1212.966913] env[63515]: INFO nova.virt.block_device [None req-862e3b90-0ef2-4d13-a294-feaef7d27492 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] [instance: c8dbf542-70cb-45be-9308-5e12307c0be6] Attempting to driver detach volume b9b77a26-5a59-41f2-b991-e09802684a66 from mountpoint /dev/sdb [ 1212.967249] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-862e3b90-0ef2-4d13-a294-feaef7d27492 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] [instance: c8dbf542-70cb-45be-9308-5e12307c0be6] Volume detach. Driver type: vmdk {{(pid=63515) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1212.967463] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-862e3b90-0ef2-4d13-a294-feaef7d27492 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] [instance: c8dbf542-70cb-45be-9308-5e12307c0be6] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-243592', 'volume_id': 'b9b77a26-5a59-41f2-b991-e09802684a66', 'name': 'volume-b9b77a26-5a59-41f2-b991-e09802684a66', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'c8dbf542-70cb-45be-9308-5e12307c0be6', 'attached_at': '', 'detached_at': '', 'volume_id': 'b9b77a26-5a59-41f2-b991-e09802684a66', 'serial': 'b9b77a26-5a59-41f2-b991-e09802684a66'} {{(pid=63515) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1212.968372] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25ddbac9-3d63-4d3a-9f6f-59779c4e298d {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1212.989943] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb21f2a4-c215-453c-8f88-bb33102774e6 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1212.996540] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0895bc4c-90a8-428d-8afc-93d3869381b2 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1213.015712] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef4475a8-510d-439c-86a7-c02c0338c633 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1213.029362] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-862e3b90-0ef2-4d13-a294-feaef7d27492 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] The volume has not been displaced from its original location: [datastore2] volume-b9b77a26-5a59-41f2-b991-e09802684a66/volume-b9b77a26-5a59-41f2-b991-e09802684a66.vmdk. No consolidation needed. {{(pid=63515) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1213.034656] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-862e3b90-0ef2-4d13-a294-feaef7d27492 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] [instance: c8dbf542-70cb-45be-9308-5e12307c0be6] Reconfiguring VM instance instance-0000006b to detach disk 2001 {{(pid=63515) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1213.034947] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3e05f8d0-0d27-443e-9af0-4151c974b825 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1213.052935] env[63515]: DEBUG oslo_vmware.api [None req-862e3b90-0ef2-4d13-a294-feaef7d27492 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Waiting for the task: (returnval){ [ 1213.052935] env[63515]: value = "task-1112075" [ 1213.052935] env[63515]: _type = "Task" [ 1213.052935] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1213.061496] env[63515]: DEBUG oslo_vmware.api [None req-862e3b90-0ef2-4d13-a294-feaef7d27492 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Task: {'id': task-1112075, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1213.227260] env[63515]: DEBUG nova.scheduler.client.report [None req-8fa768a4-1e0e-4664-bc77-1d533a4f42d6 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1213.562420] env[63515]: DEBUG oslo_vmware.api [None req-862e3b90-0ef2-4d13-a294-feaef7d27492 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Task: {'id': task-1112075, 'name': ReconfigVM_Task, 'duration_secs': 0.249634} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1213.562780] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-862e3b90-0ef2-4d13-a294-feaef7d27492 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] [instance: c8dbf542-70cb-45be-9308-5e12307c0be6] Reconfigured VM instance instance-0000006b to detach disk 2001 {{(pid=63515) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1213.567299] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2018db73-2f83-4bf3-b9e3-da1c7014fb04 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1213.582473] env[63515]: DEBUG oslo_vmware.api [None req-862e3b90-0ef2-4d13-a294-feaef7d27492 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Waiting for the task: (returnval){ [ 1213.582473] env[63515]: value = "task-1112076" [ 1213.582473] env[63515]: _type = "Task" [ 1213.582473] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1213.592254] env[63515]: DEBUG oslo_vmware.api [None req-862e3b90-0ef2-4d13-a294-feaef7d27492 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Task: {'id': task-1112076, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1213.732324] env[63515]: DEBUG oslo_concurrency.lockutils [None req-8fa768a4-1e0e-4664-bc77-1d533a4f42d6 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.152s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1213.734468] env[63515]: DEBUG oslo_concurrency.lockutils [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 4.725s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1213.765959] env[63515]: INFO nova.network.neutron [None req-8fa768a4-1e0e-4664-bc77-1d533a4f42d6 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 24a79a6c-5210-4778-a6ed-482bfe14933c] Updating port f338b564-7636-4839-b034-73c3233d7da6 with attributes {'binding:host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 1214.092555] env[63515]: DEBUG oslo_vmware.api [None req-862e3b90-0ef2-4d13-a294-feaef7d27492 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Task: {'id': task-1112076, 'name': ReconfigVM_Task, 'duration_secs': 0.135599} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1214.092923] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-862e3b90-0ef2-4d13-a294-feaef7d27492 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] [instance: c8dbf542-70cb-45be-9308-5e12307c0be6] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-243592', 'volume_id': 'b9b77a26-5a59-41f2-b991-e09802684a66', 'name': 'volume-b9b77a26-5a59-41f2-b991-e09802684a66', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'c8dbf542-70cb-45be-9308-5e12307c0be6', 'attached_at': '', 'detached_at': '', 'volume_id': 'b9b77a26-5a59-41f2-b991-e09802684a66', 'serial': 'b9b77a26-5a59-41f2-b991-e09802684a66'} {{(pid=63515) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1214.634100] env[63515]: DEBUG nova.objects.instance [None req-862e3b90-0ef2-4d13-a294-feaef7d27492 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Lazy-loading 'flavor' on Instance uuid c8dbf542-70cb-45be-9308-5e12307c0be6 {{(pid=63515) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1214.756181] env[63515]: WARNING nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Instance e13da90c-28e6-43d2-99b7-19c5095954ca is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 1214.756343] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Instance c8dbf542-70cb-45be-9308-5e12307c0be6 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63515) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1214.756468] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Instance 24a79a6c-5210-4778-a6ed-482bfe14933c actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63515) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1214.756636] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Total usable vcpus: 48, total allocated vcpus: 2 {{(pid=63515) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1214.756768] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=896MB phys_disk=200GB used_disk=2GB total_vcpus=48 used_vcpus=2 pci_stats=[] {{(pid=63515) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1214.802740] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa519702-f3d6-4c56-8729-6c8a18dafbc0 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1214.810382] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eba6c052-6bdd-465b-8759-89ed47b73afb {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1214.840070] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15cbe32b-96fb-487f-be9a-589881d6d243 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1214.846914] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6264bc96-e7be-4dd2-8cb5-54507347a3db {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1214.859516] env[63515]: DEBUG nova.compute.provider_tree [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1215.228070] env[63515]: DEBUG nova.compute.manager [req-3c822b6a-1fe7-4911-95ee-ae2c1985069a req-a25e2f51-c45c-42ac-bced-b9909e85a90b service nova] [instance: 24a79a6c-5210-4778-a6ed-482bfe14933c] Received event network-vif-plugged-f338b564-7636-4839-b034-73c3233d7da6 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1215.228306] env[63515]: DEBUG oslo_concurrency.lockutils [req-3c822b6a-1fe7-4911-95ee-ae2c1985069a req-a25e2f51-c45c-42ac-bced-b9909e85a90b service nova] Acquiring lock "24a79a6c-5210-4778-a6ed-482bfe14933c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1215.228511] env[63515]: DEBUG oslo_concurrency.lockutils [req-3c822b6a-1fe7-4911-95ee-ae2c1985069a req-a25e2f51-c45c-42ac-bced-b9909e85a90b service nova] Lock "24a79a6c-5210-4778-a6ed-482bfe14933c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1215.228686] env[63515]: DEBUG oslo_concurrency.lockutils [req-3c822b6a-1fe7-4911-95ee-ae2c1985069a req-a25e2f51-c45c-42ac-bced-b9909e85a90b service nova] Lock "24a79a6c-5210-4778-a6ed-482bfe14933c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1215.228870] env[63515]: DEBUG nova.compute.manager [req-3c822b6a-1fe7-4911-95ee-ae2c1985069a req-a25e2f51-c45c-42ac-bced-b9909e85a90b service nova] [instance: 24a79a6c-5210-4778-a6ed-482bfe14933c] No waiting events found dispatching network-vif-plugged-f338b564-7636-4839-b034-73c3233d7da6 {{(pid=63515) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1215.229110] env[63515]: WARNING nova.compute.manager [req-3c822b6a-1fe7-4911-95ee-ae2c1985069a req-a25e2f51-c45c-42ac-bced-b9909e85a90b service nova] [instance: 24a79a6c-5210-4778-a6ed-482bfe14933c] Received unexpected event network-vif-plugged-f338b564-7636-4839-b034-73c3233d7da6 for instance with vm_state shelved_offloaded and task_state spawning. [ 1215.309782] env[63515]: DEBUG oslo_concurrency.lockutils [None req-8fa768a4-1e0e-4664-bc77-1d533a4f42d6 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Acquiring lock "refresh_cache-24a79a6c-5210-4778-a6ed-482bfe14933c" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1215.310016] env[63515]: DEBUG oslo_concurrency.lockutils [None req-8fa768a4-1e0e-4664-bc77-1d533a4f42d6 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Acquired lock "refresh_cache-24a79a6c-5210-4778-a6ed-482bfe14933c" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1215.310211] env[63515]: DEBUG nova.network.neutron [None req-8fa768a4-1e0e-4664-bc77-1d533a4f42d6 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 24a79a6c-5210-4778-a6ed-482bfe14933c] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1215.362815] env[63515]: DEBUG nova.scheduler.client.report [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1215.641075] env[63515]: DEBUG oslo_concurrency.lockutils [None req-862e3b90-0ef2-4d13-a294-feaef7d27492 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Lock "c8dbf542-70cb-45be-9308-5e12307c0be6" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.212s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1215.867042] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=63515) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1215.867042] env[63515]: DEBUG oslo_concurrency.lockutils [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.133s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1215.867389] env[63515]: DEBUG oslo_concurrency.lockutils [None req-e870b2fc-bda2-43f3-9911-405eca0785b8 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 6.473s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1215.867627] env[63515]: DEBUG oslo_concurrency.lockutils [None req-e870b2fc-bda2-43f3-9911-405eca0785b8 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1215.891234] env[63515]: INFO nova.scheduler.client.report [None req-e870b2fc-bda2-43f3-9911-405eca0785b8 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Deleted allocations for instance e13da90c-28e6-43d2-99b7-19c5095954ca [ 1216.043946] env[63515]: DEBUG nova.network.neutron [None req-8fa768a4-1e0e-4664-bc77-1d533a4f42d6 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 24a79a6c-5210-4778-a6ed-482bfe14933c] Updating instance_info_cache with network_info: [{"id": "f338b564-7636-4839-b034-73c3233d7da6", "address": "fa:16:3e:bc:5a:12", "network": {"id": "5758c885-e1c9-45c8-86aa-27a802101f52", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-734458138-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.204", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b18d4daacbc84e758f9f9ae4f3f09d28", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69351262-8d39-441a-85ba-3a78df436d17", "external-id": "nsx-vlan-transportzone-205", "segmentation_id": 205, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf338b564-76", "ovs_interfaceid": "f338b564-7636-4839-b034-73c3233d7da6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1216.399877] env[63515]: DEBUG oslo_concurrency.lockutils [None req-e870b2fc-bda2-43f3-9911-405eca0785b8 tempest-ServerActionsTestOtherA-1321861311 tempest-ServerActionsTestOtherA-1321861311-project-member] Lock "e13da90c-28e6-43d2-99b7-19c5095954ca" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 9.862s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1216.547069] env[63515]: DEBUG oslo_concurrency.lockutils [None req-8fa768a4-1e0e-4664-bc77-1d533a4f42d6 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Releasing lock "refresh_cache-24a79a6c-5210-4778-a6ed-482bfe14933c" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1216.574762] env[63515]: DEBUG nova.virt.hardware [None req-8fa768a4-1e0e-4664-bc77-1d533a4f42d6 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T02:52:14Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='01b9b77006ee7b8779d5256dff529b29',container_format='bare',created_at=2024-10-03T03:04:55Z,direct_url=,disk_format='vmdk',id=af53d566-21dc-48bf-aacf-f6a96a111935,min_disk=1,min_ram=0,name='tempest-ServerActionsTestOtherB-server-1251030405-shelved',owner='b18d4daacbc84e758f9f9ae4f3f09d28',properties=ImageMetaProps,protected=,size=31667712,status='active',tags=,updated_at=2024-10-03T03:05:10Z,virtual_size=,visibility=), allow threads: False {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1216.575036] env[63515]: DEBUG nova.virt.hardware [None req-8fa768a4-1e0e-4664-bc77-1d533a4f42d6 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Flavor limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1216.575207] env[63515]: DEBUG nova.virt.hardware [None req-8fa768a4-1e0e-4664-bc77-1d533a4f42d6 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Image limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1216.575392] env[63515]: DEBUG nova.virt.hardware [None req-8fa768a4-1e0e-4664-bc77-1d533a4f42d6 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Flavor pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1216.575544] env[63515]: DEBUG nova.virt.hardware [None req-8fa768a4-1e0e-4664-bc77-1d533a4f42d6 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Image pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1216.575697] env[63515]: DEBUG nova.virt.hardware [None req-8fa768a4-1e0e-4664-bc77-1d533a4f42d6 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1216.575912] env[63515]: DEBUG nova.virt.hardware [None req-8fa768a4-1e0e-4664-bc77-1d533a4f42d6 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1216.576093] env[63515]: DEBUG nova.virt.hardware [None req-8fa768a4-1e0e-4664-bc77-1d533a4f42d6 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1216.576273] env[63515]: DEBUG nova.virt.hardware [None req-8fa768a4-1e0e-4664-bc77-1d533a4f42d6 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Got 1 possible topologies {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1216.576440] env[63515]: DEBUG nova.virt.hardware [None req-8fa768a4-1e0e-4664-bc77-1d533a4f42d6 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1216.576618] env[63515]: DEBUG nova.virt.hardware [None req-8fa768a4-1e0e-4664-bc77-1d533a4f42d6 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1216.577525] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5a5c687-8e87-4ae2-adf8-e20a28c44a65 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1216.585811] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-885c6e8e-4b52-4bc7-a7a2-f53752cc0f04 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1216.600766] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-8fa768a4-1e0e-4664-bc77-1d533a4f42d6 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 24a79a6c-5210-4778-a6ed-482bfe14933c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:bc:5a:12', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '69351262-8d39-441a-85ba-3a78df436d17', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f338b564-7636-4839-b034-73c3233d7da6', 'vif_model': 'vmxnet3'}] {{(pid=63515) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1216.608343] env[63515]: DEBUG oslo.service.loopingcall [None req-8fa768a4-1e0e-4664-bc77-1d533a4f42d6 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1216.608602] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 24a79a6c-5210-4778-a6ed-482bfe14933c] Creating VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1216.608818] env[63515]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-02640737-c8e2-47f3-bc31-f81bdff21c8a {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1216.627892] env[63515]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1216.627892] env[63515]: value = "task-1112077" [ 1216.627892] env[63515]: _type = "Task" [ 1216.627892] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1216.635660] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1112077, 'name': CreateVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1216.684343] env[63515]: DEBUG oslo_concurrency.lockutils [None req-dc87d49d-dcd9-4981-8dc8-1f724cc8296c tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Acquiring lock "c8dbf542-70cb-45be-9308-5e12307c0be6" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1216.684768] env[63515]: DEBUG oslo_concurrency.lockutils [None req-dc87d49d-dcd9-4981-8dc8-1f724cc8296c tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Lock "c8dbf542-70cb-45be-9308-5e12307c0be6" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1216.684986] env[63515]: DEBUG oslo_concurrency.lockutils [None req-dc87d49d-dcd9-4981-8dc8-1f724cc8296c tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Acquiring lock "c8dbf542-70cb-45be-9308-5e12307c0be6-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1216.685322] env[63515]: DEBUG oslo_concurrency.lockutils [None req-dc87d49d-dcd9-4981-8dc8-1f724cc8296c tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Lock "c8dbf542-70cb-45be-9308-5e12307c0be6-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1216.685591] env[63515]: DEBUG oslo_concurrency.lockutils [None req-dc87d49d-dcd9-4981-8dc8-1f724cc8296c tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Lock "c8dbf542-70cb-45be-9308-5e12307c0be6-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1216.688737] env[63515]: INFO nova.compute.manager [None req-dc87d49d-dcd9-4981-8dc8-1f724cc8296c tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] [instance: c8dbf542-70cb-45be-9308-5e12307c0be6] Terminating instance [ 1216.691380] env[63515]: DEBUG nova.compute.manager [None req-dc87d49d-dcd9-4981-8dc8-1f724cc8296c tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] [instance: c8dbf542-70cb-45be-9308-5e12307c0be6] Start destroying the instance on the hypervisor. {{(pid=63515) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1216.691702] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-dc87d49d-dcd9-4981-8dc8-1f724cc8296c tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] [instance: c8dbf542-70cb-45be-9308-5e12307c0be6] Destroying instance {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1216.692993] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d37ec4c-39ad-4acd-9a94-53dd3f724471 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1216.702051] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-dc87d49d-dcd9-4981-8dc8-1f724cc8296c tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] [instance: c8dbf542-70cb-45be-9308-5e12307c0be6] Powering off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1216.702395] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-071c47f2-b015-4fac-933a-13c7a0fa79e5 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1216.708639] env[63515]: DEBUG oslo_vmware.api [None req-dc87d49d-dcd9-4981-8dc8-1f724cc8296c tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Waiting for the task: (returnval){ [ 1216.708639] env[63515]: value = "task-1112078" [ 1216.708639] env[63515]: _type = "Task" [ 1216.708639] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1216.716885] env[63515]: DEBUG oslo_vmware.api [None req-dc87d49d-dcd9-4981-8dc8-1f724cc8296c tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Task: {'id': task-1112078, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1216.864570] env[63515]: DEBUG oslo_service.periodic_task [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=63515) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1217.137897] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1112077, 'name': CreateVM_Task, 'duration_secs': 0.333164} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1217.138153] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 24a79a6c-5210-4778-a6ed-482bfe14933c] Created VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1217.138882] env[63515]: DEBUG oslo_concurrency.lockutils [None req-8fa768a4-1e0e-4664-bc77-1d533a4f42d6 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/af53d566-21dc-48bf-aacf-f6a96a111935" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1217.139015] env[63515]: DEBUG oslo_concurrency.lockutils [None req-8fa768a4-1e0e-4664-bc77-1d533a4f42d6 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Acquired lock "[datastore1] devstack-image-cache_base/af53d566-21dc-48bf-aacf-f6a96a111935" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1217.139396] env[63515]: DEBUG oslo_concurrency.lockutils [None req-8fa768a4-1e0e-4664-bc77-1d533a4f42d6 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/af53d566-21dc-48bf-aacf-f6a96a111935" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1217.139654] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5d0ff6a2-0cd5-401f-a6d9-5ec06a70efea {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1217.145039] env[63515]: DEBUG oslo_vmware.api [None req-8fa768a4-1e0e-4664-bc77-1d533a4f42d6 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Waiting for the task: (returnval){ [ 1217.145039] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]5224c172-2e61-502b-4921-93758cfd0172" [ 1217.145039] env[63515]: _type = "Task" [ 1217.145039] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1217.153150] env[63515]: DEBUG oslo_vmware.api [None req-8fa768a4-1e0e-4664-bc77-1d533a4f42d6 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]5224c172-2e61-502b-4921-93758cfd0172, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1217.220845] env[63515]: DEBUG oslo_vmware.api [None req-dc87d49d-dcd9-4981-8dc8-1f724cc8296c tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Task: {'id': task-1112078, 'name': PowerOffVM_Task, 'duration_secs': 0.2419} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1217.221177] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-dc87d49d-dcd9-4981-8dc8-1f724cc8296c tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] [instance: c8dbf542-70cb-45be-9308-5e12307c0be6] Powered off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1217.221354] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-dc87d49d-dcd9-4981-8dc8-1f724cc8296c tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] [instance: c8dbf542-70cb-45be-9308-5e12307c0be6] Unregistering the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1217.221617] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-1b8527b2-73b4-459a-a25b-306f021f98cc {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1217.260289] env[63515]: DEBUG nova.compute.manager [req-ce3ae472-ca9d-4ab4-a974-58788a2c9fa7 req-2fdb9bd2-43ba-4f53-ba3f-678e9af7ddfe service nova] [instance: 24a79a6c-5210-4778-a6ed-482bfe14933c] Received event network-changed-f338b564-7636-4839-b034-73c3233d7da6 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1217.260804] env[63515]: DEBUG nova.compute.manager [req-ce3ae472-ca9d-4ab4-a974-58788a2c9fa7 req-2fdb9bd2-43ba-4f53-ba3f-678e9af7ddfe service nova] [instance: 24a79a6c-5210-4778-a6ed-482bfe14933c] Refreshing instance network info cache due to event network-changed-f338b564-7636-4839-b034-73c3233d7da6. {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1217.260938] env[63515]: DEBUG oslo_concurrency.lockutils [req-ce3ae472-ca9d-4ab4-a974-58788a2c9fa7 req-2fdb9bd2-43ba-4f53-ba3f-678e9af7ddfe service nova] Acquiring lock "refresh_cache-24a79a6c-5210-4778-a6ed-482bfe14933c" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1217.261543] env[63515]: DEBUG oslo_concurrency.lockutils [req-ce3ae472-ca9d-4ab4-a974-58788a2c9fa7 req-2fdb9bd2-43ba-4f53-ba3f-678e9af7ddfe service nova] Acquired lock "refresh_cache-24a79a6c-5210-4778-a6ed-482bfe14933c" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1217.261543] env[63515]: DEBUG nova.network.neutron [req-ce3ae472-ca9d-4ab4-a974-58788a2c9fa7 req-2fdb9bd2-43ba-4f53-ba3f-678e9af7ddfe service nova] [instance: 24a79a6c-5210-4778-a6ed-482bfe14933c] Refreshing network info cache for port f338b564-7636-4839-b034-73c3233d7da6 {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1217.282895] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-dc87d49d-dcd9-4981-8dc8-1f724cc8296c tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] [instance: c8dbf542-70cb-45be-9308-5e12307c0be6] Unregistered the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1217.282895] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-dc87d49d-dcd9-4981-8dc8-1f724cc8296c tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] [instance: c8dbf542-70cb-45be-9308-5e12307c0be6] Deleting contents of the VM from datastore datastore1 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1217.282895] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-dc87d49d-dcd9-4981-8dc8-1f724cc8296c tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Deleting the datastore file [datastore1] c8dbf542-70cb-45be-9308-5e12307c0be6 {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1217.283760] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b39fab8b-d432-472d-96e5-35ef4f66f1d1 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1217.289768] env[63515]: DEBUG oslo_vmware.api [None req-dc87d49d-dcd9-4981-8dc8-1f724cc8296c tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Waiting for the task: (returnval){ [ 1217.289768] env[63515]: value = "task-1112080" [ 1217.289768] env[63515]: _type = "Task" [ 1217.289768] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1217.298212] env[63515]: DEBUG oslo_vmware.api [None req-dc87d49d-dcd9-4981-8dc8-1f724cc8296c tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Task: {'id': task-1112080, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1217.655192] env[63515]: DEBUG oslo_concurrency.lockutils [None req-8fa768a4-1e0e-4664-bc77-1d533a4f42d6 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Releasing lock "[datastore1] devstack-image-cache_base/af53d566-21dc-48bf-aacf-f6a96a111935" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1217.655614] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-8fa768a4-1e0e-4664-bc77-1d533a4f42d6 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 24a79a6c-5210-4778-a6ed-482bfe14933c] Processing image af53d566-21dc-48bf-aacf-f6a96a111935 {{(pid=63515) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1217.655987] env[63515]: DEBUG oslo_concurrency.lockutils [None req-8fa768a4-1e0e-4664-bc77-1d533a4f42d6 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/af53d566-21dc-48bf-aacf-f6a96a111935/af53d566-21dc-48bf-aacf-f6a96a111935.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1217.656266] env[63515]: DEBUG oslo_concurrency.lockutils [None req-8fa768a4-1e0e-4664-bc77-1d533a4f42d6 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Acquired lock "[datastore1] devstack-image-cache_base/af53d566-21dc-48bf-aacf-f6a96a111935/af53d566-21dc-48bf-aacf-f6a96a111935.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1217.656537] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-8fa768a4-1e0e-4664-bc77-1d533a4f42d6 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1217.656796] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-9b56f308-3317-4b9a-8d7a-4013aaa5299f {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1217.664677] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-8fa768a4-1e0e-4664-bc77-1d533a4f42d6 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1217.664873] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-8fa768a4-1e0e-4664-bc77-1d533a4f42d6 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63515) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1217.665557] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4c4c26e9-6f93-40df-bfc7-e940ff856fd3 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1217.670382] env[63515]: DEBUG oslo_vmware.api [None req-8fa768a4-1e0e-4664-bc77-1d533a4f42d6 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Waiting for the task: (returnval){ [ 1217.670382] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]52eec25e-bd83-45f5-1818-71da5cbe3d85" [ 1217.670382] env[63515]: _type = "Task" [ 1217.670382] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1217.677295] env[63515]: DEBUG oslo_vmware.api [None req-8fa768a4-1e0e-4664-bc77-1d533a4f42d6 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52eec25e-bd83-45f5-1818-71da5cbe3d85, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1217.802756] env[63515]: DEBUG oslo_vmware.api [None req-dc87d49d-dcd9-4981-8dc8-1f724cc8296c tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Task: {'id': task-1112080, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.177525} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1217.803099] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-dc87d49d-dcd9-4981-8dc8-1f724cc8296c tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Deleted the datastore file {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1217.803149] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-dc87d49d-dcd9-4981-8dc8-1f724cc8296c tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] [instance: c8dbf542-70cb-45be-9308-5e12307c0be6] Deleted contents of the VM from datastore datastore1 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1217.804058] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-dc87d49d-dcd9-4981-8dc8-1f724cc8296c tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] [instance: c8dbf542-70cb-45be-9308-5e12307c0be6] Instance destroyed {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1217.804058] env[63515]: INFO nova.compute.manager [None req-dc87d49d-dcd9-4981-8dc8-1f724cc8296c tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] [instance: c8dbf542-70cb-45be-9308-5e12307c0be6] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1217.804058] env[63515]: DEBUG oslo.service.loopingcall [None req-dc87d49d-dcd9-4981-8dc8-1f724cc8296c tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1217.804058] env[63515]: DEBUG nova.compute.manager [-] [instance: c8dbf542-70cb-45be-9308-5e12307c0be6] Deallocating network for instance {{(pid=63515) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1217.804282] env[63515]: DEBUG nova.network.neutron [-] [instance: c8dbf542-70cb-45be-9308-5e12307c0be6] deallocate_for_instance() {{(pid=63515) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1218.144473] env[63515]: DEBUG nova.network.neutron [req-ce3ae472-ca9d-4ab4-a974-58788a2c9fa7 req-2fdb9bd2-43ba-4f53-ba3f-678e9af7ddfe service nova] [instance: 24a79a6c-5210-4778-a6ed-482bfe14933c] Updated VIF entry in instance network info cache for port f338b564-7636-4839-b034-73c3233d7da6. {{(pid=63515) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1218.144473] env[63515]: DEBUG nova.network.neutron [req-ce3ae472-ca9d-4ab4-a974-58788a2c9fa7 req-2fdb9bd2-43ba-4f53-ba3f-678e9af7ddfe service nova] [instance: 24a79a6c-5210-4778-a6ed-482bfe14933c] Updating instance_info_cache with network_info: [{"id": "f338b564-7636-4839-b034-73c3233d7da6", "address": "fa:16:3e:bc:5a:12", "network": {"id": "5758c885-e1c9-45c8-86aa-27a802101f52", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-734458138-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.204", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b18d4daacbc84e758f9f9ae4f3f09d28", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69351262-8d39-441a-85ba-3a78df436d17", "external-id": "nsx-vlan-transportzone-205", "segmentation_id": 205, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf338b564-76", "ovs_interfaceid": "f338b564-7636-4839-b034-73c3233d7da6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1218.184455] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-8fa768a4-1e0e-4664-bc77-1d533a4f42d6 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 24a79a6c-5210-4778-a6ed-482bfe14933c] Preparing fetch location {{(pid=63515) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 1218.185309] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-8fa768a4-1e0e-4664-bc77-1d533a4f42d6 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 24a79a6c-5210-4778-a6ed-482bfe14933c] Fetch image to [datastore1] OSTACK_IMG_c076be39-a467-49ad-a1d3-b011c6655f80/OSTACK_IMG_c076be39-a467-49ad-a1d3-b011c6655f80.vmdk {{(pid=63515) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 1218.185653] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-8fa768a4-1e0e-4664-bc77-1d533a4f42d6 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 24a79a6c-5210-4778-a6ed-482bfe14933c] Downloading stream optimized image af53d566-21dc-48bf-aacf-f6a96a111935 to [datastore1] OSTACK_IMG_c076be39-a467-49ad-a1d3-b011c6655f80/OSTACK_IMG_c076be39-a467-49ad-a1d3-b011c6655f80.vmdk on the data store datastore1 as vApp {{(pid=63515) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 1218.185987] env[63515]: DEBUG nova.virt.vmwareapi.images [None req-8fa768a4-1e0e-4664-bc77-1d533a4f42d6 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 24a79a6c-5210-4778-a6ed-482bfe14933c] Downloading image file data af53d566-21dc-48bf-aacf-f6a96a111935 to the ESX as VM named 'OSTACK_IMG_c076be39-a467-49ad-a1d3-b011c6655f80' {{(pid=63515) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 1218.270245] env[63515]: DEBUG oslo_vmware.rw_handles [None req-8fa768a4-1e0e-4664-bc77-1d533a4f42d6 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 1218.270245] env[63515]: value = "resgroup-9" [ 1218.270245] env[63515]: _type = "ResourcePool" [ 1218.270245] env[63515]: }. {{(pid=63515) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 1218.270845] env[63515]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-a7c01515-5d0b-4fa9-aa15-59aecf1b28d5 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1218.291740] env[63515]: DEBUG oslo_vmware.rw_handles [None req-8fa768a4-1e0e-4664-bc77-1d533a4f42d6 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Lease: (returnval){ [ 1218.291740] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]527163ae-ffbf-7504-e6ff-01721c761d84" [ 1218.291740] env[63515]: _type = "HttpNfcLease" [ 1218.291740] env[63515]: } obtained for vApp import into resource pool (val){ [ 1218.291740] env[63515]: value = "resgroup-9" [ 1218.291740] env[63515]: _type = "ResourcePool" [ 1218.291740] env[63515]: }. {{(pid=63515) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 1218.292410] env[63515]: DEBUG oslo_vmware.api [None req-8fa768a4-1e0e-4664-bc77-1d533a4f42d6 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Waiting for the lease: (returnval){ [ 1218.292410] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]527163ae-ffbf-7504-e6ff-01721c761d84" [ 1218.292410] env[63515]: _type = "HttpNfcLease" [ 1218.292410] env[63515]: } to be ready. {{(pid=63515) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1218.300029] env[63515]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1218.300029] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]527163ae-ffbf-7504-e6ff-01721c761d84" [ 1218.300029] env[63515]: _type = "HttpNfcLease" [ 1218.300029] env[63515]: } is initializing. {{(pid=63515) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1218.647677] env[63515]: DEBUG oslo_concurrency.lockutils [req-ce3ae472-ca9d-4ab4-a974-58788a2c9fa7 req-2fdb9bd2-43ba-4f53-ba3f-678e9af7ddfe service nova] Releasing lock "refresh_cache-24a79a6c-5210-4778-a6ed-482bfe14933c" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1218.800842] env[63515]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1218.800842] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]527163ae-ffbf-7504-e6ff-01721c761d84" [ 1218.800842] env[63515]: _type = "HttpNfcLease" [ 1218.800842] env[63515]: } is ready. {{(pid=63515) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1218.801152] env[63515]: DEBUG oslo_vmware.rw_handles [None req-8fa768a4-1e0e-4664-bc77-1d533a4f42d6 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1218.801152] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]527163ae-ffbf-7504-e6ff-01721c761d84" [ 1218.801152] env[63515]: _type = "HttpNfcLease" [ 1218.801152] env[63515]: }. {{(pid=63515) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 1218.801856] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2018921-e334-4955-b9bf-46241692bb32 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1218.809233] env[63515]: DEBUG oslo_vmware.rw_handles [None req-8fa768a4-1e0e-4664-bc77-1d533a4f42d6 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5288525c-9620-00c8-adbc-aa9d1746cb3a/disk-0.vmdk from lease info. {{(pid=63515) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1218.809557] env[63515]: DEBUG oslo_vmware.rw_handles [None req-8fa768a4-1e0e-4664-bc77-1d533a4f42d6 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Creating HTTP connection to write to file with size = 31667712 and URL = https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5288525c-9620-00c8-adbc-aa9d1746cb3a/disk-0.vmdk. {{(pid=63515) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 1218.866930] env[63515]: DEBUG nova.network.neutron [-] [instance: c8dbf542-70cb-45be-9308-5e12307c0be6] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1218.876160] env[63515]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-26fd0093-7e5e-4e18-bda7-5bb8bc023f54 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1219.293602] env[63515]: DEBUG nova.compute.manager [req-dc7fa1b6-98fa-4b1b-a267-555a1954e565 req-b67b32af-2dbc-4c97-99ff-86eba4040ddc service nova] [instance: c8dbf542-70cb-45be-9308-5e12307c0be6] Received event network-vif-deleted-5569cb8c-d9c8-4cb5-abdc-5890ad7d7ede {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1219.369800] env[63515]: INFO nova.compute.manager [-] [instance: c8dbf542-70cb-45be-9308-5e12307c0be6] Took 1.57 seconds to deallocate network for instance. [ 1219.876971] env[63515]: DEBUG oslo_concurrency.lockutils [None req-dc87d49d-dcd9-4981-8dc8-1f724cc8296c tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1219.877388] env[63515]: DEBUG oslo_concurrency.lockutils [None req-dc87d49d-dcd9-4981-8dc8-1f724cc8296c tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1219.877599] env[63515]: DEBUG nova.objects.instance [None req-dc87d49d-dcd9-4981-8dc8-1f724cc8296c tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Lazy-loading 'resources' on Instance uuid c8dbf542-70cb-45be-9308-5e12307c0be6 {{(pid=63515) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1219.975626] env[63515]: DEBUG oslo_vmware.rw_handles [None req-8fa768a4-1e0e-4664-bc77-1d533a4f42d6 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Completed reading data from the image iterator. {{(pid=63515) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 1219.975869] env[63515]: DEBUG oslo_vmware.rw_handles [None req-8fa768a4-1e0e-4664-bc77-1d533a4f42d6 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5288525c-9620-00c8-adbc-aa9d1746cb3a/disk-0.vmdk. {{(pid=63515) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1219.976887] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd6bd4e9-8e90-4a35-9d17-217c761bcdfe {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1219.983740] env[63515]: DEBUG oslo_vmware.rw_handles [None req-8fa768a4-1e0e-4664-bc77-1d533a4f42d6 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5288525c-9620-00c8-adbc-aa9d1746cb3a/disk-0.vmdk is in state: ready. {{(pid=63515) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1219.983959] env[63515]: DEBUG oslo_vmware.rw_handles [None req-8fa768a4-1e0e-4664-bc77-1d533a4f42d6 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Releasing lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5288525c-9620-00c8-adbc-aa9d1746cb3a/disk-0.vmdk. {{(pid=63515) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 1219.984240] env[63515]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-9c2cb35e-26f9-4cc8-a40d-b6be1bca98b6 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1220.182902] env[63515]: DEBUG oslo_vmware.rw_handles [None req-8fa768a4-1e0e-4664-bc77-1d533a4f42d6 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Closed VMDK write handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5288525c-9620-00c8-adbc-aa9d1746cb3a/disk-0.vmdk. {{(pid=63515) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 1220.183138] env[63515]: INFO nova.virt.vmwareapi.images [None req-8fa768a4-1e0e-4664-bc77-1d533a4f42d6 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 24a79a6c-5210-4778-a6ed-482bfe14933c] Downloaded image file data af53d566-21dc-48bf-aacf-f6a96a111935 [ 1220.183949] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-812be7ec-5f88-4ed1-99dc-94cfc96e3ec6 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1220.198852] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-96957b43-0d66-4d66-b3b1-50c6363c967f {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1220.220540] env[63515]: INFO nova.virt.vmwareapi.images [None req-8fa768a4-1e0e-4664-bc77-1d533a4f42d6 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 24a79a6c-5210-4778-a6ed-482bfe14933c] The imported VM was unregistered [ 1220.222930] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-8fa768a4-1e0e-4664-bc77-1d533a4f42d6 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 24a79a6c-5210-4778-a6ed-482bfe14933c] Caching image {{(pid=63515) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 1220.223184] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-8fa768a4-1e0e-4664-bc77-1d533a4f42d6 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Creating directory with path [datastore1] devstack-image-cache_base/af53d566-21dc-48bf-aacf-f6a96a111935 {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1220.223441] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-78841607-f34a-4a21-812b-0848ddc7f6fb {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1220.233812] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-8fa768a4-1e0e-4664-bc77-1d533a4f42d6 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Created directory with path [datastore1] devstack-image-cache_base/af53d566-21dc-48bf-aacf-f6a96a111935 {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1220.233961] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-8fa768a4-1e0e-4664-bc77-1d533a4f42d6 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Moving virtual disk from [datastore1] OSTACK_IMG_c076be39-a467-49ad-a1d3-b011c6655f80/OSTACK_IMG_c076be39-a467-49ad-a1d3-b011c6655f80.vmdk to [datastore1] devstack-image-cache_base/af53d566-21dc-48bf-aacf-f6a96a111935/af53d566-21dc-48bf-aacf-f6a96a111935.vmdk. {{(pid=63515) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 1220.234223] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-c1ac966f-1eef-4ee3-a237-d08a67595a6f {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1220.240335] env[63515]: DEBUG oslo_vmware.api [None req-8fa768a4-1e0e-4664-bc77-1d533a4f42d6 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Waiting for the task: (returnval){ [ 1220.240335] env[63515]: value = "task-1112083" [ 1220.240335] env[63515]: _type = "Task" [ 1220.240335] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1220.248934] env[63515]: DEBUG oslo_vmware.api [None req-8fa768a4-1e0e-4664-bc77-1d533a4f42d6 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Task: {'id': task-1112083, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1220.427251] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e11bb28e-4030-49c5-91ef-0d8f25c5974d {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1220.435576] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9fb68d0-b8f6-4f52-93d6-276524e16a7d {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1220.466308] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42b91046-b45f-44f8-9457-9d63d7c11e10 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1220.473962] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01f19b56-eb68-4bcb-b639-6b77582871de {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1220.489336] env[63515]: DEBUG nova.compute.provider_tree [None req-dc87d49d-dcd9-4981-8dc8-1f724cc8296c tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1220.752880] env[63515]: DEBUG oslo_vmware.api [None req-8fa768a4-1e0e-4664-bc77-1d533a4f42d6 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Task: {'id': task-1112083, 'name': MoveVirtualDisk_Task} progress is 24%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1220.993184] env[63515]: DEBUG nova.scheduler.client.report [None req-dc87d49d-dcd9-4981-8dc8-1f724cc8296c tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1221.254037] env[63515]: DEBUG oslo_vmware.api [None req-8fa768a4-1e0e-4664-bc77-1d533a4f42d6 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Task: {'id': task-1112083, 'name': MoveVirtualDisk_Task} progress is 46%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1221.498632] env[63515]: DEBUG oslo_concurrency.lockutils [None req-dc87d49d-dcd9-4981-8dc8-1f724cc8296c tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.621s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1221.521963] env[63515]: INFO nova.scheduler.client.report [None req-dc87d49d-dcd9-4981-8dc8-1f724cc8296c tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Deleted allocations for instance c8dbf542-70cb-45be-9308-5e12307c0be6 [ 1221.755880] env[63515]: DEBUG oslo_vmware.api [None req-8fa768a4-1e0e-4664-bc77-1d533a4f42d6 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Task: {'id': task-1112083, 'name': MoveVirtualDisk_Task} progress is 71%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1222.031184] env[63515]: DEBUG oslo_concurrency.lockutils [None req-dc87d49d-dcd9-4981-8dc8-1f724cc8296c tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Lock "c8dbf542-70cb-45be-9308-5e12307c0be6" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.346s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1222.254373] env[63515]: DEBUG oslo_vmware.api [None req-8fa768a4-1e0e-4664-bc77-1d533a4f42d6 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Task: {'id': task-1112083, 'name': MoveVirtualDisk_Task} progress is 94%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1222.753516] env[63515]: DEBUG oslo_vmware.api [None req-8fa768a4-1e0e-4664-bc77-1d533a4f42d6 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Task: {'id': task-1112083, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.187567} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1222.753812] env[63515]: INFO nova.virt.vmwareapi.ds_util [None req-8fa768a4-1e0e-4664-bc77-1d533a4f42d6 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Moved virtual disk from [datastore1] OSTACK_IMG_c076be39-a467-49ad-a1d3-b011c6655f80/OSTACK_IMG_c076be39-a467-49ad-a1d3-b011c6655f80.vmdk to [datastore1] devstack-image-cache_base/af53d566-21dc-48bf-aacf-f6a96a111935/af53d566-21dc-48bf-aacf-f6a96a111935.vmdk. [ 1222.754014] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-8fa768a4-1e0e-4664-bc77-1d533a4f42d6 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 24a79a6c-5210-4778-a6ed-482bfe14933c] Cleaning up location [datastore1] OSTACK_IMG_c076be39-a467-49ad-a1d3-b011c6655f80 {{(pid=63515) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 1222.754190] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-8fa768a4-1e0e-4664-bc77-1d533a4f42d6 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Deleting the datastore file [datastore1] OSTACK_IMG_c076be39-a467-49ad-a1d3-b011c6655f80 {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1222.754435] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-5d3061a5-b32e-4dac-a429-f07800a342e0 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1222.760704] env[63515]: DEBUG oslo_vmware.api [None req-8fa768a4-1e0e-4664-bc77-1d533a4f42d6 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Waiting for the task: (returnval){ [ 1222.760704] env[63515]: value = "task-1112084" [ 1222.760704] env[63515]: _type = "Task" [ 1222.760704] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1222.768693] env[63515]: DEBUG oslo_vmware.api [None req-8fa768a4-1e0e-4664-bc77-1d533a4f42d6 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Task: {'id': task-1112084, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1223.270925] env[63515]: DEBUG oslo_vmware.api [None req-8fa768a4-1e0e-4664-bc77-1d533a4f42d6 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Task: {'id': task-1112084, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.038725} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1223.271265] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-8fa768a4-1e0e-4664-bc77-1d533a4f42d6 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Deleted the datastore file {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1223.271390] env[63515]: DEBUG oslo_concurrency.lockutils [None req-8fa768a4-1e0e-4664-bc77-1d533a4f42d6 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Releasing lock "[datastore1] devstack-image-cache_base/af53d566-21dc-48bf-aacf-f6a96a111935/af53d566-21dc-48bf-aacf-f6a96a111935.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1223.271634] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-8fa768a4-1e0e-4664-bc77-1d533a4f42d6 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/af53d566-21dc-48bf-aacf-f6a96a111935/af53d566-21dc-48bf-aacf-f6a96a111935.vmdk to [datastore1] 24a79a6c-5210-4778-a6ed-482bfe14933c/24a79a6c-5210-4778-a6ed-482bfe14933c.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1223.271894] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-1996580f-9f80-49db-8b4c-1ec703e577be {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1223.278892] env[63515]: DEBUG oslo_vmware.api [None req-8fa768a4-1e0e-4664-bc77-1d533a4f42d6 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Waiting for the task: (returnval){ [ 1223.278892] env[63515]: value = "task-1112085" [ 1223.278892] env[63515]: _type = "Task" [ 1223.278892] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1223.287182] env[63515]: DEBUG oslo_vmware.api [None req-8fa768a4-1e0e-4664-bc77-1d533a4f42d6 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Task: {'id': task-1112085, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1223.727094] env[63515]: DEBUG oslo_concurrency.lockutils [None req-7a41c677-8af5-4616-8502-88e4da1d930f tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Acquiring lock "d63a2a0d-2736-4d1f-a9fb-ec105747dcc3" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1223.727466] env[63515]: DEBUG oslo_concurrency.lockutils [None req-7a41c677-8af5-4616-8502-88e4da1d930f tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Lock "d63a2a0d-2736-4d1f-a9fb-ec105747dcc3" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1223.790472] env[63515]: DEBUG oslo_vmware.api [None req-8fa768a4-1e0e-4664-bc77-1d533a4f42d6 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Task: {'id': task-1112085, 'name': CopyVirtualDisk_Task} progress is 24%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1224.230262] env[63515]: DEBUG nova.compute.manager [None req-7a41c677-8af5-4616-8502-88e4da1d930f tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] [instance: d63a2a0d-2736-4d1f-a9fb-ec105747dcc3] Starting instance... {{(pid=63515) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1224.291019] env[63515]: DEBUG oslo_vmware.api [None req-8fa768a4-1e0e-4664-bc77-1d533a4f42d6 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Task: {'id': task-1112085, 'name': CopyVirtualDisk_Task} progress is 46%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1224.754628] env[63515]: DEBUG oslo_concurrency.lockutils [None req-7a41c677-8af5-4616-8502-88e4da1d930f tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1224.754934] env[63515]: DEBUG oslo_concurrency.lockutils [None req-7a41c677-8af5-4616-8502-88e4da1d930f tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1224.756710] env[63515]: INFO nova.compute.claims [None req-7a41c677-8af5-4616-8502-88e4da1d930f tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] [instance: d63a2a0d-2736-4d1f-a9fb-ec105747dcc3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1224.792478] env[63515]: DEBUG oslo_vmware.api [None req-8fa768a4-1e0e-4664-bc77-1d533a4f42d6 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Task: {'id': task-1112085, 'name': CopyVirtualDisk_Task} progress is 71%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1225.292817] env[63515]: DEBUG oslo_vmware.api [None req-8fa768a4-1e0e-4664-bc77-1d533a4f42d6 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Task: {'id': task-1112085, 'name': CopyVirtualDisk_Task} progress is 94%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1225.792488] env[63515]: DEBUG oslo_vmware.api [None req-8fa768a4-1e0e-4664-bc77-1d533a4f42d6 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Task: {'id': task-1112085, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.142327} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1225.792762] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-8fa768a4-1e0e-4664-bc77-1d533a4f42d6 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/af53d566-21dc-48bf-aacf-f6a96a111935/af53d566-21dc-48bf-aacf-f6a96a111935.vmdk to [datastore1] 24a79a6c-5210-4778-a6ed-482bfe14933c/24a79a6c-5210-4778-a6ed-482bfe14933c.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1225.793527] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-420499b3-766d-4bf7-8f2a-7d6b2567dd2b {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1225.816559] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-8fa768a4-1e0e-4664-bc77-1d533a4f42d6 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 24a79a6c-5210-4778-a6ed-482bfe14933c] Reconfiguring VM instance instance-0000006f to attach disk [datastore1] 24a79a6c-5210-4778-a6ed-482bfe14933c/24a79a6c-5210-4778-a6ed-482bfe14933c.vmdk or device None with type streamOptimized {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1225.817456] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9c480159-3573-43bf-87ea-e771782ac0d0 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1225.831458] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8fc6973-5938-4952-a20d-54a418c82ab5 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1225.838256] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36c7921e-8f16-4792-a942-fb77bafbbdea {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1225.841864] env[63515]: DEBUG oslo_vmware.api [None req-8fa768a4-1e0e-4664-bc77-1d533a4f42d6 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Waiting for the task: (returnval){ [ 1225.841864] env[63515]: value = "task-1112086" [ 1225.841864] env[63515]: _type = "Task" [ 1225.841864] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1225.869401] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-256926d9-7cd4-440a-af51-95e42b9a9908 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1225.874595] env[63515]: DEBUG oslo_vmware.api [None req-8fa768a4-1e0e-4664-bc77-1d533a4f42d6 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Task: {'id': task-1112086, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1225.879021] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4bacb9f9-9368-4692-8c85-b761f6627aef {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1225.891530] env[63515]: DEBUG nova.compute.provider_tree [None req-7a41c677-8af5-4616-8502-88e4da1d930f tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1226.353827] env[63515]: DEBUG oslo_vmware.api [None req-8fa768a4-1e0e-4664-bc77-1d533a4f42d6 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Task: {'id': task-1112086, 'name': ReconfigVM_Task} progress is 14%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1226.394974] env[63515]: DEBUG nova.scheduler.client.report [None req-7a41c677-8af5-4616-8502-88e4da1d930f tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1226.852290] env[63515]: DEBUG oslo_vmware.api [None req-8fa768a4-1e0e-4664-bc77-1d533a4f42d6 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Task: {'id': task-1112086, 'name': ReconfigVM_Task, 'duration_secs': 0.682822} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1226.852514] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-8fa768a4-1e0e-4664-bc77-1d533a4f42d6 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 24a79a6c-5210-4778-a6ed-482bfe14933c] Reconfigured VM instance instance-0000006f to attach disk [datastore1] 24a79a6c-5210-4778-a6ed-482bfe14933c/24a79a6c-5210-4778-a6ed-482bfe14933c.vmdk or device None with type streamOptimized {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1226.853178] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c9a737b2-f7ba-434a-aca3-100546de9360 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1226.859430] env[63515]: DEBUG oslo_vmware.api [None req-8fa768a4-1e0e-4664-bc77-1d533a4f42d6 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Waiting for the task: (returnval){ [ 1226.859430] env[63515]: value = "task-1112087" [ 1226.859430] env[63515]: _type = "Task" [ 1226.859430] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1226.867383] env[63515]: DEBUG oslo_vmware.api [None req-8fa768a4-1e0e-4664-bc77-1d533a4f42d6 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Task: {'id': task-1112087, 'name': Rename_Task} progress is 5%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1226.899401] env[63515]: DEBUG oslo_concurrency.lockutils [None req-7a41c677-8af5-4616-8502-88e4da1d930f tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.144s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1226.899922] env[63515]: DEBUG nova.compute.manager [None req-7a41c677-8af5-4616-8502-88e4da1d930f tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] [instance: d63a2a0d-2736-4d1f-a9fb-ec105747dcc3] Start building networks asynchronously for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1227.369118] env[63515]: DEBUG oslo_vmware.api [None req-8fa768a4-1e0e-4664-bc77-1d533a4f42d6 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Task: {'id': task-1112087, 'name': Rename_Task, 'duration_secs': 0.132558} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1227.369448] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-8fa768a4-1e0e-4664-bc77-1d533a4f42d6 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 24a79a6c-5210-4778-a6ed-482bfe14933c] Powering on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1227.369719] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-984b5927-922d-434b-a48c-600b1fab2e55 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1227.376684] env[63515]: DEBUG oslo_vmware.api [None req-8fa768a4-1e0e-4664-bc77-1d533a4f42d6 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Waiting for the task: (returnval){ [ 1227.376684] env[63515]: value = "task-1112088" [ 1227.376684] env[63515]: _type = "Task" [ 1227.376684] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1227.382670] env[63515]: DEBUG oslo_vmware.api [None req-8fa768a4-1e0e-4664-bc77-1d533a4f42d6 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Task: {'id': task-1112088, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1227.405206] env[63515]: DEBUG nova.compute.utils [None req-7a41c677-8af5-4616-8502-88e4da1d930f tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Using /dev/sd instead of None {{(pid=63515) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1227.406584] env[63515]: DEBUG nova.compute.manager [None req-7a41c677-8af5-4616-8502-88e4da1d930f tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] [instance: d63a2a0d-2736-4d1f-a9fb-ec105747dcc3] Allocating IP information in the background. {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1227.406757] env[63515]: DEBUG nova.network.neutron [None req-7a41c677-8af5-4616-8502-88e4da1d930f tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] [instance: d63a2a0d-2736-4d1f-a9fb-ec105747dcc3] allocate_for_instance() {{(pid=63515) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1227.445665] env[63515]: DEBUG nova.policy [None req-7a41c677-8af5-4616-8502-88e4da1d930f tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'cb4f41c6bfb94aa8ab60135e1d729a6a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '97f2e119708746038f38359572c99438', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63515) authorize /opt/stack/nova/nova/policy.py:201}} [ 1227.765770] env[63515]: DEBUG nova.network.neutron [None req-7a41c677-8af5-4616-8502-88e4da1d930f tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] [instance: d63a2a0d-2736-4d1f-a9fb-ec105747dcc3] Successfully created port: 8e9d8cfb-9749-419b-b009-786e95b2f419 {{(pid=63515) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1227.885059] env[63515]: DEBUG oslo_vmware.api [None req-8fa768a4-1e0e-4664-bc77-1d533a4f42d6 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Task: {'id': task-1112088, 'name': PowerOnVM_Task} progress is 100%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1227.909877] env[63515]: DEBUG nova.compute.manager [None req-7a41c677-8af5-4616-8502-88e4da1d930f tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] [instance: d63a2a0d-2736-4d1f-a9fb-ec105747dcc3] Start building block device mappings for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1228.385181] env[63515]: DEBUG oslo_vmware.api [None req-8fa768a4-1e0e-4664-bc77-1d533a4f42d6 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Task: {'id': task-1112088, 'name': PowerOnVM_Task, 'duration_secs': 0.522803} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1228.385479] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-8fa768a4-1e0e-4664-bc77-1d533a4f42d6 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 24a79a6c-5210-4778-a6ed-482bfe14933c] Powered on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1228.480649] env[63515]: DEBUG nova.compute.manager [None req-8fa768a4-1e0e-4664-bc77-1d533a4f42d6 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 24a79a6c-5210-4778-a6ed-482bfe14933c] Checking state {{(pid=63515) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1228.481580] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b2550b4-9840-43d7-af3a-91e1d0393b58 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1228.919936] env[63515]: DEBUG nova.compute.manager [None req-7a41c677-8af5-4616-8502-88e4da1d930f tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] [instance: d63a2a0d-2736-4d1f-a9fb-ec105747dcc3] Start spawning the instance on the hypervisor. {{(pid=63515) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1228.940704] env[63515]: DEBUG nova.virt.hardware [None req-7a41c677-8af5-4616-8502-88e4da1d930f tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T02:52:14Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T02:51:56Z,direct_url=,disk_format='vmdk',id=8a120570-cb06-4099-b262-554ca0ad15c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b569255cc43e42e7ae3f2b4ad37c6ef4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T02:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1228.940974] env[63515]: DEBUG nova.virt.hardware [None req-7a41c677-8af5-4616-8502-88e4da1d930f tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Flavor limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1228.941163] env[63515]: DEBUG nova.virt.hardware [None req-7a41c677-8af5-4616-8502-88e4da1d930f tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Image limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1228.941354] env[63515]: DEBUG nova.virt.hardware [None req-7a41c677-8af5-4616-8502-88e4da1d930f tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Flavor pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1228.941507] env[63515]: DEBUG nova.virt.hardware [None req-7a41c677-8af5-4616-8502-88e4da1d930f tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Image pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1228.941661] env[63515]: DEBUG nova.virt.hardware [None req-7a41c677-8af5-4616-8502-88e4da1d930f tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1228.941967] env[63515]: DEBUG nova.virt.hardware [None req-7a41c677-8af5-4616-8502-88e4da1d930f tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1228.942171] env[63515]: DEBUG nova.virt.hardware [None req-7a41c677-8af5-4616-8502-88e4da1d930f tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1228.942366] env[63515]: DEBUG nova.virt.hardware [None req-7a41c677-8af5-4616-8502-88e4da1d930f tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Got 1 possible topologies {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1228.942537] env[63515]: DEBUG nova.virt.hardware [None req-7a41c677-8af5-4616-8502-88e4da1d930f tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1228.942716] env[63515]: DEBUG nova.virt.hardware [None req-7a41c677-8af5-4616-8502-88e4da1d930f tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1228.943642] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e2fdeed-0c7a-439e-aa2a-6f54b8f8468f {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1228.951834] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3453f50-721a-44da-bc8a-595815609f73 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1228.997614] env[63515]: DEBUG oslo_concurrency.lockutils [None req-8fa768a4-1e0e-4664-bc77-1d533a4f42d6 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Lock "24a79a6c-5210-4778-a6ed-482bfe14933c" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 21.197s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1229.126920] env[63515]: DEBUG nova.compute.manager [req-f13f5af3-1268-4c7e-8405-2de5b2ff42d1 req-621193d6-009f-45b5-9c14-23d129afd89e service nova] [instance: d63a2a0d-2736-4d1f-a9fb-ec105747dcc3] Received event network-vif-plugged-8e9d8cfb-9749-419b-b009-786e95b2f419 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1229.127255] env[63515]: DEBUG oslo_concurrency.lockutils [req-f13f5af3-1268-4c7e-8405-2de5b2ff42d1 req-621193d6-009f-45b5-9c14-23d129afd89e service nova] Acquiring lock "d63a2a0d-2736-4d1f-a9fb-ec105747dcc3-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1229.127421] env[63515]: DEBUG oslo_concurrency.lockutils [req-f13f5af3-1268-4c7e-8405-2de5b2ff42d1 req-621193d6-009f-45b5-9c14-23d129afd89e service nova] Lock "d63a2a0d-2736-4d1f-a9fb-ec105747dcc3-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1229.127594] env[63515]: DEBUG oslo_concurrency.lockutils [req-f13f5af3-1268-4c7e-8405-2de5b2ff42d1 req-621193d6-009f-45b5-9c14-23d129afd89e service nova] Lock "d63a2a0d-2736-4d1f-a9fb-ec105747dcc3-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1229.127766] env[63515]: DEBUG nova.compute.manager [req-f13f5af3-1268-4c7e-8405-2de5b2ff42d1 req-621193d6-009f-45b5-9c14-23d129afd89e service nova] [instance: d63a2a0d-2736-4d1f-a9fb-ec105747dcc3] No waiting events found dispatching network-vif-plugged-8e9d8cfb-9749-419b-b009-786e95b2f419 {{(pid=63515) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1229.128351] env[63515]: WARNING nova.compute.manager [req-f13f5af3-1268-4c7e-8405-2de5b2ff42d1 req-621193d6-009f-45b5-9c14-23d129afd89e service nova] [instance: d63a2a0d-2736-4d1f-a9fb-ec105747dcc3] Received unexpected event network-vif-plugged-8e9d8cfb-9749-419b-b009-786e95b2f419 for instance with vm_state building and task_state spawning. [ 1229.493110] env[63515]: DEBUG oslo_concurrency.lockutils [None req-7516c657-2678-40c1-9786-3e48e1357d97 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Acquiring lock "24a79a6c-5210-4778-a6ed-482bfe14933c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1229.493523] env[63515]: DEBUG oslo_concurrency.lockutils [None req-7516c657-2678-40c1-9786-3e48e1357d97 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Lock "24a79a6c-5210-4778-a6ed-482bfe14933c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1229.493751] env[63515]: DEBUG oslo_concurrency.lockutils [None req-7516c657-2678-40c1-9786-3e48e1357d97 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Acquiring lock "24a79a6c-5210-4778-a6ed-482bfe14933c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1229.493945] env[63515]: DEBUG oslo_concurrency.lockutils [None req-7516c657-2678-40c1-9786-3e48e1357d97 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Lock "24a79a6c-5210-4778-a6ed-482bfe14933c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1229.494173] env[63515]: DEBUG oslo_concurrency.lockutils [None req-7516c657-2678-40c1-9786-3e48e1357d97 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Lock "24a79a6c-5210-4778-a6ed-482bfe14933c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1229.496350] env[63515]: INFO nova.compute.manager [None req-7516c657-2678-40c1-9786-3e48e1357d97 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 24a79a6c-5210-4778-a6ed-482bfe14933c] Terminating instance [ 1229.498218] env[63515]: DEBUG nova.compute.manager [None req-7516c657-2678-40c1-9786-3e48e1357d97 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 24a79a6c-5210-4778-a6ed-482bfe14933c] Start destroying the instance on the hypervisor. {{(pid=63515) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1229.498413] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-7516c657-2678-40c1-9786-3e48e1357d97 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 24a79a6c-5210-4778-a6ed-482bfe14933c] Destroying instance {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1229.499426] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-edc23df8-4622-4882-91da-760088d13617 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1229.507322] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-7516c657-2678-40c1-9786-3e48e1357d97 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 24a79a6c-5210-4778-a6ed-482bfe14933c] Powering off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1229.507549] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-afeb4f33-5042-4ca8-ae4a-1ad532d72af1 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1229.513468] env[63515]: DEBUG oslo_vmware.api [None req-7516c657-2678-40c1-9786-3e48e1357d97 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Waiting for the task: (returnval){ [ 1229.513468] env[63515]: value = "task-1112089" [ 1229.513468] env[63515]: _type = "Task" [ 1229.513468] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1229.520830] env[63515]: DEBUG oslo_vmware.api [None req-7516c657-2678-40c1-9786-3e48e1357d97 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Task: {'id': task-1112089, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1229.684094] env[63515]: DEBUG nova.network.neutron [None req-7a41c677-8af5-4616-8502-88e4da1d930f tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] [instance: d63a2a0d-2736-4d1f-a9fb-ec105747dcc3] Successfully updated port: 8e9d8cfb-9749-419b-b009-786e95b2f419 {{(pid=63515) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1229.707071] env[63515]: DEBUG nova.compute.manager [req-52073ded-29c8-49c7-8972-658f8beae202 req-ca4df250-ec34-40be-876c-6c4a25793a42 service nova] [instance: d63a2a0d-2736-4d1f-a9fb-ec105747dcc3] Received event network-changed-8e9d8cfb-9749-419b-b009-786e95b2f419 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1229.707291] env[63515]: DEBUG nova.compute.manager [req-52073ded-29c8-49c7-8972-658f8beae202 req-ca4df250-ec34-40be-876c-6c4a25793a42 service nova] [instance: d63a2a0d-2736-4d1f-a9fb-ec105747dcc3] Refreshing instance network info cache due to event network-changed-8e9d8cfb-9749-419b-b009-786e95b2f419. {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1229.707512] env[63515]: DEBUG oslo_concurrency.lockutils [req-52073ded-29c8-49c7-8972-658f8beae202 req-ca4df250-ec34-40be-876c-6c4a25793a42 service nova] Acquiring lock "refresh_cache-d63a2a0d-2736-4d1f-a9fb-ec105747dcc3" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1229.707658] env[63515]: DEBUG oslo_concurrency.lockutils [req-52073ded-29c8-49c7-8972-658f8beae202 req-ca4df250-ec34-40be-876c-6c4a25793a42 service nova] Acquired lock "refresh_cache-d63a2a0d-2736-4d1f-a9fb-ec105747dcc3" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1229.707837] env[63515]: DEBUG nova.network.neutron [req-52073ded-29c8-49c7-8972-658f8beae202 req-ca4df250-ec34-40be-876c-6c4a25793a42 service nova] [instance: d63a2a0d-2736-4d1f-a9fb-ec105747dcc3] Refreshing network info cache for port 8e9d8cfb-9749-419b-b009-786e95b2f419 {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1230.023316] env[63515]: DEBUG oslo_vmware.api [None req-7516c657-2678-40c1-9786-3e48e1357d97 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Task: {'id': task-1112089, 'name': PowerOffVM_Task, 'duration_secs': 0.202426} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1230.023586] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-7516c657-2678-40c1-9786-3e48e1357d97 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 24a79a6c-5210-4778-a6ed-482bfe14933c] Powered off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1230.023759] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-7516c657-2678-40c1-9786-3e48e1357d97 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 24a79a6c-5210-4778-a6ed-482bfe14933c] Unregistering the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1230.024026] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f7131fdc-0309-4550-a2e6-25815048a639 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1230.089735] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-7516c657-2678-40c1-9786-3e48e1357d97 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 24a79a6c-5210-4778-a6ed-482bfe14933c] Unregistered the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1230.090014] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-7516c657-2678-40c1-9786-3e48e1357d97 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 24a79a6c-5210-4778-a6ed-482bfe14933c] Deleting contents of the VM from datastore datastore1 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1230.090314] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-7516c657-2678-40c1-9786-3e48e1357d97 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Deleting the datastore file [datastore1] 24a79a6c-5210-4778-a6ed-482bfe14933c {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1230.090604] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d6ab2251-4621-4b35-904a-9138437d6ad3 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1230.096444] env[63515]: DEBUG oslo_vmware.api [None req-7516c657-2678-40c1-9786-3e48e1357d97 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Waiting for the task: (returnval){ [ 1230.096444] env[63515]: value = "task-1112091" [ 1230.096444] env[63515]: _type = "Task" [ 1230.096444] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1230.103787] env[63515]: DEBUG oslo_vmware.api [None req-7516c657-2678-40c1-9786-3e48e1357d97 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Task: {'id': task-1112091, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1230.189739] env[63515]: DEBUG oslo_concurrency.lockutils [None req-7a41c677-8af5-4616-8502-88e4da1d930f tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Acquiring lock "refresh_cache-d63a2a0d-2736-4d1f-a9fb-ec105747dcc3" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1230.241918] env[63515]: DEBUG nova.network.neutron [req-52073ded-29c8-49c7-8972-658f8beae202 req-ca4df250-ec34-40be-876c-6c4a25793a42 service nova] [instance: d63a2a0d-2736-4d1f-a9fb-ec105747dcc3] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1230.317920] env[63515]: DEBUG nova.network.neutron [req-52073ded-29c8-49c7-8972-658f8beae202 req-ca4df250-ec34-40be-876c-6c4a25793a42 service nova] [instance: d63a2a0d-2736-4d1f-a9fb-ec105747dcc3] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1230.605934] env[63515]: DEBUG oslo_vmware.api [None req-7516c657-2678-40c1-9786-3e48e1357d97 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Task: {'id': task-1112091, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.139177} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1230.606303] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-7516c657-2678-40c1-9786-3e48e1357d97 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Deleted the datastore file {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1230.606343] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-7516c657-2678-40c1-9786-3e48e1357d97 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 24a79a6c-5210-4778-a6ed-482bfe14933c] Deleted contents of the VM from datastore datastore1 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1230.606500] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-7516c657-2678-40c1-9786-3e48e1357d97 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 24a79a6c-5210-4778-a6ed-482bfe14933c] Instance destroyed {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1230.606674] env[63515]: INFO nova.compute.manager [None req-7516c657-2678-40c1-9786-3e48e1357d97 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] [instance: 24a79a6c-5210-4778-a6ed-482bfe14933c] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1230.606912] env[63515]: DEBUG oslo.service.loopingcall [None req-7516c657-2678-40c1-9786-3e48e1357d97 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1230.607136] env[63515]: DEBUG nova.compute.manager [-] [instance: 24a79a6c-5210-4778-a6ed-482bfe14933c] Deallocating network for instance {{(pid=63515) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1230.607247] env[63515]: DEBUG nova.network.neutron [-] [instance: 24a79a6c-5210-4778-a6ed-482bfe14933c] deallocate_for_instance() {{(pid=63515) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1230.820596] env[63515]: DEBUG oslo_concurrency.lockutils [req-52073ded-29c8-49c7-8972-658f8beae202 req-ca4df250-ec34-40be-876c-6c4a25793a42 service nova] Releasing lock "refresh_cache-d63a2a0d-2736-4d1f-a9fb-ec105747dcc3" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1230.820968] env[63515]: DEBUG oslo_concurrency.lockutils [None req-7a41c677-8af5-4616-8502-88e4da1d930f tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Acquired lock "refresh_cache-d63a2a0d-2736-4d1f-a9fb-ec105747dcc3" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1230.821145] env[63515]: DEBUG nova.network.neutron [None req-7a41c677-8af5-4616-8502-88e4da1d930f tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] [instance: d63a2a0d-2736-4d1f-a9fb-ec105747dcc3] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1231.358703] env[63515]: DEBUG nova.network.neutron [None req-7a41c677-8af5-4616-8502-88e4da1d930f tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] [instance: d63a2a0d-2736-4d1f-a9fb-ec105747dcc3] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1231.481155] env[63515]: DEBUG nova.network.neutron [None req-7a41c677-8af5-4616-8502-88e4da1d930f tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] [instance: d63a2a0d-2736-4d1f-a9fb-ec105747dcc3] Updating instance_info_cache with network_info: [{"id": "8e9d8cfb-9749-419b-b009-786e95b2f419", "address": "fa:16:3e:59:53:ff", "network": {"id": "d6da2650-9ba9-4ef6-8035-e58b635a64b2", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-391564968-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "97f2e119708746038f38359572c99438", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "043ca97b-0fca-4b54-8be6-027123fa76d0", "external-id": "nsx-vlan-transportzone-702", "segmentation_id": 702, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8e9d8cfb-97", "ovs_interfaceid": "8e9d8cfb-9749-419b-b009-786e95b2f419", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1231.586485] env[63515]: DEBUG nova.network.neutron [-] [instance: 24a79a6c-5210-4778-a6ed-482bfe14933c] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1231.731877] env[63515]: DEBUG nova.compute.manager [req-7b075692-22b6-447d-af7c-cfa163383800 req-5a9fd549-a034-4226-988b-602014c27f5f service nova] [instance: 24a79a6c-5210-4778-a6ed-482bfe14933c] Received event network-vif-deleted-f338b564-7636-4839-b034-73c3233d7da6 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1231.984276] env[63515]: DEBUG oslo_concurrency.lockutils [None req-7a41c677-8af5-4616-8502-88e4da1d930f tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Releasing lock "refresh_cache-d63a2a0d-2736-4d1f-a9fb-ec105747dcc3" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1231.984612] env[63515]: DEBUG nova.compute.manager [None req-7a41c677-8af5-4616-8502-88e4da1d930f tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] [instance: d63a2a0d-2736-4d1f-a9fb-ec105747dcc3] Instance network_info: |[{"id": "8e9d8cfb-9749-419b-b009-786e95b2f419", "address": "fa:16:3e:59:53:ff", "network": {"id": "d6da2650-9ba9-4ef6-8035-e58b635a64b2", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-391564968-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "97f2e119708746038f38359572c99438", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "043ca97b-0fca-4b54-8be6-027123fa76d0", "external-id": "nsx-vlan-transportzone-702", "segmentation_id": 702, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8e9d8cfb-97", "ovs_interfaceid": "8e9d8cfb-9749-419b-b009-786e95b2f419", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1231.985061] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-7a41c677-8af5-4616-8502-88e4da1d930f tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] [instance: d63a2a0d-2736-4d1f-a9fb-ec105747dcc3] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:59:53:ff', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '043ca97b-0fca-4b54-8be6-027123fa76d0', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '8e9d8cfb-9749-419b-b009-786e95b2f419', 'vif_model': 'vmxnet3'}] {{(pid=63515) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1231.992484] env[63515]: DEBUG oslo.service.loopingcall [None req-7a41c677-8af5-4616-8502-88e4da1d930f tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1231.993029] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d63a2a0d-2736-4d1f-a9fb-ec105747dcc3] Creating VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1231.993272] env[63515]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-8cf96dbf-53c5-460a-8a13-9139ac8041c8 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1232.016592] env[63515]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1232.016592] env[63515]: value = "task-1112092" [ 1232.016592] env[63515]: _type = "Task" [ 1232.016592] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1232.024735] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1112092, 'name': CreateVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1232.090620] env[63515]: INFO nova.compute.manager [-] [instance: 24a79a6c-5210-4778-a6ed-482bfe14933c] Took 1.48 seconds to deallocate network for instance. [ 1232.527803] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1112092, 'name': CreateVM_Task, 'duration_secs': 0.294329} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1232.527968] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d63a2a0d-2736-4d1f-a9fb-ec105747dcc3] Created VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1232.528715] env[63515]: DEBUG oslo_concurrency.lockutils [None req-7a41c677-8af5-4616-8502-88e4da1d930f tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1232.528929] env[63515]: DEBUG oslo_concurrency.lockutils [None req-7a41c677-8af5-4616-8502-88e4da1d930f tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1232.529314] env[63515]: DEBUG oslo_concurrency.lockutils [None req-7a41c677-8af5-4616-8502-88e4da1d930f tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1232.529588] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6b3c9b89-d7ad-4fa1-a7fa-a529a90d2bc8 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1232.534168] env[63515]: DEBUG oslo_vmware.api [None req-7a41c677-8af5-4616-8502-88e4da1d930f tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Waiting for the task: (returnval){ [ 1232.534168] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]52c85c25-d97c-0e1b-063e-35478891d22b" [ 1232.534168] env[63515]: _type = "Task" [ 1232.534168] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1232.542127] env[63515]: DEBUG oslo_vmware.api [None req-7a41c677-8af5-4616-8502-88e4da1d930f tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52c85c25-d97c-0e1b-063e-35478891d22b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1232.597313] env[63515]: DEBUG oslo_concurrency.lockutils [None req-7516c657-2678-40c1-9786-3e48e1357d97 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1232.597606] env[63515]: DEBUG oslo_concurrency.lockutils [None req-7516c657-2678-40c1-9786-3e48e1357d97 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1232.597873] env[63515]: DEBUG nova.objects.instance [None req-7516c657-2678-40c1-9786-3e48e1357d97 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Lazy-loading 'resources' on Instance uuid 24a79a6c-5210-4778-a6ed-482bfe14933c {{(pid=63515) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1233.044587] env[63515]: DEBUG oslo_vmware.api [None req-7a41c677-8af5-4616-8502-88e4da1d930f tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52c85c25-d97c-0e1b-063e-35478891d22b, 'name': SearchDatastore_Task, 'duration_secs': 0.008476} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1233.044937] env[63515]: DEBUG oslo_concurrency.lockutils [None req-7a41c677-8af5-4616-8502-88e4da1d930f tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1233.045205] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-7a41c677-8af5-4616-8502-88e4da1d930f tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] [instance: d63a2a0d-2736-4d1f-a9fb-ec105747dcc3] Processing image 8a120570-cb06-4099-b262-554ca0ad15c5 {{(pid=63515) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1233.045440] env[63515]: DEBUG oslo_concurrency.lockutils [None req-7a41c677-8af5-4616-8502-88e4da1d930f tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1233.045593] env[63515]: DEBUG oslo_concurrency.lockutils [None req-7a41c677-8af5-4616-8502-88e4da1d930f tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1233.045775] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-7a41c677-8af5-4616-8502-88e4da1d930f tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1233.046043] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2f12c8f6-f982-48e9-912f-a5f1ae55fa44 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1233.053730] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-7a41c677-8af5-4616-8502-88e4da1d930f tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1233.053916] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-7a41c677-8af5-4616-8502-88e4da1d930f tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63515) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1233.054625] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6aa119cd-227a-4fff-b395-7d0d92a3fde6 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1233.059286] env[63515]: DEBUG oslo_vmware.api [None req-7a41c677-8af5-4616-8502-88e4da1d930f tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Waiting for the task: (returnval){ [ 1233.059286] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]52404e5b-418c-83df-26cf-c1483df44f3a" [ 1233.059286] env[63515]: _type = "Task" [ 1233.059286] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1233.066803] env[63515]: DEBUG oslo_vmware.api [None req-7a41c677-8af5-4616-8502-88e4da1d930f tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52404e5b-418c-83df-26cf-c1483df44f3a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1233.141978] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d81c50de-68b9-419d-8ba2-ac5941a9f267 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1233.148969] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12b7431e-3ca7-4b1c-b34f-2c68d51e7715 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1233.179591] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5dd331cc-f3d0-4205-af5f-1073bb57941f {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1233.187871] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c55e018d-3931-444f-9b0b-d7e4ac57e2a8 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1233.200429] env[63515]: DEBUG nova.compute.provider_tree [None req-7516c657-2678-40c1-9786-3e48e1357d97 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1233.569403] env[63515]: DEBUG oslo_vmware.api [None req-7a41c677-8af5-4616-8502-88e4da1d930f tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52404e5b-418c-83df-26cf-c1483df44f3a, 'name': SearchDatastore_Task, 'duration_secs': 0.008238} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1233.570203] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8fcf6cca-e62e-4660-8c40-a598fe6c7f49 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1233.574893] env[63515]: DEBUG oslo_vmware.api [None req-7a41c677-8af5-4616-8502-88e4da1d930f tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Waiting for the task: (returnval){ [ 1233.574893] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]52b14a9d-28cc-f043-1f68-853284344638" [ 1233.574893] env[63515]: _type = "Task" [ 1233.574893] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1233.582218] env[63515]: DEBUG oslo_vmware.api [None req-7a41c677-8af5-4616-8502-88e4da1d930f tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52b14a9d-28cc-f043-1f68-853284344638, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1233.703479] env[63515]: DEBUG nova.scheduler.client.report [None req-7516c657-2678-40c1-9786-3e48e1357d97 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1234.085175] env[63515]: DEBUG oslo_vmware.api [None req-7a41c677-8af5-4616-8502-88e4da1d930f tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52b14a9d-28cc-f043-1f68-853284344638, 'name': SearchDatastore_Task, 'duration_secs': 0.008536} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1234.085496] env[63515]: DEBUG oslo_concurrency.lockutils [None req-7a41c677-8af5-4616-8502-88e4da1d930f tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1234.085759] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-7a41c677-8af5-4616-8502-88e4da1d930f tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk to [datastore1] d63a2a0d-2736-4d1f-a9fb-ec105747dcc3/d63a2a0d-2736-4d1f-a9fb-ec105747dcc3.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1234.086092] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-55611ba4-fedf-4f9c-802e-4ebc69dc4b57 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1234.092318] env[63515]: DEBUG oslo_vmware.api [None req-7a41c677-8af5-4616-8502-88e4da1d930f tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Waiting for the task: (returnval){ [ 1234.092318] env[63515]: value = "task-1112093" [ 1234.092318] env[63515]: _type = "Task" [ 1234.092318] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1234.099692] env[63515]: DEBUG oslo_vmware.api [None req-7a41c677-8af5-4616-8502-88e4da1d930f tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Task: {'id': task-1112093, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1234.208865] env[63515]: DEBUG oslo_concurrency.lockutils [None req-7516c657-2678-40c1-9786-3e48e1357d97 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.611s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1234.231250] env[63515]: INFO nova.scheduler.client.report [None req-7516c657-2678-40c1-9786-3e48e1357d97 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Deleted allocations for instance 24a79a6c-5210-4778-a6ed-482bfe14933c [ 1234.602609] env[63515]: DEBUG oslo_vmware.api [None req-7a41c677-8af5-4616-8502-88e4da1d930f tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Task: {'id': task-1112093, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.430812} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1234.602870] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-7a41c677-8af5-4616-8502-88e4da1d930f tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk to [datastore1] d63a2a0d-2736-4d1f-a9fb-ec105747dcc3/d63a2a0d-2736-4d1f-a9fb-ec105747dcc3.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1234.603102] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-7a41c677-8af5-4616-8502-88e4da1d930f tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] [instance: d63a2a0d-2736-4d1f-a9fb-ec105747dcc3] Extending root virtual disk to 1048576 {{(pid=63515) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1234.603358] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-f82bb134-7041-463c-a1ec-3b3a8b321265 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1234.609257] env[63515]: DEBUG oslo_vmware.api [None req-7a41c677-8af5-4616-8502-88e4da1d930f tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Waiting for the task: (returnval){ [ 1234.609257] env[63515]: value = "task-1112095" [ 1234.609257] env[63515]: _type = "Task" [ 1234.609257] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1234.615807] env[63515]: DEBUG oslo_vmware.api [None req-7a41c677-8af5-4616-8502-88e4da1d930f tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Task: {'id': task-1112095, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1234.739561] env[63515]: DEBUG oslo_concurrency.lockutils [None req-7516c657-2678-40c1-9786-3e48e1357d97 tempest-ServerActionsTestOtherB-316162520 tempest-ServerActionsTestOtherB-316162520-project-member] Lock "24a79a6c-5210-4778-a6ed-482bfe14933c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.246s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1235.120366] env[63515]: DEBUG oslo_vmware.api [None req-7a41c677-8af5-4616-8502-88e4da1d930f tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Task: {'id': task-1112095, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.058778} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1235.120678] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-7a41c677-8af5-4616-8502-88e4da1d930f tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] [instance: d63a2a0d-2736-4d1f-a9fb-ec105747dcc3] Extended root virtual disk {{(pid=63515) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1235.121395] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc5aa967-57aa-4f7a-b37a-5d6fbc4f7154 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1235.141979] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-7a41c677-8af5-4616-8502-88e4da1d930f tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] [instance: d63a2a0d-2736-4d1f-a9fb-ec105747dcc3] Reconfiguring VM instance instance-00000071 to attach disk [datastore1] d63a2a0d-2736-4d1f-a9fb-ec105747dcc3/d63a2a0d-2736-4d1f-a9fb-ec105747dcc3.vmdk or device None with type sparse {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1235.142212] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-02eec467-8c2b-409a-ac0f-ab6b9ad75e8a {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1235.160505] env[63515]: DEBUG oslo_vmware.api [None req-7a41c677-8af5-4616-8502-88e4da1d930f tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Waiting for the task: (returnval){ [ 1235.160505] env[63515]: value = "task-1112096" [ 1235.160505] env[63515]: _type = "Task" [ 1235.160505] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1235.167804] env[63515]: DEBUG oslo_vmware.api [None req-7a41c677-8af5-4616-8502-88e4da1d930f tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Task: {'id': task-1112096, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1235.670344] env[63515]: DEBUG oslo_vmware.api [None req-7a41c677-8af5-4616-8502-88e4da1d930f tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Task: {'id': task-1112096, 'name': ReconfigVM_Task, 'duration_secs': 0.299931} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1235.670636] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-7a41c677-8af5-4616-8502-88e4da1d930f tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] [instance: d63a2a0d-2736-4d1f-a9fb-ec105747dcc3] Reconfigured VM instance instance-00000071 to attach disk [datastore1] d63a2a0d-2736-4d1f-a9fb-ec105747dcc3/d63a2a0d-2736-4d1f-a9fb-ec105747dcc3.vmdk or device None with type sparse {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1235.671276] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e13c74e4-57ad-4eff-b0ec-6dddcc7123f3 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1235.677283] env[63515]: DEBUG oslo_vmware.api [None req-7a41c677-8af5-4616-8502-88e4da1d930f tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Waiting for the task: (returnval){ [ 1235.677283] env[63515]: value = "task-1112097" [ 1235.677283] env[63515]: _type = "Task" [ 1235.677283] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1235.684224] env[63515]: DEBUG oslo_vmware.api [None req-7a41c677-8af5-4616-8502-88e4da1d930f tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Task: {'id': task-1112097, 'name': Rename_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1236.186732] env[63515]: DEBUG oslo_vmware.api [None req-7a41c677-8af5-4616-8502-88e4da1d930f tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Task: {'id': task-1112097, 'name': Rename_Task, 'duration_secs': 0.126868} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1236.187117] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-7a41c677-8af5-4616-8502-88e4da1d930f tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] [instance: d63a2a0d-2736-4d1f-a9fb-ec105747dcc3] Powering on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1236.187279] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ead99d3a-8cd0-4552-ad77-8c494e2b61bd {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1236.193800] env[63515]: DEBUG oslo_vmware.api [None req-7a41c677-8af5-4616-8502-88e4da1d930f tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Waiting for the task: (returnval){ [ 1236.193800] env[63515]: value = "task-1112098" [ 1236.193800] env[63515]: _type = "Task" [ 1236.193800] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1236.200793] env[63515]: DEBUG oslo_vmware.api [None req-7a41c677-8af5-4616-8502-88e4da1d930f tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Task: {'id': task-1112098, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1236.704208] env[63515]: DEBUG oslo_vmware.api [None req-7a41c677-8af5-4616-8502-88e4da1d930f tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Task: {'id': task-1112098, 'name': PowerOnVM_Task, 'duration_secs': 0.428076} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1236.704475] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-7a41c677-8af5-4616-8502-88e4da1d930f tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] [instance: d63a2a0d-2736-4d1f-a9fb-ec105747dcc3] Powered on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1236.704683] env[63515]: INFO nova.compute.manager [None req-7a41c677-8af5-4616-8502-88e4da1d930f tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] [instance: d63a2a0d-2736-4d1f-a9fb-ec105747dcc3] Took 7.78 seconds to spawn the instance on the hypervisor. [ 1236.704866] env[63515]: DEBUG nova.compute.manager [None req-7a41c677-8af5-4616-8502-88e4da1d930f tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] [instance: d63a2a0d-2736-4d1f-a9fb-ec105747dcc3] Checking state {{(pid=63515) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1236.705634] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e47e4a29-9b67-42bf-8871-d45487d05f06 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1237.222016] env[63515]: INFO nova.compute.manager [None req-7a41c677-8af5-4616-8502-88e4da1d930f tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] [instance: d63a2a0d-2736-4d1f-a9fb-ec105747dcc3] Took 12.48 seconds to build instance. [ 1237.724745] env[63515]: DEBUG oslo_concurrency.lockutils [None req-7a41c677-8af5-4616-8502-88e4da1d930f tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Lock "d63a2a0d-2736-4d1f-a9fb-ec105747dcc3" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.997s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1238.275474] env[63515]: DEBUG nova.compute.manager [req-d20894b6-e337-42b0-9a15-c23f0b44234d req-19babe16-0570-44f3-b946-7d9fa0ab379a service nova] [instance: d63a2a0d-2736-4d1f-a9fb-ec105747dcc3] Received event network-changed-8e9d8cfb-9749-419b-b009-786e95b2f419 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1238.275735] env[63515]: DEBUG nova.compute.manager [req-d20894b6-e337-42b0-9a15-c23f0b44234d req-19babe16-0570-44f3-b946-7d9fa0ab379a service nova] [instance: d63a2a0d-2736-4d1f-a9fb-ec105747dcc3] Refreshing instance network info cache due to event network-changed-8e9d8cfb-9749-419b-b009-786e95b2f419. {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1238.275897] env[63515]: DEBUG oslo_concurrency.lockutils [req-d20894b6-e337-42b0-9a15-c23f0b44234d req-19babe16-0570-44f3-b946-7d9fa0ab379a service nova] Acquiring lock "refresh_cache-d63a2a0d-2736-4d1f-a9fb-ec105747dcc3" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1238.276287] env[63515]: DEBUG oslo_concurrency.lockutils [req-d20894b6-e337-42b0-9a15-c23f0b44234d req-19babe16-0570-44f3-b946-7d9fa0ab379a service nova] Acquired lock "refresh_cache-d63a2a0d-2736-4d1f-a9fb-ec105747dcc3" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1238.276530] env[63515]: DEBUG nova.network.neutron [req-d20894b6-e337-42b0-9a15-c23f0b44234d req-19babe16-0570-44f3-b946-7d9fa0ab379a service nova] [instance: d63a2a0d-2736-4d1f-a9fb-ec105747dcc3] Refreshing network info cache for port 8e9d8cfb-9749-419b-b009-786e95b2f419 {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1239.088852] env[63515]: DEBUG nova.network.neutron [req-d20894b6-e337-42b0-9a15-c23f0b44234d req-19babe16-0570-44f3-b946-7d9fa0ab379a service nova] [instance: d63a2a0d-2736-4d1f-a9fb-ec105747dcc3] Updated VIF entry in instance network info cache for port 8e9d8cfb-9749-419b-b009-786e95b2f419. {{(pid=63515) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1239.089209] env[63515]: DEBUG nova.network.neutron [req-d20894b6-e337-42b0-9a15-c23f0b44234d req-19babe16-0570-44f3-b946-7d9fa0ab379a service nova] [instance: d63a2a0d-2736-4d1f-a9fb-ec105747dcc3] Updating instance_info_cache with network_info: [{"id": "8e9d8cfb-9749-419b-b009-786e95b2f419", "address": "fa:16:3e:59:53:ff", "network": {"id": "d6da2650-9ba9-4ef6-8035-e58b635a64b2", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-391564968-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.183", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "97f2e119708746038f38359572c99438", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "043ca97b-0fca-4b54-8be6-027123fa76d0", "external-id": "nsx-vlan-transportzone-702", "segmentation_id": 702, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8e9d8cfb-97", "ovs_interfaceid": "8e9d8cfb-9749-419b-b009-786e95b2f419", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1239.591682] env[63515]: DEBUG oslo_concurrency.lockutils [req-d20894b6-e337-42b0-9a15-c23f0b44234d req-19babe16-0570-44f3-b946-7d9fa0ab379a service nova] Releasing lock "refresh_cache-d63a2a0d-2736-4d1f-a9fb-ec105747dcc3" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1257.534787] env[63515]: DEBUG oslo_service.periodic_task [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=63515) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1260.535610] env[63515]: DEBUG oslo_service.periodic_task [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=63515) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1262.534848] env[63515]: DEBUG oslo_service.periodic_task [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=63515) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1262.535250] env[63515]: DEBUG nova.compute.manager [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=63515) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10545}} [ 1263.535397] env[63515]: DEBUG oslo_service.periodic_task [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=63515) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1263.535766] env[63515]: DEBUG oslo_service.periodic_task [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=63515) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1264.535469] env[63515]: DEBUG oslo_service.periodic_task [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=63515) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1264.535866] env[63515]: DEBUG nova.compute.manager [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Starting heal instance info cache {{(pid=63515) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9926}} [ 1265.541527] env[63515]: DEBUG nova.compute.manager [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Didn't find any instances for network info cache update. {{(pid=63515) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10012}} [ 1265.541895] env[63515]: DEBUG oslo_service.periodic_task [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=63515) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1266.536270] env[63515]: DEBUG oslo_service.periodic_task [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=63515) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1268.535255] env[63515]: DEBUG oslo_service.periodic_task [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Running periodic task ComputeManager.update_available_resource {{(pid=63515) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1269.038649] env[63515]: DEBUG oslo_concurrency.lockutils [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1269.038894] env[63515]: DEBUG oslo_concurrency.lockutils [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1269.039085] env[63515]: DEBUG oslo_concurrency.lockutils [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1269.039248] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=63515) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1269.040173] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d80eced-5159-42bc-a6f4-39fa75127b1d {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1269.048575] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78e4af0e-d7f9-484b-8b65-969fa221fefa {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1269.062095] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c147c5b9-df56-4c97-9d80-aa08812523ef {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1269.067915] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-108fd72e-4918-4d21-896e-38a093ad4fbb {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1269.096949] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181295MB free_disk=171GB free_vcpus=48 pci_devices=None {{(pid=63515) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1269.097119] env[63515]: DEBUG oslo_concurrency.lockutils [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1269.097286] env[63515]: DEBUG oslo_concurrency.lockutils [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1270.121959] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Instance d63a2a0d-2736-4d1f-a9fb-ec105747dcc3 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63515) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1270.122257] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Total usable vcpus: 48, total allocated vcpus: 1 {{(pid=63515) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1270.122355] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=704MB phys_disk=200GB used_disk=1GB total_vcpus=48 used_vcpus=1 pci_stats=[] {{(pid=63515) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1270.147206] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18f90252-25ab-468c-b4bd-592f436c3429 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1270.154881] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-666f4099-393c-4ae1-9489-d040d9ca90a5 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1270.184727] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fbeb68ca-1d00-4a38-bd42-0cdc33fee08d {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1270.191778] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c215bde6-ef4d-4e65-83a2-df951840e4e4 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1270.204508] env[63515]: DEBUG nova.compute.provider_tree [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1270.707447] env[63515]: DEBUG nova.scheduler.client.report [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1271.212398] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=63515) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1271.212762] env[63515]: DEBUG oslo_concurrency.lockutils [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.115s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1275.067070] env[63515]: DEBUG oslo_concurrency.lockutils [None req-80a706e9-7c0d-4bcd-8ac1-a607239f95b2 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Acquiring lock "d63a2a0d-2736-4d1f-a9fb-ec105747dcc3" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1275.067364] env[63515]: DEBUG oslo_concurrency.lockutils [None req-80a706e9-7c0d-4bcd-8ac1-a607239f95b2 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Lock "d63a2a0d-2736-4d1f-a9fb-ec105747dcc3" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1275.570956] env[63515]: DEBUG nova.compute.utils [None req-80a706e9-7c0d-4bcd-8ac1-a607239f95b2 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Using /dev/sd instead of None {{(pid=63515) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1276.073857] env[63515]: DEBUG oslo_concurrency.lockutils [None req-80a706e9-7c0d-4bcd-8ac1-a607239f95b2 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Lock "d63a2a0d-2736-4d1f-a9fb-ec105747dcc3" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.006s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1277.140418] env[63515]: DEBUG oslo_concurrency.lockutils [None req-80a706e9-7c0d-4bcd-8ac1-a607239f95b2 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Acquiring lock "d63a2a0d-2736-4d1f-a9fb-ec105747dcc3" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1277.140798] env[63515]: DEBUG oslo_concurrency.lockutils [None req-80a706e9-7c0d-4bcd-8ac1-a607239f95b2 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Lock "d63a2a0d-2736-4d1f-a9fb-ec105747dcc3" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1277.140932] env[63515]: INFO nova.compute.manager [None req-80a706e9-7c0d-4bcd-8ac1-a607239f95b2 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] [instance: d63a2a0d-2736-4d1f-a9fb-ec105747dcc3] Attaching volume e15a1aaa-f784-4ec1-b01d-ee7b55fe66ee to /dev/sdb [ 1277.173146] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5dd1efe-d414-489d-8888-049904b87561 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1277.183313] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c46d63ab-aea0-425c-82ec-cc924f28e8f0 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1277.200218] env[63515]: DEBUG nova.virt.block_device [None req-80a706e9-7c0d-4bcd-8ac1-a607239f95b2 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] [instance: d63a2a0d-2736-4d1f-a9fb-ec105747dcc3] Updating existing volume attachment record: 51183a4c-da71-4e06-87c8-dc9e651ff0de {{(pid=63515) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1281.746995] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-80a706e9-7c0d-4bcd-8ac1-a607239f95b2 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] [instance: d63a2a0d-2736-4d1f-a9fb-ec105747dcc3] Volume attach. Driver type: vmdk {{(pid=63515) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1281.747277] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-80a706e9-7c0d-4bcd-8ac1-a607239f95b2 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] [instance: d63a2a0d-2736-4d1f-a9fb-ec105747dcc3] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-243603', 'volume_id': 'e15a1aaa-f784-4ec1-b01d-ee7b55fe66ee', 'name': 'volume-e15a1aaa-f784-4ec1-b01d-ee7b55fe66ee', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'd63a2a0d-2736-4d1f-a9fb-ec105747dcc3', 'attached_at': '', 'detached_at': '', 'volume_id': 'e15a1aaa-f784-4ec1-b01d-ee7b55fe66ee', 'serial': 'e15a1aaa-f784-4ec1-b01d-ee7b55fe66ee'} {{(pid=63515) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1281.748167] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ab1158e-47c6-4d3b-84a4-f18c68ca32c4 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1281.764200] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-adc000f2-1c65-4ebf-bf43-942955128296 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1281.789225] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-80a706e9-7c0d-4bcd-8ac1-a607239f95b2 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] [instance: d63a2a0d-2736-4d1f-a9fb-ec105747dcc3] Reconfiguring VM instance instance-00000071 to attach disk [datastore1] volume-e15a1aaa-f784-4ec1-b01d-ee7b55fe66ee/volume-e15a1aaa-f784-4ec1-b01d-ee7b55fe66ee.vmdk or device None with type thin {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1281.789549] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-87e82181-bd46-41ac-a843-ac03af4bce99 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1281.807607] env[63515]: DEBUG oslo_vmware.api [None req-80a706e9-7c0d-4bcd-8ac1-a607239f95b2 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Waiting for the task: (returnval){ [ 1281.807607] env[63515]: value = "task-1112101" [ 1281.807607] env[63515]: _type = "Task" [ 1281.807607] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1281.818304] env[63515]: DEBUG oslo_vmware.api [None req-80a706e9-7c0d-4bcd-8ac1-a607239f95b2 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Task: {'id': task-1112101, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1282.318421] env[63515]: DEBUG oslo_vmware.api [None req-80a706e9-7c0d-4bcd-8ac1-a607239f95b2 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Task: {'id': task-1112101, 'name': ReconfigVM_Task, 'duration_secs': 0.336541} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1282.318831] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-80a706e9-7c0d-4bcd-8ac1-a607239f95b2 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] [instance: d63a2a0d-2736-4d1f-a9fb-ec105747dcc3] Reconfigured VM instance instance-00000071 to attach disk [datastore1] volume-e15a1aaa-f784-4ec1-b01d-ee7b55fe66ee/volume-e15a1aaa-f784-4ec1-b01d-ee7b55fe66ee.vmdk or device None with type thin {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1282.323397] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-78b28ec1-bf0f-4cfb-b7c8-5a41ac8cd048 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1282.337475] env[63515]: DEBUG oslo_vmware.api [None req-80a706e9-7c0d-4bcd-8ac1-a607239f95b2 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Waiting for the task: (returnval){ [ 1282.337475] env[63515]: value = "task-1112102" [ 1282.337475] env[63515]: _type = "Task" [ 1282.337475] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1282.344921] env[63515]: DEBUG oslo_vmware.api [None req-80a706e9-7c0d-4bcd-8ac1-a607239f95b2 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Task: {'id': task-1112102, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1282.846988] env[63515]: DEBUG oslo_vmware.api [None req-80a706e9-7c0d-4bcd-8ac1-a607239f95b2 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Task: {'id': task-1112102, 'name': ReconfigVM_Task, 'duration_secs': 0.144058} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1282.847372] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-80a706e9-7c0d-4bcd-8ac1-a607239f95b2 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] [instance: d63a2a0d-2736-4d1f-a9fb-ec105747dcc3] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-243603', 'volume_id': 'e15a1aaa-f784-4ec1-b01d-ee7b55fe66ee', 'name': 'volume-e15a1aaa-f784-4ec1-b01d-ee7b55fe66ee', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'd63a2a0d-2736-4d1f-a9fb-ec105747dcc3', 'attached_at': '', 'detached_at': '', 'volume_id': 'e15a1aaa-f784-4ec1-b01d-ee7b55fe66ee', 'serial': 'e15a1aaa-f784-4ec1-b01d-ee7b55fe66ee'} {{(pid=63515) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1283.882957] env[63515]: DEBUG nova.objects.instance [None req-80a706e9-7c0d-4bcd-8ac1-a607239f95b2 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Lazy-loading 'flavor' on Instance uuid d63a2a0d-2736-4d1f-a9fb-ec105747dcc3 {{(pid=63515) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1284.389705] env[63515]: DEBUG oslo_concurrency.lockutils [None req-80a706e9-7c0d-4bcd-8ac1-a607239f95b2 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Lock "d63a2a0d-2736-4d1f-a9fb-ec105747dcc3" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.249s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1284.572437] env[63515]: DEBUG oslo_concurrency.lockutils [None req-ab7acecb-46d2-4bbe-9297-cd7eeab86519 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Acquiring lock "d63a2a0d-2736-4d1f-a9fb-ec105747dcc3" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1284.572840] env[63515]: DEBUG oslo_concurrency.lockutils [None req-ab7acecb-46d2-4bbe-9297-cd7eeab86519 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Lock "d63a2a0d-2736-4d1f-a9fb-ec105747dcc3" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.001s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1285.075922] env[63515]: INFO nova.compute.manager [None req-ab7acecb-46d2-4bbe-9297-cd7eeab86519 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] [instance: d63a2a0d-2736-4d1f-a9fb-ec105747dcc3] Detaching volume e15a1aaa-f784-4ec1-b01d-ee7b55fe66ee [ 1285.105580] env[63515]: INFO nova.virt.block_device [None req-ab7acecb-46d2-4bbe-9297-cd7eeab86519 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] [instance: d63a2a0d-2736-4d1f-a9fb-ec105747dcc3] Attempting to driver detach volume e15a1aaa-f784-4ec1-b01d-ee7b55fe66ee from mountpoint /dev/sdb [ 1285.105822] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-ab7acecb-46d2-4bbe-9297-cd7eeab86519 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] [instance: d63a2a0d-2736-4d1f-a9fb-ec105747dcc3] Volume detach. Driver type: vmdk {{(pid=63515) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1285.106023] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-ab7acecb-46d2-4bbe-9297-cd7eeab86519 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] [instance: d63a2a0d-2736-4d1f-a9fb-ec105747dcc3] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-243603', 'volume_id': 'e15a1aaa-f784-4ec1-b01d-ee7b55fe66ee', 'name': 'volume-e15a1aaa-f784-4ec1-b01d-ee7b55fe66ee', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'd63a2a0d-2736-4d1f-a9fb-ec105747dcc3', 'attached_at': '', 'detached_at': '', 'volume_id': 'e15a1aaa-f784-4ec1-b01d-ee7b55fe66ee', 'serial': 'e15a1aaa-f784-4ec1-b01d-ee7b55fe66ee'} {{(pid=63515) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1285.106890] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8bca81a-6405-4660-93d1-5cc0b92a13b4 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1285.127424] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f61241f1-4151-4700-870a-c4e8b49cab89 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1285.133726] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bdb1e06c-a4fb-4ec8-ad44-5cd487bbd541 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1285.153420] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0114e111-1645-4e77-9cc6-3dc53b3e7eeb {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1285.166810] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-ab7acecb-46d2-4bbe-9297-cd7eeab86519 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] The volume has not been displaced from its original location: [datastore1] volume-e15a1aaa-f784-4ec1-b01d-ee7b55fe66ee/volume-e15a1aaa-f784-4ec1-b01d-ee7b55fe66ee.vmdk. No consolidation needed. {{(pid=63515) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1285.171878] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-ab7acecb-46d2-4bbe-9297-cd7eeab86519 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] [instance: d63a2a0d-2736-4d1f-a9fb-ec105747dcc3] Reconfiguring VM instance instance-00000071 to detach disk 2001 {{(pid=63515) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1285.172144] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-52bcc3cb-72da-40e3-99f2-c8bbed743722 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1285.188809] env[63515]: DEBUG oslo_vmware.api [None req-ab7acecb-46d2-4bbe-9297-cd7eeab86519 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Waiting for the task: (returnval){ [ 1285.188809] env[63515]: value = "task-1112103" [ 1285.188809] env[63515]: _type = "Task" [ 1285.188809] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1285.195865] env[63515]: DEBUG oslo_vmware.api [None req-ab7acecb-46d2-4bbe-9297-cd7eeab86519 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Task: {'id': task-1112103, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1285.697675] env[63515]: DEBUG oslo_vmware.api [None req-ab7acecb-46d2-4bbe-9297-cd7eeab86519 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Task: {'id': task-1112103, 'name': ReconfigVM_Task, 'duration_secs': 0.223749} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1285.697958] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-ab7acecb-46d2-4bbe-9297-cd7eeab86519 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] [instance: d63a2a0d-2736-4d1f-a9fb-ec105747dcc3] Reconfigured VM instance instance-00000071 to detach disk 2001 {{(pid=63515) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1285.702533] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-39f23ac5-c9fd-48d3-b89f-2db3ae77bc3e {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1285.716654] env[63515]: DEBUG oslo_vmware.api [None req-ab7acecb-46d2-4bbe-9297-cd7eeab86519 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Waiting for the task: (returnval){ [ 1285.716654] env[63515]: value = "task-1112104" [ 1285.716654] env[63515]: _type = "Task" [ 1285.716654] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1285.723981] env[63515]: DEBUG oslo_vmware.api [None req-ab7acecb-46d2-4bbe-9297-cd7eeab86519 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Task: {'id': task-1112104, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1286.226084] env[63515]: DEBUG oslo_vmware.api [None req-ab7acecb-46d2-4bbe-9297-cd7eeab86519 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Task: {'id': task-1112104, 'name': ReconfigVM_Task, 'duration_secs': 0.125194} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1286.226398] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-ab7acecb-46d2-4bbe-9297-cd7eeab86519 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] [instance: d63a2a0d-2736-4d1f-a9fb-ec105747dcc3] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-243603', 'volume_id': 'e15a1aaa-f784-4ec1-b01d-ee7b55fe66ee', 'name': 'volume-e15a1aaa-f784-4ec1-b01d-ee7b55fe66ee', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'd63a2a0d-2736-4d1f-a9fb-ec105747dcc3', 'attached_at': '', 'detached_at': '', 'volume_id': 'e15a1aaa-f784-4ec1-b01d-ee7b55fe66ee', 'serial': 'e15a1aaa-f784-4ec1-b01d-ee7b55fe66ee'} {{(pid=63515) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1286.766338] env[63515]: DEBUG nova.objects.instance [None req-ab7acecb-46d2-4bbe-9297-cd7eeab86519 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Lazy-loading 'flavor' on Instance uuid d63a2a0d-2736-4d1f-a9fb-ec105747dcc3 {{(pid=63515) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1287.773631] env[63515]: DEBUG oslo_concurrency.lockutils [None req-ab7acecb-46d2-4bbe-9297-cd7eeab86519 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Lock "d63a2a0d-2736-4d1f-a9fb-ec105747dcc3" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.201s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1288.800202] env[63515]: DEBUG oslo_concurrency.lockutils [None req-36f67330-2ca1-46f0-a6b6-75b0944b372a tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Acquiring lock "d63a2a0d-2736-4d1f-a9fb-ec105747dcc3" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1288.800652] env[63515]: DEBUG oslo_concurrency.lockutils [None req-36f67330-2ca1-46f0-a6b6-75b0944b372a tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Lock "d63a2a0d-2736-4d1f-a9fb-ec105747dcc3" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1288.800716] env[63515]: DEBUG oslo_concurrency.lockutils [None req-36f67330-2ca1-46f0-a6b6-75b0944b372a tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Acquiring lock "d63a2a0d-2736-4d1f-a9fb-ec105747dcc3-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1288.800875] env[63515]: DEBUG oslo_concurrency.lockutils [None req-36f67330-2ca1-46f0-a6b6-75b0944b372a tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Lock "d63a2a0d-2736-4d1f-a9fb-ec105747dcc3-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1288.801067] env[63515]: DEBUG oslo_concurrency.lockutils [None req-36f67330-2ca1-46f0-a6b6-75b0944b372a tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Lock "d63a2a0d-2736-4d1f-a9fb-ec105747dcc3-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1288.803290] env[63515]: INFO nova.compute.manager [None req-36f67330-2ca1-46f0-a6b6-75b0944b372a tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] [instance: d63a2a0d-2736-4d1f-a9fb-ec105747dcc3] Terminating instance [ 1288.805069] env[63515]: DEBUG nova.compute.manager [None req-36f67330-2ca1-46f0-a6b6-75b0944b372a tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] [instance: d63a2a0d-2736-4d1f-a9fb-ec105747dcc3] Start destroying the instance on the hypervisor. {{(pid=63515) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1288.805274] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-36f67330-2ca1-46f0-a6b6-75b0944b372a tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] [instance: d63a2a0d-2736-4d1f-a9fb-ec105747dcc3] Destroying instance {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1288.806103] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-946df646-9f53-4751-a53e-0535c16e2fdf {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1288.813987] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-36f67330-2ca1-46f0-a6b6-75b0944b372a tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] [instance: d63a2a0d-2736-4d1f-a9fb-ec105747dcc3] Powering off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1288.814233] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-fac01265-fb26-452d-b6ae-7ad69746b808 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1288.819659] env[63515]: DEBUG oslo_vmware.api [None req-36f67330-2ca1-46f0-a6b6-75b0944b372a tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Waiting for the task: (returnval){ [ 1288.819659] env[63515]: value = "task-1112105" [ 1288.819659] env[63515]: _type = "Task" [ 1288.819659] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1288.827187] env[63515]: DEBUG oslo_vmware.api [None req-36f67330-2ca1-46f0-a6b6-75b0944b372a tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Task: {'id': task-1112105, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1289.330220] env[63515]: DEBUG oslo_vmware.api [None req-36f67330-2ca1-46f0-a6b6-75b0944b372a tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Task: {'id': task-1112105, 'name': PowerOffVM_Task, 'duration_secs': 0.168067} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1289.330419] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-36f67330-2ca1-46f0-a6b6-75b0944b372a tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] [instance: d63a2a0d-2736-4d1f-a9fb-ec105747dcc3] Powered off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1289.330588] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-36f67330-2ca1-46f0-a6b6-75b0944b372a tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] [instance: d63a2a0d-2736-4d1f-a9fb-ec105747dcc3] Unregistering the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1289.330862] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a4f267a7-8e69-4a69-8af4-d65d00598e60 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1289.388839] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-36f67330-2ca1-46f0-a6b6-75b0944b372a tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] [instance: d63a2a0d-2736-4d1f-a9fb-ec105747dcc3] Unregistered the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1289.389130] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-36f67330-2ca1-46f0-a6b6-75b0944b372a tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] [instance: d63a2a0d-2736-4d1f-a9fb-ec105747dcc3] Deleting contents of the VM from datastore datastore1 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1289.389357] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-36f67330-2ca1-46f0-a6b6-75b0944b372a tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Deleting the datastore file [datastore1] d63a2a0d-2736-4d1f-a9fb-ec105747dcc3 {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1289.389582] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a9d1d5a6-d5c4-49fb-bf6f-bd1b70320b9f {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1289.397650] env[63515]: DEBUG oslo_vmware.api [None req-36f67330-2ca1-46f0-a6b6-75b0944b372a tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Waiting for the task: (returnval){ [ 1289.397650] env[63515]: value = "task-1112107" [ 1289.397650] env[63515]: _type = "Task" [ 1289.397650] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1289.405217] env[63515]: DEBUG oslo_vmware.api [None req-36f67330-2ca1-46f0-a6b6-75b0944b372a tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Task: {'id': task-1112107, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1289.907268] env[63515]: DEBUG oslo_vmware.api [None req-36f67330-2ca1-46f0-a6b6-75b0944b372a tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Task: {'id': task-1112107, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.129013} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1289.907627] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-36f67330-2ca1-46f0-a6b6-75b0944b372a tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Deleted the datastore file {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1289.907741] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-36f67330-2ca1-46f0-a6b6-75b0944b372a tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] [instance: d63a2a0d-2736-4d1f-a9fb-ec105747dcc3] Deleted contents of the VM from datastore datastore1 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1289.907872] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-36f67330-2ca1-46f0-a6b6-75b0944b372a tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] [instance: d63a2a0d-2736-4d1f-a9fb-ec105747dcc3] Instance destroyed {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1289.908056] env[63515]: INFO nova.compute.manager [None req-36f67330-2ca1-46f0-a6b6-75b0944b372a tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] [instance: d63a2a0d-2736-4d1f-a9fb-ec105747dcc3] Took 1.10 seconds to destroy the instance on the hypervisor. [ 1289.908303] env[63515]: DEBUG oslo.service.loopingcall [None req-36f67330-2ca1-46f0-a6b6-75b0944b372a tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1289.908505] env[63515]: DEBUG nova.compute.manager [-] [instance: d63a2a0d-2736-4d1f-a9fb-ec105747dcc3] Deallocating network for instance {{(pid=63515) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1289.908587] env[63515]: DEBUG nova.network.neutron [-] [instance: d63a2a0d-2736-4d1f-a9fb-ec105747dcc3] deallocate_for_instance() {{(pid=63515) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1290.514761] env[63515]: DEBUG nova.compute.manager [req-d257340a-8a46-492f-8ac5-529e401841c5 req-81ce0824-a70e-43aa-ab41-d3f4358fc7bb service nova] [instance: d63a2a0d-2736-4d1f-a9fb-ec105747dcc3] Received event network-vif-deleted-8e9d8cfb-9749-419b-b009-786e95b2f419 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1290.514969] env[63515]: INFO nova.compute.manager [req-d257340a-8a46-492f-8ac5-529e401841c5 req-81ce0824-a70e-43aa-ab41-d3f4358fc7bb service nova] [instance: d63a2a0d-2736-4d1f-a9fb-ec105747dcc3] Neutron deleted interface 8e9d8cfb-9749-419b-b009-786e95b2f419; detaching it from the instance and deleting it from the info cache [ 1290.515189] env[63515]: DEBUG nova.network.neutron [req-d257340a-8a46-492f-8ac5-529e401841c5 req-81ce0824-a70e-43aa-ab41-d3f4358fc7bb service nova] [instance: d63a2a0d-2736-4d1f-a9fb-ec105747dcc3] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1290.989010] env[63515]: DEBUG nova.network.neutron [-] [instance: d63a2a0d-2736-4d1f-a9fb-ec105747dcc3] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1291.018147] env[63515]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-42b79373-f474-402a-a141-2d33d3925ad0 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1291.027407] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c91d672-46d3-4425-801c-585fce7dbe5c {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1291.050019] env[63515]: DEBUG nova.compute.manager [req-d257340a-8a46-492f-8ac5-529e401841c5 req-81ce0824-a70e-43aa-ab41-d3f4358fc7bb service nova] [instance: d63a2a0d-2736-4d1f-a9fb-ec105747dcc3] Detach interface failed, port_id=8e9d8cfb-9749-419b-b009-786e95b2f419, reason: Instance d63a2a0d-2736-4d1f-a9fb-ec105747dcc3 could not be found. {{(pid=63515) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1291.491761] env[63515]: INFO nova.compute.manager [-] [instance: d63a2a0d-2736-4d1f-a9fb-ec105747dcc3] Took 1.58 seconds to deallocate network for instance. [ 1291.998825] env[63515]: DEBUG oslo_concurrency.lockutils [None req-36f67330-2ca1-46f0-a6b6-75b0944b372a tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1291.999228] env[63515]: DEBUG oslo_concurrency.lockutils [None req-36f67330-2ca1-46f0-a6b6-75b0944b372a tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1291.999379] env[63515]: DEBUG nova.objects.instance [None req-36f67330-2ca1-46f0-a6b6-75b0944b372a tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Lazy-loading 'resources' on Instance uuid d63a2a0d-2736-4d1f-a9fb-ec105747dcc3 {{(pid=63515) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1292.520283] env[63515]: DEBUG nova.scheduler.client.report [None req-36f67330-2ca1-46f0-a6b6-75b0944b372a tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Refreshing inventories for resource provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 1292.535700] env[63515]: DEBUG nova.scheduler.client.report [None req-36f67330-2ca1-46f0-a6b6-75b0944b372a tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Updating ProviderTree inventory for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 1292.535925] env[63515]: DEBUG nova.compute.provider_tree [None req-36f67330-2ca1-46f0-a6b6-75b0944b372a tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Updating inventory in ProviderTree for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1292.546301] env[63515]: DEBUG nova.scheduler.client.report [None req-36f67330-2ca1-46f0-a6b6-75b0944b372a tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Refreshing aggregate associations for resource provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a, aggregates: None {{(pid=63515) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 1292.563134] env[63515]: DEBUG nova.scheduler.client.report [None req-36f67330-2ca1-46f0-a6b6-75b0944b372a tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Refreshing trait associations for resource provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a, traits: COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=63515) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 1292.588719] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c55b95d-205e-42a9-886d-b0336d1fe2ef {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1292.596011] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-636ea845-1dc3-4920-9abb-b37af97ba062 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1292.625702] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f27f192-8471-4848-afe4-8ebead0f21bc {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1292.632409] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fdf7bdc6-7d62-4c00-a204-03eecba99e93 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1292.644814] env[63515]: DEBUG nova.compute.provider_tree [None req-36f67330-2ca1-46f0-a6b6-75b0944b372a tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1293.148272] env[63515]: DEBUG nova.scheduler.client.report [None req-36f67330-2ca1-46f0-a6b6-75b0944b372a tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1293.653631] env[63515]: DEBUG oslo_concurrency.lockutils [None req-36f67330-2ca1-46f0-a6b6-75b0944b372a tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.654s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1293.676264] env[63515]: INFO nova.scheduler.client.report [None req-36f67330-2ca1-46f0-a6b6-75b0944b372a tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Deleted allocations for instance d63a2a0d-2736-4d1f-a9fb-ec105747dcc3 [ 1294.184997] env[63515]: DEBUG oslo_concurrency.lockutils [None req-36f67330-2ca1-46f0-a6b6-75b0944b372a tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Lock "d63a2a0d-2736-4d1f-a9fb-ec105747dcc3" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.384s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1295.778510] env[63515]: DEBUG oslo_concurrency.lockutils [None req-7940c6a0-845c-481e-9b68-923f5a059a92 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Acquiring lock "b9b28202-1599-4f11-8d92-4942f90160ff" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1295.778817] env[63515]: DEBUG oslo_concurrency.lockutils [None req-7940c6a0-845c-481e-9b68-923f5a059a92 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Lock "b9b28202-1599-4f11-8d92-4942f90160ff" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1296.280640] env[63515]: DEBUG nova.compute.manager [None req-7940c6a0-845c-481e-9b68-923f5a059a92 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] [instance: b9b28202-1599-4f11-8d92-4942f90160ff] Starting instance... {{(pid=63515) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1296.799923] env[63515]: DEBUG oslo_concurrency.lockutils [None req-7940c6a0-845c-481e-9b68-923f5a059a92 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1296.800219] env[63515]: DEBUG oslo_concurrency.lockutils [None req-7940c6a0-845c-481e-9b68-923f5a059a92 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1296.801748] env[63515]: INFO nova.compute.claims [None req-7940c6a0-845c-481e-9b68-923f5a059a92 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] [instance: b9b28202-1599-4f11-8d92-4942f90160ff] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1297.835128] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a181297-5d65-43d3-b264-ad37b73ed66a {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1297.842699] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9da6df3d-ae0d-40f2-9ce1-c711183b14b5 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1297.871794] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32876a72-8528-4e50-9f46-5d67614429c5 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1297.878571] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f45dcdf-a951-49eb-8ad4-2ec403865b79 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1297.890979] env[63515]: DEBUG nova.compute.provider_tree [None req-7940c6a0-845c-481e-9b68-923f5a059a92 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1298.393747] env[63515]: DEBUG nova.scheduler.client.report [None req-7940c6a0-845c-481e-9b68-923f5a059a92 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1298.899407] env[63515]: DEBUG oslo_concurrency.lockutils [None req-7940c6a0-845c-481e-9b68-923f5a059a92 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.099s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1298.899927] env[63515]: DEBUG nova.compute.manager [None req-7940c6a0-845c-481e-9b68-923f5a059a92 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] [instance: b9b28202-1599-4f11-8d92-4942f90160ff] Start building networks asynchronously for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1299.404829] env[63515]: DEBUG nova.compute.utils [None req-7940c6a0-845c-481e-9b68-923f5a059a92 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Using /dev/sd instead of None {{(pid=63515) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1299.406604] env[63515]: DEBUG nova.compute.manager [None req-7940c6a0-845c-481e-9b68-923f5a059a92 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] [instance: b9b28202-1599-4f11-8d92-4942f90160ff] Allocating IP information in the background. {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1299.406814] env[63515]: DEBUG nova.network.neutron [None req-7940c6a0-845c-481e-9b68-923f5a059a92 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] [instance: b9b28202-1599-4f11-8d92-4942f90160ff] allocate_for_instance() {{(pid=63515) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1299.453304] env[63515]: DEBUG nova.policy [None req-7940c6a0-845c-481e-9b68-923f5a059a92 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'cb4f41c6bfb94aa8ab60135e1d729a6a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '97f2e119708746038f38359572c99438', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=63515) authorize /opt/stack/nova/nova/policy.py:201}} [ 1299.722227] env[63515]: DEBUG nova.network.neutron [None req-7940c6a0-845c-481e-9b68-923f5a059a92 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] [instance: b9b28202-1599-4f11-8d92-4942f90160ff] Successfully created port: 6eaab063-d958-4b1e-8613-d4d8acb07d09 {{(pid=63515) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1299.911852] env[63515]: DEBUG nova.compute.manager [None req-7940c6a0-845c-481e-9b68-923f5a059a92 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] [instance: b9b28202-1599-4f11-8d92-4942f90160ff] Start building block device mappings for instance. {{(pid=63515) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1300.921069] env[63515]: DEBUG nova.compute.manager [None req-7940c6a0-845c-481e-9b68-923f5a059a92 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] [instance: b9b28202-1599-4f11-8d92-4942f90160ff] Start spawning the instance on the hypervisor. {{(pid=63515) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1300.944483] env[63515]: DEBUG nova.virt.hardware [None req-7940c6a0-845c-481e-9b68-923f5a059a92 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-03T02:52:14Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-03T02:51:56Z,direct_url=,disk_format='vmdk',id=8a120570-cb06-4099-b262-554ca0ad15c5,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b569255cc43e42e7ae3f2b4ad37c6ef4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-03T02:51:57Z,virtual_size=,visibility=), allow threads: False {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1300.944724] env[63515]: DEBUG nova.virt.hardware [None req-7940c6a0-845c-481e-9b68-923f5a059a92 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Flavor limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1300.944896] env[63515]: DEBUG nova.virt.hardware [None req-7940c6a0-845c-481e-9b68-923f5a059a92 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Image limits 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1300.945100] env[63515]: DEBUG nova.virt.hardware [None req-7940c6a0-845c-481e-9b68-923f5a059a92 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Flavor pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1300.945255] env[63515]: DEBUG nova.virt.hardware [None req-7940c6a0-845c-481e-9b68-923f5a059a92 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Image pref 0:0:0 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1300.945404] env[63515]: DEBUG nova.virt.hardware [None req-7940c6a0-845c-481e-9b68-923f5a059a92 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=63515) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1300.945620] env[63515]: DEBUG nova.virt.hardware [None req-7940c6a0-845c-481e-9b68-923f5a059a92 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1300.945779] env[63515]: DEBUG nova.virt.hardware [None req-7940c6a0-845c-481e-9b68-923f5a059a92 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1300.946051] env[63515]: DEBUG nova.virt.hardware [None req-7940c6a0-845c-481e-9b68-923f5a059a92 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Got 1 possible topologies {{(pid=63515) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1300.946260] env[63515]: DEBUG nova.virt.hardware [None req-7940c6a0-845c-481e-9b68-923f5a059a92 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1300.946442] env[63515]: DEBUG nova.virt.hardware [None req-7940c6a0-845c-481e-9b68-923f5a059a92 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=63515) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1300.947780] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27badcd6-c37d-4d78-b883-0471bdd36919 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1300.955528] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45961dd3-9d79-4367-9615-e9692863e0bb {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1301.081613] env[63515]: DEBUG nova.compute.manager [req-c3e1900e-b4c7-4041-9c0d-b12590fcacac req-37e87816-930e-4865-8580-66f705fcfa39 service nova] [instance: b9b28202-1599-4f11-8d92-4942f90160ff] Received event network-vif-plugged-6eaab063-d958-4b1e-8613-d4d8acb07d09 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1301.081870] env[63515]: DEBUG oslo_concurrency.lockutils [req-c3e1900e-b4c7-4041-9c0d-b12590fcacac req-37e87816-930e-4865-8580-66f705fcfa39 service nova] Acquiring lock "b9b28202-1599-4f11-8d92-4942f90160ff-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1301.082099] env[63515]: DEBUG oslo_concurrency.lockutils [req-c3e1900e-b4c7-4041-9c0d-b12590fcacac req-37e87816-930e-4865-8580-66f705fcfa39 service nova] Lock "b9b28202-1599-4f11-8d92-4942f90160ff-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1301.082272] env[63515]: DEBUG oslo_concurrency.lockutils [req-c3e1900e-b4c7-4041-9c0d-b12590fcacac req-37e87816-930e-4865-8580-66f705fcfa39 service nova] Lock "b9b28202-1599-4f11-8d92-4942f90160ff-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1301.082437] env[63515]: DEBUG nova.compute.manager [req-c3e1900e-b4c7-4041-9c0d-b12590fcacac req-37e87816-930e-4865-8580-66f705fcfa39 service nova] [instance: b9b28202-1599-4f11-8d92-4942f90160ff] No waiting events found dispatching network-vif-plugged-6eaab063-d958-4b1e-8613-d4d8acb07d09 {{(pid=63515) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1301.082598] env[63515]: WARNING nova.compute.manager [req-c3e1900e-b4c7-4041-9c0d-b12590fcacac req-37e87816-930e-4865-8580-66f705fcfa39 service nova] [instance: b9b28202-1599-4f11-8d92-4942f90160ff] Received unexpected event network-vif-plugged-6eaab063-d958-4b1e-8613-d4d8acb07d09 for instance with vm_state building and task_state spawning. [ 1301.167407] env[63515]: DEBUG nova.network.neutron [None req-7940c6a0-845c-481e-9b68-923f5a059a92 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] [instance: b9b28202-1599-4f11-8d92-4942f90160ff] Successfully updated port: 6eaab063-d958-4b1e-8613-d4d8acb07d09 {{(pid=63515) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1301.669158] env[63515]: DEBUG oslo_concurrency.lockutils [None req-7940c6a0-845c-481e-9b68-923f5a059a92 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Acquiring lock "refresh_cache-b9b28202-1599-4f11-8d92-4942f90160ff" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1301.669314] env[63515]: DEBUG oslo_concurrency.lockutils [None req-7940c6a0-845c-481e-9b68-923f5a059a92 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Acquired lock "refresh_cache-b9b28202-1599-4f11-8d92-4942f90160ff" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1301.669465] env[63515]: DEBUG nova.network.neutron [None req-7940c6a0-845c-481e-9b68-923f5a059a92 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] [instance: b9b28202-1599-4f11-8d92-4942f90160ff] Building network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1302.200421] env[63515]: DEBUG nova.network.neutron [None req-7940c6a0-845c-481e-9b68-923f5a059a92 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] [instance: b9b28202-1599-4f11-8d92-4942f90160ff] Instance cache missing network info. {{(pid=63515) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1302.318668] env[63515]: DEBUG nova.network.neutron [None req-7940c6a0-845c-481e-9b68-923f5a059a92 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] [instance: b9b28202-1599-4f11-8d92-4942f90160ff] Updating instance_info_cache with network_info: [{"id": "6eaab063-d958-4b1e-8613-d4d8acb07d09", "address": "fa:16:3e:11:91:4d", "network": {"id": "d6da2650-9ba9-4ef6-8035-e58b635a64b2", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-391564968-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "97f2e119708746038f38359572c99438", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "043ca97b-0fca-4b54-8be6-027123fa76d0", "external-id": "nsx-vlan-transportzone-702", "segmentation_id": 702, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6eaab063-d9", "ovs_interfaceid": "6eaab063-d958-4b1e-8613-d4d8acb07d09", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1302.821740] env[63515]: DEBUG oslo_concurrency.lockutils [None req-7940c6a0-845c-481e-9b68-923f5a059a92 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Releasing lock "refresh_cache-b9b28202-1599-4f11-8d92-4942f90160ff" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1302.822105] env[63515]: DEBUG nova.compute.manager [None req-7940c6a0-845c-481e-9b68-923f5a059a92 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] [instance: b9b28202-1599-4f11-8d92-4942f90160ff] Instance network_info: |[{"id": "6eaab063-d958-4b1e-8613-d4d8acb07d09", "address": "fa:16:3e:11:91:4d", "network": {"id": "d6da2650-9ba9-4ef6-8035-e58b635a64b2", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-391564968-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "97f2e119708746038f38359572c99438", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "043ca97b-0fca-4b54-8be6-027123fa76d0", "external-id": "nsx-vlan-transportzone-702", "segmentation_id": 702, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6eaab063-d9", "ovs_interfaceid": "6eaab063-d958-4b1e-8613-d4d8acb07d09", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=63515) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1302.822558] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-7940c6a0-845c-481e-9b68-923f5a059a92 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] [instance: b9b28202-1599-4f11-8d92-4942f90160ff] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:11:91:4d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '043ca97b-0fca-4b54-8be6-027123fa76d0', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '6eaab063-d958-4b1e-8613-d4d8acb07d09', 'vif_model': 'vmxnet3'}] {{(pid=63515) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1302.830042] env[63515]: DEBUG oslo.service.loopingcall [None req-7940c6a0-845c-481e-9b68-923f5a059a92 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1302.830257] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b9b28202-1599-4f11-8d92-4942f90160ff] Creating VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1302.830476] env[63515]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a8400ba5-6685-490e-96d2-0a42179d67fe {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1302.851724] env[63515]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1302.851724] env[63515]: value = "task-1112108" [ 1302.851724] env[63515]: _type = "Task" [ 1302.851724] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1302.858788] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1112108, 'name': CreateVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1303.107543] env[63515]: DEBUG nova.compute.manager [req-0d65eebf-809b-4aee-9602-c3680440aade req-c9c27ce4-f0de-4ca9-9b8a-a757bf034000 service nova] [instance: b9b28202-1599-4f11-8d92-4942f90160ff] Received event network-changed-6eaab063-d958-4b1e-8613-d4d8acb07d09 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1303.107691] env[63515]: DEBUG nova.compute.manager [req-0d65eebf-809b-4aee-9602-c3680440aade req-c9c27ce4-f0de-4ca9-9b8a-a757bf034000 service nova] [instance: b9b28202-1599-4f11-8d92-4942f90160ff] Refreshing instance network info cache due to event network-changed-6eaab063-d958-4b1e-8613-d4d8acb07d09. {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1303.107978] env[63515]: DEBUG oslo_concurrency.lockutils [req-0d65eebf-809b-4aee-9602-c3680440aade req-c9c27ce4-f0de-4ca9-9b8a-a757bf034000 service nova] Acquiring lock "refresh_cache-b9b28202-1599-4f11-8d92-4942f90160ff" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1303.108179] env[63515]: DEBUG oslo_concurrency.lockutils [req-0d65eebf-809b-4aee-9602-c3680440aade req-c9c27ce4-f0de-4ca9-9b8a-a757bf034000 service nova] Acquired lock "refresh_cache-b9b28202-1599-4f11-8d92-4942f90160ff" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1303.108391] env[63515]: DEBUG nova.network.neutron [req-0d65eebf-809b-4aee-9602-c3680440aade req-c9c27ce4-f0de-4ca9-9b8a-a757bf034000 service nova] [instance: b9b28202-1599-4f11-8d92-4942f90160ff] Refreshing network info cache for port 6eaab063-d958-4b1e-8613-d4d8acb07d09 {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1303.361689] env[63515]: DEBUG oslo_vmware.api [-] Task: {'id': task-1112108, 'name': CreateVM_Task, 'duration_secs': 0.396436} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1303.362103] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b9b28202-1599-4f11-8d92-4942f90160ff] Created VM on the ESX host {{(pid=63515) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1303.369232] env[63515]: DEBUG oslo_concurrency.lockutils [None req-7940c6a0-845c-481e-9b68-923f5a059a92 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1303.369394] env[63515]: DEBUG oslo_concurrency.lockutils [None req-7940c6a0-845c-481e-9b68-923f5a059a92 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1303.369715] env[63515]: DEBUG oslo_concurrency.lockutils [None req-7940c6a0-845c-481e-9b68-923f5a059a92 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1303.369979] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-59777d06-e40e-465d-ab47-8fedde83f0cc {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1303.374793] env[63515]: DEBUG oslo_vmware.api [None req-7940c6a0-845c-481e-9b68-923f5a059a92 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Waiting for the task: (returnval){ [ 1303.374793] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]526e97cd-ed49-a86b-b980-3f5c6059b528" [ 1303.374793] env[63515]: _type = "Task" [ 1303.374793] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1303.384304] env[63515]: DEBUG oslo_vmware.api [None req-7940c6a0-845c-481e-9b68-923f5a059a92 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]526e97cd-ed49-a86b-b980-3f5c6059b528, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1303.796567] env[63515]: DEBUG nova.network.neutron [req-0d65eebf-809b-4aee-9602-c3680440aade req-c9c27ce4-f0de-4ca9-9b8a-a757bf034000 service nova] [instance: b9b28202-1599-4f11-8d92-4942f90160ff] Updated VIF entry in instance network info cache for port 6eaab063-d958-4b1e-8613-d4d8acb07d09. {{(pid=63515) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1303.796935] env[63515]: DEBUG nova.network.neutron [req-0d65eebf-809b-4aee-9602-c3680440aade req-c9c27ce4-f0de-4ca9-9b8a-a757bf034000 service nova] [instance: b9b28202-1599-4f11-8d92-4942f90160ff] Updating instance_info_cache with network_info: [{"id": "6eaab063-d958-4b1e-8613-d4d8acb07d09", "address": "fa:16:3e:11:91:4d", "network": {"id": "d6da2650-9ba9-4ef6-8035-e58b635a64b2", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-391564968-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "97f2e119708746038f38359572c99438", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "043ca97b-0fca-4b54-8be6-027123fa76d0", "external-id": "nsx-vlan-transportzone-702", "segmentation_id": 702, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6eaab063-d9", "ovs_interfaceid": "6eaab063-d958-4b1e-8613-d4d8acb07d09", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1303.884600] env[63515]: DEBUG oslo_vmware.api [None req-7940c6a0-845c-481e-9b68-923f5a059a92 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]526e97cd-ed49-a86b-b980-3f5c6059b528, 'name': SearchDatastore_Task, 'duration_secs': 0.010881} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1303.884896] env[63515]: DEBUG oslo_concurrency.lockutils [None req-7940c6a0-845c-481e-9b68-923f5a059a92 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1303.885164] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-7940c6a0-845c-481e-9b68-923f5a059a92 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] [instance: b9b28202-1599-4f11-8d92-4942f90160ff] Processing image 8a120570-cb06-4099-b262-554ca0ad15c5 {{(pid=63515) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1303.885406] env[63515]: DEBUG oslo_concurrency.lockutils [None req-7940c6a0-845c-481e-9b68-923f5a059a92 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1303.885556] env[63515]: DEBUG oslo_concurrency.lockutils [None req-7940c6a0-845c-481e-9b68-923f5a059a92 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1303.885737] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-7940c6a0-845c-481e-9b68-923f5a059a92 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1303.885996] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-9a135eb6-eaeb-4fb4-9ed3-88ba8d57e263 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1303.893846] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-7940c6a0-845c-481e-9b68-923f5a059a92 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=63515) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1303.894043] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-7940c6a0-845c-481e-9b68-923f5a059a92 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=63515) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1303.894729] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-182cff16-8e6f-4883-8b2e-0f2421f61ed6 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1303.899512] env[63515]: DEBUG oslo_vmware.api [None req-7940c6a0-845c-481e-9b68-923f5a059a92 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Waiting for the task: (returnval){ [ 1303.899512] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]522168e7-90a6-ace5-d01d-f38bbdade1ab" [ 1303.899512] env[63515]: _type = "Task" [ 1303.899512] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1303.906413] env[63515]: DEBUG oslo_vmware.api [None req-7940c6a0-845c-481e-9b68-923f5a059a92 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]522168e7-90a6-ace5-d01d-f38bbdade1ab, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1304.300035] env[63515]: DEBUG oslo_concurrency.lockutils [req-0d65eebf-809b-4aee-9602-c3680440aade req-c9c27ce4-f0de-4ca9-9b8a-a757bf034000 service nova] Releasing lock "refresh_cache-b9b28202-1599-4f11-8d92-4942f90160ff" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1304.409221] env[63515]: DEBUG oslo_vmware.api [None req-7940c6a0-845c-481e-9b68-923f5a059a92 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]522168e7-90a6-ace5-d01d-f38bbdade1ab, 'name': SearchDatastore_Task, 'duration_secs': 0.008319} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1304.410038] env[63515]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5e801648-4374-40ce-b562-4236f42bebab {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1304.415340] env[63515]: DEBUG oslo_vmware.api [None req-7940c6a0-845c-481e-9b68-923f5a059a92 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Waiting for the task: (returnval){ [ 1304.415340] env[63515]: value = "session[52fac58a-0b80-4804-f1ea-b28295510846]52aaa42a-d472-afee-9c28-b93dd4278aa4" [ 1304.415340] env[63515]: _type = "Task" [ 1304.415340] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1304.422416] env[63515]: DEBUG oslo_vmware.api [None req-7940c6a0-845c-481e-9b68-923f5a059a92 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52aaa42a-d472-afee-9c28-b93dd4278aa4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1304.925866] env[63515]: DEBUG oslo_vmware.api [None req-7940c6a0-845c-481e-9b68-923f5a059a92 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Task: {'id': session[52fac58a-0b80-4804-f1ea-b28295510846]52aaa42a-d472-afee-9c28-b93dd4278aa4, 'name': SearchDatastore_Task, 'duration_secs': 0.009394} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1304.926143] env[63515]: DEBUG oslo_concurrency.lockutils [None req-7940c6a0-845c-481e-9b68-923f5a059a92 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1304.926405] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-7940c6a0-845c-481e-9b68-923f5a059a92 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk to [datastore1] b9b28202-1599-4f11-8d92-4942f90160ff/b9b28202-1599-4f11-8d92-4942f90160ff.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1304.926651] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b25c892a-55ef-49f1-abab-639bad10581b {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1304.932632] env[63515]: DEBUG oslo_vmware.api [None req-7940c6a0-845c-481e-9b68-923f5a059a92 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Waiting for the task: (returnval){ [ 1304.932632] env[63515]: value = "task-1112109" [ 1304.932632] env[63515]: _type = "Task" [ 1304.932632] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1304.939785] env[63515]: DEBUG oslo_vmware.api [None req-7940c6a0-845c-481e-9b68-923f5a059a92 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Task: {'id': task-1112109, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1305.443135] env[63515]: DEBUG oslo_vmware.api [None req-7940c6a0-845c-481e-9b68-923f5a059a92 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Task: {'id': task-1112109, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.436745} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1305.443522] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-7940c6a0-845c-481e-9b68-923f5a059a92 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8a120570-cb06-4099-b262-554ca0ad15c5/8a120570-cb06-4099-b262-554ca0ad15c5.vmdk to [datastore1] b9b28202-1599-4f11-8d92-4942f90160ff/b9b28202-1599-4f11-8d92-4942f90160ff.vmdk {{(pid=63515) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1305.443598] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-7940c6a0-845c-481e-9b68-923f5a059a92 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] [instance: b9b28202-1599-4f11-8d92-4942f90160ff] Extending root virtual disk to 1048576 {{(pid=63515) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1305.443824] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-320113bb-b111-4391-a2b6-6ece832d0ff7 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1305.449885] env[63515]: DEBUG oslo_vmware.api [None req-7940c6a0-845c-481e-9b68-923f5a059a92 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Waiting for the task: (returnval){ [ 1305.449885] env[63515]: value = "task-1112110" [ 1305.449885] env[63515]: _type = "Task" [ 1305.449885] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1305.456564] env[63515]: DEBUG oslo_vmware.api [None req-7940c6a0-845c-481e-9b68-923f5a059a92 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Task: {'id': task-1112110, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1305.959713] env[63515]: DEBUG oslo_vmware.api [None req-7940c6a0-845c-481e-9b68-923f5a059a92 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Task: {'id': task-1112110, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.066919} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1305.959991] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-7940c6a0-845c-481e-9b68-923f5a059a92 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] [instance: b9b28202-1599-4f11-8d92-4942f90160ff] Extended root virtual disk {{(pid=63515) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1305.960741] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-553eef46-3853-4100-818e-223d9cdc2247 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1305.981814] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-7940c6a0-845c-481e-9b68-923f5a059a92 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] [instance: b9b28202-1599-4f11-8d92-4942f90160ff] Reconfiguring VM instance instance-00000072 to attach disk [datastore1] b9b28202-1599-4f11-8d92-4942f90160ff/b9b28202-1599-4f11-8d92-4942f90160ff.vmdk or device None with type sparse {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1305.982051] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3eb6d7b5-d622-4577-bb07-3f17cd861646 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1306.001071] env[63515]: DEBUG oslo_vmware.api [None req-7940c6a0-845c-481e-9b68-923f5a059a92 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Waiting for the task: (returnval){ [ 1306.001071] env[63515]: value = "task-1112111" [ 1306.001071] env[63515]: _type = "Task" [ 1306.001071] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1306.008332] env[63515]: DEBUG oslo_vmware.api [None req-7940c6a0-845c-481e-9b68-923f5a059a92 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Task: {'id': task-1112111, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1306.512415] env[63515]: DEBUG oslo_vmware.api [None req-7940c6a0-845c-481e-9b68-923f5a059a92 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Task: {'id': task-1112111, 'name': ReconfigVM_Task, 'duration_secs': 0.288832} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1306.512801] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-7940c6a0-845c-481e-9b68-923f5a059a92 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] [instance: b9b28202-1599-4f11-8d92-4942f90160ff] Reconfigured VM instance instance-00000072 to attach disk [datastore1] b9b28202-1599-4f11-8d92-4942f90160ff/b9b28202-1599-4f11-8d92-4942f90160ff.vmdk or device None with type sparse {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1306.513298] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-529edad4-1877-4d78-b611-63dd3ed1ba88 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1306.519765] env[63515]: DEBUG oslo_vmware.api [None req-7940c6a0-845c-481e-9b68-923f5a059a92 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Waiting for the task: (returnval){ [ 1306.519765] env[63515]: value = "task-1112112" [ 1306.519765] env[63515]: _type = "Task" [ 1306.519765] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1306.526740] env[63515]: DEBUG oslo_vmware.api [None req-7940c6a0-845c-481e-9b68-923f5a059a92 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Task: {'id': task-1112112, 'name': Rename_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1307.029417] env[63515]: DEBUG oslo_vmware.api [None req-7940c6a0-845c-481e-9b68-923f5a059a92 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Task: {'id': task-1112112, 'name': Rename_Task, 'duration_secs': 0.146153} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1307.029786] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-7940c6a0-845c-481e-9b68-923f5a059a92 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] [instance: b9b28202-1599-4f11-8d92-4942f90160ff] Powering on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1307.030050] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3eee0e70-e154-4422-9f97-c404fd8f0488 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1307.036652] env[63515]: DEBUG oslo_vmware.api [None req-7940c6a0-845c-481e-9b68-923f5a059a92 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Waiting for the task: (returnval){ [ 1307.036652] env[63515]: value = "task-1112113" [ 1307.036652] env[63515]: _type = "Task" [ 1307.036652] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1307.043715] env[63515]: DEBUG oslo_vmware.api [None req-7940c6a0-845c-481e-9b68-923f5a059a92 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Task: {'id': task-1112113, 'name': PowerOnVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1307.547043] env[63515]: DEBUG oslo_vmware.api [None req-7940c6a0-845c-481e-9b68-923f5a059a92 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Task: {'id': task-1112113, 'name': PowerOnVM_Task, 'duration_secs': 0.436317} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1307.547458] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-7940c6a0-845c-481e-9b68-923f5a059a92 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] [instance: b9b28202-1599-4f11-8d92-4942f90160ff] Powered on the VM {{(pid=63515) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1307.547559] env[63515]: INFO nova.compute.manager [None req-7940c6a0-845c-481e-9b68-923f5a059a92 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] [instance: b9b28202-1599-4f11-8d92-4942f90160ff] Took 6.63 seconds to spawn the instance on the hypervisor. [ 1307.547709] env[63515]: DEBUG nova.compute.manager [None req-7940c6a0-845c-481e-9b68-923f5a059a92 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] [instance: b9b28202-1599-4f11-8d92-4942f90160ff] Checking state {{(pid=63515) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1307.548462] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b02273e-b35b-4caa-89c7-cf5ea32217f9 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1308.067371] env[63515]: INFO nova.compute.manager [None req-7940c6a0-845c-481e-9b68-923f5a059a92 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] [instance: b9b28202-1599-4f11-8d92-4942f90160ff] Took 11.28 seconds to build instance. [ 1308.570503] env[63515]: DEBUG oslo_concurrency.lockutils [None req-7940c6a0-845c-481e-9b68-923f5a059a92 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Lock "b9b28202-1599-4f11-8d92-4942f90160ff" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 12.791s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1308.841689] env[63515]: DEBUG nova.compute.manager [req-188a366b-8382-4751-8789-964be81d1e69 req-0bd1cb2b-d768-4dc9-8282-cba2be78335f service nova] [instance: b9b28202-1599-4f11-8d92-4942f90160ff] Received event network-changed-6eaab063-d958-4b1e-8613-d4d8acb07d09 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1308.841689] env[63515]: DEBUG nova.compute.manager [req-188a366b-8382-4751-8789-964be81d1e69 req-0bd1cb2b-d768-4dc9-8282-cba2be78335f service nova] [instance: b9b28202-1599-4f11-8d92-4942f90160ff] Refreshing instance network info cache due to event network-changed-6eaab063-d958-4b1e-8613-d4d8acb07d09. {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1308.841774] env[63515]: DEBUG oslo_concurrency.lockutils [req-188a366b-8382-4751-8789-964be81d1e69 req-0bd1cb2b-d768-4dc9-8282-cba2be78335f service nova] Acquiring lock "refresh_cache-b9b28202-1599-4f11-8d92-4942f90160ff" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1308.841951] env[63515]: DEBUG oslo_concurrency.lockutils [req-188a366b-8382-4751-8789-964be81d1e69 req-0bd1cb2b-d768-4dc9-8282-cba2be78335f service nova] Acquired lock "refresh_cache-b9b28202-1599-4f11-8d92-4942f90160ff" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1308.842131] env[63515]: DEBUG nova.network.neutron [req-188a366b-8382-4751-8789-964be81d1e69 req-0bd1cb2b-d768-4dc9-8282-cba2be78335f service nova] [instance: b9b28202-1599-4f11-8d92-4942f90160ff] Refreshing network info cache for port 6eaab063-d958-4b1e-8613-d4d8acb07d09 {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1309.558836] env[63515]: DEBUG nova.network.neutron [req-188a366b-8382-4751-8789-964be81d1e69 req-0bd1cb2b-d768-4dc9-8282-cba2be78335f service nova] [instance: b9b28202-1599-4f11-8d92-4942f90160ff] Updated VIF entry in instance network info cache for port 6eaab063-d958-4b1e-8613-d4d8acb07d09. {{(pid=63515) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1309.559301] env[63515]: DEBUG nova.network.neutron [req-188a366b-8382-4751-8789-964be81d1e69 req-0bd1cb2b-d768-4dc9-8282-cba2be78335f service nova] [instance: b9b28202-1599-4f11-8d92-4942f90160ff] Updating instance_info_cache with network_info: [{"id": "6eaab063-d958-4b1e-8613-d4d8acb07d09", "address": "fa:16:3e:11:91:4d", "network": {"id": "d6da2650-9ba9-4ef6-8035-e58b635a64b2", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-391564968-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.157", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "97f2e119708746038f38359572c99438", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "043ca97b-0fca-4b54-8be6-027123fa76d0", "external-id": "nsx-vlan-transportzone-702", "segmentation_id": 702, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6eaab063-d9", "ovs_interfaceid": "6eaab063-d958-4b1e-8613-d4d8acb07d09", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1310.062033] env[63515]: DEBUG oslo_concurrency.lockutils [req-188a366b-8382-4751-8789-964be81d1e69 req-0bd1cb2b-d768-4dc9-8282-cba2be78335f service nova] Releasing lock "refresh_cache-b9b28202-1599-4f11-8d92-4942f90160ff" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1320.212869] env[63515]: DEBUG oslo_service.periodic_task [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=63515) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1322.535573] env[63515]: DEBUG oslo_service.periodic_task [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=63515) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1323.535895] env[63515]: DEBUG oslo_service.periodic_task [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=63515) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1323.536272] env[63515]: DEBUG oslo_service.periodic_task [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=63515) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1323.536272] env[63515]: DEBUG nova.compute.manager [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=63515) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10545}} [ 1325.535967] env[63515]: DEBUG oslo_service.periodic_task [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=63515) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1325.536373] env[63515]: DEBUG oslo_service.periodic_task [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=63515) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1326.531939] env[63515]: DEBUG oslo_service.periodic_task [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=63515) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1327.038039] env[63515]: DEBUG oslo_service.periodic_task [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=63515) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1327.038436] env[63515]: DEBUG nova.compute.manager [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Starting heal instance info cache {{(pid=63515) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9926}} [ 1327.038436] env[63515]: DEBUG nova.compute.manager [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Rebuilding the list of instances to heal {{(pid=63515) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9930}} [ 1327.568103] env[63515]: DEBUG oslo_concurrency.lockutils [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Acquiring lock "refresh_cache-b9b28202-1599-4f11-8d92-4942f90160ff" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1327.568258] env[63515]: DEBUG oslo_concurrency.lockutils [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Acquired lock "refresh_cache-b9b28202-1599-4f11-8d92-4942f90160ff" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1327.568403] env[63515]: DEBUG nova.network.neutron [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] [instance: b9b28202-1599-4f11-8d92-4942f90160ff] Forcefully refreshing network info cache for instance {{(pid=63515) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2004}} [ 1327.568559] env[63515]: DEBUG nova.objects.instance [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Lazy-loading 'info_cache' on Instance uuid b9b28202-1599-4f11-8d92-4942f90160ff {{(pid=63515) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1329.274989] env[63515]: DEBUG nova.network.neutron [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] [instance: b9b28202-1599-4f11-8d92-4942f90160ff] Updating instance_info_cache with network_info: [{"id": "6eaab063-d958-4b1e-8613-d4d8acb07d09", "address": "fa:16:3e:11:91:4d", "network": {"id": "d6da2650-9ba9-4ef6-8035-e58b635a64b2", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-391564968-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.157", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "97f2e119708746038f38359572c99438", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "043ca97b-0fca-4b54-8be6-027123fa76d0", "external-id": "nsx-vlan-transportzone-702", "segmentation_id": 702, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6eaab063-d9", "ovs_interfaceid": "6eaab063-d958-4b1e-8613-d4d8acb07d09", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1329.777375] env[63515]: DEBUG oslo_concurrency.lockutils [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Releasing lock "refresh_cache-b9b28202-1599-4f11-8d92-4942f90160ff" {{(pid=63515) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1329.777615] env[63515]: DEBUG nova.compute.manager [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] [instance: b9b28202-1599-4f11-8d92-4942f90160ff] Updated the network info_cache for instance {{(pid=63515) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9997}} [ 1329.777816] env[63515]: DEBUG oslo_service.periodic_task [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Running periodic task ComputeManager.update_available_resource {{(pid=63515) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1330.281609] env[63515]: DEBUG oslo_concurrency.lockutils [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1330.282050] env[63515]: DEBUG oslo_concurrency.lockutils [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1330.282050] env[63515]: DEBUG oslo_concurrency.lockutils [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1330.282198] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=63515) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1330.283090] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05f4173d-3c20-4a65-a3e3-2faddcf4e2b6 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1330.291488] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00219a55-d61d-473c-89d5-7df59e62e058 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1330.304795] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6512dda3-0d65-4583-8198-f12f3decc480 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1330.310703] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6daeb637-aabd-4844-ba33-c006980bfc7c {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1330.338897] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181369MB free_disk=171GB free_vcpus=48 pci_devices=None {{(pid=63515) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1330.339101] env[63515]: DEBUG oslo_concurrency.lockutils [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1330.339249] env[63515]: DEBUG oslo_concurrency.lockutils [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1331.362340] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Instance b9b28202-1599-4f11-8d92-4942f90160ff actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=63515) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1331.362628] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Total usable vcpus: 48, total allocated vcpus: 1 {{(pid=63515) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1331.362685] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=704MB phys_disk=200GB used_disk=1GB total_vcpus=48 used_vcpus=1 pci_stats=[] {{(pid=63515) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1331.387060] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96cbefcd-a6a9-4a17-9793-ced70e6d68a1 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1331.394665] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca1e4bf6-f292-4390-ad32-5f94c0532e07 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1331.423342] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0401d964-bb47-442f-afc6-375e1119d99b {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1331.429880] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90e83a95-71f6-4ae9-bc23-8fac93fbd940 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1331.442342] env[63515]: DEBUG nova.compute.provider_tree [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1331.945766] env[63515]: DEBUG nova.scheduler.client.report [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1332.450979] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=63515) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1332.451401] env[63515]: DEBUG oslo_concurrency.lockutils [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.112s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1333.449940] env[63515]: DEBUG oslo_service.periodic_task [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=63515) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1345.854171] env[63515]: DEBUG oslo_concurrency.lockutils [None req-3bc1118a-79e3-42ba-ade2-463b9e3960f5 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Acquiring lock "b9b28202-1599-4f11-8d92-4942f90160ff" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1345.854504] env[63515]: DEBUG oslo_concurrency.lockutils [None req-3bc1118a-79e3-42ba-ade2-463b9e3960f5 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Lock "b9b28202-1599-4f11-8d92-4942f90160ff" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.001s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1346.358163] env[63515]: DEBUG nova.compute.utils [None req-3bc1118a-79e3-42ba-ade2-463b9e3960f5 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Using /dev/sd instead of None {{(pid=63515) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1346.861652] env[63515]: DEBUG oslo_concurrency.lockutils [None req-3bc1118a-79e3-42ba-ade2-463b9e3960f5 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Lock "b9b28202-1599-4f11-8d92-4942f90160ff" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.007s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1347.921348] env[63515]: DEBUG oslo_concurrency.lockutils [None req-3bc1118a-79e3-42ba-ade2-463b9e3960f5 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Acquiring lock "b9b28202-1599-4f11-8d92-4942f90160ff" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1347.921648] env[63515]: DEBUG oslo_concurrency.lockutils [None req-3bc1118a-79e3-42ba-ade2-463b9e3960f5 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Lock "b9b28202-1599-4f11-8d92-4942f90160ff" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1347.921860] env[63515]: INFO nova.compute.manager [None req-3bc1118a-79e3-42ba-ade2-463b9e3960f5 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] [instance: b9b28202-1599-4f11-8d92-4942f90160ff] Attaching volume 39fb96cb-6c42-419b-9820-6808110dfa5f to /dev/sdb [ 1347.951603] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-462c3761-2a49-468a-9638-fc99d77a479a {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1347.958654] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64eeefbb-e9b3-407c-ab2d-334374a7238c {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1347.970814] env[63515]: DEBUG nova.virt.block_device [None req-3bc1118a-79e3-42ba-ade2-463b9e3960f5 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] [instance: b9b28202-1599-4f11-8d92-4942f90160ff] Updating existing volume attachment record: 3b8ce779-f6b4-4e51-8def-a0cabb2dabbb {{(pid=63515) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1352.513755] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-3bc1118a-79e3-42ba-ade2-463b9e3960f5 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] [instance: b9b28202-1599-4f11-8d92-4942f90160ff] Volume attach. Driver type: vmdk {{(pid=63515) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1352.514064] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-3bc1118a-79e3-42ba-ade2-463b9e3960f5 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] [instance: b9b28202-1599-4f11-8d92-4942f90160ff] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-243605', 'volume_id': '39fb96cb-6c42-419b-9820-6808110dfa5f', 'name': 'volume-39fb96cb-6c42-419b-9820-6808110dfa5f', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'b9b28202-1599-4f11-8d92-4942f90160ff', 'attached_at': '', 'detached_at': '', 'volume_id': '39fb96cb-6c42-419b-9820-6808110dfa5f', 'serial': '39fb96cb-6c42-419b-9820-6808110dfa5f'} {{(pid=63515) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1352.514978] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01c2b31b-9a1b-4cc3-8e22-28ffe33ad69d {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1352.531248] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b4174c4-9885-4d29-9604-52139bbc0999 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1352.554714] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-3bc1118a-79e3-42ba-ade2-463b9e3960f5 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] [instance: b9b28202-1599-4f11-8d92-4942f90160ff] Reconfiguring VM instance instance-00000072 to attach disk [datastore1] volume-39fb96cb-6c42-419b-9820-6808110dfa5f/volume-39fb96cb-6c42-419b-9820-6808110dfa5f.vmdk or device None with type thin {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1352.554962] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8531341f-c1fa-4132-a189-f68ca4e3ed53 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1352.572918] env[63515]: DEBUG oslo_vmware.api [None req-3bc1118a-79e3-42ba-ade2-463b9e3960f5 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Waiting for the task: (returnval){ [ 1352.572918] env[63515]: value = "task-1112116" [ 1352.572918] env[63515]: _type = "Task" [ 1352.572918] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1352.582784] env[63515]: DEBUG oslo_vmware.api [None req-3bc1118a-79e3-42ba-ade2-463b9e3960f5 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Task: {'id': task-1112116, 'name': ReconfigVM_Task} progress is 6%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1353.084990] env[63515]: DEBUG oslo_vmware.api [None req-3bc1118a-79e3-42ba-ade2-463b9e3960f5 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Task: {'id': task-1112116, 'name': ReconfigVM_Task} progress is 99%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1353.584797] env[63515]: DEBUG oslo_vmware.api [None req-3bc1118a-79e3-42ba-ade2-463b9e3960f5 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Task: {'id': task-1112116, 'name': ReconfigVM_Task} progress is 99%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1354.085704] env[63515]: DEBUG oslo_vmware.api [None req-3bc1118a-79e3-42ba-ade2-463b9e3960f5 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Task: {'id': task-1112116, 'name': ReconfigVM_Task, 'duration_secs': 1.324736} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1354.085989] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-3bc1118a-79e3-42ba-ade2-463b9e3960f5 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] [instance: b9b28202-1599-4f11-8d92-4942f90160ff] Reconfigured VM instance instance-00000072 to attach disk [datastore1] volume-39fb96cb-6c42-419b-9820-6808110dfa5f/volume-39fb96cb-6c42-419b-9820-6808110dfa5f.vmdk or device None with type thin {{(pid=63515) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1354.090584] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6b9b4028-b267-4561-90a4-b8771926398a {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1354.105016] env[63515]: DEBUG oslo_vmware.api [None req-3bc1118a-79e3-42ba-ade2-463b9e3960f5 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Waiting for the task: (returnval){ [ 1354.105016] env[63515]: value = "task-1112117" [ 1354.105016] env[63515]: _type = "Task" [ 1354.105016] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1354.112617] env[63515]: DEBUG oslo_vmware.api [None req-3bc1118a-79e3-42ba-ade2-463b9e3960f5 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Task: {'id': task-1112117, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1354.615229] env[63515]: DEBUG oslo_vmware.api [None req-3bc1118a-79e3-42ba-ade2-463b9e3960f5 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Task: {'id': task-1112117, 'name': ReconfigVM_Task, 'duration_secs': 0.14666} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1354.615656] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-3bc1118a-79e3-42ba-ade2-463b9e3960f5 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] [instance: b9b28202-1599-4f11-8d92-4942f90160ff] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-243605', 'volume_id': '39fb96cb-6c42-419b-9820-6808110dfa5f', 'name': 'volume-39fb96cb-6c42-419b-9820-6808110dfa5f', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'b9b28202-1599-4f11-8d92-4942f90160ff', 'attached_at': '', 'detached_at': '', 'volume_id': '39fb96cb-6c42-419b-9820-6808110dfa5f', 'serial': '39fb96cb-6c42-419b-9820-6808110dfa5f'} {{(pid=63515) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1355.649447] env[63515]: DEBUG nova.objects.instance [None req-3bc1118a-79e3-42ba-ade2-463b9e3960f5 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Lazy-loading 'flavor' on Instance uuid b9b28202-1599-4f11-8d92-4942f90160ff {{(pid=63515) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1356.155573] env[63515]: DEBUG oslo_concurrency.lockutils [None req-3bc1118a-79e3-42ba-ade2-463b9e3960f5 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Lock "b9b28202-1599-4f11-8d92-4942f90160ff" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 8.234s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1356.348231] env[63515]: DEBUG oslo_concurrency.lockutils [None req-77d60859-bea3-4d73-a67b-041011cd988d tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Acquiring lock "b9b28202-1599-4f11-8d92-4942f90160ff" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1356.348493] env[63515]: DEBUG oslo_concurrency.lockutils [None req-77d60859-bea3-4d73-a67b-041011cd988d tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Lock "b9b28202-1599-4f11-8d92-4942f90160ff" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1356.851076] env[63515]: INFO nova.compute.manager [None req-77d60859-bea3-4d73-a67b-041011cd988d tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] [instance: b9b28202-1599-4f11-8d92-4942f90160ff] Detaching volume 39fb96cb-6c42-419b-9820-6808110dfa5f [ 1356.881271] env[63515]: INFO nova.virt.block_device [None req-77d60859-bea3-4d73-a67b-041011cd988d tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] [instance: b9b28202-1599-4f11-8d92-4942f90160ff] Attempting to driver detach volume 39fb96cb-6c42-419b-9820-6808110dfa5f from mountpoint /dev/sdb [ 1356.881669] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-77d60859-bea3-4d73-a67b-041011cd988d tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] [instance: b9b28202-1599-4f11-8d92-4942f90160ff] Volume detach. Driver type: vmdk {{(pid=63515) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1356.881717] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-77d60859-bea3-4d73-a67b-041011cd988d tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] [instance: b9b28202-1599-4f11-8d92-4942f90160ff] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-243605', 'volume_id': '39fb96cb-6c42-419b-9820-6808110dfa5f', 'name': 'volume-39fb96cb-6c42-419b-9820-6808110dfa5f', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'b9b28202-1599-4f11-8d92-4942f90160ff', 'attached_at': '', 'detached_at': '', 'volume_id': '39fb96cb-6c42-419b-9820-6808110dfa5f', 'serial': '39fb96cb-6c42-419b-9820-6808110dfa5f'} {{(pid=63515) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1356.882644] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f27d412-55e0-4728-8816-1a255a1f1092 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1356.903944] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43c9a692-cc77-416f-bf7d-92d718dc9e5d {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1356.910516] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e27c719-5b6a-4313-ab90-9c3bba59237e {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1356.930714] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc58badb-17b4-453c-9a2a-4011170cda40 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1356.944360] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-77d60859-bea3-4d73-a67b-041011cd988d tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] The volume has not been displaced from its original location: [datastore1] volume-39fb96cb-6c42-419b-9820-6808110dfa5f/volume-39fb96cb-6c42-419b-9820-6808110dfa5f.vmdk. No consolidation needed. {{(pid=63515) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1356.949614] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-77d60859-bea3-4d73-a67b-041011cd988d tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] [instance: b9b28202-1599-4f11-8d92-4942f90160ff] Reconfiguring VM instance instance-00000072 to detach disk 2001 {{(pid=63515) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1356.949848] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0bc47423-0073-4a43-8b41-19520c05e36c {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1356.966773] env[63515]: DEBUG oslo_vmware.api [None req-77d60859-bea3-4d73-a67b-041011cd988d tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Waiting for the task: (returnval){ [ 1356.966773] env[63515]: value = "task-1112118" [ 1356.966773] env[63515]: _type = "Task" [ 1356.966773] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1356.973868] env[63515]: DEBUG oslo_vmware.api [None req-77d60859-bea3-4d73-a67b-041011cd988d tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Task: {'id': task-1112118, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1357.475724] env[63515]: DEBUG oslo_vmware.api [None req-77d60859-bea3-4d73-a67b-041011cd988d tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Task: {'id': task-1112118, 'name': ReconfigVM_Task, 'duration_secs': 0.190882} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1357.476017] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-77d60859-bea3-4d73-a67b-041011cd988d tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] [instance: b9b28202-1599-4f11-8d92-4942f90160ff] Reconfigured VM instance instance-00000072 to detach disk 2001 {{(pid=63515) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1357.480563] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-060dfa95-63ca-4588-9456-c7253e37d6ca {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1357.494221] env[63515]: DEBUG oslo_vmware.api [None req-77d60859-bea3-4d73-a67b-041011cd988d tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Waiting for the task: (returnval){ [ 1357.494221] env[63515]: value = "task-1112119" [ 1357.494221] env[63515]: _type = "Task" [ 1357.494221] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1357.501221] env[63515]: DEBUG oslo_vmware.api [None req-77d60859-bea3-4d73-a67b-041011cd988d tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Task: {'id': task-1112119, 'name': ReconfigVM_Task} progress is 5%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1358.004191] env[63515]: DEBUG oslo_vmware.api [None req-77d60859-bea3-4d73-a67b-041011cd988d tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Task: {'id': task-1112119, 'name': ReconfigVM_Task, 'duration_secs': 0.129473} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1358.004569] env[63515]: DEBUG nova.virt.vmwareapi.volumeops [None req-77d60859-bea3-4d73-a67b-041011cd988d tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] [instance: b9b28202-1599-4f11-8d92-4942f90160ff] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-243605', 'volume_id': '39fb96cb-6c42-419b-9820-6808110dfa5f', 'name': 'volume-39fb96cb-6c42-419b-9820-6808110dfa5f', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'b9b28202-1599-4f11-8d92-4942f90160ff', 'attached_at': '', 'detached_at': '', 'volume_id': '39fb96cb-6c42-419b-9820-6808110dfa5f', 'serial': '39fb96cb-6c42-419b-9820-6808110dfa5f'} {{(pid=63515) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1358.545107] env[63515]: DEBUG nova.objects.instance [None req-77d60859-bea3-4d73-a67b-041011cd988d tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Lazy-loading 'flavor' on Instance uuid b9b28202-1599-4f11-8d92-4942f90160ff {{(pid=63515) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1359.553487] env[63515]: DEBUG oslo_concurrency.lockutils [None req-77d60859-bea3-4d73-a67b-041011cd988d tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Lock "b9b28202-1599-4f11-8d92-4942f90160ff" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.205s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1360.587515] env[63515]: DEBUG oslo_concurrency.lockutils [None req-ae8ea796-0c43-484e-a090-8da42880d869 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Acquiring lock "b9b28202-1599-4f11-8d92-4942f90160ff" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1360.587877] env[63515]: DEBUG oslo_concurrency.lockutils [None req-ae8ea796-0c43-484e-a090-8da42880d869 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Lock "b9b28202-1599-4f11-8d92-4942f90160ff" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1360.587947] env[63515]: DEBUG oslo_concurrency.lockutils [None req-ae8ea796-0c43-484e-a090-8da42880d869 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Acquiring lock "b9b28202-1599-4f11-8d92-4942f90160ff-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1360.588149] env[63515]: DEBUG oslo_concurrency.lockutils [None req-ae8ea796-0c43-484e-a090-8da42880d869 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Lock "b9b28202-1599-4f11-8d92-4942f90160ff-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1360.588328] env[63515]: DEBUG oslo_concurrency.lockutils [None req-ae8ea796-0c43-484e-a090-8da42880d869 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Lock "b9b28202-1599-4f11-8d92-4942f90160ff-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1360.590663] env[63515]: INFO nova.compute.manager [None req-ae8ea796-0c43-484e-a090-8da42880d869 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] [instance: b9b28202-1599-4f11-8d92-4942f90160ff] Terminating instance [ 1360.592446] env[63515]: DEBUG nova.compute.manager [None req-ae8ea796-0c43-484e-a090-8da42880d869 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] [instance: b9b28202-1599-4f11-8d92-4942f90160ff] Start destroying the instance on the hypervisor. {{(pid=63515) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1360.592648] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-ae8ea796-0c43-484e-a090-8da42880d869 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] [instance: b9b28202-1599-4f11-8d92-4942f90160ff] Destroying instance {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1360.593485] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c187c1e-dd86-47a9-ae2a-c1b4db7553a9 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1360.600554] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-ae8ea796-0c43-484e-a090-8da42880d869 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] [instance: b9b28202-1599-4f11-8d92-4942f90160ff] Powering off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1360.600780] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c63c159a-a44c-4df2-9e40-22cfb29e20cb {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1360.606866] env[63515]: DEBUG oslo_vmware.api [None req-ae8ea796-0c43-484e-a090-8da42880d869 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Waiting for the task: (returnval){ [ 1360.606866] env[63515]: value = "task-1112120" [ 1360.606866] env[63515]: _type = "Task" [ 1360.606866] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1360.615060] env[63515]: DEBUG oslo_vmware.api [None req-ae8ea796-0c43-484e-a090-8da42880d869 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Task: {'id': task-1112120, 'name': PowerOffVM_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1361.117012] env[63515]: DEBUG oslo_vmware.api [None req-ae8ea796-0c43-484e-a090-8da42880d869 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Task: {'id': task-1112120, 'name': PowerOffVM_Task, 'duration_secs': 0.176299} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1361.117313] env[63515]: DEBUG nova.virt.vmwareapi.vm_util [None req-ae8ea796-0c43-484e-a090-8da42880d869 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] [instance: b9b28202-1599-4f11-8d92-4942f90160ff] Powered off the VM {{(pid=63515) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1361.117490] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-ae8ea796-0c43-484e-a090-8da42880d869 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] [instance: b9b28202-1599-4f11-8d92-4942f90160ff] Unregistering the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1361.117743] env[63515]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-53f13c54-07a1-4958-bc9c-3e01e5b5bb09 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1361.182055] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-ae8ea796-0c43-484e-a090-8da42880d869 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] [instance: b9b28202-1599-4f11-8d92-4942f90160ff] Unregistered the VM {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1361.182278] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-ae8ea796-0c43-484e-a090-8da42880d869 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] [instance: b9b28202-1599-4f11-8d92-4942f90160ff] Deleting contents of the VM from datastore datastore1 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1361.182472] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-ae8ea796-0c43-484e-a090-8da42880d869 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Deleting the datastore file [datastore1] b9b28202-1599-4f11-8d92-4942f90160ff {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1361.182723] env[63515]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f7dcfa53-6433-45d2-8bd9-2be50e3bef11 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1361.188930] env[63515]: DEBUG oslo_vmware.api [None req-ae8ea796-0c43-484e-a090-8da42880d869 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Waiting for the task: (returnval){ [ 1361.188930] env[63515]: value = "task-1112122" [ 1361.188930] env[63515]: _type = "Task" [ 1361.188930] env[63515]: } to complete. {{(pid=63515) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1361.195813] env[63515]: DEBUG oslo_vmware.api [None req-ae8ea796-0c43-484e-a090-8da42880d869 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Task: {'id': task-1112122, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1361.698630] env[63515]: DEBUG oslo_vmware.api [None req-ae8ea796-0c43-484e-a090-8da42880d869 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Task: {'id': task-1112122, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.134494} completed successfully. {{(pid=63515) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1361.699074] env[63515]: DEBUG nova.virt.vmwareapi.ds_util [None req-ae8ea796-0c43-484e-a090-8da42880d869 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Deleted the datastore file {{(pid=63515) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1361.699149] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-ae8ea796-0c43-484e-a090-8da42880d869 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] [instance: b9b28202-1599-4f11-8d92-4942f90160ff] Deleted contents of the VM from datastore datastore1 {{(pid=63515) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1361.699277] env[63515]: DEBUG nova.virt.vmwareapi.vmops [None req-ae8ea796-0c43-484e-a090-8da42880d869 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] [instance: b9b28202-1599-4f11-8d92-4942f90160ff] Instance destroyed {{(pid=63515) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1361.699461] env[63515]: INFO nova.compute.manager [None req-ae8ea796-0c43-484e-a090-8da42880d869 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] [instance: b9b28202-1599-4f11-8d92-4942f90160ff] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1361.699707] env[63515]: DEBUG oslo.service.loopingcall [None req-ae8ea796-0c43-484e-a090-8da42880d869 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=63515) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1361.699896] env[63515]: DEBUG nova.compute.manager [-] [instance: b9b28202-1599-4f11-8d92-4942f90160ff] Deallocating network for instance {{(pid=63515) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1361.699991] env[63515]: DEBUG nova.network.neutron [-] [instance: b9b28202-1599-4f11-8d92-4942f90160ff] deallocate_for_instance() {{(pid=63515) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1362.102877] env[63515]: DEBUG nova.compute.manager [req-864fc240-01ca-4921-89ef-47b5883fc859 req-b85b1319-33c2-44fe-8a00-2062e3485d1d service nova] [instance: b9b28202-1599-4f11-8d92-4942f90160ff] Received event network-vif-deleted-6eaab063-d958-4b1e-8613-d4d8acb07d09 {{(pid=63515) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1362.102941] env[63515]: INFO nova.compute.manager [req-864fc240-01ca-4921-89ef-47b5883fc859 req-b85b1319-33c2-44fe-8a00-2062e3485d1d service nova] [instance: b9b28202-1599-4f11-8d92-4942f90160ff] Neutron deleted interface 6eaab063-d958-4b1e-8613-d4d8acb07d09; detaching it from the instance and deleting it from the info cache [ 1362.103124] env[63515]: DEBUG nova.network.neutron [req-864fc240-01ca-4921-89ef-47b5883fc859 req-b85b1319-33c2-44fe-8a00-2062e3485d1d service nova] [instance: b9b28202-1599-4f11-8d92-4942f90160ff] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1362.582462] env[63515]: DEBUG nova.network.neutron [-] [instance: b9b28202-1599-4f11-8d92-4942f90160ff] Updating instance_info_cache with network_info: [] {{(pid=63515) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1362.606040] env[63515]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-5c523dcf-0aac-4aab-8890-c74a0cd21628 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1362.616228] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f59996f6-f8c7-40ec-894d-63023c03b88a {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1362.639154] env[63515]: DEBUG nova.compute.manager [req-864fc240-01ca-4921-89ef-47b5883fc859 req-b85b1319-33c2-44fe-8a00-2062e3485d1d service nova] [instance: b9b28202-1599-4f11-8d92-4942f90160ff] Detach interface failed, port_id=6eaab063-d958-4b1e-8613-d4d8acb07d09, reason: Instance b9b28202-1599-4f11-8d92-4942f90160ff could not be found. {{(pid=63515) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1363.085810] env[63515]: INFO nova.compute.manager [-] [instance: b9b28202-1599-4f11-8d92-4942f90160ff] Took 1.39 seconds to deallocate network for instance. [ 1363.592203] env[63515]: DEBUG oslo_concurrency.lockutils [None req-ae8ea796-0c43-484e-a090-8da42880d869 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1363.592491] env[63515]: DEBUG oslo_concurrency.lockutils [None req-ae8ea796-0c43-484e-a090-8da42880d869 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1363.592753] env[63515]: DEBUG nova.objects.instance [None req-ae8ea796-0c43-484e-a090-8da42880d869 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Lazy-loading 'resources' on Instance uuid b9b28202-1599-4f11-8d92-4942f90160ff {{(pid=63515) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1364.128702] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b25644b2-b5b7-4467-bbe3-be53d74e25fd {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1364.136257] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f379eca9-1bd5-4776-a111-ee4d7eef3702 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1364.166208] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-686151c9-4503-477c-9563-c93801ebf966 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1364.173486] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0aaf5d3e-da72-4249-825b-5ae848842457 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1364.186588] env[63515]: DEBUG nova.compute.provider_tree [None req-ae8ea796-0c43-484e-a090-8da42880d869 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1364.689759] env[63515]: DEBUG nova.scheduler.client.report [None req-ae8ea796-0c43-484e-a090-8da42880d869 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1365.194544] env[63515]: DEBUG oslo_concurrency.lockutils [None req-ae8ea796-0c43-484e-a090-8da42880d869 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.602s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1365.215834] env[63515]: INFO nova.scheduler.client.report [None req-ae8ea796-0c43-484e-a090-8da42880d869 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Deleted allocations for instance b9b28202-1599-4f11-8d92-4942f90160ff [ 1365.723998] env[63515]: DEBUG oslo_concurrency.lockutils [None req-ae8ea796-0c43-484e-a090-8da42880d869 tempest-AttachVolumeNegativeTest-524267820 tempest-AttachVolumeNegativeTest-524267820-project-member] Lock "b9b28202-1599-4f11-8d92-4942f90160ff" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.136s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1378.535445] env[63515]: DEBUG oslo_service.periodic_task [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=63515) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1381.536334] env[63515]: DEBUG oslo_service.periodic_task [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Running periodic task ComputeManager._cleanup_expired_console_auth_tokens {{(pid=63515) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1384.037034] env[63515]: DEBUG oslo_service.periodic_task [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=63515) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1384.037432] env[63515]: DEBUG oslo_service.periodic_task [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=63515) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1384.037432] env[63515]: DEBUG nova.compute.manager [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=63515) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10545}} [ 1384.535670] env[63515]: DEBUG oslo_service.periodic_task [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=63515) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1385.535850] env[63515]: DEBUG oslo_service.periodic_task [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=63515) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1386.535401] env[63515]: DEBUG oslo_service.periodic_task [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=63515) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1387.530627] env[63515]: DEBUG oslo_service.periodic_task [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=63515) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1388.535341] env[63515]: DEBUG oslo_service.periodic_task [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=63515) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1388.535841] env[63515]: DEBUG nova.compute.manager [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Starting heal instance info cache {{(pid=63515) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9926}} [ 1388.535841] env[63515]: DEBUG nova.compute.manager [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Rebuilding the list of instances to heal {{(pid=63515) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9930}} [ 1389.038793] env[63515]: DEBUG nova.compute.manager [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Didn't find any instances for network info cache update. {{(pid=63515) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10012}} [ 1389.039039] env[63515]: DEBUG oslo_service.periodic_task [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Running periodic task ComputeManager.update_available_resource {{(pid=63515) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1389.542590] env[63515]: DEBUG oslo_concurrency.lockutils [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1389.542957] env[63515]: DEBUG oslo_concurrency.lockutils [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1389.542957] env[63515]: DEBUG oslo_concurrency.lockutils [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1389.543090] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=63515) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1389.543985] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32be1d41-3741-49ff-9292-eed6ee09865a {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1389.553419] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6feaa0e1-4f6b-4615-9915-332574ef94e3 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1389.566784] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ebda24e0-3ba4-4ec0-803a-c0906fa95ef9 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1389.572574] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-287b108f-fb1c-483d-9256-41e327c70c9a {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1389.600238] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181365MB free_disk=171GB free_vcpus=48 pci_devices=None {{(pid=63515) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1389.600371] env[63515]: DEBUG oslo_concurrency.lockutils [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1389.600560] env[63515]: DEBUG oslo_concurrency.lockutils [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1390.739281] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Total usable vcpus: 48, total allocated vcpus: 0 {{(pid=63515) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1390.739628] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=512MB phys_disk=200GB used_disk=0GB total_vcpus=48 used_vcpus=0 pci_stats=[] {{(pid=63515) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1390.753164] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd59a2a4-3bcd-44c0-9c11-d274565a43a9 {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1390.760532] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22722e68-94dd-433a-b216-5ff84d29b00d {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1390.790096] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f8e4e7d-17c7-4a53-89b8-d7b5203c6e1a {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1390.796294] env[63515]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6458c39-2561-4ad6-8f5e-b8b4aa443f8c {{(pid=63515) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1390.808685] env[63515]: DEBUG nova.compute.provider_tree [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Inventory has not changed in ProviderTree for provider: 2bd86232-2b6f-44d5-9057-1a3a6b27185a {{(pid=63515) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1391.311903] env[63515]: DEBUG nova.scheduler.client.report [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Inventory has not changed for provider 2bd86232-2b6f-44d5-9057-1a3a6b27185a based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 171, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=63515) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1391.816751] env[63515]: DEBUG nova.compute.resource_tracker [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=63515) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1391.817125] env[63515]: DEBUG oslo_concurrency.lockutils [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.216s {{(pid=63515) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1394.536155] env[63515]: DEBUG oslo_service.periodic_task [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Running periodic task ComputeManager._run_pending_deletes {{(pid=63515) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1394.536553] env[63515]: DEBUG nova.compute.manager [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] Cleaning up deleted instances {{(pid=63515) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11226}} [ 1395.043760] env[63515]: DEBUG nova.compute.manager [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] There are 12 instances to clean {{(pid=63515) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11235}} [ 1395.044031] env[63515]: DEBUG nova.compute.manager [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] [instance: b9b28202-1599-4f11-8d92-4942f90160ff] Instance has had 0 of 5 cleanup attempts {{(pid=63515) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1395.548750] env[63515]: DEBUG nova.compute.manager [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] [instance: d63a2a0d-2736-4d1f-a9fb-ec105747dcc3] Instance has had 0 of 5 cleanup attempts {{(pid=63515) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1396.052293] env[63515]: DEBUG nova.compute.manager [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] [instance: bd3a77ee-a78b-4719-9353-307ae8b166b6] Instance has had 0 of 5 cleanup attempts {{(pid=63515) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1396.556330] env[63515]: DEBUG nova.compute.manager [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] [instance: 24a79a6c-5210-4778-a6ed-482bfe14933c] Instance has had 0 of 5 cleanup attempts {{(pid=63515) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1397.060058] env[63515]: DEBUG nova.compute.manager [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] [instance: 9a864a51-7df3-4afc-8be7-2794ef2dcf1a] Instance has had 0 of 5 cleanup attempts {{(pid=63515) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1397.563861] env[63515]: DEBUG nova.compute.manager [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] [instance: 7b0df8d3-1b14-4e00-90c0-3dab3efc153c] Instance has had 0 of 5 cleanup attempts {{(pid=63515) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1398.066974] env[63515]: DEBUG nova.compute.manager [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] [instance: c9e07a6c-fd14-4071-8c69-6ece62e9109a] Instance has had 0 of 5 cleanup attempts {{(pid=63515) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1398.570595] env[63515]: DEBUG nova.compute.manager [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] [instance: c8dbf542-70cb-45be-9308-5e12307c0be6] Instance has had 0 of 5 cleanup attempts {{(pid=63515) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1399.074554] env[63515]: DEBUG nova.compute.manager [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] [instance: ea2e0810-c9f1-49a6-8589-7db402847aea] Instance has had 0 of 5 cleanup attempts {{(pid=63515) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1399.578463] env[63515]: DEBUG nova.compute.manager [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] [instance: 6336f10d-97b3-4413-a828-823a27492215] Instance has had 0 of 5 cleanup attempts {{(pid=63515) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1400.082256] env[63515]: DEBUG nova.compute.manager [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] [instance: aa83d035-9c47-434d-ab1a-140e6b3110b4] Instance has had 0 of 5 cleanup attempts {{(pid=63515) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1400.585477] env[63515]: DEBUG nova.compute.manager [None req-e5b9f5ef-9492-4900-a7ae-d40a95327b5f None None] [instance: e13da90c-28e6-43d2-99b7-19c5095954ca] Instance has had 0 of 5 cleanup attempts {{(pid=63515) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}}